--- license: cc-by-nc-sa-4.0 --- # OpenDV-YouTube This is the dataset repository of `OpenDV-YouTube` language annotations, including `context` and `command`. For more details, please refer to GenAD project and OpenDV-YouTube. ## Usage To use the annotations, you need to first download and prepare the data as instructed in OpenDV-YouTube. You can use the following code to load in annotations respectively. ```python import json # for train full_annos = [] for split_id in range(10): split = json.load(open("10hz_YouTube_train_split{}.json".format(str(split_id)), "r")) full_annos.extend(split) # for val val_annos = json.load(open("10hz_YouTube_val.json", "r")) ``` Annotations will be loaded in `full_annos` as a list where each element contains annotations for one video clip. All elements in the list are dictionaries of the following structure. ``` { "cmd": -- command, i.e. the command of the ego vehicle in the video clip. "blip": -- context, i.e. the BLIP description of the center frame in the video clip. "folder": -- the relative path from the processed OpenDV-YouTube dataset root to the image folder of the video clip. "first_frame": -- the filename of the first frame in the clip. Note that this file is included in the video clip. "last_frame": -- the filename of the last frame in the clip. Note that this file is included in the video clip. } ```