Frame²
This repo contains the scripts and data related to the Frame² dataset.
The dataset comprises of two JSON Lines files. One for visual objects data/VO.jsonl and another for textual annotation.
Visual annotations
The visual annotations file consists of a list of objects similar to this:
{
"episode": 1,
"objectId": 313136,
"objectTimespan": [850.36, 852.76],
"frame": "People_by_origin",
"frameElement": "Person",
"boundingBoxes": [
[850.36, 120.0, 22.0, 617.0, 444.0],
...
]
}
The episode
field identifies the Pedro Pelo Mundo episode where that object appears.
The objectTimespan
is a tuple that representing the start and end miliseconds of the video where that object appears.
frame
and frameElement
are the actual FrameNet entities that the object represents.
Finally, boundingBoxes
is an array of variable size (with at least one element). Each element is a 5-tuple representing a fixed time point where that bounding box appears and the four other numbers to represent the box itself. It's a tuple of (milisecond, x, y, width, heigh)
. Where the video resolution is 854 x 480 (or 480p).
How to cite
Frederico Belcavello, Tiago Timponi Torrent, Ely E. Matos, Adriana S. Pagano, Maucha Gamonal, Natalia Sigiliano, Lívia Vicente Dutra, Helen de Andrade Abreu, Mairon Samagaio, Mariane Carvalho, Franciany Campos, Gabrielly Azalim, Bruna Mazzei, Mateus Fonseca de Oliveira, Ana Carolina Loçasso Luz, Lívia Pádua Ruiz, Júlia Bellei, Amanda Pestana, Josiane Costa, et al.. 2024. Frame2: A FrameNet-based Multimodal Dataset for Tackling Text-image Interactions in Video. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), pages 7429–7437, Torino, Italia. ELRA and ICCL.
License
This dataset is shared under a CC BY-NC 4.0 DEED license.
- Downloads last month
- 0