|
--- |
|
language: en |
|
license: mit |
|
size_categories: |
|
- 1K<n<10K |
|
task_categories: |
|
- image-to-text |
|
- visual-question-answering |
|
tags: |
|
- spatial |
|
- dialogue |
|
- visual-grounding |
|
dataset_info: |
|
features: |
|
- name: instance_id |
|
dtype: int32 |
|
- name: scene_key |
|
dtype: string |
|
- name: listener_view_image |
|
dtype: image |
|
- name: speaker_view_image |
|
dtype: image |
|
- name: human_speaker_message |
|
dtype: string |
|
- name: speaker_elapsed_time |
|
dtype: float32 |
|
- name: positions |
|
dtype: string |
|
- name: listener_target_bbox |
|
dtype: string |
|
- name: listener_distractor_0_bbox |
|
dtype: string |
|
- name: listener_distractor_1_bbox |
|
dtype: string |
|
- name: speaker_target_bbox |
|
dtype: string |
|
- name: speaker_distractor_0_bbox |
|
dtype: string |
|
- name: speaker_distractor_1_bbox |
|
dtype: string |
|
- name: human_listener_message |
|
dtype: string |
|
- name: listener_elapsed_time |
|
dtype: float32 |
|
- name: type |
|
dtype: string |
|
splits: |
|
- name: validation |
|
num_bytes: 6561385869.0 |
|
num_examples: 2970 |
|
download_size: 6419735381 |
|
dataset_size: 6561385869.0 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: validation |
|
path: data/validation-* |
|
--- |
|
|
|
# Dataset Card for Multi-Agent Referential Communication Dataset |
|
|
|
<div align="center"> |
|
<img src="assets/main.png" alt="Example scene" width="400"/> |
|
|
|
*Example scene showing the speaker (left) and listener (right) views.* |
|
</div> |
|
|
|
## Dataset Details |
|
|
|
### Dataset Description |
|
|
|
This dataset contains spatial dialogue data for multi-agent referential communication tasks in 3D environments. It includes pairs of images showing speaker and listener views within photorealistic indoor scenes, along with natural language descriptions of target object locations. |
|
|
|
The key feature of this dataset is that it captures communication between two agents with different physical perspectives in a shared 3D space. Each agent has their own unique viewpoint of the scene, requiring them to consider each other's perspectives when generating and interpreting spatial references. |
|
|
|
### Dataset Summary |
|
|
|
- **Size**: 2,970 dialogue instances across 1,485 scenes |
|
- **Total Scenes Generated**: 27,504 scenes (24,644 train, 1,485 validation, 1,375 test) |
|
- **Task Type**: Referential communication between embodied agents |
|
- **Language(s)**: English |
|
- **License**: MIT |
|
- **Curated by**: University of California, Berkeley |
|
- **Time per Task**: Median 33.0s for speakers, 10.5s for listeners |
|
|
|
## Dataset Structure |
|
|
|
Each instance contains: |
|
- Speaker view image (1024x1024 resolution) |
|
- Listener view image (1024x1024 resolution) |
|
- Natural language referring expression from human speaker |
|
- Target object location |
|
- Listener object selection |
|
- Scene metadata including: |
|
- Agent positions and orientations |
|
- Referent placement method (random vs adversarial) |
|
- Base environment identifier |
|
|
|
## Dataset Creation |
|
|
|
1. Base environments from ScanNet++ (450 high-quality 3D indoor environments) |
|
2. Scene generation process: |
|
- Place two agents with controlled relative orientations (0° to 180°) |
|
- Place 3 referent objects using either random or adversarial placement |
|
- Render images from each agent's perspective |
|
- Apply quality filtering using GPT-4V |
|
|
|
## Citation |
|
|
|
**BibTeX:** |
|
``` |
|
@article{tang2024grounding, |
|
title={Grounding Language in Multi-Perspective Referential Communication}, |
|
author={Tang, Zineng and Mao, Lingjun and Suhr, Alane}, |
|
journal={EMNLP}, |
|
year={2024} |
|
} |
|
``` |
|
|
|
## Dataset Card Contact |
|
|
|
Contact the authors at terran@berkeley.edu |