Update README.md
Browse files
README.md
CHANGED
@@ -12,146 +12,124 @@ tags:
|
|
12 |
- visual-grounding
|
13 |
---
|
14 |
|
15 |
-
# Dataset Card for Dataset
|
16 |
-
|
17 |
-
<!-- Provide a quick summary of the dataset. -->
|
18 |
-
|
19 |
-
|
20 |
|
21 |
## Dataset Details
|
22 |
|
23 |
### Dataset Description
|
24 |
|
25 |
-
|
26 |
|
27 |
-
|
28 |
-
This dataset contains spatial dialogue data for visual grounding tasks. It includes pairs of images showing speaker and listener views, along with natural language descriptions of spatial locations and target object positions.
|
29 |
|
30 |
### Dataset Summary
|
31 |
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
- **
|
36 |
-
- **
|
37 |
-
- **
|
38 |
-
- **
|
39 |
-
- **License:** apache-2.0
|
40 |
-
|
41 |
-
### Dataset Sources [optional]
|
42 |
-
|
43 |
-
<!-- Provide the basic links for the dataset. -->
|
44 |
-
|
45 |
-
- **Repository:** [More Information Needed]
|
46 |
-
- **Paper [optional]:** [More Information Needed]
|
47 |
-
- **Demo [optional]:** [More Information Needed]
|
48 |
|
49 |
## Uses
|
50 |
|
51 |
-
<!-- Address questions around how the dataset is intended to be used. -->
|
52 |
-
|
53 |
### Direct Use
|
54 |
|
55 |
-
|
56 |
-
|
57 |
-
|
|
|
|
|
|
|
58 |
|
59 |
### Out-of-Scope Use
|
60 |
|
61 |
-
|
62 |
-
|
63 |
-
|
|
|
64 |
|
65 |
## Dataset Structure
|
66 |
|
67 |
-
|
68 |
-
|
69 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
70 |
|
71 |
## Dataset Creation
|
72 |
|
73 |
### Curation Rationale
|
74 |
|
75 |
-
|
76 |
-
|
77 |
-
[More Information Needed]
|
78 |
|
79 |
### Source Data
|
80 |
|
81 |
-
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
|
82 |
-
|
83 |
#### Data Collection and Processing
|
84 |
|
85 |
-
|
86 |
-
|
87 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
88 |
|
89 |
#### Who are the source data producers?
|
90 |
|
91 |
-
|
92 |
-
|
93 |
-
|
94 |
-
|
95 |
-
### Annotations [optional]
|
96 |
-
|
97 |
-
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
|
98 |
-
|
99 |
-
#### Annotation process
|
100 |
|
101 |
-
|
102 |
|
103 |
-
|
104 |
-
|
105 |
-
#### Who are the annotators?
|
106 |
-
|
107 |
-
<!-- This section describes the people or systems who created the annotations. -->
|
108 |
-
|
109 |
-
[More Information Needed]
|
110 |
-
|
111 |
-
#### Personal and Sensitive Information
|
112 |
-
|
113 |
-
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
|
114 |
-
|
115 |
-
[More Information Needed]
|
116 |
|
117 |
## Bias, Risks, and Limitations
|
118 |
|
119 |
-
|
120 |
-
|
121 |
-
|
|
|
|
|
|
|
122 |
|
123 |
### Recommendations
|
124 |
|
125 |
-
|
126 |
-
|
127 |
-
|
|
|
|
|
128 |
|
129 |
-
## Citation
|
130 |
-
|
131 |
-
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
|
132 |
|
133 |
**BibTeX:**
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
134 |
|
135 |
-
|
136 |
-
|
137 |
-
**APA:**
|
138 |
-
|
139 |
-
[More Information Needed]
|
140 |
-
|
141 |
-
## Glossary [optional]
|
142 |
-
|
143 |
-
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
|
144 |
-
|
145 |
-
[More Information Needed]
|
146 |
-
|
147 |
-
## More Information [optional]
|
148 |
-
|
149 |
-
[More Information Needed]
|
150 |
-
|
151 |
-
## Dataset Card Authors [optional]
|
152 |
|
153 |
-
|
154 |
|
155 |
-
##
|
156 |
|
157 |
-
|
|
|
12 |
- visual-grounding
|
13 |
---
|
14 |
|
15 |
+
# Dataset Card for Multi-Agent Referential Communication Dataset
|
|
|
|
|
|
|
|
|
16 |
|
17 |
## Dataset Details
|
18 |
|
19 |
### Dataset Description
|
20 |
|
21 |
+
This dataset contains spatial dialogue data for multi-agent referential communication tasks in 3D environments. It includes pairs of images showing speaker and listener views within photorealistic indoor scenes, along with natural language descriptions of target object locations.
|
22 |
|
23 |
+
The key feature of this dataset is that it captures communication between two agents with different physical perspectives in a shared 3D space. Each agent has their own unique viewpoint of the scene, requiring them to consider each other's perspectives when generating and interpreting spatial references.
|
|
|
24 |
|
25 |
### Dataset Summary
|
26 |
|
27 |
+
- **Size**: 2,970 dialogue instances across 1,485 scenes
|
28 |
+
- **Total Scenes Generated**: 27,504 scenes (24,644 train, 1,485 validation, 1,375 test)
|
29 |
+
- **Task Type**: Referential communication between embodied agents
|
30 |
+
- **Language(s)**: English
|
31 |
+
- **License**: Apache-2.0
|
32 |
+
- **Curated by**: University of California, Berkeley
|
33 |
+
- **Time per Task**: Median 33.0s for speakers, 10.5s for listeners
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
|
35 |
## Uses
|
36 |
|
|
|
|
|
37 |
### Direct Use
|
38 |
|
39 |
+
The dataset is designed for:
|
40 |
+
- Training and evaluating referring expression generation models
|
41 |
+
- Training and evaluating visual question answering systems
|
42 |
+
- Studying human spatial language use in multi-perspective scenarios
|
43 |
+
- Developing embodied AI systems that can communicate about shared environments
|
44 |
+
- Research on perspective-taking in language generation and comprehension
|
45 |
|
46 |
### Out-of-Scope Use
|
47 |
|
48 |
+
The dataset should not be used for:
|
49 |
+
- Training systems to navigate or manipulate physical environments
|
50 |
+
- Training general-purpose vision-language models without consideration of perspective
|
51 |
+
- Applications requiring real-time interaction or dialogue (dataset contains single-turn interactions only)
|
52 |
|
53 |
## Dataset Structure
|
54 |
|
55 |
+
Each instance contains:
|
56 |
+
- Speaker view image (1280x720 resolution)
|
57 |
+
- Listener view image (1280x720 resolution)
|
58 |
+
- Natural language referring expression from speaker
|
59 |
+
- Target object location
|
60 |
+
- Listener object selection
|
61 |
+
- Scene metadata including:
|
62 |
+
- Agent positions and orientations
|
63 |
+
- Field of view overlap measurements
|
64 |
+
- Referent placement method (random vs adversarial)
|
65 |
+
- Base environment identifier
|
66 |
|
67 |
## Dataset Creation
|
68 |
|
69 |
### Curation Rationale
|
70 |
|
71 |
+
The dataset was created to study how humans and AI systems handle referential communication when embodied in different physical perspectives within a shared environment. The multi-perspective nature of the task reflects real-world scenarios where agents must coordinate despite having different viewpoints.
|
|
|
|
|
72 |
|
73 |
### Source Data
|
74 |
|
|
|
|
|
75 |
#### Data Collection and Processing
|
76 |
|
77 |
+
1. Base environments from ScanNet++ (450 high-quality 3D indoor environments)
|
78 |
+
2. Scene generation process:
|
79 |
+
- Place two agents with controlled relative orientations (0° to 180°)
|
80 |
+
- Place 3 referent objects using either random or adversarial placement
|
81 |
+
- Render images from each agent's perspective
|
82 |
+
- Apply quality filtering using GPT-4V
|
83 |
+
3. Human data collection:
|
84 |
+
- 194 qualified crowdworkers on Prolific
|
85 |
+
- Speakers describe target object location
|
86 |
+
- Listeners select object based on description
|
87 |
+
- 3 listener judgments per description
|
88 |
|
89 |
#### Who are the source data producers?
|
90 |
|
91 |
+
- Base 3D environments: ScanNet++ dataset
|
92 |
+
- Referring expressions: English-speaking crowdworkers from the United States
|
93 |
+
- Quality filtering: Automated GPT-4V system
|
94 |
+
- Scene generation: Automated system with physics simulation
|
|
|
|
|
|
|
|
|
|
|
95 |
|
96 |
+
### Personal and Sensitive Information
|
97 |
|
98 |
+
The dataset does not contain personally identifiable information. Crowdworker data was checked to exclude private information and offensive content.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
99 |
|
100 |
## Bias, Risks, and Limitations
|
101 |
|
102 |
+
- Limited to indoor environments from ScanNet++
|
103 |
+
- English language only
|
104 |
+
- Single-turn interactions only (no dialogue)
|
105 |
+
- Restricted to specific object types (spheres)
|
106 |
+
- May reflect cultural biases in spatial language use
|
107 |
+
- Limited demographic diversity of crowdworkers
|
108 |
|
109 |
### Recommendations
|
110 |
|
111 |
+
- Consider cultural and linguistic differences in spatial language when using the dataset
|
112 |
+
- Account for perspective differences when developing models
|
113 |
+
- Evaluate performance across different relative orientations and referent placements
|
114 |
+
- Consider expanding to multi-turn dialogue in future work
|
115 |
+
- Test for biases in spatial language use across different demographics
|
116 |
|
117 |
+
## Citation
|
|
|
|
|
118 |
|
119 |
**BibTeX:**
|
120 |
+
```
|
121 |
+
@article{tang2024grounding,
|
122 |
+
title={Grounding Language in Multi-Perspective Referential Communication},
|
123 |
+
author={Tang, Zineng and Mao, Lingjun and Suhr, Alane},
|
124 |
+
journal={EMNLP},
|
125 |
+
year={2024}
|
126 |
+
}
|
127 |
+
```
|
128 |
|
129 |
+
## Dataset Card Contact
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
130 |
|
131 |
+
Contact the authors at {terran, lingjun, suhr}@berkeley.edu
|
132 |
|
133 |
+
## More Information
|
134 |
|
135 |
+
Code, models, and dataset available at: https://github.com/zinengtang/MulAgentRef
|