|
--- |
|
license: cc-by-4.0 |
|
--- |
|
|
|
# Dataset Card for [Dataset Name] |
|
|
|
## Table of Contents |
|
- [Table of Contents](#table-of-contents) |
|
- [Dataset Description](#dataset-description) |
|
- [Dataset Summary](#dataset-summary) |
|
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) |
|
- [Languages](#languages) |
|
- [Dataset Structure](#dataset-structure) |
|
- [Data Instances](#data-instances) |
|
- [Data Fields](#data-fields) |
|
- [Data Splits](#data-splits) |
|
- [Dataset Creation](#dataset-creation) |
|
- [Curation Rationale](#curation-rationale) |
|
- [Source Data](#source-data) |
|
- [Annotations](#annotations) |
|
- [Personal and Sensitive Information](#personal-and-sensitive-information) |
|
- [Considerations for Using the Data](#considerations-for-using-the-data) |
|
- [Social Impact of Dataset](#social-impact-of-dataset) |
|
- [Discussion of Biases](#discussion-of-biases) |
|
- [Other Known Limitations](#other-known-limitations) |
|
- [Additional Information](#additional-information) |
|
- [Dataset Curators](#dataset-curators) |
|
- [Licensing Information](#licensing-information) |
|
- [Citation Information](#citation-information) |
|
- [Contributions](#contributions) |
|
|
|
## Dataset Description |
|
|
|
- **Homepage:** [https://google.github.io/localized-narratives/(https://google.github.io/localized-narratives/) |
|
- **Repository:**: [https://github.com/google/localized-narratives](https://github.com/google/localized-narratives) |
|
- **Paper:** [Connecting Vision and Language with Localized Narratives](https://arxiv.org/pdf/1912.03098.pdf) |
|
- **Leaderboard:** |
|
- **Point of Contact:** |
|
|
|
### Dataset Summary |
|
|
|
Localized Narratives, a new form of multimodal image annotations connecting vision and language. |
|
We ask annotators to describe an image with their voice while simultaneously hovering their mouse over the region they are describing. |
|
Since the voice and the mouse pointer are synchronized, we can localize every single word in the description. |
|
This dense visual grounding takes the form of a mouse trace segment per word and is unique to our data. |
|
We annotated 849k images with Localized Narratives: the whole COCO, Flickr30k, and ADE20K datasets, and 671k images of Open Images, all of which we make publicly available. |
|
|
|
As of now, there is only the `OpenImages` subset, but feel free to contribute the other subset of Localized Narratives! |
|
|
|
`OpenImages_captions` is similar to the `OpenImages` subset. The differences are that captions are groupped per image (images can have multiple captions). For this subset, `timed_caption`, `traces` and `voice_recording` are not available. |
|
|
|
### Supported Tasks and Leaderboards |
|
|
|
[More Information Needed] |
|
|
|
### Languages |
|
|
|
[More Information Needed] |
|
|
|
## Dataset Structure |
|
|
|
### Data Instances |
|
|
|
Each instance has the following structure: |
|
``` |
|
{ |
|
dataset_id: 'mscoco_val2017', |
|
image_id: '137576', |
|
annotator_id: 93, |
|
caption: 'In this image there are group of cows standing and eating th...', |
|
timed_caption: [{'utterance': 'In this', 'start_time': 0.0, 'end_time': 0.4}, ...], |
|
traces: [[{'x': 0.2086, 'y': -0.0533, 't': 0.022}, ...], ...], |
|
voice_recording: 'coco_val/coco_val_137576_93.ogg' |
|
} |
|
``` |
|
|
|
### Data Fields |
|
|
|
Each line represents one Localized Narrative annotation on one image by one annotator and has the following fields: |
|
|
|
- `dataset_id`: String identifying the dataset and split where the image belongs, e.g. mscoco_val2017. |
|
- `image_id` String identifier of the image, as specified on each dataset. |
|
- `annotator_id` Integer number uniquely identifying each annotator. |
|
- `caption` Image caption as a string of characters. |
|
- `timed_caption` List of timed utterances, i.e. {utterance, start_time, end_time} where utterance is a word (or group of words) and (start_time, end_time) is the time during which it was spoken, with respect to the start of the recording. |
|
- `traces` List of trace segments, one between each time the mouse pointer enters the image and goes away from it. Each trace segment is represented as a list of timed points, i.e. {x, y, t}, where x and y are the normalized image coordinates (with origin at the top-left corner of the image) and t is the time in seconds since the start of the recording. Please note that the coordinates can go a bit beyond the image, i.e. <0 or >1, as we recorded the mouse traces including a small band around the image. |
|
- `voice_recording` Relative URL path with respect to https://storage.googleapis.com/localized-narratives/voice-recordings where to find the voice recording (in OGG format) for that particular image. |
|
|
|
### Data Splits |
|
|
|
[More Information Needed] |
|
|
|
## Dataset Creation |
|
|
|
### Curation Rationale |
|
|
|
[More Information Needed] |
|
|
|
### Source Data |
|
|
|
#### Initial Data Collection and Normalization |
|
|
|
[More Information Needed] |
|
|
|
#### Who are the source language producers? |
|
|
|
[More Information Needed] |
|
|
|
### Annotations |
|
|
|
#### Annotation process |
|
|
|
[More Information Needed] |
|
|
|
#### Who are the annotators? |
|
|
|
[More Information Needed] |
|
|
|
### Personal and Sensitive Information |
|
|
|
[More Information Needed] |
|
|
|
## Considerations for Using the Data |
|
|
|
### Social Impact of Dataset |
|
|
|
[More Information Needed] |
|
|
|
### Discussion of Biases |
|
|
|
[More Information Needed] |
|
|
|
### Other Known Limitations |
|
|
|
[More Information Needed] |
|
|
|
## Additional Information |
|
|
|
### Dataset Curators |
|
|
|
[More Information Needed] |
|
|
|
### Licensing Information |
|
|
|
[More Information Needed] |
|
|
|
### Citation Information |
|
|
|
[More Information Needed] |
|
|
|
### Contributions |
|
|
|
Thanks to [@VictorSanh](https://github.com/VictorSanh) for adding this dataset. |
|
|