Datasets:
Tasks:
Object Detection
Modalities:
Image
Languages:
English
Size:
10K<n<100K
ArXiv:
Libraries:
FiftyOne
annotations_creators: [] | |
language: en | |
size_categories: | |
- 10K<n<100K | |
task_categories: | |
- object-detection | |
task_ids: [] | |
pretty_name: GQA-35k | |
tags: | |
- fiftyone | |
- image | |
- object-detection | |
dataset_summary: ' | |
This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 35000 samples. | |
## Installation | |
If you haven''t already, install FiftyOne: | |
```bash | |
pip install -U fiftyone | |
``` | |
## Usage | |
```python | |
import fiftyone as fo | |
import fiftyone.utils.huggingface as fouh | |
# Load the dataset | |
# Note: other available arguments include ''max_samples'', etc | |
dataset = fouh.load_from_hub("Voxel51/GQA-Scene-Graph") | |
# Launch the App | |
session = fo.launch_app(dataset) | |
``` | |
' | |
# Dataset Card for GQA-35k | |
 | |
The GQA (Visual Reasoning in the Real World) dataset is a large-scale visual question answering dataset that includes scene graph annotations for each image. | |
This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 35000 samples. | |
Note: This is a 35,000 sample subset which does not contain questions, only the scene graph annotations as detection-level attributes. | |
You can find the recipe notebook for creating the dataset [here](https://colab.research.google.com/drive/1IjyvUSFuRtW2c5ErzSnz1eB9syKm0vo4?usp=sharing) | |
## Installation | |
If you haven't already, install FiftyOne: | |
```bash | |
pip install -U fiftyone | |
``` | |
## Usage | |
```python | |
import fiftyone as fo | |
import fiftyone.utils.huggingface as fouh | |
# Load the dataset | |
# Note: other available arguments include 'max_samples', etc | |
dataset = fouh.load_from_hub("Voxel51/GQA-Scene-Graph") | |
# Launch the App | |
session = fo.launch_app(dataset) | |
``` | |
## Dataset Details | |
### Dataset Description | |
## Scene Graph Annotations | |
- Each of the 113K images in GQA is associated with a detailed scene graph describing the objects, attributes and relations present. | |
- The scene graphs are based on a cleaner version of the Visual Genome scene graphs. | |
- For each image, the scene graph is provided as a dictionary (sceneGraph) containing: | |
- Image metadata like width, height, location, weather | |
- A dictionary (objects) mapping each object ID to its name, bounding box coordinates, attributes, and relations[6] | |
- Relations are represented as triples specifying the predicate (e.g. "holding", "on", "left of") and the target object ID[6] | |
- **Curated by:** Drew Hudson & Christopher Manning | |
- **Shared by:** [Harpreet Sahota](https://x.com/datascienceharp), Hacker-in-Residence at Voxel51 | |
- **Language(s) (NLP):** en | |
- **License:** | |
- GQA annotations (scene graphs, questions, programs) licensed under CC BY 4.0 | |
- Images sourced from Visual Genome may have different licensing terms | |
### Dataset Sources | |
- **Repository:** https://cs.stanford.edu/people/dorarad/gqa/ | |
- **Paper :** https://arxiv.org/pdf/1902.09506 | |
- **Demo:** https://cs.stanford.edu/people/dorarad/gqa/vis.html | |
## Dataset Structure | |
Here's the information presented as a markdown table: | |
| Field | Type | Description | | |
|-------|------|-------------| | |
| location | str | Optional. The location of the image, e.g. kitchen, beach. | | |
| weather | str | Optional. The weather in the image, e.g. sunny, cloudy. | | |
| objects | dict | A dictionary from objectId to its object. | | |
| object | dict | A visual element in the image (node). | | |
| name | str | The name of the object, e.g. person, apple or sky. | | |
| x | int | Horizontal position of the object bounding box (top left). | | |
| y | int | Vertical position of the object bounding box (top left). | | |
| w | int | The object bounding box width in pixels. | | |
| h | int | The object bounding box height in pixels. | | |
| attributes | [str] | A list of all the attributes of the object, e.g. blue, small, running. | | |
| relations | [dict] | A list of all outgoing relations (edges) from the object (source). | | |
| relation | dict | A triple representing the relation between source and target objects. | | |
Note: I've used non-breaking spaces (` `) to indent the nested fields in the 'Field' column to represent the hierarchy. This helps to visually distinguish the nested structure within the table. | |
## Citation | |
**BibTeX:** | |
```bibtex | |
@article{Hudson_2019, | |
title={GQA: A New Dataset for Real-World Visual Reasoning and Compositional Question Answering}, | |
ISBN={9781728132938}, | |
url={http://dx.doi.org/10.1109/CVPR.2019.00686}, | |
DOI={10.1109/cvpr.2019.00686}, | |
journal={2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)}, | |
publisher={IEEE}, | |
author={Hudson, Drew A. and Manning, Christopher D.}, | |
year={2019}, | |
month={Jun} | |
} | |
``` |