license: cc0-1.0
dataset_info:
features:
- name: uuid
dtype: string
- name: image
dtype: image
- name: reference
dtype: string
splits:
- name: train
num_bytes: 3513634757
num_examples: 1000
- name: val
num_bytes: 866045160
num_examples: 250
- name: test
num_bytes: 1722946225
num_examples: 500
download_size: 6101723890
dataset_size: 6102626142
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: val
path: data/val-*
- split: test
path: data/test-*
task_categories:
- image-to-text
language:
- en
tags:
- art
size_categories:
- 1K<n<10K
Dataset Card for docent
This dataset contains works of art with expert-written detailed descriptions from the U.S. National Gallery of Art, published as part of DOCENT. It was introduced in "PoSh: Using Scene Graphs To Guide LLMs-as-a-Judge For Detailed Image Descriptions". You can find a full description of its collection methodology in the paper: https://arxiv.org/abs/2510.19060.
Dataset Details
- Language: English
- License: CC-0
Dataset Sources
- Images: images in the public domain, from the U.S. National Gallery of Art
- Reference Descriptions: expert-written references from the U.S. National Gallery of Art, published as part of their Open Data Initiative (https://github.com/NationalGalleryOfArt/opendata)
- Repository: https://github.com/amith-ananthram/posh
- Paper: https://arxiv.org/abs/2510.19060
Uses
The intended use of this dataset is as a benchmark for evaluating detailed image description, in particular for artwork. It contains three splits: a training set of 1,000 images, a validation set of 250 images and a test set of 500 images. When evaluating model generations, we recommend reporting PoSh scores (https://github.com/amith-ananthram/posh) or using a replicable metric that produces stronger correlations with the judgments in https://huggingface.co/datasets/amitha/docent-eval-coarse.
Dataset Structure
Each row in the dataset corresponds to a work of art.
- uuid: a unique identifier for work of art
- image: an image of the work of art (useful for multimodal metrics)
- reference: an expert-written reference description of this artwork from the U.S. National Gallery of Art
Dataset Creation
Curation Rationale
This dataset was collected to evaluate detailed image description, especially for artwork.
Source Data
The images/artwork are all in the public domain and provided by the U.S. National Gallery of Art.
The expert written references were published by the U.S. National Gallery of Art as part of their Open Data Initiative (https://github.com/NationalGalleryOfArt/opendata).
Annotations
Annotation process
The expert written reference descriptions were composed according to the U.S. National Gallery of Art's Accessibility Guidelines: https://www.nga.gov/visit/accessibility/collection-image-descriptions.
Who are the annotators?
An expert in art history from the U.S. National Gallery of Art.
Bias, Risks, and Limitations
While this work aims to benefit accessibility applications for blind and low-vision users (as reference descriptions were written according to the U.S. National Gallery of Art's Accessibility Guidelines: https://www.nga.gov/visit/accessibility/collection-image-descriptions), we acknowledge that it assumes a one-size-fits-all approach to assistive text. Ideally, such a benchmark would include different styles of accessibility text more representative of diverse user needs. However, it is our hope that by including reference descriptions that are extremely detailed, models that perform well in this more challenging setting will be able to adapt to a wide number of description needs.
Additionally, as with other computer vision systems, this work could theoretically be applied to surveillance contexts, but our focus on detailed description does not introduce novel privacy risks beyond those inherent to existing image analysis technologies. The primary intended application—-improving accessibility—-aligns with beneficial societal outcomes.
Citation
BibTeX:
@misc{ananthram2025poshusingscenegraphs, title={PoSh: Using Scene Graphs To Guide LLMs-as-a-Judge For Detailed Image Descriptions}, author={Amith Ananthram and Elias Stengel-Eskin and Lorena A. Bradford and Julia Demarest and Adam Purvis and Keith Krut and Robert Stein and Rina Elster Pantalony and Mohit Bansal and Kathleen McKeown}, year={2025}, eprint={2510.19060}, archivePrefix={arXiv}, primaryClass={cs.CV}, url={https://arxiv.org/abs/2510.19060}, }
APA:
Ananthram, A., Stengel-Eskin, E., Bradford, L.A., Demarest, J., Purvis, A., Krut, K., Stein, R., Pantalony, R.E., Bansal, M., McKeown, K. (2025). PoSh: Using Scene Graphs To Guide LLMs-as-a-Judge For Detailed Image Descriptions. arXiv preprint arXiv:2510.19060.
Dataset Card Authors
Amith Ananthram