Metadata Acquisition¶
Todo
Introductory narrative (2-3 sentences): What metadata was collected or computed for the stimuli beyond basic image properties? Why is this metadata useful (e.g., for controlling confounds, building feature spaces, analyzing category structure)?
Image-level Metadata¶
Todo
Document all metadata fields that were collected or computed per image. For each field or group of fields, describe:
What it is and how it was obtained (manual annotation, automated tool, model extraction, database lookup)
Software / model / API used (with version)
Units or value range
Candidate metadata categories to cover:
Low-level Visual Properties¶
Todo
E.g., luminance, contrast, spatial frequency, color statistics. How were these computed? Which tool or script?
Semantic Annotations¶
Todo
E.g., object categories, scene types, number of objects, animacy. Were these human-annotated, from an existing dataset (COCO, ImageNet labels), or model-derived?
Model-derived Features¶
Todo
E.g., CLIP embeddings, DNN feature vectors, caption embeddings. Which models and layers? How are they stored (separate files, columns in stimuli.tsv)?
Other Metadata¶
Todo
Any other metadata: image source/provenance, licensing info per image, NSFW scores, aesthetic scores, etc.
Metadata File Format¶
Todo
Where does the metadata live? Is it all in stimuli/stimuli.tsv, or
are there separate files for different metadata types (e.g., embeddings
as .npy)? Document file paths and formats.
Cross-reference Stimulus Data for the full stimulus file organization.