Ambiguous Images With Human Judgments for Robust Visual Event Classification
Sanders, Kate, Kriz, Reno, Liu, Anqi, Van Durme, Benjamin
–arXiv.org Artificial Intelligence
Contemporary vision benchmarks predominantly consider tasks on which humans can achieve near-perfect performance. However, humans are frequently presented with visual data that they cannot classify with 100% certainty, and models trained on standard vision benchmarks achieve low performance when evaluated on this data. To address this issue, we introduce a procedure for creating datasets of ambiguous images and use it to produce SQUID-E ("Squidy"), a collection of noisy images extracted from videos. All images are annotated with ground truth values and a test set is annotated with human uncertainty judgments. We use this dataset to characterize human uncertainty in vision tasks and evaluate existing visual event classification models. Experimental results suggest that existing vision models are not sufficiently equipped to provide meaningful outputs for ambiguous images and that datasets of this nature can be used to assess and improve such models through model training and direct evaluation of model calibration. These findings motivate large-scale ambiguous dataset creation and further research focusing on noisy visual data.
arXiv.org Artificial Intelligence
Oct-22-2022
- Country:
- Asia
- North America
- Canada > Ontario
- Toronto (0.14)
- United States > New York
- New York County > New York City (0.04)
- Canada > Ontario
- Genre:
- Research Report > New Finding (0.66)
- Industry:
- Health & Medicine > Therapeutic Area (0.46)
- Leisure & Entertainment (1.00)
- Technology:
- Information Technology
- Artificial Intelligence
- Machine Learning
- Neural Networks (0.68)
- Statistical Learning (0.93)
- Natural Language (1.00)
- Representation & Reasoning (1.00)
- Vision (1.00)
- Machine Learning
- Data Science > Data Mining (1.00)
- Sensing and Signal Processing > Image Processing (1.00)
- Artificial Intelligence
- Information Technology