Perspective

Leveraging Human Understanding for Identifying and Characterizing Image Atypicality

Conference Paper (2023)
Author(s)

Shahin Sharifi Noorian (TU Delft - Web Information Systems)

Sihang Qiu (Hunan Institute of Advanced Technology)

Burcu Sayin (UniversitĂ  degli Studi di Trento)

Agathe Balayn (TU Delft - Web Information Systems)

Ujwal Gadiraju (TU Delft - Web Information Systems)

Jie Yang (TU Delft - Web Information Systems)

Alessandro Bozzon (TU Delft - Human-Centred Artificial Intelligence)

Research Group
Web Information Systems
DOI related publication
https://doi.org/10.1145/3581641.3584096
More Info
expand_more
Publication Year
2023
Language
English
Research Group
Web Information Systems
Pages (from-to)
650-663
ISBN (print)
979-8-4007-0106-1
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

High-quality data plays a vital role in developing reliable image classification models. Despite that, what makes an image difficult to classify remains an unstudied topic. This paper provides a first-of-its-kind, model-agnostic characterization of image atypicality based on human understanding. We consider the setting of image classification "in the wild", where a large number of unlabeled images are accessible, and introduce a scalable and effective human computation approach for proactive identification and characterization of atypical images. Our approach consists of i) an image atypicality identification and characterization task that presents to the human worker both a local view of visually similar images and a global view of images from the class of interest and ii) an automatic image sampling method that selects a diverse set of atypical images based on both visual and semantic features. We demonstrate the effectiveness and cost-efficiency of our approach through controlled crowdsourcing experiments and provide a characterization of image atypicality based on human annotations of 10K images. We showcase the utility of the identified atypical images by testing state-of-the-art image classification services against such images and provide an in-depth comparative analysis of the alignment between human- and machine-perceived image atypicality. Our findings have important implications for developing and deploying reliable image classification systems.