Perspective

Leveraging Human Understanding for Identifying and Characterizing Image Atypicality

Conference Paper (2023)
Author(s)

S. Sharifi Noorian (TU Delft - Web Information Systems)

Sihang Qiu (Hunan Institute of Advanced Technology)

Burcu Sayin (Università degli Studi di Trento)

A.M.A. Balayn (TU Delft - Web Information Systems)

Ujwal Gadiraju (TU Delft - Web Information Systems)

J Yang (TU Delft - Web Information Systems)

A. Bozzon (TU Delft - Human-Centred Artificial Intelligence)

Research Group
Web Information Systems
Copyright
© 2023 S. Sharifi Noorian, S. Qiu, Burcu Sayin, A.M.A. Balayn, Ujwal Gadiraju, J. Yang, A. Bozzon
DOI related publication
https://doi.org/10.1145/3581641.3584096
More Info
expand_more
Publication Year
2023
Language
English
Copyright
© 2023 S. Sharifi Noorian, S. Qiu, Burcu Sayin, A.M.A. Balayn, Ujwal Gadiraju, J. Yang, A. Bozzon
Research Group
Web Information Systems
Pages (from-to)
650-663
ISBN (print)
979-8-4007-0106-1
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

High-quality data plays a vital role in developing reliable image classification models. Despite that, what makes an image difficult to classify remains an unstudied topic. This paper provides a first-of-its-kind, model-agnostic characterization of image atypicality based on human understanding. We consider the setting of image classification "in the wild", where a large number of unlabeled images are accessible, and introduce a scalable and effective human computation approach for proactive identification and characterization of atypical images. Our approach consists of i) an image atypicality identification and characterization task that presents to the human worker both a local view of visually similar images and a global view of images from the class of interest and ii) an automatic image sampling method that selects a diverse set of atypical images based on both visual and semantic features. We demonstrate the effectiveness and cost-efficiency of our approach through controlled crowdsourcing experiments and provide a characterization of image atypicality based on human annotations of 10K images. We showcase the utility of the identified atypical images by testing state-of-the-art image classification services against such images and provide an in-depth comparative analysis of the alignment between human- and machine-perceived image atypicality. Our findings have important implications for developing and deploying reliable image classification systems.