Annotator-Centric Active Learning for Subjective NLP Tasks

Conference Paper (2024)
Author(s)

M.T. van der Meer (Idiap Research Institute, Universiteit Leiden)

Neele Falk (University of Stuttgart)

Pradeep Murukannaiah (TU Delft - Interactive Intelligence)

E. Liscio (TU Delft - Interactive Intelligence)

Research Group
Interactive Intelligence
DOI related publication
https://doi.org/10.18653/v1/2024.emnlp-main.1031
More Info
expand_more
Publication Year
2024
Language
English
Research Group
Interactive Intelligence
Pages (from-to)
18537-18555
ISBN (electronic)
9798891761643
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Active Learning (AL) addresses the high costs of collecting human annotations by strategically annotating the most informative samples.However, for subjective NLP tasks, incorporating a wide range of perspectives in the annotation process is crucial to capture the variability in human judgments.We introduce Annotator-Centric Active Learning (ACAL), which incorporates an annotator selection strategy following data sampling.Our objective is two-fold: (1) to efficiently approximate the full diversity of human judgments, and (2) to assess model performance using annotator-centric metrics, which value minority and majority perspectives equally.We experiment with multiple annotator selection strategies across seven subjective NLP tasks, employing both traditional and novel, human-centered evaluation metrics.Our findings indicate that ACAL improves data efficiency and excels in annotator-centric performance evaluations.However, its success depends on the availability of a sufficiently large and diverse pool of annotators to sample from.