BERT Rankers are Brittle

A Study using Adversarial Document Perturbations

Conference Paper (2022)
Author(s)

Yumeng Wang (L3S)

Lijun Lyu (L3S)

Avishek Anand (TU Delft - Web Information Systems)

Research Group
Web Information Systems
Copyright
© 2022 Yumeng Wang, Lijun Lyu, A. Anand
DOI related publication
https://doi.org/10.1145/3539813.3545122
More Info
expand_more
Publication Year
2022
Language
English
Copyright
© 2022 Yumeng Wang, Lijun Lyu, A. Anand
Research Group
Web Information Systems
Pages (from-to)
115-120
ISBN (electronic)
978-1-4503-9412-3
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Contextual ranking models based on BERT are now well established for a wide range of passage and document ranking tasks. However, the robustness of BERT-based ranking models under adversarial inputs is under-explored. In this paper, we argue that BERT-rankers are not immune to adversarial attacks targeting retrieved documents given a query. Firstly, we propose algorithms for adversarial perturbation of both highly relevant and non-relevant documents using gradient-based optimization methods. The aim of our algorithms is to add/replace a small number of tokens to a highly relevant or non-relevant document to cause a large rank demotion or promotion. Our experiments show that a small number of tokens can already result in a large change in the rank of a document. Moreover, we find that BERT-rankers heavily rely on the document start/head for relevance prediction, making the initial part of the document more susceptible to adversarial attacks. More interestingly, we find a small set of recurring adversarial words that when added to documents result in successful rank demotion/promotion of any relevant/non-relevant document respectively. Finally, our adversarial tokens also show particular topic preferences within and across datasets, exposing potential biases from BERT pre-training or downstream datasets.

Files

3539813.3545122.pdf
(pdf | 1.25 Mb)
- Embargo expired in 01-07-2023
License info not available