A Security Risk Taxonomy for Prompt-Based Interaction With Large Language Models

Review (2024)
Author(s)

Erik Derner (ELLIS Alicante, Czech Technical University)

Kristina Batistic (Independent researcher)

Jan Zahalka (Czech Technical University)

Robert Babuška (TU Delft - Learning & Autonomous Control, Czech Technical University)

Research Group
Learning & Autonomous Control
DOI related publication
https://doi.org/10.1109/ACCESS.2024.3450388
More Info
expand_more
Publication Year
2024
Language
English
Research Group
Learning & Autonomous Control
Volume number
12
Pages (from-to)
126176-126187
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

As large language models (LLMs) permeate more and more applications, an assessment of their associated security risks becomes increasingly necessary. The potential for exploitation by malicious actors, ranging from disinformation to data breaches and reputation damage, is substantial. This paper addresses a gap in current research by specifically focusing on security risks posed by LLMs within the prompt-based interaction scheme, which extends beyond the widely covered ethical and societal implications. Our work proposes a taxonomy of security risks along the user-model communication pipeline and categorizes the attacks by target and attack type alongside the commonly used confidentiality, integrity, and availability (CIA) triad. The taxonomy is reinforced with specific attack examples to showcase the real-world impact of these risks. Through this taxonomy, we aim to inform the development of robust and secure LLM applications, enhancing their safety and trustworthiness.