The Cross-evaluation of Machine Learning-based Network Intrusion Detection Systems

Journal Article (2022)
Author(s)

Giovanni Apruzzese (Universität Liechtenstein)

Luca Pajola (Università degli Studi di Padova)

M. Conti (TU Delft - Cyber Security, Università degli Studi di Padova)

Research Group
Cyber Security
Copyright
© 2022 Giovanni Apruzzese, Luca Pajola, M. Conti
DOI related publication
https://doi.org/10.1109/TNSM.2022.3157344
More Info
expand_more
Publication Year
2022
Language
English
Copyright
© 2022 Giovanni Apruzzese, Luca Pajola, M. Conti
Research Group
Cyber Security
Bibliographical Note
Green Open Access added to TU Delft Institutional Repository 'You share, we take care!' - Taverne project https://www.openaccess.nl/en/you-share-we-take-care Otherwise as indicated in the copyright section: the publisher is the copyright holder of this work and the author uses the Dutch legislation to make this work public.@en
Issue number
4
Volume number
19
Pages (from-to)
5152 - 5169
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Enhancing Network Intrusion Detection Systems (NIDS) with supervised Machine Learning (ML) is tough. ML-NIDS must be trained and evaluated, operations requiring data where benign and malicious samples are clearly labeled. Such labels demand costly expert knowledge, resulting in a lack of real deployments, as well as on papers always relying on the same outdated data. The situation improved recently, as some efforts disclosed their labeled datasets. However, most past works used such datasets just as a 'yet another' testbed, overlooking the added potential provided by such availability. In contrast, we promote using such existing labeled data to cross-evaluate ML-NIDS. Such approach received only limited attention and, due to its complexity, requires a dedicated treatment. We hence propose the first cross-evaluation model. Our model highlights the broader range of realistic use-cases that can be assessed via cross-evaluations, allowing the discovery of still unknown qualities of state-of-the-art ML-NIDS. For instance, their detection surface can be extended - at no additional labeling cost. However, conducting such cross-evaluations is challenging. Hence, we propose the first framework, XeNIDS, for reliable cross-evaluations based on Network Flows. By using XeNIDS on six well-known datasets, we demonstrate the concealed potential, but also the risks, of cross-evaluations of ML-NIDS.

Files

The_Cross_Evaluation_of_Machin... (pdf)
(pdf | 2.32 Mb)
- Embargo expired in 01-07-2023
License info not available