Adversarially Robust Decision Trees Against User-Specified Threat Models

Master Thesis (2020)
Author(s)

D.A. Vos (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

Sicco Verwer – Mentor (TU Delft - Cyber Security)

R.L. Lagendijk – Graduation committee member (TU Delft - Cyber Security)

Marco Loog – Graduation committee member (TU Delft - Pattern Recognition and Bioinformatics)

Faculty
Electrical Engineering, Mathematics and Computer Science
Copyright
© 2020 Daniël Vos
More Info
expand_more
Publication Year
2020
Language
English
Copyright
© 2020 Daniël Vos
Graduation Date
01-07-2020
Awarding Institution
Delft University of Technology
Programme
Computer Science | Cyber Security
Faculty
Electrical Engineering, Mathematics and Computer Science
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

In the present day we use machine learning for sensitive tasks that require models to be both understandable and robust. Although traditional models such as decision trees are understandable, they suffer from adversarial attacks. When a decision tree is used to differentiate between a user's benign and malicious behavior, an adversarial attack allows the user to effectively evade the model by perturbing the inputs the model receives. We can use algorithms that take adversarial attacks into account to fit trees that are more robust. In this work we propose an algorithm that is two orders of magnitudes faster and scores 4.3% better on accuracy against adversaries moving all samples than the state-of-the-art work while accepting an intuitive and permissible threat model. Where previous threat models were limited to distance norms, we allow each feature to be perturbed with a user-specified threat model specifying either a maximum distance or constraints on the direction of perturbation. Additionally we introduce two hyperparameters rho and phi that can control the trade-off between accuracy vs robustness and accuracy vs fairness respectively. Using the hyperparameters we can train models with less than 5% difference in false positive rate between population groups while scoring on average 2.4% higher on accuracy against adversarial attacks. Lastly, we show that our decision trees perform similarly to more complex random forests of fair and robust decision trees.

Files

Thesis_Daniel_Vos.pdf
(pdf | 2.6 Mb)
License info not available