Revisiting SVM Training

Optimizing SVM Hyperparameter tuning using early stopping in the SMO algorithm

Master Thesis (2025)
Author(s)

I. Dekker (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

Marcel JT Reinders – Graduation committee member (TU Delft - Pattern Recognition and Bioinformatics)

Tom J. Viering – Mentor (TU Delft - Pattern Recognition and Bioinformatics)

O.T. Turan – Mentor (TU Delft - Pattern Recognition and Bioinformatics)

I.B.C.M. Barcelos Carneiro M Rocha – Graduation committee member (TU Delft - Applied Mechanics)

Faculty
Electrical Engineering, Mathematics and Computer Science
More Info
expand_more
Publication Year
2025
Language
English
Graduation Date
24-03-2025
Awarding Institution
Delft University of Technology
Programme
Computer Engineering
Faculty
Electrical Engineering, Mathematics and Computer Science
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Support Vector Machines (SVMs) are widely used in various domains, with their performance heavily dependent on hyperparameter selection. However, hyperparameter tuning is computationally demanding due to the SVM training complexity, which is at best $O(n^2)$, where $n$ represents the number of training samples. To mitigate this challenge, we propose integrating a validation-based early stopping criterion into the Sequential Minimal Optimization (SMO) algorithm to enhance tuning efficiency.

We evaluate this approach within Random Search and Successive Halving frameworks, aiming to reduce tuning runtime while preserving model performance. We introduce a composite score function to facilitate a balanced assessment of accuracy and efficiency. Our empirical analysis reveals that incorporating early stopping into SMO significantly reduces hyperparameter tuning time under RS but provides limited benefits in successive halving, given its inherent efficiency. Additionally, while dataset characteristics influence the effectiveness of early stopping, we found evidence that dimensionality does not. We also observe that frequent early stopping objective assessments introduce computational overhead, which can offset runtime improvements. Reducing assessment frequency alleviates this issue, but it diminishes the effectiveness of early stopping.

Our findings highlight the potential of early stopping in SMO for optimizing SVM hyperparameter tuning, particularly within random search-based approaches. They also identify trade-offs in assessment frequency and dataset-specific factors.

Files

License info not available