Exploiting Asset Similarities in Sample Efficient Portfolio Optimization

Bayesian Optimization with Learned Partitioning Structures

Master Thesis (2025)
Author(s)

N. Riemens (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

L.J.J. van Iersel – Mentor (TU Delft - Discrete Mathematics and Optimization)

G.N.J.C. Bierkens – Graduation committee member (TU Delft - Statistics)

Bálint Négyesi – Graduation committee member (Ortec Finance)

Faculty
Electrical Engineering, Mathematics and Computer Science
More Info
expand_more
Publication Year
2025
Language
English
Graduation Date
16-12-2025
Awarding Institution
Delft University of Technology
Programme
['Applied Mathematics']
Faculty
Electrical Engineering, Mathematics and Computer Science
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Multi-objective portfolio optimization problems with complex objectives and expensive function evaluations can be solved through the sample efficient solution method of Bayesian optimization. In realistic settings, inclusion of many assets coupled with a limited evaluation budget negatively impacts the standard ARD Gaussian process surrogate model, as in high-dimensional environments this kernel no longer permits meaningful inference without a vast number of observations. The ARD kernel overestimates problem complexity, as financial assets display significant structure or dependence, for instance by belonging to the same asset class. This work presents a methodology to learn and exploit these asset similarities through a Mahalanobis kernel. Within this kernel, we treat the partitioning structure as a latent variable, and learn this structure through a coordinate-wise greedy optimization approach based on k-fold cross-validation using previously observed data points. We examine our proposed methodology based on ability to retrieve synthetic similarities, surrogate predictive and uncertainty quantification performance, along with attained hypervolume during Bayesian optimization. Our results show we successfully find similarities both in the synthetic and authentic settings, leading to improved predictive performance of Gaussian processes equipped with the learned Mahalanobis kernel. Furthermore, we showcase the importance of correct uncertainty quantification for Bayesian optimization, and achieve improved results for models averaged over multiple structures - thus taking model uncertainty into account - compared to the use of a single locally optimal partitioning. For the moderate dimensionalities considered within this work, the ARD kernel achieves competitive hypervolume results. However, this work indicates the exploitation of similarity structures within portfolio optimization is a promising avenue, and we provide multiple ways forward to capture the full potential the main ideas presented have to offer.

Files

License info not available
warning

File under embargo until 16-12-2026