Bayesian deep learning

Insights in the Bayesian paradigm for deep learning

Master Thesis (2023)
Author(s)

W.R. Schipper (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

Aad van der Vaart – Mentor (TU Delft - Statistics)

A. Heinlein – Graduation committee member (TU Delft - Numerical Analysis)

Faculty
Electrical Engineering, Mathematics and Computer Science
Copyright
© 2023 Wieger Schipper
More Info
expand_more
Publication Year
2023
Language
English
Copyright
© 2023 Wieger Schipper
Graduation Date
30-08-2023
Awarding Institution
Delft University of Technology
Programme
Applied Mathematics | Stochastics
Faculty
Electrical Engineering, Mathematics and Computer Science
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

In this thesis, we study a particle method for Bayesian deep learning. In particular, we look at the estimation of the parameters of an ensemble of Bayesian neural networks by means of this particle method, called Stein variational gradient descent (SVGD). This method iteratively updates a collection of parameters and it has the property that its update directions are chosen such that they optimally decrease the Kullback-Leibler divergence. We also study gradient flows of probability measures and show how gradient flows corresponding to functionals on the space of probability measures can induce particle flows. We formulate SVGD as a method in this space. In the regime of infinite particles we show results about convergence of SVGD. An existing convergence result for SVGD can be extended by showing that the probability measures, governing the collection of SVGD particles, are uniformly tight. We give conditions under which this holds.

Files

MSc_Thesis_WR_Schipper.pdf
(pdf | 8.73 Mb)
License info not available