Tensor-based Kernel Machines with Structured Inducing Points for Large and High-Dimensional Data

Journal Article (2023)
Author(s)

F. Wesel (TU Delft - Team Kim Batselier)

Kim Batselier (TU Delft - Team Kim Batselier)

Research Group
Team Kim Batselier
Copyright
© 2023 F. Wesel, K. Batselier
More Info
expand_more
Publication Year
2023
Language
English
Copyright
© 2023 F. Wesel, K. Batselier
Research Group
Team Kim Batselier
Volume number
206
Pages (from-to)
8308-8320
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Kernel machines are one of the most studied family of methods in machine learning. In the exact setting, training requires to instantiate the kernel matrix, thereby prohibiting their application to large-sampled data. One popular kernel approximation strategy which allows to tackle large-sampled data consists in interpolating product kernels on a set of grid-structured inducing points. However, since the number of model parameters increases exponentially with the dimensionality of the data, these methods are limited to small-dimensional datasets. In this work we lift this limitation entirely by placing inducing points on a grid and constraining the primal weights to be a low-rank Canonical Polyadic Decomposition. We derive a block coordinate descent algorithm that efficiently exploits grid-structured inducing points. The computational complexity of the algorithm scales linearly both in the number of samples and in the dimensionality of the data for any product kernel. We demonstrate the performance of our algorithm on large-scale and high-dimensional data, achieving state-of-the art results on a laptop computer. Our results show that grid-structured approaches can work in higher-dimensional problems.

Files

Wesel23a.pdf
(pdf | 2.41 Mb)
License info not available