Frequency learning for structured CNN filters with Gaussian fractional derivatives

Master Thesis (2021)
Author(s)

N.L. Saldanha (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

Silvia L. Pintea – Mentor (TU Delft - Pattern Recognition and Bioinformatics)

N. Tomen – Mentor (TU Delft - Pattern Recognition and Bioinformatics)

J.C. Gemert – Graduation committee member (TU Delft - Pattern Recognition and Bioinformatics)

Frans Oliehoek – Coach (TU Delft - Interactive Intelligence)

Faculty
Electrical Engineering, Mathematics and Computer Science
Copyright
© 2021 Nikhil Saldanha
More Info
expand_more
Publication Year
2021
Language
English
Copyright
© 2021 Nikhil Saldanha
Graduation Date
21-07-2021
Awarding Institution
Delft University of Technology
Programme
Computer Science | Data Science and Technology
Faculty
Electrical Engineering, Mathematics and Computer Science
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

A structured CNN filter basis allows incorporating priors about natural image statistics and thus require less training examples to learn, saving valuable annotation time. Here, we build on the Gaussian derivative CNN filter basis that learn both the orientation and scale of the filters. However, this Gaussian filter basis definition depends on a predetermined derivative order, which typically results in fixed frequency responses for the basis functions whereas the optimal frequency of the filters should depend on the data and the downstream learning task. We show that by learning the order of the basis we can accurately learn the frequency of the filters, and hence adapt to the optimal frequencies for the underlying task. We investigate the well-founded mathematical formulation of fractional derivatives to adapt the filter frequencies during training. Our formulation leads to parameter savings and data efficiency when compared to the standard CNNs and the Gaussian derivative CNN filter networks that we build on.

Files

TUD_Report.pdf
(pdf | 9.58 Mb)
License info not available