Scale Learning in Scale-Equivariant Convolutional Networks
Mark Basting (Student TU Delft)
Robert Jan Bruintjes (TU Delft - Pattern Recognition and Bioinformatics)
Thaddäus Wiedemer (Eberhard Karls Universität Tübingen, Max Planck Institute for Intelligent Systems)
Matthias Kümmerer (Eberhard Karls Universität Tübingen)
Matthias Bethge (Tübingen AI Center, Eberhard Karls Universität Tübingen)
Jan van Gemert (TU Delft - Pattern Recognition and Bioinformatics)
More Info
expand_more
Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.
Abstract
Objects can take up an arbitrary number of pixels in an image: Objects come in different sizes, and, photographs of these objects may be taken at various distances to the camera. These pixel size variations are problematic for CNNs, causing them to learn separate filters for scaled variants of the same objects which prevents learning across scales. This is addressed by scale-equivariant approaches that share features across a set of pre-determined fixed internal scales. These works, however, give little information about how to best choose the internal scales when the underlying distribution of sizes, or scale distribution, in the dataset, is unknown. In this work we investigate learning the internal scales distribution in scale-equivariant CNNs, allowing them to adapt to unknown data scale distributions. We show that our method can learn the internal scales on various data scale distributions and can adapt the internal scales in current scale-equivariant approaches.