TRIDENT

Transductive Variational Inference of Decoupled Latent Variables for Few Shot Classification

Master Thesis (2022)
Author(s)

A.R. Singh (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

H. Jamali Rad – Mentor (TU Delft - Pattern Recognition and Bioinformatics)

J.C. van Gemert – Graduation committee member (TU Delft - Pattern Recognition and Bioinformatics)

Geert Leus – Graduation committee member (TU Delft - Signal Processing Systems)

Faculty
Electrical Engineering, Mathematics and Computer Science
Copyright
© 2022 Anuj Singh
More Info
expand_more
Publication Year
2022
Language
English
Copyright
© 2022 Anuj Singh
Graduation Date
26-08-2022
Awarding Institution
Delft University of Technology
Programme
Computer Science | Bioinformatics
Faculty
Electrical Engineering, Mathematics and Computer Science
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

The versatility to learn from a handful of samples is the hallmark of human intelligence. Few-shot learning is an endeavour to transcend this capability down to machines. Inspired by the promise and power of probabilistic deep learning, we propose a novel variational inference network for few-shot classification (coined as TRIDENT) to decouple the representation of an image into semantic and label latent variables, and simultaneously infer them in an intertwined fashion. To induce task-awareness, as part of the inference mechanics of TRIDENT, we exploit information across both query and support images of a few-shot task using a novel built-in attention-based transductive feature extraction module (we call AttFEX). Our extensive experimental results corroborate the efficacy of TRIDENT and demonstrate that, using the simplest of backbones, it sets a new state-of-the-art in the most commonly adopted datasets miniImageNet and tieredImageNet (offering up to 4% and 5% improvements, respectively), as well as for the recent challenging cross-domain miniImagenet --> CUB scenario offering a significant margin (up to 20% improvement) beyond the best existing cross-domain baselines.

Files

Masters_Thesis.pdf
(pdf | 22.7 Mb)
License info not available