Lookahead adversarial learning for near real-time semantic segmentation

Journal Article (2021)
Author(s)

Hadi Jamali-Rad (TU Delft - Pattern Recognition and Bioinformatics, Shell Technology Center Amsterdam, Amsterdam)

Attila Szabo (Shell Technology Center Amsterdam, Amsterdam, Universiteit van Amsterdam)

Research Group
Pattern Recognition and Bioinformatics
Copyright
© 2021 H. Jamali-Rad, Attila Szabó
DOI related publication
https://doi.org/10.1016/j.cviu.2021.103271
More Info
expand_more
Publication Year
2021
Language
English
Copyright
© 2021 H. Jamali-Rad, Attila Szabó
Research Group
Pattern Recognition and Bioinformatics
Volume number
212
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Semantic segmentation is one of the most fundamental problems in computer vision with significant impact on a wide variety of applications. Adversarial learning is shown to be an effective approach for improving semantic segmentation quality by enforcing higher-level pixel correlations and structural information. However, state-of-the-art semantic segmentation models cannot be easily plugged into an adversarial setting because they are not designed to accommodate convergence and stability issues in adversarial networks. We bridge this gap by building a conditional adversarial network with a state-of-the-art segmentation model (DeepLabv3+) at its core. To battle the stability issues, we introduce a novel lookahead adversarial learning (LoAd) approach with an embedded label map aggregation module. We focus on semantic segmentation models that run fast at inference for near real-time field applications. Through extensive experimentation, we demonstrate that the proposed solution can alleviate divergence issues in an adversarial semantic segmentation setting and results in considerable performance improvements (+5% in some classes) on the baseline for three standard datasets.