An Attention Module for Convolutional Neural Networks

Conference Paper (2021)
Author(s)

B. Baozhou (TU Delft - Computer Engineering)

H. Hofstee (IBM Austin, TU Delft - Computer Engineering)

Jinho Lee (Yonsei University)

Z Al-Ars (TU Delft - Computer Engineering)

Research Group
Computer Engineering
Copyright
© 2021 B. Zhu, H.P. Hofstee, Jinho Lee, Z. Al-Ars
DOI related publication
https://doi.org/10.1007/978-3-030-86362-3_14
More Info
expand_more
Publication Year
2021
Language
English
Copyright
© 2021 B. Zhu, H.P. Hofstee, Jinho Lee, Z. Al-Ars
Research Group
Computer Engineering
Volume number
12891
Pages (from-to)
167-178
ISBN (print)
9783030863616
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Attention mechanism has been regarded as an advanced technique to capture long-range feature interactions and to boost the representation capability for convolutional neural networks. However, we found two ignored problems in current attentional activations-based models: the approximation problem and the insufficient capacity problem of the attention maps. To solve the two problems together, we initially propose an attention module for convolutional neural networks by developing an AW-convolution, where the shape of attention maps matches that of the weights rather than the activations. Our proposed attention module is a complementary method to previous attention-based schemes, such as those that apply the attention mechanism to explore the relationship between channel-wise and spatial features. Experiments on several datasets for image classification and object detection tasks show the effectiveness of our proposed attention module. In particular, our proposed attention module achieves 1.00 % Top-1 accuracy improvement on ImageNet classification over a ResNet101 baseline and 0.63 COCO-style Average Precision improvement on the COCO object detection on top of a Faster R-CNN baseline with the backbone of ResNet101-FPN. When integrating with the previous attentional activations-based models, our proposed attention module can further increase their Top-1 accuracy on ImageNet classification by up to 0.57 % and COCO-style Average Precision on the COCO object detection by up to 0.45. Code and pre-trained models will be publicly available.

Files

An_Attention_Module_for_Convol... (pdf)
(pdf | 0.361 Mb)
- Embargo expired in 08-04-2022
License info not available