Using Phase Instead of Optical Flow for Action Recognition

Conference Paper (2019)
Authors

Omar Hommos (Student TU Delft)

Silvia Pintea (TU Delft - Pattern Recognition and Bioinformatics)

Pascal S.M. Mettes (Universiteit van Amsterdam)

Jan van Gemert (TU Delft - Pattern Recognition and Bioinformatics)

Research Group
Pattern Recognition and Bioinformatics
Copyright
© 2019 Omar Hommos, S. Pintea, Pascal S.M. Mettes, J.C. van Gemert
To reference this document use:
https://doi.org/10.1007/978-3-030-11024-6_51
More Info
expand_more
Publication Year
2019
Language
English
Copyright
© 2019 Omar Hommos, S. Pintea, Pascal S.M. Mettes, J.C. van Gemert
Research Group
Pattern Recognition and Bioinformatics
Bibliographical Note
Green Open Access added to TU Delft Institutional Repository ‘You share, we take care!’ – Taverne project https://www.openaccess.nl/en/you-share-we-take-care Otherwise as indicated in the copyright section: the publisher is the copyright holder of this work and the author uses the Dutch legislation to make this work public.@en
Volume number
11134
Pages (from-to)
678-691
ISBN (print)
978-303011023-9
ISBN (electronic)
978-3-030-11024-6
DOI:
https://doi.org/10.1007/978-3-030-11024-6_51
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Currently, the most common motion representation for action recognition is optical flow. Optical flow is based on particle tracking which adheres to a Lagrangian perspective on dynamics. In contrast to the Lagrangian perspective, the Eulerian model of dynamics does not track, but describes local changes. For video, an Eulerian phase-based motion representation, using complex steerable filters, has been successfully employed recently for motion magnification and video frame interpolation. Inspired by these previous works, here, we proposes learning Eulerian motion representations in a deep architecture for action recognition. We learn filters in the complex domain in an end-to-end manner. We design these complex filters to resemble complex Gabor filters, typically employed for phase-information extraction. We propose a phase-information extraction module, based on these complex filters, that can be used in any network architecture for extracting Eulerian representations. We experimentally analyze the added value of Eulerian motion representations, as extracted by our proposed phase extraction module, and compare with existing motion representations based on optical flow, on the UCF101 dataset.

Files

License info not available