Towards Single Camera Human 3D-Kinematics

Journal Article (2022)
Author(s)

M. Bittner (TU Delft - Biomechatronics & Human-Machine Control, Vicarious Perception Technologies)

W. Yang (Student TU Delft)

Xucong Zhang (TU Delft - Pattern Recognition and Bioinformatics)

Ajay Seth (TU Delft - Biomechatronics & Human-Machine Control)

Jan van Gemert (TU Delft - Pattern Recognition and Bioinformatics)

FCT van der Helm (TU Delft - Biomechatronics & Human-Machine Control)

Research Group
Biomechatronics & Human-Machine Control
Copyright
© 2022 M. Bittner, W. Yang, X. Zhang, A. Seth, J.C. van Gemert, F.C.T. van der Helm
DOI related publication
https://doi.org/10.3390/s23010341
More Info
expand_more
Publication Year
2022
Language
English
Copyright
© 2022 M. Bittner, W. Yang, X. Zhang, A. Seth, J.C. van Gemert, F.C.T. van der Helm
Research Group
Biomechatronics & Human-Machine Control
Issue number
1
Volume number
23
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Markerless estimation of 3D Kinematics has the great potential to clinically diagnose and monitor movement disorders without referrals to expensive motion capture labs; however, current approaches are limited by performing multiple de-coupled steps to estimate the kinematics of a person from videos. Most current techniques work in a multi-step approach by first detecting the pose of the body and then fitting a musculoskeletal model to the data for accurate kinematic estimation. Errors in training data of the pose detection algorithms, model scaling, as well the requirement of multiple cameras limit the use of these techniques in a clinical setting. Our goal is to pave the way toward fast, easily applicable and accurate 3D kinematic estimation . To this end, we propose a novel approach for direct 3D human kinematic estimation D3KE from videos using deep neural networks. Our experiments demonstrate that the proposed end-to-end training is robust and outperforms 2D and 3D markerless motion capture based kinematic estimation pipelines in terms of joint angles error by a large margin (35% from 5.44 to 3.54 degrees). We show that D3KE is superior to the multi-step approach and can run at video framerate speeds. This technology shows the potential for clinical analysis from mobile devices in the future.