No-Audio Multimodal Speech Detection in Crowded Social Settings task at MediaEval 2018

Conference Paper (2018)
Author(s)

Laura Cabrera-Quiros (TU Delft - Pattern Recognition and Bioinformatics, Instituto Tecnologico de Costa Rica)

E. Gedik (TU Delft - Pattern Recognition and Bioinformatics)

H.S. Hung (TU Delft - Pattern Recognition and Bioinformatics)

Research Group
Pattern Recognition and Bioinformatics
Copyright
© 2018 L.C. Cabrera Quiros, E. Gedik, H.S. Hung
More Info
expand_more
Publication Year
2018
Language
English
Copyright
© 2018 L.C. Cabrera Quiros, E. Gedik, H.S. Hung
Research Group
Pattern Recognition and Bioinformatics
Pages (from-to)
1-3
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

This overview paper provides a description of the automatic Human Behaviour Analysis (HBA) task for the MediaEval 2018. In its first edition, the HBA task focuses on analyzing one of the most basic elements of social behavior: the estimation of speaking status. Task participants are provided with cropped videos of individuals while interacting freely during a crowded mingle event that
was captured by an overhead camera. Each individual is also wearing a badge-like device hung around the neck recording tri-axial acceleration.
The goal of this task is to automatically estimate if a person is speaking or not using these two alternative modalities. In contrast to conventional speech detection approaches, no audio is used for this task. Instead, the automatic estimation system must exploit the natural human movements that accompany speech. The task seeks to achieve competitive estimation performance
compared to audio-based systems by exploiting the multi-modal aspects of the problem.

Files

License info not available