Cross-modal approach for conversational well-being monitoring with multi-sensory earables
Chulhong Min (Nokia Bell Labs)
Alessandro Montanari (Nokia Bell Labs)
Akhil Mathur (Nokia Bell Labs)
Seungchul Lee (Korea Advanced Institute of Science and Technology)
Fahim Kawsar (TU Delft - Knowledge and Intelligence Design)
More Info
expand_more
Abstract
We propose a cross-modal approach for conversational well-being monitoring with a multi-sensory earable. It consists of motion, audio, and BLE models on earables. Using the IMU sensor, the microphone, and BLE scanning, the models detect speaking activities, stress and emotion, and participants in the conversation, respectively. We discuss the feasibility in qualifying conversations with our purpose-built cross-modal model in an energy-efficient and privacy-preserving way. With the cross-modal model, we develop a mobile application that qualifies on-going conversations and provides personalised feedback on social well-being.