Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/121695
Title: Multimodal detection and classification of head movements in face-to-face conversations : exploring models, features and their interaction
Authors: Agirrezabal, Manex
Paggio, Patrizia
Navarretta, Costanza
Jongejan, Bart
Keywords: Human-computer interaction
Multimodal communication
Machine learning
Nonverbal communication
Computer vision
Issue Date: 2023
Citation: Agirrezabal, M., Paggio, P., Navarretta, C., Jongejan, B. (2023). Multimodal Detection and Classification of Head Movements in Face-to-Face Conversations: Exploring Models, Features and Their Interaction. Proceedings of Gespin 2023, Nijmegen, Netherlands. 1-6.
Abstract: In this work we perform multimodal detection and classification of head movements from face to face video conversation data. We have experimented with different models and feature sets and provided some insight on the effect of independent features, but also how their interaction can enhance a head movement classifier. Used features include nose, neck and mid hip position coordinates and their derivatives together with acoustic features, namely, intensity and pitch of the speaker on focus. Results show that when input features are sufficiently processed by interacting with each other, a linear classifier can reach a similar performance to a more complex non-linear neural model with several hidden layers. Our best models achieve state-of-the-art performance in the detection task, measured by macro-averaged F1 score.
URI: https://www.um.edu.mt/library/oar/handle/123456789/121695
Appears in Collections:Scholarly Works - InsLin



Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.