Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/103812
Title: Automatic detection and classification of head movements in face-to-face conversations
Authors: Paggio, Patrizia
Gatt, Albert
Klinge, Roman
Keywords: Human-computer interaction
Corpora (Linguistics)
Speech processing systems
Automatic speech recognition
Visual perception
Face perception
Issue Date: 2020
Publisher: European Language Resources Association (ELRA)
Citation: Paggio, P., Agirrezabal, M., Jongejan, B., & Navarretta, C. (2020, May). Automatic detection and classification of head movements in face-to-face conversations. Proceedings of LREC2020 Workshop" People in language, vision and the mind (ONION 2020), France. 15-21.
Abstract: This paper presents an approach to automatic head movement detection and classification in data from a corpus of video-recorded face-toface conversations in Danish involving 12 different speakers. A number of classifiers were trained with different combinations of visual, acoustic and word features and tested in a leave-one-out cross validation scenario. The visual movement features were extracted from the raw video data using OpenPose, the acoustic ones from the sound files using Praat, and the word features from the transcriptions. The best results were obtained by a Multilayer Perceptron classifier, which reached an average 0.68 F1 score across the 12 speakers for head movement detection, and 0.40 for head movement classification given four different classes. In both cases, the classifier outperformed a simple most frequent class baseline, a more advanced baseline only relying on velocity features, and linear classifiers using different combinations of features
URI: https://www.um.edu.mt/library/oar/handle/123456789/103812
Appears in Collections:Scholarly Works - InsLin



Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.