Detecting head movements in video-recorded dyadic conversations
Publikation: Bidrag til bog/antologi/rapport › Konferencebidrag i proceedings › Forskning › fagfællebedømt
Standard
Detecting head movements in video-recorded dyadic conversations. / Paggio, Patrizia; Jongejan, Bart; Agirrezabal, Manex; Navarretta, Costanza.
Proceedings of the International Conference on Multimodal Interaction: Adjunct. New York : Association for Computing Machinery, 2018. s. 1-6.Publikation: Bidrag til bog/antologi/rapport › Konferencebidrag i proceedings › Forskning › fagfællebedømt
Harvard
APA
Vancouver
Author
Bibtex
}
RIS
TY - GEN
T1 - Detecting head movements in video-recorded dyadic conversations
AU - Paggio, Patrizia
AU - Jongejan, Bart
AU - Agirrezabal, Manex
AU - Navarretta, Costanza
PY - 2018
Y1 - 2018
N2 - This paper is about the automatic recognition of head movements in videos of face-to-face dyadic conversations. We present an approach where recognition of head movements is casted as a multimodal frame classification problem based on visual and acoustic features. The visual features include velocity, acceleration, and jerk values associated with head movements, while the acoustic ones are pitch and intensity measurements from the co-occuring speech. We present the results obtained by training and testing a number of classifiers on manually annotated data from two conversations. The best performing classifier, a Multilayer Perceptron trained using all the features, obtains 0.75 accuracy and outperforms the mono-modal baseline classifier.
AB - This paper is about the automatic recognition of head movements in videos of face-to-face dyadic conversations. We present an approach where recognition of head movements is casted as a multimodal frame classification problem based on visual and acoustic features. The visual features include velocity, acceleration, and jerk values associated with head movements, while the acoustic ones are pitch and intensity measurements from the co-occuring speech. We present the results obtained by training and testing a number of classifiers on manually annotated data from two conversations. The best performing classifier, a Multilayer Perceptron trained using all the features, obtains 0.75 accuracy and outperforms the mono-modal baseline classifier.
UR - https://dl.acm.org/citation.cfm?doid=3281151.3281152
U2 - 10.1145/3281151.3281152
DO - 10.1145/3281151.3281152
M3 - Article in proceedings
SN - 978-1-4503-6002-9
SP - 1
EP - 6
BT - Proceedings of the International Conference on Multimodal Interaction: Adjunct
PB - Association for Computing Machinery
CY - New York
ER -
ID: 209096029