Oct. 14, 2011 to Oct. 17, 2011
DOI Bookmark: http://doi.ieeecomputersociety.org/10.1109/KSE.2011.49
Laughter has been determined as an important social signal that can predict emotional information of users. This paper presents an extension of a previous study that discovers underlying affect in Filipino laughter using audio features, a posed laughter database and categorical labels. For this study, analysis of visual (facial points) and audio (voice) information from a spontaneous laughter corpus with dimensional labels was explored. Laughter instances from a three test subject made up the corpus. Audio features extracted from the instances included prosodic features such as pitch, energy, intensity, formants (F1, F2 and F3), pitch contours, and thirteen Mel Frequency Cepstral Coefficients. Visual features included 170 facial distances taken from 68 facial points. Machine learning experiments were then performed in which Support Vector Machines -- Regression yielded the lowest mean absolute error rate of 0.0506 for the facial dataset. Other classifiers used were Linear Regression and Multilayer Perceptron.
Laughter, Audio Signals, Video Signal, Affect/Emotion Recognition, Emphatic Computing
Christopher Galvan, David Manangan, Michael Sanchez, Jason Wong, Jocelynn Cu, "Audiovisual Affect Recognition in Spontaneous Filipino Laughter", KSE, 2011, Knowledge and Systems Engineering, International Conference on, Knowledge and Systems Engineering, International Conference on 2011, pp. 266-271, doi:10.1109/KSE.2011.49