1 |
Interpersonal Synchrony: From Social Perception to Social Interaction
|
|
|
|
In: Social Signal Processing ; https://hal-pasteur.archives-ouvertes.fr/pasteur-02070422 ; Edited by Judee K. Burgoon, University of Arizona Nadia Magnenat-Thalmann, Université de Genève Maja Pantic, Imperial College London Alessandro Vinciarelli, University of Glasgow. Social Signal Processing, Cambridge University Press, pp.202-212, 2017, Social Signal Processing, 9781316676202. ⟨10.1017/9781316676202.015⟩ ; https://www.cambridge.org/core/books/social-signal-processing/interpersonal-synchrony-from-social-perception-to-social-interaction/50D491B6C3AB7767858C80CF612C28A5 (2017)
|
|
BASE
|
|
Show details
|
|
2 |
Fera 2015 - second facial expression recognition and analysis challenge
|
|
|
|
In: http://www.cs.nott.ac.uk/%7Epszmv/Documents/FERA2015.pdf (2015)
|
|
BASE
|
|
Show details
|
|
3 |
M.: A semi-automatic methodology for facial landmark annotation. In
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/sagonas_cvpr_2013_amfg_w.pdf (2013)
|
|
BASE
|
|
Show details
|
|
4 |
The first facial expression recognition and analysis challenge
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/pdf17.pdf (2011)
|
|
BASE
|
|
Show details
|
|
5 |
Audiovisual discrimination between speech and laughter: Why and when visual information might help
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/petridispantic_2011_tmm.pdf (2011)
|
|
Abstract:
Abstract—Past research on automatic laughter classification/detection has focused mainly on audio-based approaches. Here we present an audiovisual approach to distinguishing laughter from speech, and we show that integrating the information from audio and video channels may lead to improved performance over single-modal approaches. Both audio and visual channels consist of two streams (cues), facial expressions and head pose for video and cepstral and prosodic features for audio. Two types of experiments were performed: 1) subject-independent cross-validation on the AMI dataset and 2) cross-database experiments on the AMI and SAL datasets. We experimented with different combinations of cues with the most informative being the combination of facial expressions, cepstral, and prosodic features. Our results suggest that the performance of the audiovisual approach is better on average than single-modal approaches. The addition of visual information produces better results when it comes to female subjects. When the training conditions are less diverse in terms of head movements than the testing conditions (training on the SAL dataset, testing on the AMI dataset), then no improvement was observed with the addition of visual information. On the other hand, when the training conditions are similar (cross validation on the AMI dataset), or more diverse (training on the AMI dataset, testing on the SAL dataset), in terms of head movements than is the case in the testing conditions, an absolute increase of about 3% in the F1 rate for laughter is reported when visual information is added to audio information. Index Terms—Human behavior analysis, laughter-versusspeech discrimination, neural networks, principal components analysis (PCA). I.
|
|
URL: http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.371.1291 http://ibug.doc.ic.ac.uk/media/uploads/documents/petridispantic_2011_tmm.pdf
|
|
BASE
|
|
Hide details
|
|
6 |
Classifying laughter and speech using audio-visual feature prediction
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/ICASSP-2010-PetridisEtAl-CAMERA.pdf (2010)
|
|
BASE
|
|
Show details
|
|
7 |
Static vs. Dynamic Modeling of Human Nonverbal Behavior from Multiple Cues and Modalities
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/ICMI-2009-PetridisEtAl-CAMERA.pdf (2009)
|
|
BASE
|
|
Show details
|
|
8 |
A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/PAMI-AVemotionSurvey-CAMERA.pdf (2009)
|
|
BASE
|
|
Show details
|
|
10 |
Social Signal Processing: Survey of an Emerging Domain
|
|
|
|
In: http://www.idiap.ch/~vincia/papers/sspsurvey.pdf (2008)
|
|
BASE
|
|
Show details
|
|
11 |
Social Signal Processing: State-of-the-art and future perspectives of an emerging domain
|
|
|
|
In: http://www.idiap.ch/~vincia/papers/bravetopic.pdf (2008)
|
|
BASE
|
|
Show details
|
|
12 |
Social Signal Processing: State-of-the-art and future perspectives of an emerging domain
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/ACM-MM-2008-VinciarelliEtAl-CAMERA.pdf (2008)
|
|
BASE
|
|
Show details
|
|
13 |
Social Signal Processing: Survey of an Emerging Domain
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/IVCJ-SSPsurvey-FINAL.pdf (2008)
|
|
BASE
|
|
Show details
|
|
14 |
Fusion of audio and visual cues for laughter detection
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/CIVR-2008-PetridisPantic-CAMERA.pdf (2008)
|
|
BASE
|
|
Show details
|
|
15 |
Facial Action Recognition for Facial Expression Analysis from Static Face Images
|
|
|
|
In: http://www.kbs.twi.tudelft.nl/People/Staff/M.Pantic/SMCB-2004-final.pdf (2004)
|
|
BASE
|
|
Show details
|
|
16 |
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
|
|
|
|
In: http://s.i-techonline.com/Book/Face-Recognition/ISBN978-3-902613-03-5-fr20.pdf
|
|
BASE
|
|
Show details
|
|
17 |
VISUAL-ONLY DISCRIMINATION BETWEEN NATIVE AND NON-NATIVE SPEECH
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/georgakisetal_visualonlynativevsnonnative.pdf
|
|
BASE
|
|
Show details
|
|
18 |
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
|
|
|
|
In: http://mplab.ucsd.edu/~marni/pubs/panticbartlett_2007.pdf
|
|
BASE
|
|
Show details
|
|
19 |
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/PanticBartlett-Chapter-Proof2.pdf
|
|
BASE
|
|
Show details
|
|
20 |
Social Signal Processing: The Research Agenda
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/looking@people-panticetal-revision-final.pdf
|
|
BASE
|
|
Show details
|
|
|
|