1 |
Interpersonal Synchrony: From Social Perception to Social Interaction
|
|
|
|
In: Social Signal Processing ; https://hal-pasteur.archives-ouvertes.fr/pasteur-02070422 ; Edited by Judee K. Burgoon, University of Arizona Nadia Magnenat-Thalmann, Université de Genève Maja Pantic, Imperial College London Alessandro Vinciarelli, University of Glasgow. Social Signal Processing, Cambridge University Press, pp.202-212, 2017, Social Signal Processing, 9781316676202. ⟨10.1017/9781316676202.015⟩ ; https://www.cambridge.org/core/books/social-signal-processing/interpersonal-synchrony-from-social-perception-to-social-interaction/50D491B6C3AB7767858C80CF612C28A5 (2017)
|
|
BASE
|
|
Show details
|
|
2 |
Fera 2015 - second facial expression recognition and analysis challenge
|
|
|
|
In: http://www.cs.nott.ac.uk/%7Epszmv/Documents/FERA2015.pdf (2015)
|
|
BASE
|
|
Show details
|
|
3 |
M.: A semi-automatic methodology for facial landmark annotation. In
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/sagonas_cvpr_2013_amfg_w.pdf (2013)
|
|
BASE
|
|
Show details
|
|
4 |
The first facial expression recognition and analysis challenge
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/pdf17.pdf (2011)
|
|
BASE
|
|
Show details
|
|
5 |
Audiovisual discrimination between speech and laughter: Why and when visual information might help
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/petridispantic_2011_tmm.pdf (2011)
|
|
BASE
|
|
Show details
|
|
6 |
Classifying laughter and speech using audio-visual feature prediction
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/ICASSP-2010-PetridisEtAl-CAMERA.pdf (2010)
|
|
Abstract:
In this study, a system that discriminates laughter from speech by modelling the relationship between audio and visual features is presented. The underlying assumption is that this relationship is different between speech and laughter. Neural networks are trained which learn the audio-to-visual and visual-to-audio features mapping for both classes. Classification of a new frame is performed via prediction. All the networks produce a prediction of the expected audio / visual features and the network with the best prediction, i.e., the model which best describes the audiovisual feature relationship, provides its label to the input frame. When trained on a simple dataset and tested on a hard dataset, the proposed approach outperforms audiovisual feature-level fusion, resulting in a 10.9 % and 6.4 % absolute increase in the F1 rate for laughter and classification rate, respectively. This indicates that classification based on prediction can produce a good model even when the available dataset is not challenging enough. Index Terms — laughter-vs-speech discrimination, audiovisual speech / laughter feature relationship, prediction-based classification 1.
|
|
URL: http://ibug.doc.ic.ac.uk/media/uploads/documents/ICASSP-2010-PetridisEtAl-CAMERA.pdf http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.370.9928
|
|
BASE
|
|
Hide details
|
|
7 |
Static vs. Dynamic Modeling of Human Nonverbal Behavior from Multiple Cues and Modalities
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/ICMI-2009-PetridisEtAl-CAMERA.pdf (2009)
|
|
BASE
|
|
Show details
|
|
8 |
A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/PAMI-AVemotionSurvey-CAMERA.pdf (2009)
|
|
BASE
|
|
Show details
|
|
10 |
Social Signal Processing: Survey of an Emerging Domain
|
|
|
|
In: http://www.idiap.ch/~vincia/papers/sspsurvey.pdf (2008)
|
|
BASE
|
|
Show details
|
|
11 |
Social Signal Processing: State-of-the-art and future perspectives of an emerging domain
|
|
|
|
In: http://www.idiap.ch/~vincia/papers/bravetopic.pdf (2008)
|
|
BASE
|
|
Show details
|
|
12 |
Social Signal Processing: State-of-the-art and future perspectives of an emerging domain
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/ACM-MM-2008-VinciarelliEtAl-CAMERA.pdf (2008)
|
|
BASE
|
|
Show details
|
|
13 |
Social Signal Processing: Survey of an Emerging Domain
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/IVCJ-SSPsurvey-FINAL.pdf (2008)
|
|
BASE
|
|
Show details
|
|
14 |
Fusion of audio and visual cues for laughter detection
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/CIVR-2008-PetridisPantic-CAMERA.pdf (2008)
|
|
BASE
|
|
Show details
|
|
15 |
Facial Action Recognition for Facial Expression Analysis from Static Face Images
|
|
|
|
In: http://www.kbs.twi.tudelft.nl/People/Staff/M.Pantic/SMCB-2004-final.pdf (2004)
|
|
BASE
|
|
Show details
|
|
16 |
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
|
|
|
|
In: http://s.i-techonline.com/Book/Face-Recognition/ISBN978-3-902613-03-5-fr20.pdf
|
|
BASE
|
|
Show details
|
|
17 |
VISUAL-ONLY DISCRIMINATION BETWEEN NATIVE AND NON-NATIVE SPEECH
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/georgakisetal_visualonlynativevsnonnative.pdf
|
|
BASE
|
|
Show details
|
|
18 |
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
|
|
|
|
In: http://mplab.ucsd.edu/~marni/pubs/panticbartlett_2007.pdf
|
|
BASE
|
|
Show details
|
|
19 |
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
|
|
|
|
In: http://www.doc.ic.ac.uk/~maja/PanticBartlett-Chapter-Proof2.pdf
|
|
BASE
|
|
Show details
|
|
20 |
Social Signal Processing: The Research Agenda
|
|
|
|
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/looking@people-panticetal-revision-final.pdf
|
|
BASE
|
|
Show details
|
|
|
|