DE eng

Search in the Catalogues and Directories

Page: 1 2
Hits 1 – 20 of 29

1
Interpersonal Synchrony: From Social Perception to Social Interaction
In: Social Signal Processing ; https://hal-pasteur.archives-ouvertes.fr/pasteur-02070422 ; Edited by Judee K. Burgoon, University of Arizona Nadia Magnenat-Thalmann, Université de Genève Maja Pantic, Imperial College London Alessandro Vinciarelli, University of Glasgow. Social Signal Processing, Cambridge University Press, pp.202-212, 2017, Social Signal Processing, 9781316676202. ⟨10.1017/9781316676202.015⟩ ; https://www.cambridge.org/core/books/social-signal-processing/interpersonal-synchrony-from-social-perception-to-social-interaction/50D491B6C3AB7767858C80CF612C28A5 (2017)
BASE
Show details
2
Fera 2015 - second facial expression recognition and analysis challenge
In: http://www.cs.nott.ac.uk/%7Epszmv/Documents/FERA2015.pdf (2015)
BASE
Show details
3
M.: A semi-automatic methodology for facial landmark annotation. In
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/sagonas_cvpr_2013_amfg_w.pdf (2013)
BASE
Show details
4
The first facial expression recognition and analysis challenge
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/pdf17.pdf (2011)
BASE
Show details
5
Audiovisual discrimination between speech and laughter: Why and when visual information might help
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/petridispantic_2011_tmm.pdf (2011)
Abstract: Abstract—Past research on automatic laughter classification/detection has focused mainly on audio-based approaches. Here we present an audiovisual approach to distinguishing laughter from speech, and we show that integrating the information from audio and video channels may lead to improved performance over single-modal approaches. Both audio and visual channels consist of two streams (cues), facial expressions and head pose for video and cepstral and prosodic features for audio. Two types of experiments were performed: 1) subject-independent cross-validation on the AMI dataset and 2) cross-database experiments on the AMI and SAL datasets. We experimented with different combinations of cues with the most informative being the combination of facial expressions, cepstral, and prosodic features. Our results suggest that the performance of the audiovisual approach is better on average than single-modal approaches. The addition of visual information produces better results when it comes to female subjects. When the training conditions are less diverse in terms of head movements than the testing conditions (training on the SAL dataset, testing on the AMI dataset), then no improvement was observed with the addition of visual information. On the other hand, when the training conditions are similar (cross validation on the AMI dataset), or more diverse (training on the AMI dataset, testing on the SAL dataset), in terms of head movements than is the case in the testing conditions, an absolute increase of about 3% in the F1 rate for laughter is reported when visual information is added to audio information. Index Terms—Human behavior analysis, laughter-versusspeech discrimination, neural networks, principal components analysis (PCA). I.
URL: http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.371.1291
http://ibug.doc.ic.ac.uk/media/uploads/documents/petridispantic_2011_tmm.pdf
BASE
Hide details
6
Classifying laughter and speech using audio-visual feature prediction
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/ICASSP-2010-PetridisEtAl-CAMERA.pdf (2010)
BASE
Show details
7
Static vs. Dynamic Modeling of Human Nonverbal Behavior from Multiple Cues and Modalities
In: http://www.doc.ic.ac.uk/~maja/ICMI-2009-PetridisEtAl-CAMERA.pdf (2009)
BASE
Show details
8
A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions
In: http://www.doc.ic.ac.uk/~maja/PAMI-AVemotionSurvey-CAMERA.pdf (2009)
BASE
Show details
9
Social Signal Processing: Survey of an Emerging Domain
BASE
Show details
10
Social Signal Processing: Survey of an Emerging Domain
In: http://www.idiap.ch/~vincia/papers/sspsurvey.pdf (2008)
BASE
Show details
11
Social Signal Processing: State-of-the-art and future perspectives of an emerging domain
In: http://www.idiap.ch/~vincia/papers/bravetopic.pdf (2008)
BASE
Show details
12
Social Signal Processing: State-of-the-art and future perspectives of an emerging domain
In: http://www.doc.ic.ac.uk/~maja/ACM-MM-2008-VinciarelliEtAl-CAMERA.pdf (2008)
BASE
Show details
13
Social Signal Processing: Survey of an Emerging Domain
In: http://www.doc.ic.ac.uk/~maja/IVCJ-SSPsurvey-FINAL.pdf (2008)
BASE
Show details
14
Fusion of audio and visual cues for laughter detection
In: http://www.doc.ic.ac.uk/~maja/CIVR-2008-PetridisPantic-CAMERA.pdf (2008)
BASE
Show details
15
Facial Action Recognition for Facial Expression Analysis from Static Face Images
In: http://www.kbs.twi.tudelft.nl/People/Staff/M.Pantic/SMCB-2004-final.pdf (2004)
BASE
Show details
16
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
In: http://s.i-techonline.com/Book/Face-Recognition/ISBN978-3-902613-03-5-fr20.pdf
BASE
Show details
17
VISUAL-ONLY DISCRIMINATION BETWEEN NATIVE AND NON-NATIVE SPEECH
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/georgakisetal_visualonlynativevsnonnative.pdf
BASE
Show details
18
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
In: http://mplab.ucsd.edu/~marni/pubs/panticbartlett_2007.pdf
BASE
Show details
19
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
In: http://www.doc.ic.ac.uk/~maja/PanticBartlett-Chapter-Proof2.pdf
BASE
Show details
20
Social Signal Processing: The Research Agenda
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/looking@people-panticetal-revision-final.pdf
BASE
Show details

Page: 1 2

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
29
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern