DE eng

Search in the Catalogues and Directories

Page: 1 2
Hits 1 – 20 of 29

1
Interpersonal Synchrony: From Social Perception to Social Interaction
In: Social Signal Processing ; https://hal-pasteur.archives-ouvertes.fr/pasteur-02070422 ; Edited by Judee K. Burgoon, University of Arizona Nadia Magnenat-Thalmann, Université de Genève Maja Pantic, Imperial College London Alessandro Vinciarelli, University of Glasgow. Social Signal Processing, Cambridge University Press, pp.202-212, 2017, Social Signal Processing, 9781316676202. ⟨10.1017/9781316676202.015⟩ ; https://www.cambridge.org/core/books/social-signal-processing/interpersonal-synchrony-from-social-perception-to-social-interaction/50D491B6C3AB7767858C80CF612C28A5 (2017)
BASE
Show details
2
Fera 2015 - second facial expression recognition and analysis challenge
In: http://www.cs.nott.ac.uk/%7Epszmv/Documents/FERA2015.pdf (2015)
BASE
Show details
3
M.: A semi-automatic methodology for facial landmark annotation. In
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/sagonas_cvpr_2013_amfg_w.pdf (2013)
BASE
Show details
4
The first facial expression recognition and analysis challenge
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/pdf17.pdf (2011)
BASE
Show details
5
Audiovisual discrimination between speech and laughter: Why and when visual information might help
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/petridispantic_2011_tmm.pdf (2011)
BASE
Show details
6
Classifying laughter and speech using audio-visual feature prediction
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/ICASSP-2010-PetridisEtAl-CAMERA.pdf (2010)
Abstract: In this study, a system that discriminates laughter from speech by modelling the relationship between audio and visual features is presented. The underlying assumption is that this relationship is different between speech and laughter. Neural networks are trained which learn the audio-to-visual and visual-to-audio features mapping for both classes. Classification of a new frame is performed via prediction. All the networks produce a prediction of the expected audio / visual features and the network with the best prediction, i.e., the model which best describes the audiovisual feature relationship, provides its label to the input frame. When trained on a simple dataset and tested on a hard dataset, the proposed approach outperforms audiovisual feature-level fusion, resulting in a 10.9 % and 6.4 % absolute increase in the F1 rate for laughter and classification rate, respectively. This indicates that classification based on prediction can produce a good model even when the available dataset is not challenging enough. Index Terms — laughter-vs-speech discrimination, audiovisual speech / laughter feature relationship, prediction-based classification 1.
URL: http://ibug.doc.ic.ac.uk/media/uploads/documents/ICASSP-2010-PetridisEtAl-CAMERA.pdf
http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.370.9928
BASE
Hide details
7
Static vs. Dynamic Modeling of Human Nonverbal Behavior from Multiple Cues and Modalities
In: http://www.doc.ic.ac.uk/~maja/ICMI-2009-PetridisEtAl-CAMERA.pdf (2009)
BASE
Show details
8
A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions
In: http://www.doc.ic.ac.uk/~maja/PAMI-AVemotionSurvey-CAMERA.pdf (2009)
BASE
Show details
9
Social Signal Processing: Survey of an Emerging Domain
BASE
Show details
10
Social Signal Processing: Survey of an Emerging Domain
In: http://www.idiap.ch/~vincia/papers/sspsurvey.pdf (2008)
BASE
Show details
11
Social Signal Processing: State-of-the-art and future perspectives of an emerging domain
In: http://www.idiap.ch/~vincia/papers/bravetopic.pdf (2008)
BASE
Show details
12
Social Signal Processing: State-of-the-art and future perspectives of an emerging domain
In: http://www.doc.ic.ac.uk/~maja/ACM-MM-2008-VinciarelliEtAl-CAMERA.pdf (2008)
BASE
Show details
13
Social Signal Processing: Survey of an Emerging Domain
In: http://www.doc.ic.ac.uk/~maja/IVCJ-SSPsurvey-FINAL.pdf (2008)
BASE
Show details
14
Fusion of audio and visual cues for laughter detection
In: http://www.doc.ic.ac.uk/~maja/CIVR-2008-PetridisPantic-CAMERA.pdf (2008)
BASE
Show details
15
Facial Action Recognition for Facial Expression Analysis from Static Face Images
In: http://www.kbs.twi.tudelft.nl/People/Staff/M.Pantic/SMCB-2004-final.pdf (2004)
BASE
Show details
16
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
In: http://s.i-techonline.com/Book/Face-Recognition/ISBN978-3-902613-03-5-fr20.pdf
BASE
Show details
17
VISUAL-ONLY DISCRIMINATION BETWEEN NATIVE AND NON-NATIVE SPEECH
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/georgakisetal_visualonlynativevsnonnative.pdf
BASE
Show details
18
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
In: http://mplab.ucsd.edu/~marni/pubs/panticbartlett_2007.pdf
BASE
Show details
19
20 Machine Analysis of Facial Expressions 1. Human Face and Its Expression
In: http://www.doc.ic.ac.uk/~maja/PanticBartlett-Chapter-Proof2.pdf
BASE
Show details
20
Social Signal Processing: The Research Agenda
In: http://ibug.doc.ic.ac.uk/media/uploads/documents/looking@people-panticetal-revision-final.pdf
BASE
Show details

Page: 1 2

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
29
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern