DE eng

Search in the Catalogues and Directories

Hits 1 – 14 of 14

1
'Foret' comme doublet rare et archaïque de 'esset' : état de la question et pistes complémentaires
In: Nemo par eloquentia. Mélanges de linguistique ancienne en hommage à Colette Bodelot ; https://hal.archives-ouvertes.fr/hal-02426929 ; Nemo par eloquentia. Mélanges de linguistique ancienne en hommage à Colette Bodelot, Presses Universitaires Blaise-Pascal, pp.271-284, 2019 (2019)
BASE
Show details
2
Transforming an embodied conversational agent into an efficient talking head : from keyframe-based animation to multimodal concatenation synthesis
Gibert, Guillaume; Olsen, Kirk N; Leung, Yvonne. - : SpringerOpen, 2015
BASE
Show details
3
Transforming an embodied conversational agent into an efficient talking head : from keyframe-based animation to multimodal concatenation synthesis
Gibert, Guillaume (R14173); Olsen, Kirk N. (R17200); Leung, Yvonne ( S2751); Stevens, Catherine J. (R8645). - : Germany, SpringerOpen, 2015
Abstract: Background: Virtual humans have become part of our everyday life (movies, internet, and computer games). Even though they are becoming more and more realistic, their speech capabilities are, most of the time, limited and not coherent and/or not synchronous with the corresponding acoustic signal. Methods: We describe a method to convert a virtual human avatar (animated through key frames and interpolation) into a more naturalistic talking head. In fact, speech articulation cannot be accurately replicated using interpolation between key frames and talking heads with good speech capabilities are derived from real speech production data. Motion capture data are commonly used to provide accurate facial motion for visible speech articulators (jaw and lips) synchronous with acoustics. To access tongue trajectories (partially occluded speech articulator), electromagnetic articulography (EMA) is often used. We recorded a large database of phonetically-balanced English sentences with synchronous EMA, motion capture data, and acoustics. An articulatory model was computed on this database to recover missing data and to provide ‘normalized’ animation (i.e., articulatory) parameters. In addition, semi-automatic segmentation was performed on the acoustic stream. A dictionary of multimodal Australian English diphones was created. It is composed of the variation of the articulatory parameters between all the successive stable allophones. Results: The avatar’s facial key frames were converted into articulatory parameters steering its speech articulators (jaw, lips and tongue). The speech production database was used to drive the Embodied Conversational Agent (ECA) and to enhance its speech capabilities. A Text-To-Auditory Visual Speech synthesizer was created based on the MaryTTS software and on the diphone dictionary derived from the speech production database. Conclusions: We describe a method to transform an ECA with generic tongue model and animation by key frames into a talking head that displays naturalistic tongue, jaw and lip motions. Thanks to a multimodal speech production database, a Text-To-Auditory Visual Speech synthesizer drives the ECA’s facial movements enhancing its speech capabilities.
Keyword: 080602 - Computer-Human Interaction; 970117 - Expanding Knowledge in Psychology and Cognitive Sciences; facial animation; multimodal speech; speech synthesis; talking heads
URL: http://handle.uws.edu.au:8081/1959.7/uws:32915
https://doi.org/10.1186/s40469-015-0007-8
BASE
Hide details
4
Control of speech-related facial movements of an avatar from video
In: Speech communication. - Amsterdam [u.a.] : Elsevier 55 (2013) 1, 135-146
OLC Linguistik
Show details
5
On-Line Learning of Lexical Items and Grammatical Constructions via Speech, Gaze and Action-Based Human-Robot Interaction
In: INTERSPEECH 2013 - 14th Annual Conference of the International Speech Communication Association ; https://hal.inria.fr/hal-02561340 ; INTERSPEECH 2013 - 14th Annual Conference of the International Speech Communication Association, Aug 2013, Lyon, France ; https://www.isca-speech.org/archive/interspeech_2013/i13_2657.html (2013)
BASE
Show details
6
Control of speech-related facial movements of an avatar from video
Gibert, Guillaume (R14173); Leung, Yvonne (R16710); Stevens, Catherine J. (R8645). - : Netherlands, Elsevier, 2013
BASE
Show details
7
Evaluating a synthetic talking head using a dual task : modality effects on speech understanding and cognitive load
Stevens, Catherine J. (R8645); Gibert, Guillaume (R14173); Leung, Yvonne (R16710). - : U.K., Academic Press, 2013
BASE
Show details
8
Speech articulator movements recorded from facing talkers using two electromagnetic articulometer systems simultaneously
Tiede, Mark; Bundgaard-Nielsen, Rikke L. (R14172); Kroos, Christian (R11604). - : U.S.A., Acoustical Society of America, 2012
BASE
Show details
9
Multimodal speech animation from electromagnetic articulography data
Gibert, Guillaume (R14173); Attina, Virginie (R14534); Tiede, Mark. - : U.S., IEEE, 2012
BASE
Show details
10
Role of form and motion information in auditory-visual speech perception of McGurk combinations and fusions
Gibert, Guillaume (R14173); Fordyce, Andrew; Stevens, Catherine J.. - : Japan, AVSP, 2010
BASE
Show details
11
Prosody for the eyes: quantifying visual prosody using guided principal component analysis
Cvejic, Erin (S25037); Kim, Jeesun (R11607); Davis, Chris (R11605). - : France, International Speech Communication Association, 2010
BASE
Show details
12
Production of Mandarin lexical tones: auditory and visual components
Attina, Virginie (R14534); Gibert, Guillaume; Vatikiotis-Bateson, Eric. - : Japan, AVSP, 2010
BASE
Show details
13
Evaluation of a virtual speech cuer
In: Workshop on Experimental Linguistics ; https://hal.archives-ouvertes.fr/hal-00366488 ; Workshop on Experimental Linguistics, Aug 2006, Athens, Greece. pp.141-144 (2006)
BASE
Show details
14
Analysis and synthesis of the 3D movements of the head, face and hand of a speaker using cued speech
In: ISSN: 0001-4966 ; EISSN: 1520-8524 ; Journal of the Acoustical Society of America ; https://hal.archives-ouvertes.fr/hal-00143622 ; Journal of the Acoustical Society of America, Acoustical Society of America, 2005, 118 (2), pp.1144-1153 (2005)
BASE
Show details

Catalogues
0
0
1
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
13
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern