1 |
Using deep neural networks to estimate tongue movements from speech face motion
|
|
|
|
BASE
|
|
Show details
|
|
2 |
Using deep neural networks to estimate tongue movements from speech face motion
|
|
|
|
BASE
|
|
Show details
|
|
5 |
Consonantal timing and release burst acoustics distinguish multiple coronal stop place distinctions in Wubuy (Australia)
|
|
|
|
BASE
|
|
Show details
|
|
6 |
Message vs. messenger effects on cross-modal matching for spoken phrases
|
|
|
|
BASE
|
|
Show details
|
|
7 |
Discrimination of Multiple Coronal Stop Contrasts in Wubuy (Australia): A Natural Referent Consonant Account
|
|
|
|
BASE
|
|
Show details
|
|
8 |
Discrimination of Multiple Coronal Stop Contrasts in Wubuy (Australia): A Natural Referent Consonant Account
|
|
|
|
BASE
|
|
Show details
|
|
9 |
Message vs. messenger effects on cross-modal matching for spoken phrases
|
|
|
|
Abstract:
A core issue in speech perception and word recognition research is the nature of information perceivers use to identify spoken utterances across indexical variations in their phonetic details, such as talker and accent differences. Separately, a crucial question in audiovisual research is the nature of information perceivers use to recognize phonetic congruency between the audio and visual (talking face) signals that arise from speaking. We combined these issues in a study examining how differences between connected speech utterances (messages) versus between talkers and accents (messenger characteristics) contribute to recognition of cross-modal articulatory congruence between audio-only (AO) and video-only (VO) components of spoken utterances. Participants heard AO phrases in their native regional English accent or another English accent, and then saw two synchronous VO displays of point-light talking faces from which they had to select the one that corresponded to the audio target. The incorrect video in each pair was either the same or a different phrase as the audio target, produced by the same or a different talker, who spoke in either the same or a different English accent. Results indicate that cross-modal articulatory correspondence is more accurately and quickly detected for message content than for messenger details, suggesting that recognising the linguistic message is more fundamental than messenger features is to cross-modal detection of audiovisual articulatory congruency. Nonetheless, messenger characteristics , especially accent, affected performance to some degree, analogous to recent findings in AO speech research.
|
|
Keyword:
200408 - Linguistic Structures (incl. Grammar; 970120 - Expanding Knowledge in Languages; articulation; Communication and Culture; Lexicon; Phonology; Semantics); speech perception
|
|
URL: http://handle.uws.edu.au:8081/1959.7/uws:32876 http://faavsp2015.ftw .at/
|
|
BASE
|
|
Hide details
|
|
10 |
Discrimination of multiple coronal stop contrasts in Wubuy (Australia) : a natural referent consonant account
|
|
|
|
BASE
|
|
Show details
|
|
11 |
Discrimination of multiple coronal stop contrasts in Wubuy (Australia): a natural referent consonant account
|
|
|
|
BASE
|
|
Show details
|
|
12 |
Singing emotionally: a study of pre-production, production, and post-production facial expressions
|
|
|
|
BASE
|
|
Show details
|
|
13 |
Singing emotionally : a study of pre-production, production, and post-production facial expressions
|
|
|
|
BASE
|
|
Show details
|
|
14 |
Articulatory basis of the apical/laminal distinction : tongue tip/body coordination in the Wubuy 4-way coronal stop contrast
|
|
|
|
BASE
|
|
Show details
|
|
15 |
Exploring nonlinear relationships between speech face motion and tongue movements using Mutual Information
|
|
|
|
BASE
|
|
Show details
|
|
20 |
Now you see it, now you don’t - frequency distribution of articulatory information reflected in speech face motion
|
|
|
|
BASE
|
|
Show details
|
|
|
|