DE eng

Search in the Catalogues and Directories

Page: 1 2
Hits 1 – 20 of 32

1
Modeling phones, keywords, topics and intents in spoken languages
Chen, Wenda. - 2021
BASE
Show details
2
A sensorimotor basis of speech communication
Bryan, Jacob. - 2019
BASE
Show details
3
Polychronization as a mechanism for language acquisition in spiking neural networks
Wang, Felix Y. - 2018
BASE
Show details
4
Training iCub robot pitch detection with recurrent neural network and LSTM
Tang, Steven. - 2018
BASE
Show details
5
Unsupervised learning of vocal tract sensory-motor synergies
BASE
Show details
6
iCub Tries to Play the Keyboard
Chang, Peixin. - 2017
BASE
Show details
7
Optimal nonlinear control and estimation using global domain linearization
BASE
Show details
8
On the effects of masking of perceptual cues in hearing-impaired ears
BASE
Show details
9
Minimum-error, energy-constrained source coding by sensory neurons
BASE
Show details
10
Autoregressive hidden Markov models and the speech signal
Bryan, Jacob. - 2014
BASE
Show details
11
Robots as language users: a computational model for pragmatic word learning
Niehaus, Logan. - 2014
BASE
Show details
12
Techniques for understanding hearing-impaired perception of consonant cues
Trevino, Andrea. - 2013
Abstract: We examine the cues used for consonant perception and the systematic behavior of normal and hearing-impaired listeners. All stimuli were presented as isolated consonant-vowel tokens, using the vowel /A/. Use of low-context stimuli, such as consonants, aids in minimizing the influence of some variable cognitive abilities (e.g., use of context, memory) across listeners, and focuses on differences in the processing or interpretation of the existing acoustic consonant cues. In a previous study on stop consonants, the 3D Deep Search (3DDS) method for the exploration of the necessary and sufficient cues for normal-hearing speech perception was introduced. Here, this method is used to isolate and analyze the perceptual cues of the naturally produced American English fricatives /S, Z, s, z, f, v, T, D/ in time, frequency, and intensity. The 3DDS analysis labels the perceptual cues of sibilant fricatives /Sa, Za, sa, za/ as a sustained frication noise preceding the vowel onset, with the acoustic cue for both /sa, za/ located between 3.8–7 kHz, and the acoustic cue for both /Sa, Za/ located between 2–4 kHz. The /Sa, Za/ utterances were also found to contain frication components above 4 kHz in natural speech that are unnecessary for correct perception, but can cause listeners to correspondingly hear /sa, za/ when the dominant cue between 2–4 kHz is removed by filtering; such cues are denoted “conflicting cues”. While unvoiced fricatives were observed to generally have a longer frication period than their voiced counterparts, duration of frication was found to be an unreliable cue for the differentiation of voiced from unvoiced fricatives. The wideband amplitude-modulation of the F2 and F3 formants at the pitch frequency F0 was found to be a defining cue for voicing. Similar to previous results with stop consonants, the robustness of fricative consonants to noise was found to be significantly correlated to the intensity of the acoustic cues that were isolated with the 3DDS method. The consonant recognition of 17 ears with sensorineural hearing loss is evaluated for fourteen consonants /p, t, k, f, s, S, b, d, g, v, z, Z, m, n/+/A/, under four speech-weighted noise conditions (0, 6, 12 [dB] SNR, quiet). For a single listener, we find that high errors can exist for a small subset of test stimuli, while performance for the majority of test stimuli can remain at ceiling. We show that hearing-impaired perception can vary across multiple tokens of the same consonant, in both noise-robustness and confusion groups. Within-consonant differences in noise-robustness are related to natural variations in intensity of the consonant cue region. Within-consonant differences in confusion groups entail that an average over multiple tokens of the same consonant results in a larger confusion group than for a single consonant token, causing the listener to appear to behave in a less systematic way. At the token level, hearing-impaired listeners are relatively consistent in their low-noise confusions; confusion groups are restricted to fewer than three confusions, on average. For each consonant token, the same confusion group is consistently observed across a population of hearing-impaired listeners. Quantifying these token differences provides insight into hearing-impaired perception of speech under noisy conditions and characterizes each listener’s hearing impairment. Auditory training programs are currently being explored as a method of improving hearing-impaired speech perception; precise knowledge of a patient’s individual differences in speech perception allows for a more accurately prescribed training program. Re-mapping or variations in the weighting of acoustic cues, due to auditory plasticity, can be examined with the detailed confusion analyses that we have developed. Although the tested tokens are noise-robust and unambiguous for normal-hearing listeners, the subtle natural variations in signal properties can lead to systematic within- consonant differences for hearing-impaired listeners. At the individual token level, a k-means clustering analysis of the confusion data shows that hearing- impaired listeners fall into similar confusion-based groups. Many of the token-dependent confusions that define these groups can also be observed for normal-hearing listeners, under higher noise levels or filtering conditions. These hearing-impaired listener groups correspond to different acoustic-cue weighting schemes, highlighting where auditory training should be most effective.
Keyword: 3D Deep Search (3dds); Aigram; Auditory Training; Confusion Matrix; Consonant; Hearing; Hearing Aids; Hearing Impaired; k means; Normal Hearing; Perception; Speech
URL: http://hdl.handle.net/2142/46591
BASE
Hide details
13
Semi-supervised learning for acoustic and prosodic modeling in speech applications
Huang, Jui Ting. - 2012
BASE
Show details
14
Acoustic model adaptation for recognition of dysarthric speech
Sharma, Harsh. - 2012
BASE
Show details
15
Computational differences between whispered and non-whispered speech
Lim, Boon Pang. - 2011
BASE
Show details
16
Autonomous learning of action-word semantics in a humanoid robot
Niehaus, Logan. - 2011
BASE
Show details
17
Statistical Model Based Multi-Microphone Speech Processing: Toward Overcoming Mismatch Problem
Kim, Lae-Hoon. - 2010
BASE
Show details
18
Estimation problems in speech and natural language
Bhat, Suma P.. - 2010
BASE
Show details
19
Extraction of pragmatic and semantic salience from spontaneous spoken English
In: Speech communication. - Amsterdam [u.a.] : Elsevier 48 (2006) 3-4, 437-462
BLLDB
OLC Linguistik
Show details
20
Cognitive state classification in a spoken tutorial dialogue system
In: Speech communication. - Amsterdam [u.a.] : Elsevier 48 (2006) 6, 616-632
BLLDB
Show details

Page: 1 2

Catalogues
2
0
1
0
0
0
0
Bibliographies
7
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
25
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern