DE eng

Search in the Catalogues and Directories

Hits 1 – 19 of 19

1
Mapping theoretical and methodological perspectives for understanding speech interface interactions ; CHI EA '19 Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems
Harte, Naomi. - 2019
BASE
Show details
2
Multimodal Continuous Turn-Taking Prediction Using Multiscale RNNs ...
BASE
Show details
3
Multimodal continuous turn-taking prediction using multiscale RNNs ; ICMI 2018 - 20th ACM International Conference on Multimodal Interaction
Harte, Naomi. - 2018
BASE
Show details
4
Survival at the museum: A cooperation experiment with emotionally expressive virtual characters ; ICMI '18 Proceedings of the 20th ACM International Conference on Multimodal Interaction
BASE
Show details
5
A longitudinal database of Irish political speech with annotations of speaker ability [<Journal>]
Cullen, Ailbhe [Verfasser]; Harte, Naomi [Sonstige]
DNB Subject Category Language
Show details
6
Towards predicting dialog acts from previous speakers' non-verbal cues ; BIBTEX 2017
Harte, Naomi. - 2017
Abstract: In studies of response times during conversational turn-taking, a modal time of 200 ms has been observed to be a universal value that exists across languages and cross-culturally. This 200 ms value is also seen as the limit of human response times to any stimulus (e.g the response time to a starting-gun in a race). It has also been shown that human language production is slow and can take up to 1500 ms to generate even a short clause. Due to these two observations, it is necessary for a person to start formulating their turns long before the end of their interlocutor?s turn. To do this we must predict elements of what a person will say in order to formulate our responses and sustain the flow of conversation. In this sense, the end of a person?s turn can be viewed as a trigger for a prepared response. This model of human language production informs incremental approaches to the design of dialog systems, where dialog options are evaluated incrementally, while the system processes user utterances. One way we can form our predictions is by reading the non-linguistic signals that are produced by our interlocutor. For example, prosodic information such as pitch inflection can be used to infer whether a question is being asked or a statement is being made. Pitch and intensity information can also be used to infer whether a backchannel is an appropriate response. These backchannel prediction models based on non-linguistic cues can be used by conversational agents to carry out more fluid interactions with users. The development of better prediction models that exploit the social signals that humans use will lead to agents that can reproduce the interaction behaviors of humans more effectively. In this analysis we look at non-verbal speaker signals that can be used to predict the appropriate dialogue act that will follow the speaker?s utterance. We define three categories of dialogue acts: (1) response (as in a response to a question),(2) statement (a general turn switch which does not include other dialog act types), and (3) backchannel (vocalizations encouraging the speaker to continues speaking). In addition we define a fourth category, no-response, which is not strictly a dialogue act but is a relevant category for agent interactions. We identify four types of non-verbal signals that can be used to predict the appropriate type of response dialogue act: inner eyebrow movement, outer eyebrow movement, blinks, and gaze. We analyze the behavior of these four signals in the vicinity of the dialogue acts.
Keyword: Predicting dialog acts
URL: http://people.tcd.ie/nharte
http://hdl.handle.net/2262/89214
BASE
Hide details
7
ViSQOL: an objective speech quality model
BASE
Show details
8
Building a Database of Political Speech Does Culture Matter in Charisma Annotations?
In: Conference papers (2014)
BASE
Show details
9
Speaker verification in score-ageing-quality classification space
In: Computer speech and language. - Amsterdam [u.a.] : Elsevier 27 (2013) 5, 1068-1084
OLC Linguistik
Show details
10
Speech intelligibility prediction using a Neurogram Similarity Index Measure
In: Speech communication. - Amsterdam [u.a.] : Elsevier 54 (2012) 2, 306-320
BLLDB
OLC Linguistik
Show details
11
Speech Intelligibility prediction using a Neurogram Similarity Index Measure
HINES, ANDREW; HARTE, NAOMI. - : Elsevier, 2012
BASE
Show details
12
Speech intelligibility from image processing
In: Speech communication. - Amsterdam [u.a.] : Elsevier 52 (2010) 9, 736-752
BLLDB
OLC Linguistik
Show details
13
Error Metrics for Impaired Auditory Nerve Responses of Different Phoneme Groups
In: Conference papers (2009)
BASE
Show details
14
Measurement of Phonemic Degradation in Sensorineural Hearing Loss using a Computational Model of the Auditory Periphery
In: Conference papers (2009)
BASE
Show details
15
Error Metrics for Impaired Auditory Nerve Responses of Different Phoneme Groups ; Interspeech 2009
BASE
Show details
16
Measurement of phonemic degradation in sensorineural hearing loss using a computational model of the auditory periphery ; IET Irish Signals and Systems Conference ISSC 2009
BASE
Show details
17
On Parsing Visual Sequences with the Hidden Markov Model
BASE
Show details
18
Discriminitive Multi-Resolution Sub-Band and Segmental Phonetic Model Combination
Harte, Naomi. - 2000
BASE
Show details
19
A Novel Model For Phoneme Recognition Using Phonetically Derived Features ...
BASE
Show details

Catalogues
0
0
3
0
1
0
0
Bibliographies
2
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
15
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern