Cue Integration in Categorical Tasks: Insights from Audio-Visual Speech Perception
نویسندگان
چکیده
Previous cue integration studies have examined continuous perceptual dimensions (e.g., size) and have shown that human cue integration is well described by a normative model in which cues are weighted in proportion to their sensory reliability, as estimated from single-cue performance. However, this normative model may not be applicable to categorical perceptual dimensions (e.g., phonemes). In tasks defined over categorical perceptual dimensions, optimal cue weights should depend not only on the sensory variance affecting the perception of each cue but also on the environmental variance inherent in each task-relevant category. Here, we present a computational and experimental investigation of cue integration in a categorical audio-visual (articulatory) speech perception task. Our results show that human performance during audio-visual phonemic labeling is qualitatively consistent with the behavior of a Bayes-optimal observer. Specifically, we show that the participants in our task are sensitive, on a trial-by-trial basis, to the sensory uncertainty associated with the auditory and visual cues, during phonemic categorization. In addition, we show that while sensory uncertainty is a significant factor in determining cue weights, it is not the only one and participants' performance is consistent with an optimal model in which environmental, within category variability also plays a role in determining cue weights. Furthermore, we show that in our task, the sensory variability affecting the visual modality during cue-combination is not well estimated from single-cue performance, but can be estimated from multi-cue performance. The findings and computational principles described here represent a principled first step towards characterizing the mechanisms underlying human cue integration in categorical tasks.
منابع مشابه
Auditory-visual speech perception in normal-hearing and cochlear-implant listeners.
The present study evaluated auditory-visual speech perception in cochlear-implant users as well as normal-hearing and simulated-implant controls to delineate relative contributions of sensory experience and cues. Auditory-only, visual-only, or auditory-visual speech perception was examined in the context of categorical perception, in which an animated face mouthing ba, da, or ga was paired with...
متن کاملImpact of cued speech on audio-visual speech integration in deaf and hearing adults
For hearing and deaf people, speech perception involves an integrative process between auditory and lip read information. In order to disambiguate information from lips, manual cue may be added (Cued Speech). We examined how audio-visual integration is affected by the presence of manual cues. To address this issue, we designed an original experiment using audio-visual McGurk stimuli produced wi...
متن کاملRelationships between Categorical Perception of Phonemes, Phoneme Awareness, and Visual Attention Span in Developmental Dyslexia
We tested the hypothesis that the categorical perception deficit of speech sounds in developmental dyslexia is related to phoneme awareness skills, whereas a visual attention (VA) span deficit constitutes an independent deficit. Phoneme awareness tasks, VA span tasks and categorical perception tasks of phoneme identification and discrimination using a d/t voicing continuum were administered to ...
متن کاملThe Effect of Motor Dependent/Independent Visual Perception Training on Visual-Motor Integration and Fine Motor Skills of 7-8-year-old Children: The Retest of Movement Hypothesis
The purpose of this study was to examine the effect of motor dependent/independent visual perception training on visual-motor integration and fine motor skills of 7-8 year old children .For this purpose, 107, 1st grade primary school students in Sabzevar were selected through purposive sampling (with equal economical and cultural status , optimum mental and physical health and full sight with...
متن کاملPerception of Synthesized Audible and Visible Speech
The research reported in this paper uses novel stimuli to study how speech perception is influenced by information presented to ear and eye. Auditory and visual sources ofinformation (syllables) were synthesized and presented in isolation or in factorial combination. A five-step contilllium between the syllables /bal and Idalwas synthesized along both auditory and visual dimensions, by varying ...
متن کامل