Lombard speech: Auditory (A), Visual (V) and AV effects
نویسندگان
چکیده
This study examined Auditory (A) and Visual (V) speech (speech-related head and face movement) as a function of noise environment. Measures of AV speech were recorded for 3 males and 1 female for 10 sentences spoken in quiet as well as four styles of background noise (Lombard speech). Auditory speech was analyzed in terms of overall intensity, duration, spectral tilt and prosodic parameters employing Fujisaki model based parameterizations of F0 contours. Visual speech was analyzed in terms of Principal Components (PC) of head and face movement. Compared to speech in quiet, Lombard speech was louder, of longer duration, had more energy at higher frequencies (particularly with babble speech) and had greater amplitude mean accent and phrase commands. Visual Lombard speech showed greater influence of the PCs associated with jaw and mouth movement, face expansion and contraction and head rotation (pitch). Lombard speech showed increased AV speech correlations between RMS speech intensity and the PCs that involved jaw and mouth movement. A similar increased correlation occurred for intensity and head rotation (pitch). For Lombard speech, all talkers showed an increased correlation between F0 and head translation (raising and lowering). Increased F0 correlations for other head movements were more idiosyncratic. These findings suggest that the relationships underlying Audio-Visual speech perception differ depending on how that speech was produced
منابع مشابه
Auditory and auditory-visual Lombard speech perception by younger and older adults
The current study examined older and younger adults’ perception of auditory and auditory-visual Lombard speech. A staircase procedure was used to estimate the SNR required for participants to achieve 50% correct auditory identification of Quiet and Lombard speech (CVC and VCV stimuli). Stimuli were then presented in auditory only (AO), visual only (VO) and auditory visual (AV) conditions in a s...
متن کاملElectrocorticography Reveals Enhanced Visual Cortex Responses to Visual Speech.
Human speech contains both auditory and visual components, processed by their respective sensory cortices. We test a simple model in which task-relevant speech information is enhanced during cortical processing. Visual speech is most important when the auditory component is uninformative. Therefore, the model predicts that visual cortex responses should be enhanced to visual-only (V) speech com...
متن کاملElectrophysiology of auditory-visual speech integration
Twenty-six native English Speakers identified auditory (A), visual (V), and congruent and incongruent auditory-visual (AV) syllables while undergoing electroencephalography (EEG) in three experiments. In Experiment 1, unimodal (A, V) and bimodal (AV) stimuli were presented in separate blocks. In Experiment 2, the same stimuli were pseudo-randomized in the same blocks, providing a replication of...
متن کاملThe Intelligibility of Lombard Speech: Communicative setting matters
Recently we reported that talkers modified their speech production strategies in noise as a function of whether their interlocutor could or could not be seen, i.e. face-to-face (FTF) or non-visual conditions (NV). Participants made greater auditory speech modifications (e.g. in terms of amplitude and F0) in NV condition, and greater visual speech modifications (in terms of inter-lip area) in FT...
متن کاملBimodal speech: early suppressive visual effects in human auditory cortex.
While everyone has experienced that seeing lip movements may improve speech perception, little is known about the neural mechanisms by which audiovisual speech information is combined. Event-related potentials (ERPs) were recorded while subjects performed an auditory recognition task among four different natural syllables randomly presented in the auditory (A), visual (V) or congruent bimodal (...
متن کامل