نتایج جستجو برای: speech acoustics

تعداد نتایج: 125685  

2013
P. Heribanová J. Polec S. Ondrušová

This paper discusses the cued speech recognition methods in videoconference. Cued speech is a specific gesture language that is used for communication between deaf people. We define the criteria for sentence intelligibility according to answers of testing subjects (deaf people). In our tests we use 30 sample videos coded by H.264 codec with various bit-rates and various speed of cued speech. Ad...

2015
Kevin D. Roon Mark K. Tiede Katherine Dawson Douglas H. Whalen

Studies on the relationship between eyebrow movement and other aspects of speech production have focused on large, discrete movements of the eyebrows. Using integrated optical and electromagnetic point tracking, we measured eyebrow movements relative to the skull with a high level of precision. These data in combination with a correlational analysis method that accommodates varying phasing betw...

Journal: :Computer Speech & Language 2016
Yan Tang Martin Cooke Cassia Valentini-Botinhao

Evaluating the predictions of objective intelligibility metrics for modified and synthetic speech Yan Tang a,b,∗, Martin Cooke c,a, Cassia Valentini-Botinhao d a Language and Speech Laboratory, Universidad del País Vasco, Vitoria, Spain b Acoustics Research Centre, University of Salford, UK c Ikerbasque (Basque Science Foundation), Bilbao, Spain d Centre for Speech Technology Research, Universi...

Journal: :The Journal of the Acoustical Society of America 2012
Takayuki Arai

Several vocal-tract models were reviewed, with special focus given to the sliding vocal-tract model [T. Arai, Acoust. Sci. Technol. 27(6), 384-388 (2006)]. All of the models have been shown to be excellent tools for teaching acoustics and speech science to elementary through university level students. The sliding three-tube model is based on Fant's three-tube model [G. Fant, Acoustic Theory of ...

2015
Laura Colantoni Jeffrey Steele Paola Escudero

acoustic analysis consonants. See entries ‘acoustics’ under approximants/fricatives/laterals/nasals/ rhotics/stops measurement. See under Praat preparing data for analysis, 123–124 spectrograms and waveforms, 136, 137 vowels. See vowels: acoustics allophones acquiring allophonic distributions, 202, 213 acquiring TL phonemes that are L1 allophones, 148, 217, 241–244 learning a new allophonic dis...

Journal: :Speech Communication 2005
Björn Granström David House

Prosody in a single speaking style – often read speech – has been studied extensively in acoustic speech. During the past few years we have expanded our interest in two directions: 1.) Prosody in expressive speech communication and 2.) Prosody as an audiovisual expression. Understanding the interactions between visual expressions (primarily in the face) and the acoustics of the corresponding sp...

Journal: :Acoustics 2023

Objective speech intelligibility estimations undertaken in natural acoustics communications (NAS) scenarios require the utilization of a source that approximates acoustic characteristics human talker. Only limited number special sources conform to specifications relevant guidelines are available market; however, they can be deemed expensive by professional practitioners and other users. Non-spe...

2000
Todd A. Stephenson Hervé Bourlard Samy Bengio Andrew C. Morris

Current technology for automatic speech recognition (ASR) uses hidden Markov models (HMMs) that recognize spoken speech using the acoustic signal. However, no use is made of the causes of the acoustic signal: the articulators. We present here a dynamic Bayesian network (DBN) model that utilizes an additional variable for representing the state of the articulators. A particular strength of the s...

2013
Claudia Canevari Leonardo Badino Alessandro D'Ausilio Luciano Fadiga Giorgio Metta

Classical models of speech consider an antero-posterior distinction between perceptive and productive functions. However, the selective alteration of neural activity in speech motor centers, via transcranial magnetic stimulation, was shown to affect speech discrimination. On the automatic speech recognition (ASR) side, the recognition systems have classically relied solely on acoustic data, ach...

Speech Emotion Recognition (SER) is a new and challenging research area with a wide range of applications in man-machine interactions. The aim of a SER system is to recognize human emotion by analyzing the acoustics of speech sound. In this study, we propose Spectral Pattern features (SPs) and Harmonic Energy features (HEs) for emotion recognition. These features extracted from the spectrogram ...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید