Generating Animated Pronunciation from Speech Through Articulatory Feature Extraction
نویسندگان
چکیده
We automatically generate CG animations to express the pronunciation movement of speech through articulatory feature (AF) extraction to help learn a pronunciation. The proposed system uses MRI data to map AFs to coordinate values that are needed to generate the animations. By using magnetic resonance imaging (MRI) data, we can observe the movements of the tongue, palate, and pharynx in detail while a person utters words. AFs and coordinate values are extracted by multi-layer neural networks (MLN). Specifically, the system displays animations of the pronunciation movements of both the learner and teacher from their speech in order to show in what way the learner’s pronunciation is wrong. Learners can thus understand their wrong pronunciation and the correct pronunciation method through specific animated pronunciations. Experiments to compare MRI data with the generated animations confirmed the accuracy of articulatory features. Additionally, we verified the effectiveness of using AF to generate animation.
منابع مشابه
Real-time Visualization of English Pronunciation on an IPA Chart Based on Articulatory Feature Extraction
In recent years, Computer Assisted Pronunciation Technology (CAPT) systems have been developed that can help Japanese learners to study foreign languages. We have been developing a pronunciation training system to evaluate and correct learner's pronunciation by extracting articulatory-features (AFs). In this paper, we propose a novel pronunciation training system that can plot the place and man...
متن کاملOptimized Feature Extraction and HMMs in Subword Detectors
This paper presents methods and results for optimizing subword detectors in continuous speech. Speech detectors are useful within areas like detection-based ASR, pronunciation training, phonetic analysis, word spotting, etc. We build detectors for both articulatory features and phones by discriminative training of detector-specific MFCC filterbanks and HMMs. The resulting filterbanks are clearl...
متن کاملToward a speaker-independent visual articulatory feedback system
Context Several studies tend to show that visual articulatory feedback is useful for phonetic correction, both for speech therapy and “Computer Aided Pronunciation Training” (CAPT) [1]. In [2], we proposed a visual articulatory feedback system based on a 3D talking head used in “an augmented speech scenario”, i.e. displaying all speech articulators including the tongue and velum. In the propose...
متن کاملAdaptive articulatory feature-based conditional pronunciation modeling for speaker verification
Because of the differences in education background, accents, and so on, different persons have different ways of pronunciation. Therefore, the pronunciation patterns of individuals can be used as features for discriminating speakers. This paper exploits the pronunciation characteristics of speakers and proposes a new conditional pronunciation modeling (CPM) technique for speaker verification. T...
متن کاملArticulatory feature-based pronunciation modeling
Spoken language, especially conversational speech, is characterized by great variability in word pronunciation, including many variants that differ grossly from dictionary prototypes. This is one factor in the poor performance of automatic speech recognizers on conversational speech, and it has been very difficult to mitigate in traditional phonebased approaches to speech recognition. An altern...
متن کامل