@inbook{2c65953432554b2db23ed86519885ac6,
title = "Animated pronunciation generated from speech for pronunciation training",
abstract = "Computer-assisted pronunciation training (CAPT) was introduced for language education in recent years. CAPT scores the learner's pronunciation quality and points out wrong phonemes by using speech recognition technology. However, although the learner can thus realize that his/her speech is different from the teacher's, the learner still cannot control the articulation organs to pronounce correctly. The learner cannot understand how to correct the wrong articulatory gestures precisely. We indicate these differences by visualizing a learner's wrong pronunciation movements and the correct pronunciation movements with CG animation. We propose a system for generating animated pronunciation by estimating a learner's pronunciation movements from his/her speech automatically. The proposed system maps speech to coordinate values that are needed to generate the animations by using multi-layer neural networks (MLN). We use MRI data to generate smooth animated pronunciations. Additionally, we verify whether the vocal tract area and articulatory features are suitable as characteristics of pronunciation movement through experimental evaluation.",
keywords = "Animated Pronunciation, Articulatory Feature, Pronunciation Training, Vocal Tract Area",
author = "Yurie Iribe and Silasak Manosavan and Kouichi Katsurada and Tsuneo Nitta",
year = "2012",
doi = "10.1007/978-3-642-29934-6_8",
language = "English",
isbn = "9783642299339",
series = "Smart Innovation, Systems and Technologies",
pages = "73--82",
editor = "Jain Lakhmi and Howlett Robert and Watada Junzo and Watanabe Toyohide and Takahashi Naohisa",
booktitle = "Intelligent Interactive Multimedia",
}