作者: Wen-Kai Tai , Mau-Tsuen Yang , Guang-Yi Wang , Cheng-Chin Chiang
关键词: Artificial intelligence 、 Computer facial animation 、 Computer vision 、 Virtual reality 、 Face (geometry) 、 Mel-frequency cepstrum 、 Audio signal 、 Synchronization 、 Cepstrum 、 Computer science 、 Speech recognition 、 Hidden Markov model
摘要: In this paper, we utilized Hidden Markov Model (HMM) as a mapping mechanism between two different kinds of correlated signals. Specifically, developed voice-driven talking head system by exploiting the physical relationships shape mouth and sound that is produced. The proposed can be easily trained efficiently animated. training phase, Mel-scale Frequency Cepstral Coefficients (MFCC) were analyzed from audio signals Facial Animation Parameters (FAP) extracted video Then both features integrated to train single HMM. synthesis HMM was used correlate completely novel track FAP sequence for face with help Engine (FAE). experiments demonstrated effects on man woman, styles (speaking singing) using three languages (Chinese, English Taiwanese). possible applications are computer aided instruction, online guide, virtual conference, lip synchronization, human interaction so on.