This paper presents a new technique for synthesizing visual speech from arbitrarily given text. The technique is based on an algorithm for parameter generation from HMM with dynamic features, which has been successfully applied to text-to-speech synthesis. In the training phase, syllable HMMs are trained with visual speech parameter sequences that represent lip movements. In the synthesis phase, a sentence HMM is constructed by concatenating syllable HMMs corresponding to the phonetic transcription for the input text. Then an optimum visual speech parameter sequence is generated from the sentence HMM in an ML sense. The proposed technique can generate synchronized lip movements with speech in a unified framework. Furthermore, coarticulation is implicitly incorporated into the generated mouth shapes. As a result, synthetic lip motion becomes smooth and realistic
%0 Conference Paper
%1 Masuko1998
%A Masuko, Takashi
%A Kobayashi, Takao
%A Tamura, Masatsune
%A Masubuchi, Jun
%A Tokuda, Keiichi
%B Proceedings of the 1998 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
%C Seattle, WA, USA
%D 1998
%K 2D HMM;syllable HMM;synchronized Markov algorithm;phonetic animation;ML;coarticulation;dynamic animation;feature contour extraction;hidden features;generated generation lip models;speech motion;text-to-visual mouth movements;parameter movements;synthesis parameter phase;synthetic phase;visual sequences;computer shapes;input speech synthesis;synchronisation; synthesis;training text;lip transcription;sentence
%P 3745-3748
%R 10.1109/ICASSP.1998.679698
%T Text-to-visual speech synthesis based on parameter generation from HMM
%V 6
%X This paper presents a new technique for synthesizing visual speech from arbitrarily given text. The technique is based on an algorithm for parameter generation from HMM with dynamic features, which has been successfully applied to text-to-speech synthesis. In the training phase, syllable HMMs are trained with visual speech parameter sequences that represent lip movements. In the synthesis phase, a sentence HMM is constructed by concatenating syllable HMMs corresponding to the phonetic transcription for the input text. Then an optimum visual speech parameter sequence is generated from the sentence HMM in an ML sense. The proposed technique can generate synchronized lip movements with speech in a unified framework. Furthermore, coarticulation is implicitly incorporated into the generated mouth shapes. As a result, synthetic lip motion becomes smooth and realistic
@inproceedings{Masuko1998,
abstract = {This paper presents a new technique for synthesizing visual speech from arbitrarily given text. The technique is based on an algorithm for parameter generation from HMM with dynamic features, which has been successfully applied to text-to-speech synthesis. In the training phase, syllable HMMs are trained with visual speech parameter sequences that represent lip movements. In the synthesis phase, a sentence HMM is constructed by concatenating syllable HMMs corresponding to the phonetic transcription for the input text. Then an optimum visual speech parameter sequence is generated from the sentence HMM in an ML sense. The proposed technique can generate synchronized lip movements with speech in a unified framework. Furthermore, coarticulation is implicitly incorporated into the generated mouth shapes. As a result, synthetic lip motion becomes smooth and realistic},
added-at = {2021-02-01T10:51:23.000+0100},
address = {Seattle, WA, USA},
author = {Masuko, Takashi and Kobayashi, Takao and Tamura, Masatsune and Masubuchi, Jun and Tokuda, Keiichi},
biburl = {https://www.bibsonomy.org/bibtex/26b8b72a259d2c88b39c1f8859e435801/m-toman},
booktitle = {Proceedings of the 1998 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)},
doi = {10.1109/ICASSP.1998.679698},
file = {:pdfs/masuko_icassp_1998.pdf:PDF},
interhash = {fa8aaf06187f1a734dd9f96e8af18c9f},
intrahash = {6b8b72a259d2c88b39c1f8859e435801},
issn = {1520-6149},
keywords = {2D HMM;syllable HMM;synchronized Markov algorithm;phonetic animation;ML;coarticulation;dynamic animation;feature contour extraction;hidden features;generated generation lip models;speech motion;text-to-visual mouth movements;parameter movements;synthesis parameter phase;synthetic phase;visual sequences;computer shapes;input speech synthesis;synchronisation; synthesis;training text;lip transcription;sentence},
month = may,
owner = {schabus},
pages = {3745-3748},
timestamp = {2021-02-01T10:51:23.000+0100},
title = {Text-to-visual speech synthesis based on parameter generation from HMM},
volume = 6,
year = 1998
}