TY - GEN
T1 - Model-based lip synchronization with automatically translated synthetic voice toward a multi-modal translation system
AU - Ogata, Shin
AU - Murai, Kazumasa
AU - Nakamura, Satoshi
AU - Morishima, Shigeo
PY - 2001/1/1
Y1 - 2001/1/1
N2 - In this paper, we introduce a multi-modal English-to-Japanese and Japanese-to-English translation system that also translates the speaker's speech motion while synchronizing it to the translated speech. To retain the speaker's facial expression, we substitute only the speech organ's image with the synthesized one, which is made by a three-dimensional wire-frame model that is adaptable to any speaker. Our approach enables image synthesis and translation with an extremely small database.
AB - In this paper, we introduce a multi-modal English-to-Japanese and Japanese-to-English translation system that also translates the speaker's speech motion while synchronizing it to the translated speech. To retain the speaker's facial expression, we substitute only the speech organ's image with the synthesized one, which is made by a three-dimensional wire-frame model that is adaptable to any speaker. Our approach enables image synthesis and translation with an extremely small database.
UR - http://www.scopus.com/inward/record.url?scp=70350254571&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=70350254571&partnerID=8YFLogxK
U2 - 10.1109/ICME.2001.1237647
DO - 10.1109/ICME.2001.1237647
M3 - Conference contribution
AN - SCOPUS:70350254571
T3 - Proceedings - IEEE International Conference on Multimedia and Expo
SP - 28
EP - 31
BT - Proceedings - IEEE International Conference on Multimedia and Expo
PB - IEEE Computer Society
T2 - 2001 IEEE International Conference on Multimedia and Expo, ICME 2001
Y2 - 22 August 2001 through 25 August 2001
ER -