- 1.J.E. Ball and D.T. Ling. Spoken language processing in the Persona conversational assistant. In Proc. ESCA Workshop on Spoken Dialogue Systems, 1995.Google Scholar
- 2.L. Baum. An inequality and associated maximization technique in statistical estimation of probabilistic functions of Markov processes. Inequalities, 3:1-8, 1972.Google Scholar
- 3.C. Benoit, C. Abry, M.-A. Cathiard, T. Guiard-Marigny, and T. Lallouache. Read my lips: Where? How? When? And so.. What? In 8th Int. Congress on Event Perception and Action, Marseille, France, July 1995. Springer-Verlag.Google Scholar
- 4.M. Brand. Structure discovery in conditional probability models via an entropic prior and parameter extinction. Neural Computation (accepted 8/98), October 1997. Google ScholarDigital Library
- 5.M. Brand. Pattern discovery via entropy minimization. In Proc. Artificial Intelligence and Statistics #7, Morgan Kaufmann Publishers. January 1999.Google Scholar
- 6.M. Brand. Shadow puppetry. Submitted to Int. Conf. on Computer Vision, ICCV ' 99, 1999. Google ScholarDigital Library
- 7.C. Bregler, M. Covell, and M. Slaney. Video Rewrite: Driving visual speech with audio. In Proc. ACM SIGGRAPH '97, 1997. Google ScholarDigital Library
- 8.T. Chen and R. Rao. Audio-visual interaction in nultimedia communication. In Proc. ICASSP ' 97, 1997. Google ScholarDigital Library
- 9.M.M. Cohen and D.W. Massaro. Modeling coarticulation in synthetic visual speech. In N.M. Thalmann and D. Thalmann, editors, Models and Techniques in Computer Animation. Springer-Verlag, 1993.Google ScholarCross Ref
- 10.S. Curinga, F. Lavagetto, and F. Vignoli. Lip movement sythesis using time delay neural networks. In Proc. EUSIPCO ' 96, 1996.Google Scholar
- 11.E Ekman and W.V. Friesen. Manual for the Facial Action Coding System. Consulting Psychologists Press, Inc., Palo Alto, CA, 1978.Google Scholar
- 12.T. Ezzat and T. Poggio. MikeTalk: A talking facial display based on morphing visemes. In Proc. Computer Animation Conference, June 1998. Google ScholarDigital Library
- 13.G.D. Forney. The Viterbi algorithm. Proc. IEEE, 6:268-278, 1973.Google ScholarCross Ref
- 14.G.H. Golub and C.F. van Loan. Matrix Computations. Johns Hopkins, 1996. 3rd edition.Google Scholar
- 15.G. Hager and K. Toyama. The XVision system: A generalpurpose substrate for portable real-time vision applications. Computer Vision and Image Understanding, 69(1) pp. 23-37. 1997. Google ScholarDigital Library
- 16.H. Hermansky and N. Morgan. RASTA processing of speech. IEEE Transactions on Speech and Audio Processing, 2(4):578-589, October 1994.Google ScholarCross Ref
- 17.I. Katunobu and O. Hasegawa. An active multimodal interaction system. In Proc. ESCA Workshop on Spoken Dialogue Systems, 1995.Google Scholar
- 18.J.E Lewis. Automated lip-sync: Background and techniques. J. Visualization and Computer Animation, 2:118-122, 1991.Google ScholarCross Ref
- 19.D.F. McAllister, R.D. Rodman, and D.L. Bitzer. Speaker independence in lip synchronization. In Proc. CompuGraphics ' 97, December 1997.Google Scholar
- 20.H. McGurk and J. MacDonald. Hearing lips and seeing voices. Nature, 264:746-748, 1976.Google ScholarCross Ref
- 21.K. Stevens (MIT). Personal communication., 1998.Google Scholar
- 22.S. Morishima and H. Harashima. A media conversion from speech to facial image for intelligent man-machine interface. IEEE J. Selected Areas in Communications, 4:594-599, 1991.Google ScholarDigital Library
- 23.F.I. Parke. A parametric model for human faces. Technical Report UTEC-CSc-75-047, University of Utah, 1974.Google Scholar
- 24.F.I. Parke. A model for human faces that allows speech synchronized animation. J. Computers and Graphics, 1(1): 1- 4, 1975.Google Scholar
- 25.M. Rydfalk. CANDIDE, a parameterised face. Technical Report LiTH-ISY-I-0866, Department of Electrical Engineering, Link/Sping University, Sweden, October 1987. Java demo available at http://www.bk.isy.liu.se/candide/candemo.html.Google Scholar
- 26.L.K. Saul and M.I. Jordan. A variational principle for model-based interpolation. Technical report, MIT Center for Biological and Computational Learning, 1996.Google Scholar
- 27.E.F. Walther. Lipreading. Nelson-Hall Inc., Chicago, 1982.Google Scholar
- 28.K. Waters and T. Levergood. DECface: A system for synthetic face applications. Multimedia Tools and Applications, 1:349- 366, 1995.Google ScholarCross Ref
- 29.E. Yamamoto, S. Nakamura, and K. Shikano. Lip movement synthesis from speech based on hidden Markov models. In Proc. Int. Conf. on automatic face and gesture recognition, FG '98, pages 154-159, Nara, Japan, 1998. IEEE Computer Society. Google ScholarDigital Library
Index Terms
- Voice puppetry
Recommendations
Speech driven facial animation
PUI '01: Proceedings of the 2001 workshop on Perceptive user interfacesThe results reported in this article are an integral part of a larger project aimed at achieving perceptually realistic animations, including the individualized nuances, of three-dimensional human faces driven by speech. The audiovisual system that has ...
3D puppetry: a kinect-based interface for 3D animation
UIST '12: Proceedings of the 25th annual ACM symposium on User interface software and technologyWe present a system for producing 3D animations using physical objects (i.e., puppets) as input. Puppeteers can load 3D models of familiar rigid objects, including toys, into our system and use them as puppets for an animation. During a performance, the ...
Physically-based forehead animation including wrinkles
Physically-based animation techniques enable more realistic and accurate animation to be created. We present a fully physically-based approach for efficiently producing realistic-looking animations of facial movement, including animation of expressive ...
Comments