skip to main content
10.1145/311535.311537acmconferencesArticle/Chapter ViewAbstractPublication PagessiggraphConference Proceedingsconference-collections
Article
Free Access

Voice puppetry

Authors Info & Claims
Published:01 July 1999Publication History
First page image

References

  1. 1.J.E. Ball and D.T. Ling. Spoken language processing in the Persona conversational assistant. In Proc. ESCA Workshop on Spoken Dialogue Systems, 1995.Google ScholarGoogle Scholar
  2. 2.L. Baum. An inequality and associated maximization technique in statistical estimation of probabilistic functions of Markov processes. Inequalities, 3:1-8, 1972.Google ScholarGoogle Scholar
  3. 3.C. Benoit, C. Abry, M.-A. Cathiard, T. Guiard-Marigny, and T. Lallouache. Read my lips: Where? How? When? And so.. What? In 8th Int. Congress on Event Perception and Action, Marseille, France, July 1995. Springer-Verlag.Google ScholarGoogle Scholar
  4. 4.M. Brand. Structure discovery in conditional probability models via an entropic prior and parameter extinction. Neural Computation (accepted 8/98), October 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. 5.M. Brand. Pattern discovery via entropy minimization. In Proc. Artificial Intelligence and Statistics #7, Morgan Kaufmann Publishers. January 1999.Google ScholarGoogle Scholar
  6. 6.M. Brand. Shadow puppetry. Submitted to Int. Conf. on Computer Vision, ICCV ' 99, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. 7.C. Bregler, M. Covell, and M. Slaney. Video Rewrite: Driving visual speech with audio. In Proc. ACM SIGGRAPH '97, 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. 8.T. Chen and R. Rao. Audio-visual interaction in nultimedia communication. In Proc. ICASSP ' 97, 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. 9.M.M. Cohen and D.W. Massaro. Modeling coarticulation in synthetic visual speech. In N.M. Thalmann and D. Thalmann, editors, Models and Techniques in Computer Animation. Springer-Verlag, 1993.Google ScholarGoogle ScholarCross RefCross Ref
  10. 10.S. Curinga, F. Lavagetto, and F. Vignoli. Lip movement sythesis using time delay neural networks. In Proc. EUSIPCO ' 96, 1996.Google ScholarGoogle Scholar
  11. 11.E Ekman and W.V. Friesen. Manual for the Facial Action Coding System. Consulting Psychologists Press, Inc., Palo Alto, CA, 1978.Google ScholarGoogle Scholar
  12. 12.T. Ezzat and T. Poggio. MikeTalk: A talking facial display based on morphing visemes. In Proc. Computer Animation Conference, June 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. 13.G.D. Forney. The Viterbi algorithm. Proc. IEEE, 6:268-278, 1973.Google ScholarGoogle ScholarCross RefCross Ref
  14. 14.G.H. Golub and C.F. van Loan. Matrix Computations. Johns Hopkins, 1996. 3rd edition.Google ScholarGoogle Scholar
  15. 15.G. Hager and K. Toyama. The XVision system: A generalpurpose substrate for portable real-time vision applications. Computer Vision and Image Understanding, 69(1) pp. 23-37. 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. 16.H. Hermansky and N. Morgan. RASTA processing of speech. IEEE Transactions on Speech and Audio Processing, 2(4):578-589, October 1994.Google ScholarGoogle ScholarCross RefCross Ref
  17. 17.I. Katunobu and O. Hasegawa. An active multimodal interaction system. In Proc. ESCA Workshop on Spoken Dialogue Systems, 1995.Google ScholarGoogle Scholar
  18. 18.J.E Lewis. Automated lip-sync: Background and techniques. J. Visualization and Computer Animation, 2:118-122, 1991.Google ScholarGoogle ScholarCross RefCross Ref
  19. 19.D.F. McAllister, R.D. Rodman, and D.L. Bitzer. Speaker independence in lip synchronization. In Proc. CompuGraphics ' 97, December 1997.Google ScholarGoogle Scholar
  20. 20.H. McGurk and J. MacDonald. Hearing lips and seeing voices. Nature, 264:746-748, 1976.Google ScholarGoogle ScholarCross RefCross Ref
  21. 21.K. Stevens (MIT). Personal communication., 1998.Google ScholarGoogle Scholar
  22. 22.S. Morishima and H. Harashima. A media conversion from speech to facial image for intelligent man-machine interface. IEEE J. Selected Areas in Communications, 4:594-599, 1991.Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. 23.F.I. Parke. A parametric model for human faces. Technical Report UTEC-CSc-75-047, University of Utah, 1974.Google ScholarGoogle Scholar
  24. 24.F.I. Parke. A model for human faces that allows speech synchronized animation. J. Computers and Graphics, 1(1): 1- 4, 1975.Google ScholarGoogle Scholar
  25. 25.M. Rydfalk. CANDIDE, a parameterised face. Technical Report LiTH-ISY-I-0866, Department of Electrical Engineering, Link/Sping University, Sweden, October 1987. Java demo available at http://www.bk.isy.liu.se/candide/candemo.html.Google ScholarGoogle Scholar
  26. 26.L.K. Saul and M.I. Jordan. A variational principle for model-based interpolation. Technical report, MIT Center for Biological and Computational Learning, 1996.Google ScholarGoogle Scholar
  27. 27.E.F. Walther. Lipreading. Nelson-Hall Inc., Chicago, 1982.Google ScholarGoogle Scholar
  28. 28.K. Waters and T. Levergood. DECface: A system for synthetic face applications. Multimedia Tools and Applications, 1:349- 366, 1995.Google ScholarGoogle ScholarCross RefCross Ref
  29. 29.E. Yamamoto, S. Nakamura, and K. Shikano. Lip movement synthesis from speech based on hidden Markov models. In Proc. Int. Conf. on automatic face and gesture recognition, FG '98, pages 154-159, Nara, Japan, 1998. IEEE Computer Society. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Voice puppetry

                Recommendations

                Comments

                Login options

                Check if you have access through your login credentials or your institution to get full access on this article.

                Sign in

                PDF Format

                View or Download as a PDF file.

                PDF

                eReader

                View online with eReader.

                eReader