Talking Heads: Facial Animation Kinematics-Based Synthesis of Realistic Talking Faces


References

1.
E. Vatikiotis-Bateson, K. G. Munhall, Y. Kasahara, F. Garcia, and H. Yehia, "Characterizing audiovisual information during speech," presented at Proceedings ICSLP 96, Philadelphia, Penn, 1996.

2.
E. Vatikiotis-Bateson and H. C. Yehia, "Unified model of audible-visible speech production," presented at EuroSpeech '97: 5th European Conference on Speech Communication and Technology, Rhodes, Greece, 22-25 September, 1997, 1997.

3.
D. Stork and M. Hennecke, "Speechreading by humans and machines," in NATO-ASI Series, Series F, Computers and Systems Sciences, vol. 150. Berlin: Springer-Verlag, 1996.

4.
K. G. Munhall and E. Vatikiotis-Bateson, "The moving face during speech communication," in Hearing by Eye, Part 2: The Psychology of Speechreading and audiovisual speech, R. Campbell, B. Dodd, and D. Burnham, Eds. London: Taylor & Francis - Psychology Press, 1998, 123-129.

5.
E. Vatikiotis-Bateson, I.-M. Eigsti, S. Yano, and K. Munhall, "Eye movement of perceivers during audiovisual speech perception," Perception & Psychophysics, in press.

6.
E. Vatikiotis-Bateson and H. C. Yehia, "Unifed model of vocal tract and orofacial motion during speech," Journal of the Acoustical Society of Japan, vol. 9-3, pp. 319-320, 1997.

7.
H. C. Yehia, P. E. Rubin, and E. Vatikiotis-Bateson, "Quantitative association of orofacial and vocal-tract shapes," presented at Auditory and Visual Speech Processing Workshop 1997, Rhodes, Greece, 26-27 September, 1997, 1997.

8.
C. Bregler, M. Covell, and M. Slaney, "Video rewrite: Visual speech synthesis from video," presented at Auditory Visual Speech Processing '97, Rhodes, Greece, 25-26 September, 1997, 1997.

9.
J. C. Lucero, K. G. Munhall, E. Vatikiotis-Bateson, and V. L. Gracco, "Muscle-based modeling of facial dynamics during speech," Journal of the Acoustical Society of America, vol. 101, pp. 3175, 1997.

10.
Y. Wada, Y. Koike, E. Vatikiotis-Bateson, and M. Kawato, "A computational theory for movement pattern recognition based on optimal movement pattern generation," Biological Cybernetics, vol. 73, pp. 15-25, 1995.

11.
E. Vatikiotis-Bateson and H. Yehia, "Physiological modeling of facial motion during speech," Trans. Tech. Com. Psycho. Physio. Acoust., vol. H-96-65, pp. 1-8, 1996.

12.
E. Vatikiotis-Bateson, K. G. Munhall, M. Hirayama, Y. Kasahara, and H. Yehia, "Physiology-based synthesis of audiovisual speech," presented at 4th Speech Production Seminar: Models and Data, Autrans, France, 1996.

13.
B. K. P. Horn, "Closed-form solution of absolute orientation using unit quarternions," Journal of the Optical Society of America, vol. 4, pp. 629-642, 1987.

14.
E. Vatikiotis-Bateson and D. J. Ostry, "An analysis of the dimensionality of jaw motion in speech," Journal of Phonetics, vol. 23, pp. 101-117, 1995.

15.
T. Beier and S. Neely, "Feature-based image metamorphoshis," Computer Graphics, vol. 26, pp. 35-42, 1992.

16.
M. Vitkovich and P. Barber, " Effects of video frame rate on subjectsÕ ability to shadow one of two cometing verbal passages," Journal of Speech and Hearing Research, vol. 37, pp. 1204-1210., 1994.

17.
Y. Wada and M. Kawato, "A theory for cursive handwriting based on the minimization principle.," Biological Cybernetics, vol. 73, pp. 3-13, 1995.

18.
T. Flash and N. Hogan, "The Coordination of Arm Movements: An Experimentally Confirmed Mathematical Model," Journal of Neuroscience, vol. 5, pp. 1688-1703, 1985.

19.
E. Vatikiotis-Bateson, M. K. Tiede, Y. Wada, V. Gracco, and M. Kawato, "Phoneme extraction using via point estimation of real speech," presented at The 1994 International Conference on Spoken Language Processing (ICSLP-94), Yokohama, Japan, 1994.

20.
H. C. Yehia, P. E. Rubin, and E. Vatikiotis-Bateson, "Quantitative association of acoustic, facial, and vocal-tract shapes," Speech Communication, submitted.

21.
M. Kawato, "Motor theory of speech perception revisted from the minimum torque-change neural net-work model," presented at 8th Symposium on Future Electron Devices, Tokyo, Japan, 1989.

22.
E. Vatikiotis-Bateson, M. Hirayama, K. Honda, and M. Kawato, "The articulatory dynamics of running speech: Gestures from phonemes?," presented at The International Conference on Spoken Language Processing 1992, Banff, Canada, 1992.

23.
M. Hirayama, E. Vatikiotis-Bateson, and M. Kawato, "Physiologically based speech synthesis using neural networks," IEICE Transactions, vol. E76-A, pp. 1898-1910, 1993.

24.
F. I. Parke, "A Parametric Model for Human Faces," Salt Lake City, UT: University of Utah, 1974.

25.
M. Cohen and D. Massaro, "Synthesis of visible speech," Behavior Research Methods: Instruments & Computers, vol. 22, pp. 260-263, 1990.

26.
C. Benoît, T. Lallouache, T. Mohamadi, and C. Abry, "A set of French visemes for visual speech synthesis," in Talking machines: Theories, models, and designs, G. Bailly and C. Benoît, Eds. Amsterdam: North Holland, 1992, pp. 485-504.

27.
T. Guiard-Marigny, A. Adjoudani, and C. Benoît, "A 3-D model of the lips for visual speech synthesis," presented at Proceedings of the Second ESCA/IEEE Workshop on Speech Synthesis, New Paltz, NY, 1994.

28.
K. Waters, "A muscle model for animating three-dimensional facial expression," Computer Graphics, vol. 22, pp. 17-24, 1987.

29.
D. Terzopoulos and K. Waters, "Physically-based facial modeling, analysis, and animation," Visualization and Computer Animation, vol. 1, pp. 73-80, 1990.

30.
Y. Lee, D. Terzopoulos, and K. Waters, "Realistic modeling for facial animation," Computer Graphics, vol. 29, pp. 55-62, 1995.

31.
K. Waters and D. Terzopoulos, "Modeling and animating faces using scanned data," Visualization and Computer Animation, vol. 2, 123-128, 1991.

32.
K. Waters, "A physical model of facial tissue and muscle articulation derived from computer tomography data," presented at Visualization in Biomedical Computing, 1992.

33.
S. Morishima, H. Sera, and D. Terzopoulos, "Lips shape control with physics based muscle model," presented at Nicograph, 1996.

34.
D. W. Massaro, Speech perception by ear and by eye: A paradigm for psychological enquiry. Hillsdale, NJ: Lawrence Erlbaum Associates, 1987.

35.
D. W. Massaro, M. Tsuzaki, M. M. Cohen, A. Gesi, and R. Heridia, "Bimodal speech perception: An examination across languages.," Journal of Phonetics, vol. 21, pp. 445-478, 1993.

36.
B. LeGoff, T. Guiard-Marigny, and C. Benoît, "Analysis-synthesis and intelligibility of a talking face," in Progress in speech synthesis, J. P. H. v. Santen, R. W. Sproat, J. P. Olive, and J. Hirschberg, Eds. New York: Springer-Verlag, 1996, pp. 235-246.

37.
W. H. Sumby and I. Pollack, "Visual contribution to speech intelligibility in noise," Journal of the Acoustical Society of America, vol. 26, pp. 212-215, 1954.

 

Introduction | Audible-Visible Synthesis | Extensions to the Basic Model
Issues of Realism | Conclusion | References

 

PREVIOUS CONTENTS NEXT