Automatic Estimation of Emotional Parameters for An Intelligent Avatar

Full Text (PDF, 285KB), PP.59-66

Views: 0 Downloads: 0

Author(s)

Jing Yuan 1,* Baochen Jiang 1 Menghan Sun 1

1. School of Mechanical, Electrical & Information Engineering, Shandong University at WeiHai, WeiHai, China

* Corresponding author.

DOI: https://doi.org/10.5815/ijeme.2011.04.10

Received: 3 Jul. 2011 / Revised: 10 Aug. 2011 / Accepted: 12 Sep. 2011 / Published: 29 Oct. 2011

Index Terms

Emotion parameters estimating, speech synthesis, image processing, MPEG-4

Abstract

This paper analyzes automatic estimating of emotional parameters from 2D photos based on the MPEG-4 rules. Each affective face image which obtained from a novel picture questionnaire is divided into three parts. The deformation function which calculates the displacements of vertices meshes within the influence of facial animation parameters are discussed. This paper develops geometric and morphological image processing methodologies to lifelike facial expression synthesis. Finally a 3D audio-visual avatar is created which gives the same expression corresponding to the input image. Experimental results are given to support our point of view. The experimental results show that our system properly applied in describing emotional facial expression.

Cite This Paper

Jing Yuan,Baochen Jiang,Menghan Sun,"Automatic Estimation of Emotional Parameters for An Intelligent Avatar", IJEME, vol.1, no.4, pp.59-66, 2011. DOI: 10.5815/ijeme.2011.04.10 

Reference

[1]A. Mehrabian, "Communication without words," Psychol. Today, vol.2, pp. 53–56, 1968.

[2]D.N. Jiang, W. Zhang, L. Shen, L.H., Cai, "Prosody analysis and moeling for enottonal speech synthesis", International Conference on Acoustics, Speech and Signal Processing (ICASSP-2005), pp. 281-284, 2005.

[3]F.Parke., "Computer generated animation of faces", ACM Annual Conf., Boston, Massachusetts, United States, pp. 451-457, 1972.

[4]F. Parke, "A model for human faces that allows speech synchronized animation", Journal of Computers and Graphics , vol.1, pp. 1-4, 1975.

[5]K. Waters , D. Terzopoulos, "A physical model of facial tissue and muscle articulation", 1st Conf. on Visualization in Biomedical Computing , Atlanta , USA , pp, 22-25, May 1990.

[6]D. Terzopoulos , K. Waters. "Analysis and synthesis of facial image sequences using physical and anatomical models ", IEEE Trans on Pattern Analysis and Machine Intelligence, vol.15, pp. 569 – 579, Jun 1993.

[7]P. Ekman , W. V. Friesen, "Facial Action Coding System", Palo Alto , CA : Consulting Psychologist Press , 1978

[8]M. Escher , T. Goto , S. Kshirsagar , et al. , "User interactive MPEG-4 compatible facial animation system", International Workshop on Synthetic-Natural Hybrid Coding and Three Dimensional Imaging ( IWSNHC3DI'99) , Santorini , Greece , 1999.

[9]Z.Y. Wu, S. Zhang, L.H. Cai, and H.M..Meng, "Real-time Synthesis of Chinese Visual Speech and Facial Expressions using MPEG-4 FAP Features in a Three-dimensional Avatar", Proc. Int. Conf. on Spoken Language Processing, pp. 1802-1805, 2006.

[10]Motion Pictures Expert Group, ISO/IEC 14496-2:1999/Amd. 1: 2000(E). International Standard, Information.Technology – Coding of Audio-Visual Objects. Part 2:Visual; Amendment 1: Visual Extensions.

[11]R. Wang, W. Gao, J. Y. Ma, " An approach to robust and fastlocating lip motion", 3rd Conf. on Multimodal Interfaces,Heidelberg, vol.1948, pp. 332-339, 2000.

[12]T. Chen. "Audiovisual speech processing", IEEE Signal Processing Magazine ,vol.18, pp. 9-21, 2001.