Design and Implementation of HMM for 3D Emotion Recognition
Keywords:
Hidden Markov model, emotion recognition, facial expressionAbstract
Facial expression is one of the most useful information in human robot interaction. To improve the accuracy in 3-dimension based facial expression recognition, Hidden Markov Models (HMMs) are used to recognize the emotion from facial expressions in this study. In particular, facial expressions are measured by two parameters, which are given by previous work. The human emotions are defined as: anger, smile, normal, sadness, fear, and surprise. The referred parts in human face are selected based on the activeness during the facial expression. The activity and arousal values of each facial part are used as the observations for each hidden state in HMMs. BaumWelch algorithm is used to train the hidden Markov model. As a result, six different emotions are very efficiently recognized through the trained HMMs.
References
[1] Yl. Tian, T. Kanade and JF. Cohn, “Recognizing action units for facial expression analysis,” Pattern Analysis And Machine Intelligence, vol. 23, no. 2, pp.97–115, 2001.
[2] P. Ekman and W. V. Friesen, “The facial action coding system: A technique for the measurement of facial movement,” San Francisco, USA: Consulting Psychologists Press, 1978.
[3] M. Maierdan, “Development of an emotion recognition system based on human behaviors,” Okayama University, Master Thesis, 2013
[4] M. Maierdan, K. Watanabe and S. Maeyama, “Surfacecommon-feature descriptor of point cloud data for deep learning,” in Proc. of International Conference on Mechatronics and Automation, 2016, pp. 525–529.
[5] HW. Ng, VD. Nguyen, V. Vonikakis, and S. Winkler, “Deep learning for emotion recognition on small datasets using transfer learning,” in Proc. of International Conference on Multimodal Interaction, ACM, 2015.
[6] Q. Guo, D. Tu, J. Lei, and G. Li, “Hybrid CNN-HMM model for street view house number recognition,” in Proc. of Asian Conference on Computer Vision, 2014, pp. 303–315.
[7] M. Bartlett, P. Viola, T. Sejnowski, B. Golomb, J. Hager, P. Ekman, and J. Larsen. “Classifying facial action,” in Advances in Neural Information Procession Systems, D. Touretzky, M. Mozer and M. Hasselmo, Eds. Cambridge: MIT Press, 1996, pp.823–829.