HOME JOURNALS CONTACT

Information Technology Journal

Year: 2013 | Volume: 12 | Issue: 23 | Page No.: 7305-7311
DOI: 10.3923/itj.2013.7305.7311
Dynamic Feature Extraction for Facial Expression Recognition Based on Optical Flow
Xibin Jia, Shuangqiao Liu and David M. W. Powers

Abstract: Understanding facial expressions is a fundamental problem in affective computing that has the potential to impact both sides of a conversation with a computational agent. Currently static approaches based on techniques such as Gabor transformation represent the state of the art for identifying facial expressions but remain rather slow and unreliable. In this study we introduce and compare a dynamic technique based on optical flow versus this static Gabor baseline, as well as integrating information about static and dynamic facial characteristics into novel fused models. Rather than requiring complex Machine Learning, a simple and fast template model based on K-Means uses a Hidden Markov Model to provide context. The system is trained and evaluated on distinct subsets of the Cohn-Kanade database of adult faces which provides classifications into six basic expressions. Experimental results show that the dynamic expression feature extraction based on the optical flow for facial expression recognition has considerably improved recognition rate. Fusion of the optical flow dynamic expression features with Gabor static expression feature also increases recognition rate somewhat versus the static baseline but the hybrids tested are not competitive with the pure dynamic model.

Fulltext PDF

How to cite this article
Xibin Jia, Shuangqiao Liu and David M. W. Powers, 2013. Dynamic Feature Extraction for Facial Expression Recognition Based on Optical Flow. Information Technology Journal, 12: 7305-7311.

Keywords: Feature fusion, optical flow, local binary patterns, hidden Markov models, embodied conversational agents and affective computing

REFERENCES

  • Ali, H.B., D.M.W. Powers and R. Leibbrandt and T. Lewis, 2011. Comparison of Region Based and Weighted Principal Component Analysis and Locally Salient ICA in Terms of Facial Expression Recognition. In: Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing, Lee, R. (Ed.). Springer-Verlag, Berlin, Heidelberg, pp: 81-89


  • Bradski, G. and A. Kaebler, 2009. Open CV Learning. 3rd Edn., Tsinghua University Press, Beijing, China, pp: 350-355


  • Cootes, T.F., G.J. Edwards and C.J. Taylor, 1998. Active appearance models. Proceedings of the 5th European Conference on Computer Vision, June 2-6, 1998, Freiburg, Germany, pp: 484-498.


  • Cootes, T.F., G.V. Wheeler, K.N. Walker and C.J. Taylor, 2002. View-based active appearance models. Image Vision Comput., 20: 657-664.
    CrossRef    Direct Link    


  • Horn, B.K.P. and B.G. Schunck, 1981. Determining optical flow. Artif. Intell., 17: 185-203.
    CrossRef    Direct Link    


  • Jia, X.B., X.Y. Bao, D.M.W. Powers and Y.J. Li, 2013. Facial expression recognition based on block Gabor wavelet fusion feature. J. Converg. Inform. Technol., 8: 282-289.


  • Lanitis, C.J., T.F. Taylor and T.F. Coates, 1997. Automatic interpretation and coding of face images using flexible models. IEEE Trans. Pattern Anal. Machine Intell., 19: 743-756.
    CrossRef    


  • Lang, S.R., M.H. Luerssen and D.M.W. Powers, 2013. Automated evaluation of interest point detectors. Proceedings of the IEEE/ACIS 12th International Conference on Computer and Information Science, June 16-20, 2013, Niigata, Japan, pp: 443-447.


  • Mitchell, H.B., 2010. Data Fusion: Concepts and Ideas. Springer, New York, USA., pp: 126-128


  • Newman, W., D. Franzel, T. Matsumoto, R. Leibbrandt, T. Lewis, M. Luerssen and D.M.W. Powers, 2010. Hybrid world object tracking for a virtual teaching agent. Proceedings of the IEEE International Joint Conference on Neural Networks, July 18-30, 2010, Barcelona, Spain, pp: 2244-2252.


  • Powers, D.M.W., R.E. Leibbrandt, D. Pfitzner, M.H. Luerssen, T.W. Lewis, A. Abrahamyan and K. Stevens, 2008. Language teaching in a mixed reality games environment. Proceedings of the 1st International Conference on Pervasive Technologies Related to Assistive Environments, July 15-19, 2008, Athens, Greece -.


  • Kana, R.K., T.A. Keller, V.L. Cherkassky, N.J. Minshew and M.A. Just, 2006. Sentence comprehension in autism: Thinking in pictures with decreased functional connectivity. Brain, 129: 2484-2493.
    CrossRef    


  • Wallach, J.C. and L.G. Gibson, 2001. Mechanical behavior of a three-dimensional truss material. Int. J. Solids Struct., 3: 7181-7196.
    CrossRef    


  • Wang, J., L.J. Yin, X.Z. Wei and Y. Sun, 2006. 3D facial expression recognition based on primitive surface feature distribution. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, June 17-22, 2006, New York, USA., pp: 1399-1406.


  • Kanade, T., J.F. Cohn and Y. Tian, 2000. Comprehensive database for facial expression analysis. Proceedings of the 4th IEEE International Conference on Automatic Face and Gesture Recognition, March 26-30, 2000, Grenoble, France, pp: 46-53.

  • © Science Alert. All Rights Reserved