https://doi.org/10.31449/inf.v42i4.2037 Informatica 42 (2018) 507–514 507 Feature Extraction Trends for Intelligent Facial Expression Recognition: A Survey Sajid Ali Khan, Hafiz Syed Ahmed Qasim and Irfan Azam Department of Computer Science Shaheed Zulfikar Ali Bhutto Institute of Science and Technology, Islamabad, Pakistan E-mail: sajidalibn@gmail.com, syed.ahmed@szabist-isb.edu.pk, irfan.azam@szabist-isb.edu.pk Overview Paper Keywords: facial expressions, feature extraction, classification methods Received: November 18, 2017 Human facial expression is important means of non-verbal communication and conveys a lot more information visually than vocally. In human-machine interaction facial expression recognition plays a vital role. Still facial expression recognition through machines like computer is a difficult task. Face detection, feature extraction and expression classification are the three main stages in the process of Facial Expression Recognition (FER). This survey mainly covers the recent work on FER techniques. It especially focuses on the performance including efficiency and accuracy in face detection, feature extraction and classification methods. Povzetek: V prispevku je predstavljena primerjalna študija tehnik prepoznavanja izrazov obraza. 1 Introduction Social psychology says facial expressions are means of coordinating conversations and communication. With the advancements in artificial intelligence and pattern recognition, people started considering Facial Expression Recognition (FER) as the most important technology of intelligent human interactive interface [1]. Beside differences, the expressions of different people are still recognizable. Facial information from human face, mostly provide clues for the better depiction of user mind. This increases greatly the human-computer interaction. Scientists have been working on facial expression classification and recognition for the past few decades. Problem-solving abilities and vast applications of a particular discipline act as an inspiration for further exploration and research. The urge to make the visual data useful, is the motivation for all image processing and computer vision algorithms. The FER has same motivation in the domain of computer vision. Its applications in the HCI (Human-Computer Interaction), visual look of human, touch sensations (moods), sight and voice utilization at the same time increases its requirement and value today. Moreover, it has applications like disables emotion detection system, assistance systems for autistic system [2] for detection of pain and stress in psychological studies [3], for instructor feedback an intelligent tutoring system [4], social and emotionally intelligent robot [17] etc. These applications reveal that facial expression detection systems work in effectively unless they are bound to do so in real-time. Face detection and tracking, feature extraction and tracking, feature classification and reduction etc. are the phases involved in FER. Each individual phase uses distinct algorithms, researchers tried to classify basic six expressions using these specific algorithms. This research mainly discusses algorithms for the phases of facial expression. Algorithms like adaptive skin color are used for detection and tracking [5, 6], mean shift algorithms [6], Stereo Active Appearance Model (STAAM) [7] etc. For feature extraction and tracking some algorithms are used like Local Binary Pattern (LBP) [8], Guided Particle Swarm Optimization (GPSO) [9], Gabor feature [8] etc., and there are few algorithms used for feature reduction like Principal Component Analysis (PCA), AdaBoost [10] etc. along classifiers like support vector machine [11-13], Hidden Markov Model (HMM) [14] etc. Accuracy and efficiency are two important aspects in real time environment for FER. Efficiency includes time complexity, space complexity and computational complexity. Due to very high computational complexity (Gabor feature and mean shift algorithm) and space complexity (LBP), it becomes difficult to work in real time environment for most of the mentioned algorithms. At the level of feature extraction, tracking or reduction efficiency can be improved. There are some appropriate algorithms for real-time environment like Optical flow calculation [6], Pixel Pattern-Based Texture Feature (PPBTF) [11], Adaboost [7, 10, 11], Pyramid LBP [12], Haar classifier [8, 10, 13, 15], PCA [10]. This survey mainly covers performance aspects of FER domain. We have mentioned the current approaches for FER and our views related to limitations of these approaches with respect to its execution in real time. 2 Literature review Owusu et al. presented discussion and study about improving execution time and recognition accuracy [18]. 508 Informatica 42 (2018) 507–514 S.A Khan et al. In this work, Viola and Jones algorithm was used for face detection and Basel transformation is utilized for feature extraction. Thousands of facial features are extracted using Gabor feature extraction technique and also those features represent different facial detection patterns. To improve classification speed Adaboost Hypothesis is applied which will select few hundred features from thousands of extracted features. A three layers neural network classifier is then used to further process the selected features. JAFFE and YALE facial expression database are used to train the system and also for its testing. Combination of Basel and AdaBoost is used for the reduction of expression dataset. You will be amazed to know that Basel downsampling is never been used before for FER. So, it is an innovation for improving speed and accuracy. Proposed technique gave an accuracy of 96.83 and the average recognition rate of 92.22℅ for mentioned databases JAFFE and YALE and execution time required for 100x100 pixel sizes is 14.5ms. The results show that neutral expression has the weakest accuracy 92.23℅ in JAFFE and 86.16% in YALE. In [19], Xijian and Tjahjadi extracted spatial pyramid histogram of gradients to three-dimensional facial features. They captured both spatial and motion information of facial expression by integrated the extracted features with dense optical flow. Support vector machine was used in this study with one-to-one strategy for training and testing. Investigation on CK+ and MMI datasets proved that integrated framework gives better performance than using individual descriptors. Contribution of this paper includes an integrated framework that captures dynamic information from deformation of facial regions and also facial landmarks movements, PHOG-TOP facial feature, dense optical flow having fused weighted PHOG-TOP and proposed framework analysis using contribution of different Fabian sub-regions. Canny edge detector is used for the detection of edges. Then to enhance spatial information an image is segmented into number of 3D subregions. PHOG-TOP is employed to whole face in a video sequence and also on 4 different sub-regions (forehead, eye, mouth and nose). Optical flow is implemented in order to extract dynamic info in video sequence. Dense facial points are equally distributed on mid of the face. Grid size is responsible for the efficiency of computing the optical flow. Average accuracy rate is 83.7% on CK+ and 73.1% on MMI dataset. It is also observed that happiness and surprise are easy to detect than remaining expressions. Proposed framework has limitations of generalizing to other datasets and it is also difficult to detect expressions other than happiness and surprise. This is due to the reason of smash and expressed datasets. Another limitation is that it is unable to detect faces wearing glasses or changed hairstyle. Proposed technique by Zhang [20] includes following databases consisting of videos and images from movies and websites (AFEW, SFEW, HAPPEI, CENKI-4Kand QUT FER). Database used for FER falls in two different categories. One category consists of data collected in laboratory environment and the second category which collects data from broadcast TV and World Wide Web includes different databases. They used two lab-based facial FER databases for comparisons. They first applied video selection and segmentation process. Videos are captured from real- world environment and then segmentation is done using video splitter software. Annotator’s pre-training consists of different students. Results from these students are also tested and the cycle is repeated again and again until 96% accuracy is achieved. Clips and vectors are classified by annotators and tested by experienced members. Detection applied using Viola Jones and ASM, failure rate of both algorithms is 4.7%. In Bench Mark FER approach face detection and tracking is applied on selected images using Viola Jones and ASM algorithms then texture features and geometric features are separated using SIFT, FAP and ASM algorithms. Features from texture features are selected using mRMP. Selected features and geometric features are then transferred to feature level fusion. Next, SVM classifier with Radial Basis Function (RBF) is trained for classification of facial expressions. Six basic universal emotions and three categorized emotions (positive, negative and neutral) are extracted using this method. For detection of six basic emotions SIFT+FAP gives 70% accurate results while, detection of three categories (positive, negative, neutral) gives 65% accurate result on realistic QUT images. Realistic QUT video clips have accuracy of 52.9 % for (SIFT+FAP), SIFT 48 %, FAP 50.6% on detection of six universal emotions. Accuracy for three categorized emotions (positive, negative, neutral) is given as SIFT+FAP 62.9 %, SIFT 61% and FAP 56.3 %. On the other hand, performance on lab-base data, FEEDTUM and NIVE database are used which gives 61.0% and 82% accuracy. Classification of three categorized emotions is slightly difficult from six universal expressions. Fear and sadness is highly affected due to nature of data as compared to other expressions. Fang, Hui, et al., [21]is about automatic facial expression which extracts prominent features from videos without any preprocessing of subjective and without requiring any additional data for frames selection. Proposed technique uses machine learning methods in parallel with human reasoning to achieve dynamic changes in expressions in a better way. It is mandatory to detect facial regions first and for this purpose Viola-Jones detector is used. Face detection here is used only for the initiation of group registration. After face detection our next step is to detect features and for this purpose key feature is landmarked (eye, lips and nose). To align faces in static or dynamic data these landmark features are used to eliminate rotation and scaling effect. Through these landmarks deformations in video could be captured for further feature analysis. Traditional base algorithm is then used here to select neutral face image or first frame as template and wrap all images on it. Then global sharp model, appearance, local texture can be used as a knowledge to short the searching. Machine learning method can be used to locate optional landmarks (i.e. linear regression, graphical models). For facial sequences, group-wise Feature Extraction Trends for Intelligent... Informatica 42 (2018) 507–514 509 registration is applied. Successfully displacement between landmarks and other relative measurements like lip curvature, eye size is taken for expression recognition after landmarks tracking, and geometric features can be extracted from result. In following four regions (cheek region, eyebrow region, outer eye corner wrinkle, and forehead region) a Gabor filter is applied to extract an energy value that helps to obtain texture feature for learning expressions. These textures and geometric features are used for each video sequence. For classification of gathered data two techniques can be used 50% stratified split, one half for testing and other one is for training. The other technique contains stratified 10x10 fold cross validation which produces models using given data. Six classifiers are used J48, FRNN, VQNN, random forest, SMO-SVM and logistic and database used for this purpose is MMI. Accuracy by proposed method is 71.56%. It is noticed that happiness and surprise are easily identified by automatic classifiers and also human participants but there are difficulties in identification in the remaining expressions. Zang Wang et al., in [22] proposed that face image is usually presented in high dimensional space as a data point. Principle Component Analysis (PCA) and Linear Discriminant Analysis (LDA) methods are used for dimension reduction. Both can reveal global Euclidean structure but cannot manifold structure. Various manifold learning-based methods has been developed to get discriminant features for image detection and classification. One of the dimensional reduction methods is Local Fisher Discriminant Analysis (LFDA) but this method was not sparse like others and has no discriminant information so it was removed. A new system with Sparse LFDA (SLFDA) was introduced. From LFDA the minimum L1 normalization solution was obtained as a sparse solution. L1 minimization problem overcomes by Bregman. Therefore, Bregman method is applied to obtain sparse projection vector. Original features weight can be controlled by SLFDA. Moreover, in multimode problem SLFDA works well and the contrasting power of LFDA enhanced by it. In dimension reduction methods competitive to others SLFDA can achieve performance shown by the experiments performed on the databases (JAFFE and Yale B). The best recognition rate of SLFDA is 77.92%. As the proposed method strength is in dimension reduction and also gives reasonable interpretation of extracted features but their only focus was on supervised learning, so there is possibility to extend this approach to semi-supervised learning framework as well as to find the fast-numerical algorithm to solve L1-normalization problem. By using both appearance and geometric features performance improved but most of existing algorithms are based on geometric features such algorithms track the facial components like eyes, lips, corner etc. as well as shapes and size of the face. Happy et al. [23] discussed that major issue is landmarks selection for which a relative geometric distance-based approach described to detect landmarks. Deformable model also become famous for landmark detection but high computation cost is a hurdle for them in real-time applications. It is a learning free approach to detect eyes, nose, lips etc. in face image and mark the required region. Some salient patches are extracted in training stage and within-pair of expressions; features having maximum variants are selected. Then multi-class classifier divides these selected features into basic six classes of expressions (those are fear, sadness, happiness, disgust, surprise and anger). In near frontal image with less computational complexity the results of facial landmark system are similar to the state of art method. Accurate emotion recognition in low-resolution image is ultimate goal of effective computing so this facial landmark detection technique along with salient patches-based FER framework performance is good in different image resolutions. Accuracy rate on JAFFE database is 91.8% and 94.1% on Cohn-Kanade (CK+) database which is satisfactory result but they are just considering few facial patches not whole face and also analyzing facial features without considering facial hairs. There is possibility of improvement in performance with partially occluded images and by using different appearance features. Moreover, an un-optimized MATLAB code is used for execution time. However, to improve the computational cost and real-time expression recognition with good accuracy rate an optimal implementation is required. Ying tong et al. [24] discussed facial expressions of human and feature extraction method. Although Gabber wavelet, LBP are also used for feature extraction methods but they are time-consuming and dimensions increased significantly. In [24], the authors made binary coding for the separate block images which results a LGC statistics histogram. For identification feature they linked together the resultant histograms. In order to obtain the LGC-HD operator (LGC based on Horizontal and Diagonal gradient prior principle) more optimization is provided on the LGC operator. This reduces the computational complexity without losing the main expression information from face texture and also reduces the characteristic dimension. By using JAFFE database, the recognition average time in seconds are 90 for 8x8 block size with LGC-HD operator. Even after comparison of LBP, LBP uniform pattern, Gabor filter and LGC-HD the recognition average rate of LGC-HD is 90 %, which is higher than others. Experimental results show the weakness that either the block is larger or smaller will impact the recognition rate. The very small block number has smaller sub-block which cannot be accurately extracted. The redundancy of LGC-HD factor will affect the classification which results in inaccurate expression characteristics of large area. The common universal facial expression recognizes cross-cultural facial expression including Japanese, Chinese, European and American. Ali et. al. in [25] claims that instead of using whole face only consider facial components (eyebrows, eye, etc.) as a lot of work has been done on those and through such technique satisfactory result is gained. In the race of solving problems of facial expressions consider multiple classifier decisions instead of using single classifier decision. Moreover, neural network-based ensemble classifier is made to enhance the accuracy of classifier. 510 Informatica 42 (2018) 507–514 S.A Khan et al. Multi-objective genetic algorithm is also used. Acquisition and representation of multicultural dataset are major problems of multicultural facial expression classification. Three databases JAFFE, TFED and RadBoud used to overcome these problems. To check the presence of expression they used KNN, NB (Nave Bayes classifier) and SVM classifier. Furthermore, they used established dataset verification strategy for system performance evaluation. Four types of classifier considered for the classification of expressions those are BNN, KNN, SVM and Naïve Bayes classifier so that finally this plan worked out accurately. Experimental result 93.75% got with the combination of NNE collection, with NB predictor and using HOG descriptor. This is best in order to get satisfactory result of multicultural FER. But on few confuse facial expression still future work is needed due to visual representation, facial structure and difference in number of samples. For recognition of facial expression, theoretical description of face operations often highlights the feature shapes as a primary visual signal. Although, facial surface characteristics can also be affected by changes in facial expression. Mladen Sormaz et al., in [26] examined that in the recognition of facial expression this surface knowledge can also be used. Firstly, facial expressions from images with distinct shapes and surface characteristics are identified by the participants. Mladen Sormaz et al., said various expressions depend on properties of shape and surface. They Further elaborated that facial expression categorization is feasible in any type of image. Moreover, in order to categorize the facial expressions, they evaluated the corresponding contributions of surface and shape information. This involves a correlative method in which shape properties and surface properties both are taken from different expressions. The experimental results show that in hybrid images categorization of facial expressions basically depends on the properties of surface and shape of image. Collectively, all the data directly demonstrate that recognition of facial expression is done through significant contribution of both the surface and the shape properties. Andres Hernandez-Matamoros et al., mainly focus facial expression algorithm that significantly encounter facial image, present in color picture and segment divided into two Regions of Interest (ROI) i.e. the forehead and the mouth [27]. Both these regions are further segmented into non-overlying NXM blocks. Then dimension reduction is carried out after inserting the matrix in the Principal Component Analysis (PCA) module. Lastly, the resultant matrix generates the feature vectors. These vectors are then incorporated into the low complexity classifier, which uses the congregation and fizzy logical techniques. Though this classifier gives similar recognition rate as the high-performance classifiers but it gives least computational complexity. Results show that when feature vector from only one ROI is used in the proposed system then the recognition rate was increased to 97%. But the usage of feature vector of both ROI increases recognition rate up to 99%. It means overall 97% recognition rate in proposed system can only be achieved by clogging only one ROI. In another work, Khan et al., in [28] highlighted the importance of local descriptors like Weber Local Descriptor (WLD) and LBP for recognition of facial expressions which is robust to illumination and pose changes. They argue that the local descriptors cannot be used to store the data of face image. In order to handle this problem, they proposed a framework in which the first they extracted features from face image using WLD and LBP and then fuse both type of features. Similarly, a novel framework known as Weber Local Binary Image Cosine Transform (WLBI-CT) is proposed in [29] to recognize facial expression from multi-scale images. The results of this framework are robust to image resolution and image orientation. Recently, in [30] Munir et al., utilized Merged Binary Pattern Code (MBPC) descriptor for face feature extraction. MBPC descriptor is capable to capture the prominent face feature. In this study, the results are compared with the variants of LBP. Many other works [31]-[32] is presented in literature to describe the importance of face recognition and facial expression recognition. Feature Extraction Trends for Intelligent... Informatica 42 (2018) 507–514 511 3 Analysis table References Techniques Used Pros Cons • Viola Jones • Basel transformation • Gabor feature Extraction technique • Combination of Basel and Adaboost • 3 layers neural network classifier • Improves execution time and recognition accuracy • Basel downsampling is used here first time. • Proposed technique gives an accuracy of 96.83 and 92.22℅ • Neutral expression has the weakest accuracy Xijian et al. [19] • Multi-class support vector machine Based classifier. • PHOG-TOP facial Feature • PHOG descriptor • Canny edge detector • Integrated framework gives better performance than using individual descriptors • Average accuracy rate on CK+ 83.7% and 73.1% on MMI dataset • Limitations of generalizing to other datasets • Difficult to detect expressions other than happiness and surprise • Unable to detect faces wearing glasses or changed hairstyle. Zhang et al. [20] • Video splitter software • Annotator’s • Viola Jones and ASM • SIFT, FAP • mRMP • SVM classifier with • Radial Based Function (RBF) • Facial Expression is applied to real-world dynamic data which is quite difficult to handle. • Fear and sadness is highly affected due to nature of data as compared to other expressions • Classification of 3 categorized emotions are slightly difficult from 6 universal expressions Fang, Hui, et al. [21] • Viola jones detector • Traditional base Algorithm used • Global sharp model, appearance, local texture • Machine Learning Method • Group-wise registration • Gabor filter • 6 classifiers are used J48, FRNN, VQNN, random forest, SMO-SVM and logistic. • Extracts prominent features from videos without any preprocessing of subjective and without requiring any additional data for frames selection. • Uses machine learning method in parallel with human reasoning to achieve dynamic changes in expressions in a better way. • Happiness and surprise are easily identified by automatic classifiers and also human participants but remaining expressions faces difficulties in identification. Zang Wang et al. [22] • PCA • LDA • Manifold learning based method • LFDA • SLFDA • Dimension reduction • Interpretation of extracted features • In multimode problem SLFDA work well • The best recognition rate of SLFDA is 77.92%. • Only focused was on supervised learning so possibility to extent semi-supervised learning • Find the fast-numerical algorithm to solve L1- normalization problem Happy et al. [23] • Deformable model • Landmark detection Technique • Landmark detection • Well performance of FER in low resolution image • Not considering the whole face • Analyzing facial feature without considering the facial 512 Informatica 42 (2018) 507–514 S.A Khan et al. • Accuracy rate on JAFFE database is 91.8% and 94.1% on Cohn-Kanade (CK+) database hairs • Un-optimized MATLAB code is used Ying tong et al. [24] • LGC-HD operator • Efficient without losing the main expression information from face texture • Reduce the characteristics dimension. • Recognition average rate is 90% • Block is larger or smaller will impact the recognition rate • The LGC-HD features of each sub-block will show redundancy Ghulam Ali et al. [25] • SVM • Neural Network(NN) • K Nearest Neighbor (KNN) • Rule-based classifier • Nave Bayes classifier (NB) • Binary Neural Network (BNN) • NB predictor • HOG descriptor • Result 93.75% got with the combination of NNE collection, with NB predictor and using HOG descriptor • Solved the representation of multicultural dataset problems of diverse face expression classification • Future work is needed due to some confuse facial expression Mladen Sormaz et al. [26] • Complementary or converging methods • Hybrid image • Both shape and surface play important role in the recognition of facial expressions Andres Hernandez- Matamoros et al. [27] • Voila jones • PCA • Gaussian functions • Gabor filters • Empirical mode decomposition (EMD) • SVM (Support Vector Machine) classifier • Facial recognition rate higher than 97% • Dimension reduction, • Similar recognition performance obtained in both using YUV color space and RGB color space • ROI extraction is accurate even if the brightness changes • It only requires Full face without occlusion or partially occluded face for expression recognition 4 Conclusion and future work During communication transmission facial expressions are produced so that the images can be obtained in unmanageable condition i.e. occlusion (effect of makeup, glasses, facial hair, hijab which can also affect the rate of recognition), illumination of light, posed expressions and variations in expression etc. This paper presented a survey on current work done in the domain of FER. Some techniques of feature extraction were explained. In addition, comparisons were also done which can help other researchers to advance and polish the present methods for getting accurate and better results in future. In future we are intended to investigate the local descriptor in frequency domain for real-world FER. 5 References [1] Shuai-Shi Liu, Y. Tian and Dong Li, "New research advances of facial expression recognition," 2009 International Conference on Machine Learning and Cybernetics, Hebei, 2009, pp. 1150-1155. https://doi.org/10.1109/ICMLC.2009.5212409 [2] R. Alazrai and C. S. G. Lee, "Real-time emotion identification for socially intelligent robots," 2012 IEEE International Conference on Robotics and Automation, Saint Paul, MN, 2012, pp. 4106-4111. https://doi.org/10.1109/ICRA.2012.6224587 [3] Bee Theng Lau. “Portable real time emotion detection system for the disabled." Expert Systems with Applications. 37(9): 6561-6566, 2010. Feature Extraction Trends for Intelligent... Informatica 42 (2018) 507–514 513 https://doi.org/10.1016/j.eswa.2010.02.130 [4] X. Lu, "Image Analysis for Face recognition", [online] Available: http:/www.cse.msu.edu. [5] P. Zhao-yi, Z. Yan-hui and Z. Yu, "Real-time Facial Expression Recognition Based on Adaptive Canny Operator Edge Detection," 2010 Second International Conference on Multimedia and Information Technology, Kaifeng, 2010, pp. 154- 157. https://doi.org/10.1109/MMIT.2010.100 [6] C. Mayer et al., "A real time system for model- based interpretation of the dynamics of facial expressions," 2008 8th IEEE International Conference on Automatic Face & Gesture Recognition, Amsterdam, 2008, pp. 1-2. https://doi.org/10.1109/AFGR.2008.4813440 [7] Z. Peng, Z. Wen and Y. Zhou, "Application of Mean Shift Algorithm in Real-Time Facial Expression Recognition," 2009 International Symposium on Computer Network and Multimedia Technology, Wuhan, 2009, pp. 1-4. https://doi.org/10.1109/CNMT.2009.5374770 [8] S. L. Happy, A. George and A. Routray, "A real time facial expression classification system using Local Binary Patterns," 2012 4th International Conference on Intelligent Human Computer Interaction (IHCI), Kharagpur, 2012, pp. 1-5. https://doi.org/10.1109/IHCI.2012.6481802 [9] J. N. Bailenson, E. D. Pontikakis, I. B. Mauss, J. J. Gross, M. E. Jabon, C. A.C. Hutcherson, C. Nass, O. John, "Real-time classification of evoked emotions using facial feature tracking and physiological responses." International journal of human-computer studies. 66 (5): 303-317, 2008. https://doi.org/10.1016/j.ijhcs.2007.10.011 [10] D. Ghimire and J. Lee "Geometric feature-based facial expression recognition in image sequences using multi-class adaboost and support vector machines." Sensors, 13(6): 7714-7734, 2013. https://doi.org/10.3390/s130607714 [11] Jaewon Sung, Sangjae Lee and Daijin Kim, "A Real-Time Facial Expression Recognition using the STAAM," 18th International Conference on Pattern Recognition (ICPR'06), Hong Kong, 2006, pp. 275-278. https://doi.org/10.1109/ICPR.2006.158 [12] Lu, H-C., Y-J. Huang, and Y-W. Chen. "Real-time facial expression recognition based on pixel- pattern-based texture feature." Electronics letters. 43(17): 916-918, 2007. https://doi.org/10.1049/el:20070362 [13] R. A. Khan, A. Meyer, H. Konik, S. Bouakaz, "Framework for reliable, real-time facial expression recognition for low resolution images." Pattern Recognition Letters, 34(10): 1159-1168, 2013. https://doi.org/10.1016/j.patrec.2013.03.022 [14] B. M. Ghandi, R. Nagarajan and H. Desa, "Particle Swarm Optimization algorithm for facial emotion detection," 2009 IEEE Symposium on Industrial Electronics & Applications, Kuala Lumpur, 2009, pp. 595-599. https://doi.org/10.1109/ISIEA.2009.5356389 [15] A. Punitha, M. K. Geetha, HMM Based Real Time Facial Expression Recognition, International Journal of Emerging Technology and Advanced Engineering, 3(1), 180-185, 2013. [16] A. M. Adeshina, S. Lau and C. Loo, "Real-time facial expression recognitions: A review," 2009 Innovative Technologies in Intelligent Systems and Industrial Applications, Monash, 2009, pp. 375- 378. https://doi.org/10.1109/CITISIA.2009.5224179 [17] A. Geetha, V. Ramalingam, S. Palanivel, B. Palaniappan. "Facial expression recognition–A real time approach." Expert Systems with Applications. 36(1): 303-308, 2009. https://doi.org/10.1016/j.eswa.2007.09.002 [18] Ebenezer Owusu, Yongzhao Zhan, Qi Rong Mao "A neural-AdaBoost based facial expression recognition system." Expert Systems with Applications, 41(7): 3383-3390, 2014. https://doi.org/10.1016/j.eswa.2013.11.041 [19] F. Xijian, and T. Tjahjadi "A spatial-temporal framework based on histogram of gradients and optical flow for facial expression recognition in video sequences." Pattern Recognition, 48(11): 3407-3416, 2015. https://doi.org/10.1016/j.patcog.2015.04.025 [20] Z. Ligang, D. Tjondronegoro, and V. Chandran "Facial expression recognition experiments with data from television broadcasts and the World Wide Web." Image and Vision Computing, 32(2): 107- 119, 2014. https://doi.org/10.1016/j.imavis.2013.12.008 [21] Hui Fang, Neil Mac Parthaláin, Andrew J. Aubrey, Gary K.L. Tam, Rita Borgo, Paul L. Rosin, Philip W. Grant, David Marshall, Min Chen, "Facial expression recognition in dynamic sequences: An integrated approach." Pattern Recognition ,47(3): 1271-1281, 2014. https://doi.org/10.1016/j.patcog.2013.09.023 [22] Wang, Zhan, Qiuqi Ruan, and Gaoyun An "Facial expression recognition using sparse local Fisher discriminant analysis." Neurocomputing, 174: 756- 766, 2016. https://doi.org/10.1016/j.neucom.2015.09.083 [23] S.L. Happy, A. Routray. "Automatic facial expression recognition using features of salient facial patches." IEEE transactions on Affective Computing, 6(1): 1-12, 2015. https://doi.org/10.1109/TAFFC.2014.2386334 [24] T. Ying, R. Chen, and Y. Cheng "Facial expression recognition algorithm using LGC based on horizontal and diagonal prior principle." Optik- International Journal for Light and Electron Optics, 125(16): 4186-4189, 2014. https://doi.org/10.1016/j.ijleo.2014.04.062 [25] G. Ali, M. A. Iqbal, and Tae-Sun Choi. "Boosted NNE collections for multicultural facial expression recognition." Pattern Recognition, 55: 14-27, 2016. https://doi.org/10.1016/j.patcog.2016.01.032 514 Informatica 42 (2018) 507–514 S.A Khan et al. [26] Timothy J. Andrews, Heidi Baseler, Rob Jenkins, A. Mike Burton, Andrew W. Young. "Contributions of feature shapes and surface cues to the recognition and neural representation of facial identity." Cortex, 83: 280-291, 2016. https://doi.org/10.1016/j.cortex.2016.08.008 [27] A.H Matamoros, A. Bonarini, E.E.Hernandez, M.Nakano-Miyatake, H. Perez-Meana. "A facial expression recognition with automatic segmentation of face regions." International Conference on Intelligent Software Methodologies, Tools, and Techniques. 529-540, 2015. https://doi.org/10.1007/978-3-319-22689-7_41 [28] S.A. Khan, A. Hussain, M. Usman, "Facial expression recognition on real world face images using intelligent techniques: A survey." Optik- International Journal for Light and Electron Optics, 127(15): 6195-6203, 2016. https://doi.org/10.1016/j.ijleo.2016.04.015 [29] S.A. Khan, A. Hussain, M. Usman, "Reliable facial expression recognition for multi-scale images using weber local binary image-based cosine transform features." Multimedia Tools and Applications, 77(1): 1133–1165, 2018. https://doi.org/10.1007/s11042-016-4324-z [30] A. Munir, A. Hussain, S.A. Khan, M. Nadeem, S. Arshid, "Illumination invariant facial expression recognition using selected merged binary patterns for real world images" Optik-International Journal for Light and Electron Optics, 158: 1016-1025, 2018. https://doi.org/10.1016/j.ijleo.2018.01.003 [31] S.A. Khan, M. Ishtiaq, M. Nazir, M. Shaheen, “Face recognition under varying expressions and illumination using particle swarm optimization,” Journal of Computational Science, 28: 94-100, 2018. https://doi.org/10.1016/j.jocs.2018.08.005 [32] S.A. Khan, S. Hussain, S. Xiaoming, S. Yang "An Effective Framework for Driver Fatigue Recognition Based on Intelligent Facial Expressions Analysis", IEEE ACCESS. https://doi.org/10.1109/ACCESS.2018.2878601