Fusion of visible and thermal images for facial expression recognition

Size: px
Start display at page:

Download "Fusion of visible and thermal images for facial expression recognition"

Transcription

1 Front. Comput. Sci., 2014, 8(2): DOI /s Fusion of visible and thermal images for facial expression recognition Shangfei WANG 1,2, Shan HE 1,2,YueWU 3, Menghua HE 1,2,QiangJI 3 1 School of Computer Science and Technology, University of Science and Technology of China, Hefei , China 2 Key Lab of Computing and Communicating Software of Anhui Province, Hefei , China 3 Department of Electrical, Computer, and Systems Engineering, Rensselaer Polytechnic Institute, Troy NY , USA c Higher Education Press and Springer-Verlag Berlin Heidelberg 2014 Abstract Most present research into facial expression recognition focuses on the visible spectrum, which is sensitive to illumination change. In this paper, we focus on integrating thermal infrared data with visible spectrum images for spontaneous facial expression recognition. First, the active appearance model AAM parameters and three defined head motion features are extracted from visible spectrum images, and several thermal statistical features are extracted from infrared (IR) images. Second, feature selection is performed using the F-test statistic. Third, Bayesian networks BNs and support vector machines SVMs are proposed for both decision-level and feature-level fusion. Experiments on the natural visible and infrared facial expression (NVIE) spontaneous database show the effectiveness of the proposed methods, and demonstrate thermal IR images supplementary role for visible facial expression recognition. Keywords facial expression recognition, feature-level fusion, decision-level fusion, support vector machine, Bayesian network, thermal infrared images, visible spectrum images 1 Introduction Facial expression, a primary method by which humans display their emotions, has drawn growing attention in many research areas related to human-computer interaction and Received July 30, 2013; accepted October 14, sfwang@ustc.edu.cn psychology. In recent years, considerable progress has been made in the field of facial expression recognition using visible spectrum images and videos [1 5]. However, most of existing methods are not robust enough for employment in uncontrolled environments. Illumination change is the most important factor, because it can significantly influence the appearance of visible images. However, thermal infrared (IR) images, which record temperature distribution, are not sensitive to illumination conditions. Thus, IR-based facial expression recognition algorithms can improve recognition performance in uncontrolled environments. While IR images possesses robustness to illumination change, it also has several drawbacks that visible images do not. IR images are sensitive to the surrounding temperature and opaque to glass [6]. Combining the advantages and disadvantages of both visible and IR images for facial expression recognition is promising. However, to the best of our knowledge, little attention has been paid to facial expression recognition by fusing visible and IR images [7 9]. In this paper, we focus on integrating IR with visible images for spontaneous facial expression recognition. Two Bayesian networks (BNs) and two support vector machines (SVMs) are presented to perform decision-level and featurelevel fusion, respectively, of visible spectrum and IR images for facial expression recognition. First, features are extracted from visible spectrum and infrared images. Several studies have suggested that using both geometric and appearance features may achieve good performance [1]. Thus, the active

2 Shangfei WANG et al. Fusion of visible and thermal images for facial expression recognition 233 appearance model (AAM) parameters, which capture both appearance and shape information [10], are extracted from apex expression images. As head motion has been proved useful for facial expression recognition, we also define three head motion features [11]. For IR images, considering the temperature changes of the environment and the temperature drift of IR cameras, the relative temperature, which is difference between temperature matrices of the apex and onset images, is adopted instead of the absolute temperature. Several statistical features are extracted from temperature matrices. After feature extraction we perform feature selection using the F-test statistic. Then, the structures of fusion models are defined, and we parameterize the fusion models. Using the trained model, the true state of expression is inferred through probabilistic inference using BNs or classified by SVMs. Finally, we evaluate the proposed methods on the natural visible and infrared facial expression (NVIE) spontaneous database [12] to show their effectiveness. A review of related work is presented in Section 2. The fusion methods for facial expression recognition are explained in Section 3. Experiments are given in Section 4 followed by conclusions in Section 5. 2 Related work Recently, a number of researchers have turned their attention to spontaneous facial expression recognition. A review on this topic can be found in [1]. As mentioned in [1], only a few studies have investigated fusing information from face and head motions to improve recognition performance. The study of Littlewort et al. [13] shows that head pose and head motion contain substantial information about emotion. Cohn et al. [14] proposed a method to recognize facial action units (AUs) from a spontaneous expression video using head motions and facial features. Their study shows a high correlation between brow and head motions. Tong et al. [15] proposed a unified dynamic Bayesian network (DBN) based facial action model to discover and learn spatiotemporal relationships between rigid head motions and non-rigid facial motions. They then performed robust and reliable spontaneous facial action recognition by combining these relationships with image observations. Their study suggests that the models built-in spatiotemporal relationships between rigid head motions and nonrigid facial motions can compensate for erroneous AU measurements. Valstar et al. [16] combined the face modality and the information from head and shoulder motions to automatically distinguish between posed and spontaneous smiles. Their study indicates that humans also reliably reveal their intentions through body language, such as head movements. Gunes and Pantic [17] proposed using head motions in spontaneous conversations as an extra dimension of information for predicting emotions. All the above studies suggest that head motions are useful for improving the performance of spontaneous facial expression recognition. Facial expression recognition in the IR spectrum has received relatively little attention compared to facial expression recognition in the visible spectrum [18 23]. Hernández et al. [19] defined regional descriptors of IR images using a gray level co-occurrence matrix (GLCM). These descriptors were used to distinguish the expressions of surprise, happiness, and anger. Yoshitomi [21] extracted features by performing a two-dimensional discrete cosine transformation (2D-DCT) to transform the grayscale values of each facial block into their frequency components, and then use knn (k-nearest Neighbors) as a classifier. Khan et al. [20] tried to use thermal variations along the major facial muscles as features to discern between posed and evoked expressions. Jarlier et al. [18] investigated the thermal patterns associated with specific AUs performance. Shen et al. [23] proposed a spontaneous facial expression recognition method using thermal videos. Shen et al. s method uses Adaboost with weak classifiers of knn to distinguish facial expressions into arousal and valence dimensions. All these studies show that the information provided by infrared images is useful for recognizing facial expression. To the best of our knowledge, little attention has been paid to facial expression recognition by fusing images from the visible and infrared spectrums. Yoshitomi et al. [7] described a method for the recognition of affective states using decision-level fusion of voices, visual and infrared facial images. Firstly, visible spectrum and IR features are extracted by DCT. Then, two neural network (NN) classifiers are trained for visual and IR features respectively. And, hidden Markov models (HMMs) are adopted for emotion detection from voices. Finally, the results of the three classifiers are integrated using simple weighted voting. Yoshitomi et al. evaluate their approach on a posed expression database. They do not report the relationships between the selected features and facial expressions. Wang et al. [8] and Wang and He [9] proposed two feature-level fusion methods for fusing visible and IR images for facial expression recognition. The F-statistic [8] and multiple genetic algorithms (GAs) [9] are adopted to select visible and IR features, then a knn classifier is used to recognize happiness, fear, and disgust.

3 234 Front. Comput. Sci., 2014, 8(2): Although researchers have focused on face recognition (rather than facial expression recognition) by fusing visible and IR images for several years, and have already proven the fusion effectiveness for face recognition, the role of IR images for expression recognition have not been well explored. We cannot take for granted that the role of IR images for expression recognition is the same as that for face recognition, since face recognition and expression recognition are different tasks. A good face recognition system should be invariant to facial expression, while a good facial expression system should be face independent. Furthermore, the underlying mechanism that enables IR images to help face recognition and expression recognition are different, although both recognition tasks take advantage of the robustness of IR images to illumination changes. Specially, IR images record the temperature distributions that are formed by facial vein branches. The facial vascular network is highly characteristic of each individual, much as a fingerprint is. Thus, IR images can be used for face recognition [24]. However, the changes of emotions increase the levels of adrenaline and regulates blood flow. Redistribution of blood flow in superficial blood vessels causes abrupt changes in local skin temperature that can be recorded by IR images. The muscle movements caused by facial expressions also cause the movement of facial vein branches. These movements make IR images reflect the change of emotion and expression [25]. Fusion of visible and IR images for facial expression is relatively new, we believe that many intelligent techniques worth exploring have not been investigated. Our work makes the following three main contributions: 1) We are among the first to introduce IR to expression recognition and we introduce several new models to effectively integrate IR with visible images. It is novel for the affective computing community. 2) We are the first to propose both decision-level and feature-level fusion methods fusing visible and infrared spectrums for facial expression recognition. 3) We report the relative significance of temperature variations of different facial regions for thermal facial expression recognition. 3 Proposed methods Our proposed method for facial recognition, combining IR and visible spectrum images, consists of three parts as shown in Fig. 1. After obtaining the visible and IR images from the database, visible and IR features are extracted.featureselection is then performed by using the F-test statistic. Given the expression labels and extracted features of the training set, the parameters for BNs and SVMs are learned. During testing, the extracted features of test sample are used as the input of the learned BN and SVM models for expression recognition. Fig Visible feature extraction Head motion Block diagram of the proposed method Three features, including velocities of head motion along the x- andy-axis, and the head s rotational speed, are calculated as head motion features. First, the subject s eyes are located using an AdaBoost and Haar feature eye location method [11]. Then, head motion features are calculated according to the coordinates of the pupils using Eqs. (1) (3). C apex x C onset x Velocity x = frame apex onset, (1) frame Cy onset Velocity y = frame apex onset, (2) frame ( R apex arctan y L apex ) ( y R onset y L onset ) y R apex x L apex arctan x R onset x L onset x Velocity r =, (3) frame apex frame onset where, (L x, L y ) represents the coordinates of the left pupil, (R x, R y ) represents the right pupil, and (C x, C y )represents the center of two pupils. The apex image frame refers to the frame where the expression is most exaggerated, and the onset image refers to the frame where the expression just begins AAM C apex y Since the AAM captures information about both appearance and shape, we use AAM to extract visible features from the

4 Shangfei WANG et al. Fusion of visible and thermal images for facial expression recognition 235 apex expressional images which are the most exaggerated expression images. We use the am_tools from [26] to extract the AAM features here. 3.2 IR feature extraction To eliminate the influence of ambient temperature and the temperature drift of IR cameras, IR features are extracted from the differential temperature matrix between the apex and the onset IR images. First, four points (the centers of the eyes, the tip of the nose, and tip of the jaw) are automatically located on the apex and onset IR images using the method devised by Tong et al. [27]. Using these points, the apex and onset IR images are rotated, resized, and cropped into images with size of H W. Subsequently, the differential temperature matrix is obtained by subtracting the temperature matrix that corresponds to the normalized onset IR image from that of the normalized apex IR image, as shown in Fig. 2(b). Fig. 2 (a) Visible image used for constructing AAM; (b) diagram of thermal feature extraction where N is the size of classes, n c is the number of samples of class c, n is total number of samples, x c is the average of feature x within class c, x is the global mean of feature x,andσ 2 c is the variance within class c. Features are selected from high to low according to the calculated F-statistics of all features. 3.4 Expression recognition BN models of facial expression recognition Decision-level fusion using BN We design our BN model as shown in Fig. 3(a). The model consists of three layers: detected expression layer, intermediate expression layer, and feature layer. All nodes in the detected expression layer and intermediate expression layer are discrete nodes with N states, and each state corresponds to an expression to be recognized. The relations between the detected expressions and the visible expressions or IR expressions are established through links, which capture the uncertainty of visible results or IR results. The lowest layer in the model is the feature layer containing head motion features and AAM features extracted for visible images, and IR features. All variables in this layer are observable. The relations between the second layer and the lowest layer capture the dependencies between an expression and its visual and thermal features. Owing to the opaqueness of IR light to glass, the subarea of the eyes is not taken into consideration here (since the target may be wearing glasses). We divided the facial region into several grids with the same size of w w. For each grid, several statistical parameters, including mean, standard deviation, minimum, and maximum, are calculated as IR features. 3.3 Feature selection In order to select distinctive features for classification, the F-test statistic [28] is used for feature selection. The significance of all features can be ranked by sorting their corresponding F-test statistics in descending order. This assumes that the feature fits a normal distribution. The F-test statistic of feature x is calculated according to Eq. (4). between-group variability F-statistic = within-group variability N n c ( x c x) 2 ( c=1 n N ) =, (4) N N 1 (n c 1)σ 2 c c=1 Fig. 3 (a) Decision-level fusion model using BN; (b) feature-level fusion model using BN Given the model structure, the conditional probability distribution (CPD) associated with each node in the model needs to be learned from the training set. The database regularly labels the image expression and the intermediate layer shows the expression labels resulting from different kinds of features, the values of the intermediate layer are omitted from the training set. Therefore, we divide the parameter learning phase in two phases. In the first phase, the expression labels are taken as the value of nodes in the intermediate expression layer to learn the CPDs of nodes in the feature layer. The CPDs of the fea-

5 236 Front. Comput. Sci., 2014, 8(2): ture layer are parameterized by a multivariate Gaussian distribution. Specifically, for a visible features node, let F v denote the visible features and E v denote the value of visible expression node, and assume that the CPD of the visible features p(f v E v = k) satisfy a multivariate Gaussian distribution with corresponding mean vector μ k and covariance matrix k.the CPD of visible features can be represented by Eq. (5): exp{ (F v μ k ) T (F v μ k )} k p(f v E v = k) =. (5) (2π) n/2 1/2 Now we can learn parameters μ k and k given expression E v = k by using the maximum likelihood estimation as shown in Eqs. (6) and (7): ˆ = 1 k N k ˆμ k = 1 N k N k i=1 N k i=1 k x ik, (6) (x ik ˆμ k )(x ik ˆμ k ) T, (7) where N k is the number of samples of expression k, andx ik is the feature vector of ith samples of expression k. Subsequently, we use the features in the training set as evidence to get the intermediate results according to the CPDs of the feature nodes. In the second phase, the conditional probabilistic table (CPT) for each node in the intermediate expression layer is calculated based on the analysis of real expressions and the previous intermediate results. Specifically, for the visible expressions node: p(e v = i E = j) = N ji N j, (8) where N j is the number of samples of expression j in the training set, and N ji denotes the number of samples of expression j whose visible intermediate result is i. Given the BN model and a query image, we find the expression for the image according to Eq. (9): E = argmax p(e)p(f v, F ir E) E = argmax p(e) p(e v = k E)p(F v E v = k) E k j p(e ir = j E)p(F ir E ir = j), (9) where E v and E ir denote the states of the visible expressions node and the IR expressions node respectively, F v and F ir are the feature vectors of visible and IR images. Therefore, the true state of an expression can be derived by probabilistic inference Feature-level fusion using BN For feature-level fusion using BN, we propose to construct a simplified graphic model of a naive Bayes classifier to combine visible and IR features at the feature-level as shown in Fig. 3(b). The model consists of two nodes: the expression node and the feature node. The expression node is a discrete node with N states, where each state corresponds to an expression to be recognized. The feature node is a continuous node, representing a combined feature vector of visible features and IR features. The CPD associated with feature node is parameterized as a multivariate Gaussian distribution, and parameters are estimated using maximum likelihood estimation. In the testing phase, a new sample is recognized as the one with the maximum likelihood, i.e., E = argmax p(e)p(f E). (10) E SVM models of facial expression recognition Decision-level fusion using SVM For SVM-based decision-level fusion, two linear SVM classifiers are first adopted to estimate probabilities of expressions from visible spectrum images and IR images. The pairwise coupling is used for multi-class probability estimation with the SVM [29]. Then, the recognition results of the two modalities are combined by another linear SVM to obtain the final expression Feature-level fusion using SVM For SVM-based feature-level fusion, we directly concatenate the visible features and infrared features into a higher dimension feature vector, and feed it into a linear SVM for classification [30]. 4 Experiments and analysis 4.1 Experimental conditions To evaluate the effectiveness of the proposed methods, experiments are performed on the data chosen from the NVIE spontaneous database [12]. The NVIE spontaneous database contains both visible and IR images of more than 100 subjects. All visible expression sequences have been labeled by five students on the intensity of six basic facial expressions (happiness, sadness, surprise, fear, anger, and disgust) with a

6 Shangfei WANG et al. Fusion of visible and thermal images for facial expression recognition 237 three point scale (0, 1, or 2). The expression with the highest average intensity is used as the expression label for the visible and IR image sequences. In this paper, we choose experimental samples under the following criteria. First, the average intensity associated with a sample label must be larger than 1. Second, as three expressions (happiness, fear, and disgust) were successfully induced in most cases when the NVIE spontaneous database was constructed, the sample label should be one of happiness, fear, and disgust. Third, the sample should consist of both visible and IR image sequences. Following these criteria, we obtain a total of 535 samples from 123 subjects under different illuminations. The size of the located facial region of the IR images is normalized to The size of grids is set to 4 4. There are a total of IR features extracted. Before extracting AAM features, experimental samples are equally divided into two groups according to their illuminations. At each time, we take one group as the training set to train AAM. Sixty one landmark points are manually marked on each sample of the training set as shown in Fig. 2(a). Then samples in the other group are marked using the learned AAM. We find a total of 33 visible features including three head motion features and 30 AAM features. Our experimental results are obtained by applying a tenfold cross validation method on all samples according to the subjects. Owing to the high dimension of IR features, we perform IR feature selection. Another ten-fold cross validation is applied on the training set to choose the best dimension of IR features that achieves the highest accuracy rate on the validated set. We compare the proposed methods with that of using visible or IR features alone. 4.2 Feature analyses The F-statistics of a feature reflect that feature s importance to facial expression recognition. We select features according to the calculated F-statistics from high to low. We analyze the top 50 IR features according to the F-statistics calculated based on expression labels. As shown in Fig. 4, we can find that most of top 50 selected IR features are extracted from the mouth and cheek regions. This seems to indicate that the temperature variations of the mouth and cheek regions are the most reliable sources compared to other facial regions for facial expression recognition. We also sort the visible features according to their F-statistics. We find that head motion features are high on the selected order of visible features, which demonstrates that head motion information is highly related to facial expression. Fig. 4 Counts of features from different facial area 4.3 Experimental results of Bayesian networks Due to the poor performance by using IR features alone, we consider IR features as a supplement to visible facial expression recognition. Figs. 5(a) and 5(b) show the fusion recognition accuracy curves of the decision-level and feature-level fusion using BN with different numbers of IR features on the validation set. Bold lines show the average accuracy curves. For decision-level fusion, the average accuracy curve reaches the highest accuracy when 15 IR features are selected. For feature-level fusion, of the ten folds, the highest accuracies are reached when fewer IR features are selected. The confusion matrices of the different methods using BNs are shown in Table 1. Each cell represents the number of times the the row emotion (actual) in classified as the columnemotion(classification result). The overall recognition accuracy of decision-level fusion and feature-level fusion are 72.15% and 73.83%, respectively. The recognition accuracies are 71.59% and 46.54% for visual spectrum images and IR images, respectively, when used alone. When only IR features are used, the overall recognition accuracy is poor, especially for distinguishing fear from disgust. When only visible features are used, the performance is much better. However, it also has certain limitations, such as, it is much more likely to misclassify disgust as happiness than using IR features. Using decision-level and feature-level fusion, the overall recognition accuracy is improved, especially for fear. IR features play a supplementary role in visible facial expression recognition. And it seems that feature-level fusion is more effective than decision-level fusion. To further assess the essence of the effectiveness of the fusion models, we calculate Cohen s kappa coefficients, which is a statistical measure of inter-rating agreement or interannotation agreement for categorical items and represents that bias is greater when kappa is smaller, for the four methods: considering both correctly and incorrectly classified samples. The Cohen s kappa coefficients for the decisionlevel fusion method, feature-level fusion method, visible alone, and IR alone are 0.572, 0.597, 0.562, and 0.197,

7 238 Front. Comput. Sci., 2014, 8(2): Fig. 5 Expression recognition accuracy with different number of selected IR features using (a) BN-based decision-level fusion; (b) BN-based feature-level fusion; (c) SVM-based decision-level fusion; (d) SVM-based feature-level fusion Table 1 Confusion matrices of the methods using Bayesian networks Actual emotion Decision level classification Feature level classification Visible feature classification IR feature classification Disgust Fear Happiness Disgust Fear Happiness Disgust Fear Happiness Disgust Fear Happiness Disgust Fear Happiness Accuracy% Kappa Discrepancy respectively. We conclude that the classification accuracy is improved by adding IR features. Another useful statistics is to measure the discrepancy in the kappa scores for the three expression types. A highly discrepant model might be excellent at classifying certain expressions but poor for others. Here we calculate the standard deviation of kappa scores for the three expressions. The discrepancies for the decision-level fusion method, the feature-level fusion method, the visible alone, and the IR alone are 0.097, 0.102, 0.106, and 0.117, respectively, which again indicates the effectiveness of the fusion methods according to the low value of standard deviation. From the previous experiments, we can see that proposed methods can successfully capture the supplementary information in IR images to help improve visible facial expression recognition. 4.4 Experimental results of SVMs Figures. 5(c) and 5(d) show the accuracy curves of decisionlevel and feature-level fusion using SVMs with different numbers of IR features on the validation set. Compared with those of the BN-based methods, the fusion accuracies on the validation set of the SVM-based methods change slower as the number of selected IR features increases. For decisionlevel fusion, the average accuracy curve reaches the highest accuracy when about 55 IR features are selected. For

8 Shangfei WANG et al. Fusion of visible and thermal images for facial expression recognition 239 Table 2 Confusion matrices of methods using SVMs Actual emotion Decision level classification Feature level classification Visible feature classification IR feature classification Disgust Fear Happiness Disgust Fear Happiness Disgust Fear Happiness Disgust Fear Happiness Disgust Fear Happiness Accuracy /% Kappa Discrepancy feature-level fusion, of the ten folds, the highest accuracies are reached when fewer IR features are selected, which is similar to that of BN. The recognition results of methods using SVMs are shown in Table 2. The overall recognition accuracy of decisionlevel and feature-level fusion are 76.45% and 76.82%, respectively. The overall recognition accuracies for separately using visible and IR features are 75.33% and 52.90%, respectively. The overall recognition accuracy is improved by using decision-level and feature-level fusion, especially in the case of recognizing fear. We also calculate the kappa coefficients and discrepancies for the different methods, as shown in Table 2. From the view of both kappa coefficient and discrepancy, the performance of facial expression recognition is improved by both decision-level fusion and feature-level fusion compared to each modality. Furthermore, the performance of feature-level fusion is better that that of the decision-level fusion. The above analysis demonstrates the supplementary role of IR images in improving facial expression recognition. Comparing the SVM-based methods with the BN methods, we find that, as a typical discriminative learning method, SVM achieves better performance than BN, which is a generative model, does. Without considering the computational issues and the handling of missing data, the literature suggests that discriminative classifiers are almost superior to generative ones [31, 32]. For BN-based fusion, the accuracy is improved from 71.59%, to 72.15% and 73.83%, respectively by using decision-level and feature-level fusion. The accuracy is increased by an average of 1.40%. For SVM-based fusion, the accuracy is improved from 75.33%, to 76.45% and 76.82% by respectively using decision-level and feature-level fusion. The accuracy is increased by an average of 1.30%. The average improvement for BN-based fusion is a little higher than that of SVM. This may indicate that, as a probabilistic model, BN can effectively capture uncertainties in data and allows data from different modalities to be systematically represented and integrated. 4.5 Comparison with related works In this section we compare the performance of our proposed methods with related works. Currently, there are only three reported studies of facial expression recognition based on the fusion of IR and visible images [7 9]. The database used in [7] is not publicly available. Furthermore, the method in [7] also uses audio signal alongside IR and visible images. Therefore, it is difficult to perform a fair comparison with the method in [7] since our method uses only IR and visible images. The other two studies [8, 9] perform experiments on the NVIE database. Thus, we are able to compare results directly. Table 3 compares our results with those in [8, 9]. We can see that our methods outperform theirs in both accuracy and kappa coefficient. In [8] and [9], by introducing infrared features, the accuracies are increased by 1.90% and 1.10% respectively. For the proposed method BN-based feature-level fusion, the accuracy is increased by 2.20%; this demonstrates that our proposed methods can better capture the supplementary information in IR images. Table 3 Compared with other fusion methods Methods Accuracy /% Kappa BN Decision-level fusion BN Feature-level fusion SVM Decision-level fusion SVM Feature-level fusion F-statistic & knn [8] Multiple GAs & knn [9] Conclusions In this paper, we have proposed the fusion of visible features and IR features for facial expression recognition. Specifically, two BN and two SVM based methods are proposed to perform decision-level fusion and feature-level fusion, respectively. Our experiments on the NVIE spontaneous database show that our methods yield an average improvement of 1.35% in performance of facial expression recognition com-

9 240 Front. Comput. Sci., 2014, 8(2): pared with using only visible features, especially for fear. We found that feature-level fusion is more effect than the decision-level fusion. The results of IR feature selection indicate that the temperature variations of the mouth and cheek regions are the most reliable sources compared to other facial regions for facial expression recognition. We performed a comparative study that also demonstrats the superiority of our method over existing techniques. Current research into expression recognition from thermal infrared images has adopted either temperature statistical parameters extracted from the facial regions of interest or several hand-crafted features that are commonly used in visible spectrum [23, 33]. So far there are no special features designed for thermal infrared images. The immaturities of IR features may lessen the role of IR images for improving visible facial expression recognition. In the future, we intend to investigate specific thermal features. A potential solution is to learn features using deep learning methods. Furthermore, a larger scale expression database consisting of both visible and thermal images is needed. Besides studying fusion of different image modalities, we plan to also investigate fusion of different types of features such as fusion of Scale-invariant feature transform (SIFT) and Gabor features, and visible images recorded under different illumination conditions. Acknowledgements This paper is supported by the National Natural Science Foundation of China (Grant Nos , ), Special Innovation Project on Speech of Anhui Province ( ), Project from Anhui Science and Technology Agency(1106c ) and the Fundamental Research Funds for the Central Universities. We also acknowledge partial support from the US National Science Foundation ( ). References 1. Zeng Z, Pantic M, Roisman G, Huang T. A survey of affect recognition methods: audio, visual, and spontaneous expressions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2009, 31(1): Wang Z, Liu F, Wang L. Survey of facial expression recognition based on computer vision. Computer Engineering, 2006, 32(11): Liu X m, Tan H c, Zhang Y j. New research advances in facial expression recogntion. Journal of Image and Graphics, 2006, 11(10): Xue Y L, Mao X, Guo Y, Lv S W. The research advance of facial expression recognition in human computer interaction. Journal of Image and Graphics, 2009, 14(5): Bettadapura V. Face expression recognition and analysis: The state of the art. CoRR, 2012, abs/ Wesley A, Buddharaju P, Pienta R, Pavlidis I. A comparative analysis of thermal and visual modalities for automated facial expression recognition. Advances in Visual Computing, 2012, Yoshitomi Y,KimS,KawanoT,KilazoeT.Effect of sensor fusion for recognition of emotional states using voice, face image and thermal image of face. In: Proceedings of the 9th IEEE International Workshop on Robot and Human Interactive Communication. 2000, Wang Z, Wang S. Spontaneous facial expression recognition by using feature-level fusion of visible and thermal infrared images. In: Proceedings of the 2011 IEEE International Workshop on Machine Learning for Signal Processing. 2011, Wang S, He S. Spontaneous facial expression recognition by fusing thermal infrared and visible images. Intelligent Autonomous Systems, 2013, 194: Cootes T, Edwards G, Taylor C. Active appearance models. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2001, 23(6): Lv Y, Wang S. A spontaneous facial expression recognition method using head motion and AAM features. In: Proceedings of the 2nd World Congress on Nature and Biologically Inspired Computing. 2010, WangS,LiuZ,LvS,LvY,WuG,PengP,ChenF,WangX.Anatural visible and infrared facial expression database for expression recognition and emotion inference. IEEE Transactions on Multimedia, 2010, 12(7): Littlewort G, Whitehill J, Wu T, Butko N, Ruvolo P, Movellan J, Bartlett M. The motion in emotion a cert based approach to the fera emotion challenge. In: Proceedings of the 2011 IEEE International Conference on Automatic Face & Gesture Recognition and Workshops. 2011, Cohn J, Reed L, Ambadar Z, Xiao J, Moriyama T. Automatic analysis and recognition of brow actions and head motion in spontaneous facial behavior. In: Proceedings of the 2004 IEEE International Conference on Systems, Man and Cybernetics. 2004, Tong Y, Chen J, Ji Q. A unified probabilistic framework for spontaneous facial action modeling and understanding. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2010, 32(2): Valstar M, Gunes H, Pantic M. How to distinguish posed from spontaneous smiles using geometric features. In: Proceedings of the 9th international conference on Multimodal interfaces. 2007, Gunes H, Pantic M. Dimensional emotion prediction from spontaneous head gestures for interaction with sensitive artificial listeners. In: Proceedings of the 10th International Conference on Intelligent Virtual Agents. 2010, 6356: Jarlier S, Grandjean D, Delplanque S, N Diaye K, Cayeux I, Velazco M, Sander D, Vuilleumier P, Scherer K. Thermal analysis of facial muscles contractions. IEEE Transactions on Affective Computing, 2011, 2(1): Hernández B, Olague G, Hammoud R, Trujillo L, Romero E. Visual learning of texture descriptors for facial expression recognition in thermal imagery. Computer Vision and Image Understanding, 2007, 106(2): Khan M, Ward R, Ingleby M. Classifying pretended and evoked facial expressions of positive and negative affective states using infrared measurement of skin temperature. ACM Transactions on Applied Perception, 2009, 6(1): Article Yoshitomi Y. Facial expression recognition for speaker using thermal image processing and speech recognition system. In: Proceedings of the 10th World Scientific and Engineering Academy and Society Inter-

10 Shangfei WANG et al. Fusion of visible and thermal images for facial expression recognition 241 national Conference on Applied Computer Science. 2010, Puri C, Olson L, Pavlidis I, Levine J, Starren J. Stresscam: non-contact measurement of users emotional states through thermal imaging. In: Proceedings of the 2005 Conference on Human Factors in Computing Systems. 2005, Shen P, Wang S, Liu Z. Facial expression recognition from infrared thermal videos. Intelligent Autonomous Systems, 2013, 194: Buddharaju P, Pavlidis I, Manohar C. Face recognition beyond the visible spectrum. In: Advances in Biometrics, Springer, Pavlidis I, Levine J, Baukol P. Thermal image analysis for anxiety detection. In: Proceedings of the 2001 International Conference on Image Processing. 2001, Cootes T. am_tools. bim/software/am_ tools_doc/ 27. Tong Y, Wang Y, Zhu Z, Ji Q. Robust facial feature tracking under varying face pose and facial expression. Pattern Recognition, 2007, 40(11): Ding C. Analysis of gene expression profiles: class discovery and leaf ordering. In: Proceedings of the 6th Annual International Conference on Computational Biology. 2002, Wu T F, Lin C J, Weng R C. Probability estimates for multi-class classification by pairwise coupling. The Journal of Machine Learning Research, 2004, 5: Chang C C, Lin C J. LIBSVM: a library for support vector machines. ACM Transactions on Intelligent Systems and Technology, 2011, 2: 27:1 27:27. Software available at tw/ cjlin/libsvm 31. Long P M, Servedio R A. Discriminative learning can succeed where generative learning fails. In: Proceedings of the 19th Annual Conference on Learning Theory. 2006, 4005: Ng A Y, Jordan A. On discriminative vs. generative classifiers: a comparison of logistic regression and naive bayes. Advances in Neural Information Processing Systems, 2002, 14: Hermosilla G, Ruiz-del-Solar J, Verschae R, Correa M. A comparative study of thermal face recognition methods in unconstrained environments. Pattern Recognition, 2012, 45(7): Shangfei Wang received her MS in circuits and systems, and her PhD in signal and information processing from University of Science and Technology of China (USTC), China in 1999 and 2002, respectively. From 2004 to 2005, she was a postdoctoral research fellow in Kyushu University, Japan. She is currently an associate professor in the School of Computer Science and Technology, USTC. Dr. Wang is an IEEE member. Her research interests cover computer intelligence, affective computing, multimedia computing, information retrieval, and artificial environment design. She has authored or co-authored over 50 publications. Shan He received his BS in Computer Science from Anhui Agriculture University, China in He received his MS in Computer Science from the University of Science and Technology of China, China in His research interest is affective computing. Yue Wu is a PhD candidate in the Department of Electrical, Computer, and Systems Engineering, Rensselaer Polytechnic Institute, USA. Her research interest is computer vision. Menghua He received her BS in Information and Computation Science from Anhui University, China in She is currently pursuing her MS in Computer Science at the University of Science and Technology of China, China. Her research interesting is affective computing. Qiang Ji received his PhD in Electrical Engineering from the University of Washington, USA. He is currently a professor with the Department of Electrical, Computer, and Systems Engineering at Rensselaer Polytechnic Institute (RPI), USA. He recently served as a program director at the National Science Foundation (NSF), where he managed NSF s computer vision and machine learning programs. He also held teaching and research positions with the Beckman Institute at University of Illinois at Urbana-Champaign, the Robotics Institute at Carnegie Mellon University, the Dept. of Computer Science at University of Nevada at Reno, and the US Air Force Research Laboratory, USA. Prof. Ji currently serves as the director of the Intelligent Systems Laboratory (ISL) at RPI, USA. Prof. Ji s research interests are in computer vision, probabilis-

11 242 Front. Comput. Sci., 2014, 8(2): tic graphical models, information fusion, and their applications in various fields. He has published over 160 papers in peer-reviewed journals and conferences. His research has been supported by major governmental agencies including NSF, NIH, DARPA, ONR, ARO, and AFOSR as well as by major companies including Honda and Boeing. Prof. Ji is an editor on several related IEEE and international journals and he has served as a general chair, program chair, technical area chair, and program committee member in numerous international conferences/workshops. Prof. Ji is a fellow of IAPR.

Facial expression recognition with spatiotemporal local descriptors

Facial expression recognition with spatiotemporal local descriptors Facial expression recognition with spatiotemporal local descriptors Guoying Zhao, Matti Pietikäinen Machine Vision Group, Infotech Oulu and Department of Electrical and Information Engineering, P. O. Box

More information

Emotion Recognition using a Cauchy Naive Bayes Classifier

Emotion Recognition using a Cauchy Naive Bayes Classifier Emotion Recognition using a Cauchy Naive Bayes Classifier Abstract Recognizing human facial expression and emotion by computer is an interesting and challenging problem. In this paper we propose a method

More information

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information C. Busso, Z. Deng, S. Yildirim, M. Bulut, C. M. Lee, A. Kazemzadeh, S. Lee, U. Neumann, S. Narayanan Emotion

More information

Study on Aging Effect on Facial Expression Recognition

Study on Aging Effect on Facial Expression Recognition Study on Aging Effect on Facial Expression Recognition Nora Algaraawi, Tim Morris Abstract Automatic facial expression recognition (AFER) is an active research area in computer vision. However, aging causes

More information

Face Analysis : Identity vs. Expressions

Face Analysis : Identity vs. Expressions Hugo Mercier, 1,2 Patrice Dalle 1 Face Analysis : Identity vs. Expressions 1 IRIT - Université Paul Sabatier 118 Route de Narbonne, F-31062 Toulouse Cedex 9, France 2 Websourd Bâtiment A 99, route d'espagne

More information

Facial Event Classification with Task Oriented Dynamic Bayesian Network

Facial Event Classification with Task Oriented Dynamic Bayesian Network Facial Event Classification with Task Oriented Dynamic Bayesian Network Haisong Gu Dept. of Computer Science University of Nevada Reno haisonggu@ieee.org Qiang Ji Dept. of ECSE Rensselaer Polytechnic Institute

More information

Gender Based Emotion Recognition using Speech Signals: A Review

Gender Based Emotion Recognition using Speech Signals: A Review 50 Gender Based Emotion Recognition using Speech Signals: A Review Parvinder Kaur 1, Mandeep Kaur 2 1 Department of Electronics and Communication Engineering, Punjabi University, Patiala, India 2 Department

More information

Effect of Sensor Fusion for Recognition of Emotional States Using Voice, Face Image and Thermal Image of Face

Effect of Sensor Fusion for Recognition of Emotional States Using Voice, Face Image and Thermal Image of Face Effect of Sensor Fusion for Recognition of Emotional States Using Voice, Face Image and Thermal Image of Face Yasunari Yoshitomi 1, Sung-Ill Kim 2, Takako Kawano 3 and Tetsuro Kitazoe 1 1:Department of

More information

Dimensional Emotion Prediction from Spontaneous Head Gestures for Interaction with Sensitive Artificial Listeners

Dimensional Emotion Prediction from Spontaneous Head Gestures for Interaction with Sensitive Artificial Listeners Dimensional Emotion Prediction from Spontaneous Head Gestures for Interaction with Sensitive Artificial Listeners Hatice Gunes and Maja Pantic Department of Computing, Imperial College London 180 Queen

More information

Facial Expression Biometrics Using Tracker Displacement Features

Facial Expression Biometrics Using Tracker Displacement Features Facial Expression Biometrics Using Tracker Displacement Features Sergey Tulyakov 1, Thomas Slowe 2,ZhiZhang 1, and Venu Govindaraju 1 1 Center for Unified Biometrics and Sensors University at Buffalo,

More information

A Common Framework for Real-Time Emotion Recognition and Facial Action Unit Detection

A Common Framework for Real-Time Emotion Recognition and Facial Action Unit Detection A Common Framework for Real-Time Emotion Recognition and Facial Action Unit Detection Tobias Gehrig and Hazım Kemal Ekenel Facial Image Processing and Analysis Group, Institute for Anthropomatics Karlsruhe

More information

This is the accepted version of this article. To be published as : This is the author version published as:

This is the accepted version of this article. To be published as : This is the author version published as: QUT Digital Repository: http://eprints.qut.edu.au/ This is the author version published as: This is the accepted version of this article. To be published as : This is the author version published as: Chew,

More information

A Comparison of Collaborative Filtering Methods for Medication Reconciliation

A Comparison of Collaborative Filtering Methods for Medication Reconciliation A Comparison of Collaborative Filtering Methods for Medication Reconciliation Huanian Zheng, Rema Padman, Daniel B. Neill The H. John Heinz III College, Carnegie Mellon University, Pittsburgh, PA, 15213,

More information

Facial Expression Recognition Using Principal Component Analysis

Facial Expression Recognition Using Principal Component Analysis Facial Expression Recognition Using Principal Component Analysis Ajit P. Gosavi, S. R. Khot Abstract Expression detection is useful as a non-invasive method of lie detection and behaviour prediction. However,

More information

Neuro-Inspired Statistical. Rensselaer Polytechnic Institute National Science Foundation

Neuro-Inspired Statistical. Rensselaer Polytechnic Institute National Science Foundation Neuro-Inspired Statistical Pi Prior Model lfor Robust Visual Inference Qiang Ji Rensselaer Polytechnic Institute National Science Foundation 1 Status of Computer Vision CV has been an active area for over

More information

Formulating Emotion Perception as a Probabilistic Model with Application to Categorical Emotion Classification

Formulating Emotion Perception as a Probabilistic Model with Application to Categorical Emotion Classification Formulating Emotion Perception as a Probabilistic Model with Application to Categorical Emotion Classification Reza Lotfian and Carlos Busso Multimodal Signal Processing (MSP) lab The University of Texas

More information

A MULTIMODAL NONVERBAL HUMAN-ROBOT COMMUNICATION SYSTEM ICCB 2015

A MULTIMODAL NONVERBAL HUMAN-ROBOT COMMUNICATION SYSTEM ICCB 2015 VI International Conference on Computational Bioengineering ICCB 2015 M. Cerrolaza and S.Oller (Eds) A MULTIMODAL NONVERBAL HUMAN-ROBOT COMMUNICATION SYSTEM ICCB 2015 SALAH SALEH *, MANISH SAHU, ZUHAIR

More information

Age Estimation based on Multi-Region Convolutional Neural Network

Age Estimation based on Multi-Region Convolutional Neural Network Age Estimation based on Multi-Region Convolutional Neural Network Ting Liu, Jun Wan, Tingzhao Yu, Zhen Lei, and Stan Z. Li 1 Center for Biometrics and Security Research & National Laboratory of Pattern

More information

Audio-visual Classification and Fusion of Spontaneous Affective Data in Likelihood Space

Audio-visual Classification and Fusion of Spontaneous Affective Data in Likelihood Space 2010 International Conference on Pattern Recognition Audio-visual Classification and Fusion of Spontaneous Affective Data in Likelihood Space Mihalis A. Nicolaou, Hatice Gunes and Maja Pantic, Department

More information

Statistical and Neural Methods for Vision-based Analysis of Facial Expressions and Gender

Statistical and Neural Methods for Vision-based Analysis of Facial Expressions and Gender Proc. IEEE Int. Conf. on Systems, Man and Cybernetics (SMC 2004), Den Haag, pp. 2203-2208, IEEE omnipress 2004 Statistical and Neural Methods for Vision-based Analysis of Facial Expressions and Gender

More information

A Vision-based Affective Computing System. Jieyu Zhao Ningbo University, China

A Vision-based Affective Computing System. Jieyu Zhao Ningbo University, China A Vision-based Affective Computing System Jieyu Zhao Ningbo University, China Outline Affective Computing A Dynamic 3D Morphable Model Facial Expression Recognition Probabilistic Graphical Models Some

More information

COMPARISON BETWEEN GMM-SVM SEQUENCE KERNEL AND GMM: APPLICATION TO SPEECH EMOTION RECOGNITION

COMPARISON BETWEEN GMM-SVM SEQUENCE KERNEL AND GMM: APPLICATION TO SPEECH EMOTION RECOGNITION Journal of Engineering Science and Technology Vol. 11, No. 9 (2016) 1221-1233 School of Engineering, Taylor s University COMPARISON BETWEEN GMM-SVM SEQUENCE KERNEL AND GMM: APPLICATION TO SPEECH EMOTION

More information

Development of novel algorithm by combining Wavelet based Enhanced Canny edge Detection and Adaptive Filtering Method for Human Emotion Recognition

Development of novel algorithm by combining Wavelet based Enhanced Canny edge Detection and Adaptive Filtering Method for Human Emotion Recognition International Journal of Engineering Research and Development e-issn: 2278-067X, p-issn: 2278-800X, www.ijerd.com Volume 12, Issue 9 (September 2016), PP.67-72 Development of novel algorithm by combining

More information

Design of Palm Acupuncture Points Indicator

Design of Palm Acupuncture Points Indicator Design of Palm Acupuncture Points Indicator Wen-Yuan Chen, Shih-Yen Huang and Jian-Shie Lin Abstract The acupuncture points are given acupuncture or acupressure so to stimulate the meridians on each corresponding

More information

Active User Affect Recognition and Assistance

Active User Affect Recognition and Assistance Active User Affect Recognition and Assistance Wenhui Liao, Zhiwe Zhu, Markus Guhe*, Mike Schoelles*, Qiang Ji, and Wayne Gray* Email: jiq@rpi.edu Department of Electrical, Computer, and System Eng. *Department

More information

EECS 433 Statistical Pattern Recognition

EECS 433 Statistical Pattern Recognition EECS 433 Statistical Pattern Recognition Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1 / 19 Outline What is Pattern

More information

Blue Eyes Technology

Blue Eyes Technology Blue Eyes Technology D.D. Mondal #1, Arti Gupta *2, Tarang Soni *3, Neha Dandekar *4 1 Professor, Dept. of Electronics and Telecommunication, Sinhgad Institute of Technology and Science, Narhe, Maharastra,

More information

EMOTION RECOGNITION FROM USERS EEG SIGNALS WITH THE HELP OF STIMULUS VIDEOS

EMOTION RECOGNITION FROM USERS EEG SIGNALS WITH THE HELP OF STIMULUS VIDEOS EMOTION RECOGNITION FROM USERS EEG SIGNALS WITH THE HELP OF STIMULUS VIDEOS Yachen Zhu, Shangfei Wang School of Computer Science and Technology University of Science and Technology of China 2327, Hefei,

More information

Generalization of a Vision-Based Computational Model of Mind-Reading

Generalization of a Vision-Based Computational Model of Mind-Reading Generalization of a Vision-Based Computational Model of Mind-Reading Rana el Kaliouby and Peter Robinson Computer Laboratory, University of Cambridge, 5 JJ Thomson Avenue, Cambridge UK CB3 FD Abstract.

More information

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) Color Feature Maps (BY, RG) Color Saliency Map Input Image (I) Linear Filtering and Gaussian

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) Color Feature Maps (BY, RG) Color Saliency Map Input Image (I) Linear Filtering and Gaussian The 29th Fuzzy System Symposium (Osaka, September 9-, 3) A Fuzzy Inference Method Based on Saliency Map for Prediction Mao Wang, Yoichiro Maeda 2, Yasutake Takahashi Graduate School of Engineering, University

More information

Facial Behavior as a Soft Biometric

Facial Behavior as a Soft Biometric Facial Behavior as a Soft Biometric Abhay L. Kashyap University of Maryland, Baltimore County 1000 Hilltop Circle, Baltimore, MD 21250 abhay1@umbc.edu Sergey Tulyakov, Venu Govindaraju University at Buffalo

More information

Facial Emotion Recognition with Facial Analysis

Facial Emotion Recognition with Facial Analysis Facial Emotion Recognition with Facial Analysis İsmail Öztel, Cemil Öz Sakarya University, Faculty of Computer and Information Sciences, Computer Engineering, Sakarya, Türkiye Abstract Computer vision

More information

A Study of Facial Expression Reorganization and Local Binary Patterns

A Study of Facial Expression Reorganization and Local Binary Patterns A Study of Facial Expression Reorganization and Local Binary Patterns Poonam Verma #1, Deepshikha Rathore *2 #1 MTech Scholar,Sanghvi Innovative Academy Indore *2 Asst.Professor,Sanghvi Innovative Academy

More information

Automatic Facial Action Unit Recognition by Modeling Their Semantic And Dynamic Relationships

Automatic Facial Action Unit Recognition by Modeling Their Semantic And Dynamic Relationships Chapter 10 Automatic Facial Action Unit Recognition by Modeling Their Semantic And Dynamic Relationships Yan Tong, Wenhui Liao, and Qiang Ji Abstract A system that could automatically analyze the facial

More information

Intelligent Edge Detector Based on Multiple Edge Maps. M. Qasim, W.L. Woon, Z. Aung. Technical Report DNA # May 2012

Intelligent Edge Detector Based on Multiple Edge Maps. M. Qasim, W.L. Woon, Z. Aung. Technical Report DNA # May 2012 Intelligent Edge Detector Based on Multiple Edge Maps M. Qasim, W.L. Woon, Z. Aung Technical Report DNA #2012-10 May 2012 Data & Network Analytics Research Group (DNA) Computing and Information Science

More information

A framework for the Recognition of Human Emotion using Soft Computing models

A framework for the Recognition of Human Emotion using Soft Computing models A framework for the Recognition of Human Emotion using Soft Computing models Md. Iqbal Quraishi Dept. of Information Technology Kalyani Govt Engg. College J Pal Choudhury Dept. of Information Technology

More information

HUMAN EMOTION RECOGNITION USING THERMAL IMAGE PROCESSING AND EIGENFACES

HUMAN EMOTION RECOGNITION USING THERMAL IMAGE PROCESSING AND EIGENFACES ISSN 2277-2685 IJESR/Jan 2015/ Vol-5/Issue-1/34-40 Vishwanath Bijalwan et al./ International Journal of Engineering & Science Research HUMAN EMOTION RECOGNITION USING THERMAL IMAGE PROCESSING AND EIGENFACES

More information

Get The FACS Fast: Automated FACS face analysis benefits from the addition of velocity

Get The FACS Fast: Automated FACS face analysis benefits from the addition of velocity Get The FACS Fast: Automated FACS face analysis benefits from the addition of velocity Timothy R. Brick University of Virginia Charlottesville, VA 22904 tbrick@virginia.edu Michael D. Hunter University

More information

The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression

The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression Patrick Lucey 1,2, Jeffrey F. Cohn 1,2, Takeo Kanade 1, Jason Saragih 1, Zara Ambadar 2 Robotics

More information

Valence-arousal evaluation using physiological signals in an emotion recall paradigm. CHANEL, Guillaume, ANSARI ASL, Karim, PUN, Thierry.

Valence-arousal evaluation using physiological signals in an emotion recall paradigm. CHANEL, Guillaume, ANSARI ASL, Karim, PUN, Thierry. Proceedings Chapter Valence-arousal evaluation using physiological signals in an emotion recall paradigm CHANEL, Guillaume, ANSARI ASL, Karim, PUN, Thierry Abstract The work presented in this paper aims

More information

R Jagdeesh Kanan* et al. International Journal of Pharmacy & Technology

R Jagdeesh Kanan* et al. International Journal of Pharmacy & Technology ISSN: 0975-766X CODEN: IJPTFI Available Online through Research Article www.ijptonline.com FACIAL EMOTION RECOGNITION USING NEURAL NETWORK Kashyap Chiranjiv Devendra, Azad Singh Tomar, Pratigyna.N.Javali,

More information

TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING

TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING 134 TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING H.F.S.M.Fonseka 1, J.T.Jonathan 2, P.Sabeshan 3 and M.B.Dissanayaka 4 1 Department of Electrical And Electronic Engineering, Faculty

More information

Automated Tessellated Fundus Detection in Color Fundus Images

Automated Tessellated Fundus Detection in Color Fundus Images University of Iowa Iowa Research Online Proceedings of the Ophthalmic Medical Image Analysis International Workshop 2016 Proceedings Oct 21st, 2016 Automated Tessellated Fundus Detection in Color Fundus

More information

Comparison of Lip Image Feature Extraction Methods for Improvement of Isolated Word Recognition Rate

Comparison of Lip Image Feature Extraction Methods for Improvement of Isolated Word Recognition Rate , pp.57-61 http://dx.doi.org/10.14257/astl.2015.107.14 Comparison of Lip Image Feature Extraction Methods for Improvement of Isolated Word Recognition Rate Yong-Ki Kim 1, Jong Gwan Lim 2, Mi-Hye Kim *3

More information

doi: / _59(

doi: / _59( doi: 10.1007/978-3-642-39188-0_59(http://dx.doi.org/10.1007/978-3-642-39188-0_59) Subunit modeling for Japanese sign language recognition based on phonetically depend multi-stream hidden Markov models

More information

An assistive application identifying emotional state and executing a methodical healing process for depressive individuals.

An assistive application identifying emotional state and executing a methodical healing process for depressive individuals. An assistive application identifying emotional state and executing a methodical healing process for depressive individuals. Bandara G.M.M.B.O bhanukab@gmail.com Godawita B.M.D.T tharu9363@gmail.com Gunathilaka

More information

User Affective State Assessment for HCI Systems

User Affective State Assessment for HCI Systems Association for Information Systems AIS Electronic Library (AISeL) AMCIS 2004 Proceedings Americas Conference on Information Systems (AMCIS) 12-31-2004 Xiangyang Li University of Michigan-Dearborn Qiang

More information

ANALYSIS OF FACIAL FEATURES OF DRIVERS UNDER COGNITIVE AND VISUAL DISTRACTIONS

ANALYSIS OF FACIAL FEATURES OF DRIVERS UNDER COGNITIVE AND VISUAL DISTRACTIONS ANALYSIS OF FACIAL FEATURES OF DRIVERS UNDER COGNITIVE AND VISUAL DISTRACTIONS Nanxiang Li and Carlos Busso Multimodal Signal Processing (MSP) Laboratory Department of Electrical Engineering, The University

More information

Hierarchical Age Estimation from Unconstrained Facial Images

Hierarchical Age Estimation from Unconstrained Facial Images Hierarchical Age Estimation from Unconstrained Facial Images STIC-AmSud Jhony Kaesemodel Pontes Department of Electrical Engineering Federal University of Paraná - Supervisor: Alessandro L. Koerich (/PUCPR

More information

Affective pictures and emotion analysis of facial expressions with local binary pattern operator: Preliminary results

Affective pictures and emotion analysis of facial expressions with local binary pattern operator: Preliminary results Affective pictures and emotion analysis of facial expressions with local binary pattern operator: Preliminary results Seppo J. Laukka 1, Antti Rantanen 1, Guoying Zhao 2, Matti Taini 2, Janne Heikkilä

More information

On Shape And the Computability of Emotions X. Lu, et al.

On Shape And the Computability of Emotions X. Lu, et al. On Shape And the Computability of Emotions X. Lu, et al. MICC Reading group 10.07.2013 1 On Shape and the Computability of Emotion X. Lu, P. Suryanarayan, R. B. Adams Jr., J. Li, M. G. Newman, J. Z. Wang

More information

Improved Intelligent Classification Technique Based On Support Vector Machines

Improved Intelligent Classification Technique Based On Support Vector Machines Improved Intelligent Classification Technique Based On Support Vector Machines V.Vani Asst.Professor,Department of Computer Science,JJ College of Arts and Science,Pudukkottai. Abstract:An abnormal growth

More information

Online Speaker Adaptation of an Acoustic Model using Face Recognition

Online Speaker Adaptation of an Acoustic Model using Face Recognition Online Speaker Adaptation of an Acoustic Model using Face Recognition Pavel Campr 1, Aleš Pražák 2, Josef V. Psutka 2, and Josef Psutka 2 1 Center for Machine Perception, Department of Cybernetics, Faculty

More information

Task oriented facial behavior recognition with selective sensing

Task oriented facial behavior recognition with selective sensing Computer Vision and Image Understanding 100 (2005) 385 415 www.elsevier.com/locate/cviu Task oriented facial behavior recognition with selective sensing Haisong Gu a, Yongmian Zhang a, Qiang Ji b, * a

More information

Skin color detection for face localization in humanmachine

Skin color detection for face localization in humanmachine Research Online ECU Publications Pre. 2011 2001 Skin color detection for face localization in humanmachine communications Douglas Chai Son Lam Phung Abdesselam Bouzerdoum 10.1109/ISSPA.2001.949848 This

More information

Emotion Affective Color Transfer Using Feature Based Facial Expression Recognition

Emotion Affective Color Transfer Using Feature Based Facial Expression Recognition , pp.131-135 http://dx.doi.org/10.14257/astl.2013.39.24 Emotion Affective Color Transfer Using Feature Based Facial Expression Recognition SeungTaek Ryoo and Jae-Khun Chang School of Computer Engineering

More information

MRI Image Processing Operations for Brain Tumor Detection

MRI Image Processing Operations for Brain Tumor Detection MRI Image Processing Operations for Brain Tumor Detection Prof. M.M. Bulhe 1, Shubhashini Pathak 2, Karan Parekh 3, Abhishek Jha 4 1Assistant Professor, Dept. of Electronics and Telecommunications Engineering,

More information

Performance of Gaussian Mixture Models as a Classifier for Pathological Voice

Performance of Gaussian Mixture Models as a Classifier for Pathological Voice PAGE 65 Performance of Gaussian Mixture Models as a Classifier for Pathological Voice Jianglin Wang, Cheolwoo Jo SASPL, School of Mechatronics Changwon ational University Changwon, Gyeongnam 64-773, Republic

More information

1. INTRODUCTION. Vision based Multi-feature HGR Algorithms for HCI using ISL Page 1

1. INTRODUCTION. Vision based Multi-feature HGR Algorithms for HCI using ISL Page 1 1. INTRODUCTION Sign language interpretation is one of the HCI applications where hand gesture plays important role for communication. This chapter discusses sign language interpretation system with present

More information

Quality Assessment of Human Hand Posture Recognition System Er. ManjinderKaur M.Tech Scholar GIMET Amritsar, Department of CSE

Quality Assessment of Human Hand Posture Recognition System Er. ManjinderKaur M.Tech Scholar GIMET Amritsar, Department of CSE Quality Assessment of Human Hand Posture Recognition System Er. ManjinderKaur M.Tech Scholar GIMET Amritsar, Department of CSE mkwahla@gmail.com Astt. Prof. Prabhjit Singh Assistant Professor, Department

More information

Gray level cooccurrence histograms via learning vector quantization

Gray level cooccurrence histograms via learning vector quantization Gray level cooccurrence histograms via learning vector quantization Timo Ojala, Matti Pietikäinen and Juha Kyllönen Machine Vision and Media Processing Group, Infotech Oulu and Department of Electrical

More information

EMOTION DETECTION THROUGH SPEECH AND FACIAL EXPRESSIONS

EMOTION DETECTION THROUGH SPEECH AND FACIAL EXPRESSIONS EMOTION DETECTION THROUGH SPEECH AND FACIAL EXPRESSIONS 1 KRISHNA MOHAN KUDIRI, 2 ABAS MD SAID AND 3 M YUNUS NAYAN 1 Computer and Information Sciences, Universiti Teknologi PETRONAS, Malaysia 2 Assoc.

More information

Enhanced Facial Expressions Recognition using Modular Equable 2DPCA and Equable 2DPC

Enhanced Facial Expressions Recognition using Modular Equable 2DPCA and Equable 2DPC Enhanced Facial Expressions Recognition using Modular Equable 2DPCA and Equable 2DPC Sushma Choudhar 1, Sachin Puntambekar 2 1 Research Scholar-Digital Communication Medicaps Institute of Technology &

More information

VIDEO SURVEILLANCE AND BIOMEDICAL IMAGING Research Activities and Technology Transfer at PAVIS

VIDEO SURVEILLANCE AND BIOMEDICAL IMAGING Research Activities and Technology Transfer at PAVIS VIDEO SURVEILLANCE AND BIOMEDICAL IMAGING Research Activities and Technology Transfer at PAVIS Samuele Martelli, Alessio Del Bue, Diego Sona, Vittorio Murino Istituto Italiano di Tecnologia (IIT), Genova

More information

A Multilevel Fusion Approach for Audiovisual Emotion Recognition

A Multilevel Fusion Approach for Audiovisual Emotion Recognition A Multilevel Fusion Approach for Audiovisual Emotion Recognition Girija Chetty & Michael Wagner National Centre for Biometric Studies Faculty of Information Sciences and Engineering University of Canberra,

More information

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 5, JULY

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 5, JULY IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 5, JULY 2011 1057 A Framework for Automatic Human Emotion Classification Using Emotion Profiles Emily Mower, Student Member, IEEE,

More information

A Study on Automatic Age Estimation using a Large Database

A Study on Automatic Age Estimation using a Large Database A Study on Automatic Age Estimation using a Large Database Guodong Guo WVU Guowang Mu NCCU Yun Fu BBN Technologies Charles Dyer UW-Madison Thomas Huang UIUC Abstract In this paper we study some problems

More information

A Real-Time Human Stress Monitoring System Using Dynamic Bayesian Network

A Real-Time Human Stress Monitoring System Using Dynamic Bayesian Network A Real-Time Human Stress Monitoring System Using Dynamic Bayesian Network Wenhui Liao, Weihong Zhang, Zhiwei Zhu and Qiang Ji {liaow, zhangw9, zhuz, jiq}@rpi.edu Department of Electrical, Computer and

More information

IDENTIFICATION OF REAL TIME HAND GESTURE USING SCALE INVARIANT FEATURE TRANSFORM

IDENTIFICATION OF REAL TIME HAND GESTURE USING SCALE INVARIANT FEATURE TRANSFORM Research Article Impact Factor: 0.621 ISSN: 2319507X INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK IDENTIFICATION OF REAL TIME

More information

Recognition of facial expressions using Gabor wavelets and learning vector quantization

Recognition of facial expressions using Gabor wavelets and learning vector quantization Engineering Applications of Artificial Intelligence 21 (2008) 1056 1064 www.elsevier.com/locate/engappai Recognition of facial expressions using Gabor wavelets and learning vector quantization Shishir

More information

Bayesian Bi-Cluster Change-Point Model for Exploring Functional Brain Dynamics

Bayesian Bi-Cluster Change-Point Model for Exploring Functional Brain Dynamics Int'l Conf. Bioinformatics and Computational Biology BIOCOMP'18 85 Bayesian Bi-Cluster Change-Point Model for Exploring Functional Brain Dynamics Bing Liu 1*, Xuan Guo 2, and Jing Zhang 1** 1 Department

More information

Research Proposal on Emotion Recognition

Research Proposal on Emotion Recognition Research Proposal on Emotion Recognition Colin Grubb June 3, 2012 Abstract In this paper I will introduce my thesis question: To what extent can emotion recognition be improved by combining audio and visual

More information

N RISCE 2K18 ISSN International Journal of Advance Research and Innovation

N RISCE 2K18 ISSN International Journal of Advance Research and Innovation The Computer Assistance Hand Gesture Recognition system For Physically Impairment Peoples V.Veeramanikandan(manikandan.veera97@gmail.com) UG student,department of ECE,Gnanamani College of Technology. R.Anandharaj(anandhrak1@gmail.com)

More information

Facial Feature Model for Emotion Recognition Using Fuzzy Reasoning

Facial Feature Model for Emotion Recognition Using Fuzzy Reasoning Facial Feature Model for Emotion Recognition Using Fuzzy Reasoning Renan Contreras, Oleg Starostenko, Vicente Alarcon-Aquino, and Leticia Flores-Pulido CENTIA, Department of Computing, Electronics and

More information

Introduction to Computational Neuroscience

Introduction to Computational Neuroscience Introduction to Computational Neuroscience Lecture 5: Data analysis II Lesson Title 1 Introduction 2 Structure and Function of the NS 3 Windows to the Brain 4 Data analysis 5 Data analysis II 6 Single

More information

Introduction to affect computing and its applications

Introduction to affect computing and its applications Introduction to affect computing and its applications Overview What is emotion? What is affective computing + examples? Why is affective computing useful? How do we do affect computing? Some interesting

More information

Decision tree SVM model with Fisher feature selection for speech emotion recognition

Decision tree SVM model with Fisher feature selection for speech emotion recognition Sun et al. EURASIP Journal on Audio, Speech, and Music Processing (2019) 2019:2 https://doi.org/10.1186/s13636-018-0145-5 RESEARCH Decision tree SVM model with Fisher feature selection for speech emotion

More information

A Semi-supervised Approach to Perceived Age Prediction from Face Images

A Semi-supervised Approach to Perceived Age Prediction from Face Images IEICE Transactions on Information and Systems, vol.e93-d, no.10, pp.2875 2878, 2010. 1 A Semi-supervised Approach to Perceived Age Prediction from Face Images Kazuya Ueki NEC Soft, Ltd., Japan Masashi

More information

FEATURE EXTRACTION USING GAZE OF PARTICIPANTS FOR CLASSIFYING GENDER OF PEDESTRIANS IN IMAGES

FEATURE EXTRACTION USING GAZE OF PARTICIPANTS FOR CLASSIFYING GENDER OF PEDESTRIANS IN IMAGES FEATURE EXTRACTION USING GAZE OF PARTICIPANTS FOR CLASSIFYING GENDER OF PEDESTRIANS IN IMAGES Riku Matsumoto, Hiroki Yoshimura, Masashi Nishiyama, and Yoshio Iwai Department of Information and Electronics,

More information

Facial Expression Analysis for Estimating Pain in Clinical Settings

Facial Expression Analysis for Estimating Pain in Clinical Settings Facial Expression Analysis for Estimating Pain in Clinical Settings Karan Sikka University of California San Diego 9450 Gilman Drive, La Jolla, California, USA ksikka@ucsd.edu ABSTRACT Pain assessment

More information

Emotional Design. D. Norman (Emotional Design, 2004) Model with three levels. Visceral (lowest level) Behavioral (middle level) Reflective (top level)

Emotional Design. D. Norman (Emotional Design, 2004) Model with three levels. Visceral (lowest level) Behavioral (middle level) Reflective (top level) Emotional Design D. Norman (Emotional Design, 2004) Model with three levels Visceral (lowest level) Behavioral (middle level) Reflective (top level) Emotional Intelligence (EI) IQ is not the only indicator

More information

EXTRACTION OF RETINAL BLOOD VESSELS USING IMAGE PROCESSING TECHNIQUES

EXTRACTION OF RETINAL BLOOD VESSELS USING IMAGE PROCESSING TECHNIQUES EXTRACTION OF RETINAL BLOOD VESSELS USING IMAGE PROCESSING TECHNIQUES T.HARI BABU 1, Y.RATNA KUMAR 2 1 (PG Scholar, Dept. of Electronics and Communication Engineering, College of Engineering(A), Andhra

More information

DISCRETE WAVELET PACKET TRANSFORM FOR ELECTROENCEPHALOGRAM- BASED EMOTION RECOGNITION IN THE VALENCE-AROUSAL SPACE

DISCRETE WAVELET PACKET TRANSFORM FOR ELECTROENCEPHALOGRAM- BASED EMOTION RECOGNITION IN THE VALENCE-AROUSAL SPACE DISCRETE WAVELET PACKET TRANSFORM FOR ELECTROENCEPHALOGRAM- BASED EMOTION RECOGNITION IN THE VALENCE-AROUSAL SPACE Farzana Kabir Ahmad*and Oyenuga Wasiu Olakunle Computational Intelligence Research Cluster,

More information

CPSC81 Final Paper: Facial Expression Recognition Using CNNs

CPSC81 Final Paper: Facial Expression Recognition Using CNNs CPSC81 Final Paper: Facial Expression Recognition Using CNNs Luis Ceballos Swarthmore College, 500 College Ave., Swarthmore, PA 19081 USA Sarah Wallace Swarthmore College, 500 College Ave., Swarthmore,

More information

Analysis of Speech Recognition Techniques for use in a Non-Speech Sound Recognition System

Analysis of Speech Recognition Techniques for use in a Non-Speech Sound Recognition System Analysis of Recognition Techniques for use in a Sound Recognition System Michael Cowling, Member, IEEE and Renate Sitte, Member, IEEE Griffith University Faculty of Engineering & Information Technology

More information

DEEP convolutional neural networks have gained much

DEEP convolutional neural networks have gained much Real-time emotion recognition for gaming using deep convolutional network features Sébastien Ouellet arxiv:8.37v [cs.cv] Aug 2 Abstract The goal of the present study is to explore the application of deep

More information

ABSTRACT I. INTRODUCTION

ABSTRACT I. INTRODUCTION 2018 IJSRSET Volume 4 Issue 2 Print ISSN: 2395-1990 Online ISSN : 2394-4099 National Conference on Advanced Research Trends in Information and Computing Technologies (NCARTICT-2018), Department of IT,

More information

CAN A SMILE REVEAL YOUR GENDER?

CAN A SMILE REVEAL YOUR GENDER? CAN A SMILE REVEAL YOUR GENDER? Antitza Dantcheva, Piotr Bilinski, Francois Bremond INRIA Sophia Antipolis, France JOURNEE de la BIOMETRIE 2017, Caen, 07/07/17 OUTLINE 1. Why Gender Estimation? 2. Related

More information

AUDIO-VISUAL EMOTION RECOGNITION USING AN EMOTION SPACE CONCEPT

AUDIO-VISUAL EMOTION RECOGNITION USING AN EMOTION SPACE CONCEPT 16th European Signal Processing Conference (EUSIPCO 28), Lausanne, Switzerland, August 25-29, 28, copyright by EURASIP AUDIO-VISUAL EMOTION RECOGNITION USING AN EMOTION SPACE CONCEPT Ittipan Kanluan, Michael

More information

Integrating User Affective State Assessment in Enhancing HCI: Review and Proposition

Integrating User Affective State Assessment in Enhancing HCI: Review and Proposition 192 The Open Cybernetics and Systemics Journal, 2008, 2, 192-205 Open Access Integrating User Affective State Assessment in Enhancing HCI: Review and Proposition Xiangyang Li * Department of Industrial

More information

A Unified Probabilistic Framework For Measuring The Intensity of Spontaneous Facial Action Units

A Unified Probabilistic Framework For Measuring The Intensity of Spontaneous Facial Action Units A Unified Probabilistic Framework For Measuring The Intensity of Spontaneous Facial Action Units Yongqiang Li 1, S. Mohammad Mavadati 2, Mohammad H. Mahoor and Qiang Ji Abstract Automatic facial expression

More information

Error Detection based on neural signals

Error Detection based on neural signals Error Detection based on neural signals Nir Even- Chen and Igor Berman, Electrical Engineering, Stanford Introduction Brain computer interface (BCI) is a direct communication pathway between the brain

More information

Using Bayesian Networks for Daily Activity Prediction

Using Bayesian Networks for Daily Activity Prediction Plan, Activity, and Intent Recognition: Papers from the AAAI 2013 Workshop Using Bayesian Networks for Daily Activity Prediction Ehsan Nazerfard School of Electrical Eng. and Computer Science Washington

More information

Affect Intensity Estimation using Multiple Modalities

Affect Intensity Estimation using Multiple Modalities Affect Intensity Estimation using Multiple Modalities Amol S. Patwardhan, and Gerald M. Knapp Department of Mechanical and Industrial Engineering Louisiana State University apatwa3@lsu.edu Abstract One

More information

Potential applications of affective computing in the surveillance work of CCTV operators

Potential applications of affective computing in the surveillance work of CCTV operators Loughborough University Institutional Repository Potential applications of affective computing in the surveillance work of CCTV operators This item was submitted to Loughborough University's Institutional

More information

A Deep Learning Approach for Subject Independent Emotion Recognition from Facial Expressions

A Deep Learning Approach for Subject Independent Emotion Recognition from Facial Expressions A Deep Learning Approach for Subject Independent Emotion Recognition from Facial Expressions VICTOR-EMIL NEAGOE *, ANDREI-PETRU BĂRAR *, NICU SEBE **, PAUL ROBITU * * Faculty of Electronics, Telecommunications

More information

Noise-Robust Speech Recognition Technologies in Mobile Environments

Noise-Robust Speech Recognition Technologies in Mobile Environments Noise-Robust Speech Recognition echnologies in Mobile Environments Mobile environments are highly influenced by ambient noise, which may cause a significant deterioration of speech recognition performance.

More information

Automated Image Biometrics Speeds Ultrasound Workflow

Automated Image Biometrics Speeds Ultrasound Workflow Whitepaper Automated Image Biometrics Speeds Ultrasound Workflow ACUSON SC2000 Volume Imaging Ultrasound System S. Kevin Zhou, Ph.D. Siemens Corporate Research Princeton, New Jersey USA Answers for life.

More information

Facial Expression Classification Using Convolutional Neural Network and Support Vector Machine

Facial Expression Classification Using Convolutional Neural Network and Support Vector Machine Facial Expression Classification Using Convolutional Neural Network and Support Vector Machine Valfredo Pilla Jr, André Zanellato, Cristian Bortolini, Humberto R. Gamba and Gustavo Benvenutti Borba Graduate

More information

VIDEO SALIENCY INCORPORATING SPATIOTEMPORAL CUES AND UNCERTAINTY WEIGHTING

VIDEO SALIENCY INCORPORATING SPATIOTEMPORAL CUES AND UNCERTAINTY WEIGHTING VIDEO SALIENCY INCORPORATING SPATIOTEMPORAL CUES AND UNCERTAINTY WEIGHTING Yuming Fang, Zhou Wang 2, Weisi Lin School of Computer Engineering, Nanyang Technological University, Singapore 2 Department of

More information