Abstract
The facial emotion classification is a crucial task in human behavior analysis. By taking static images, the emotion is identified from the face expression. It is one of the categories in image processing that is utilized in a variety of disciplines, including human and computer interaction. Some resources are projected to perform automatic emotion recognition, which utilizes benchmark datasets. This research work is focused on real-time dataset that is used to identify six human facial emotions that are implemented by using SVM and tree-based classifier. Experimental outcome symbolizes the top most presentation on the SVM radial basis function (RBF) kernel recognition (95.49%) when associated to the tree-based classifier.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
- Human emotion
- Face detection
- Cascade classifier
- SVM
- Random forest
- Decision tree
- Naïve Bayes
- Performance measure
1 Introduction
Human expression plays a vital role in establishing non-audio communication between human beings. Nowadays, the facial expression identification technique is gaining more and more attention from the people. Facial appearance includes key information about psychological, emotional, and even physical state of the chat. Facial appearance recognition will also create a practical impact. It has a very wide application forecast such as comprehensible interface among human and machine, humanistic mean of goods, and emotional robot. With facial appearance identification structures, the system can review the human emotions [1]. The intellectual processor will be capable to recognize, understand, and act in reaction to human intentions, expressions, and moods [2].
The human facial expression detection systems are implemented in several living places like safety or examination; they can forecast the criminal and behavior by scanning the figures of their features to facilitate and confined by the control camcorder. In addition, facial appearance identification scheme has been utilized to create the response engine that is additionally cooperative with humans. The respond engine has turn out to be more intellectual by evaluating the customer’s tone and commerce with the reactions according to their face sensation [3].
In addition, facial sensation is dominant in signed and any language identification approach that agreements with the challenge of hearing and mutilate humans. Person facial appearance identification approach has a significant contact on the diversion pasture and in addition its utilization boost the effectiveness of machines for particular medical-oriented robots and industrialized checking [4]. Usually, the automated system with facial appearance identification approach has been utilized to progress in our everyday lives.
Depression is a generally unattended health issue, unbounded by work stress, health issue, and commonly affecting majority of the student’s performance in their study. To avoid it in future, the researchers construct a real-time face expression classification approach, so the teacher can supervise the student’s mentality through classroom movement.
1.1 Categories of Facıal Emotıons
Universally six fundamental human expressions and unbiased feel are recognized globally. The real-life human facial emotions are divided into seven types of modules. The expressions are categorized as happy, fear, anger, surprise, sad, disgust, and neutral. Example pictures from online Japanese Female Facial Emotions (JAFFE) Database for six modules are given in Fig. 1. Henceforth, this paper concentrates on real-time Indian human facial expressions like fear, happy, anger, disgust, surprise, and neutral.
2 Related Work
This segment deals with the proposed work performed till now by different developers in the area of expression identification through facial language. Intentions of a number of writers are discussed in this section. Daisy and Kannan [5] projected a face appearance identification approach based on an original limited rotated local Gabor filter method. Gabor filter technique utilizes two-step feature firmness techniques such as principal component analysis (PCA) and linear discriminant analysis to choose and constrict the Gabor feature selection and smallest amount of space classification to identify the facial emotions. These techniques are valuable for a combined measurement decrease and excellent appreciation act in association with conventional entire Gabor filter method. The finest common identification achieves good recognition results for GTAVE, MIT, CMU, PIE, and real-time home databases.
Bashyal et al. [6] have projected a well-organized method for face appearance identification by using Gabor filters method as feature identification. JAFFE dataset is utilized for testing, and the obtained result is greater than 91%. Xie et al. [7] have projected spatial highest occurrence model based on numerical factors and employed elastic figure-based texture matching techniques for figure or surface-oriented human face appearance detection.
Zhang et al. [8] obtain the starting point as mixture corresponding to features of face as fixed, vibrant, position-based arithmetic, or area-based exterior characteristics. Novelist faced on fixed frames and experimental output for dimensional characters. These types of features have been identified by Gabor filters. The algorithm has presented excellent experimental outputs. In [9], the authors have published a novel technique that developed mutually geometric and texture data of facial positions. Gauss Laguerre method is employed for the identification of texture data for a variety of face emotions. The backpropagation neural network and probabilistic neural network techniques are utilized to detect the different type of X-ray images [10]. Fuzzy logic system-based aeration control approach for contaminated stream water [11]. Fuzzy chaos whale optimization and BAT integrated techniques for limitation evaluation are in sewage management [12]. Wireless rechargeable sensor network mistake techniques and immovability investigation [13].
This research exposes a variety of techniques tracked by researchers for facial emotion classification. Exhaustive investigation is carried out in additional subsections of this research work.
3 Proposed Facial Expression Recognition System
Facial language replicates the emotions of person, which disclose expensive information of one’s sentiment, feedback, etc. Properly distinguishing these emotions is a difficult task. This segment describes the facial emotion identification system.
3.1 System Architecture
Overall system architecture of the projected facial expression identification is given in Fig. 2. Four modules collect the projected system: first phase is preprocessing, face detection as second phase, facial feature extraction as third phase, and facial emotional recognition as final phase by using two techniques named as SVM and tree-based classifier.
3.2 Preprocessing
Preprocessing method is the primary phase by inflowing the frame records into the face recognition and facial expression identification system. The essential data required for most facial appearance identification method is face location. In preprocessing component, frames are reformed from 640 × 640 pixel rates to 400 × 400 pixel rates.
3.3 Face Detection Using Cascade Classifier
In some applications such as human–computer boundary, face detection, video surveillance, and facial expressions, the very first step used here is localizing and detecting the human face. Viola–Jones object recognition process supported on cascade of recognizer is worn to trace the person face within every image of the sequence of video. More distinctively, we apply 14 feature prototypes [14], which comprise four corner features, eight line total features, and two corner-surround features. These samples are balanced separately in horizontal way in order to produce a wealthy and over-complete group of characteristics. These set of characteristics can be figured out in a regular and small time irrespectively of the location as given in [15]. Face detection and extraction using cascade classifier are given in Fig. 3.
3.4 Feature Extraction for Emotion Detection
Feature information is an informative area identified from a frames or a large sequence of video. Visual information exhibits various models of characteristics that could be used to recognize or represent the relevant information it opens. The gray length method (GRLM) is a way of extracting higher-order statistical texture features. The theory and techniques behind the method are presented in [16]. A position of successive grayscale level, collinear in a given path constitutes a grayscale level run. The run length is the amount of pixels in the run, and the run length charge is the amount of times such a run occurs in a face image. The grayscale level run length of matrix (GRLM) is a four-dimensional matrix in which every factor f(x, y | θ) provides the entire number of incidences of runs of length ‘y’ at grayscale level ‘x’, in a particular direction θ.
4 Facial Emotion Recognition Using SVM and Tree-based Classifier
Machine learning (ML) contains many processes such as support vector machine (SVM), artificial neural networks (ANN), genetic algorithms (GA), Bayesian training, and probabilistic models [17], in that which we required only the accomplishment of the very first one machine learning (ML) methods on person fingerprint classification approach.
4.1 Support Vector Machine (SVM)
SVM methods is well-accepted practice for recognition in chart prototype classification [18]. The SVM is generally demoralized in kernel-based knowledge techniques. It achieves rational very important pattern classification presentation in optimization method [19]. Categorization tasks are commonly concerned with the use of testing and training data. The training data is subdivided into (x1, y1), (x2, y2), … (xm, ym) into two classes, with xi Rn enclosing an n-dimensional feature vector and yi + 1,1 enclosing class labels. SVM's goal is to create a copy that predicts the target value from the testing dataset. The hyperplane w.x + b = 0, where w Rn, b R is used to separate both classes in any space Z, is used in binary classification [20]. M = 2/||w|| gives the maximum scope.
The hyperplane is utilized to categorize the input feature space into the required target model. However, in order to fit the decision boundary in a hyperplane to make the most of distance boundary is preferred from feature data points for recognition. The sample maximum margin is given in Fig. 4.
SVM can be structured by using a variety of kernel function to improve the accuracy: polynomial, Gaussian, and sigmoidal. SVM is well appropriate for mutually unstructured and structured feature data.
4.2 Tree-Based Classifier
4.2.1 Random Forest
A random forest [RF] is a group of decision trees skilled with random features. Random forest moves as follows. Given a set of training examples, a set of random trees H is shaped such that for the kth tree in the forest, a random vector φk is produced autonomously of the past random vectors φ1…φk − 1. This vector k is then used to raise the tree resulting in a classifier hk (x, φk), where x is a feature vector. Random forest [21] is the fast and robust categorization performance that can handle multiclass problem.
4.2.2 Decısıon Tree (J48)
Decision trees are commonly used methods for pattern classification. Chi-squared automatic integration detection (CHAID) introduced in [22] and classifier 4.5 (C4.5, J48) in [17]. In this study, J48 algorithm decision tree was applied to traffic personnel hand features. J48 classifier is a standard model in C4.5 decision tree for supervised classification. In decision tree, feature collection procedure is done by information index. The information index for an exacting feature data Z at a node is calculated as
where X is the combination of instance at that exacting node and
Entropy1 of X is found as:
4.2.3 Naive Bayes (NB)
NB tree concept is a hybrid algorithm that characterizes a cross between Naive Bayes recognizer and C4.5 decision tree recognizer, and it is most excellent explained as a decision tree with nodes and branches [23]. The feature space which is classified in Naive Bayes is always independent to every other. If a is class variable and b is dependent feature space.
P1 (a) is called class probability and is provisional probability.
Bayesian theorem probability states
5 Proposed Experimental Results
In this segment, the investigational outputs acquired in facial emotion identification systems are offered. The experimentations are performed by using SVM Torch and Weka Tool. LIBSVM [24] technique to expand the mold for every sensation and the molds are utilized to experiment the presentation, and recognitions are utilized for performance reason. Weka is a whole group of Java technique methods that are utilized to bring out several data mining and machine learning algorithms [10]. We evaluate the presentation of tree methods namely NBTree, REPTree, random forest, and random tree.
5.1 Database
The experimentations are accomplished with 25 subjects and the objects being clicked by the camera in natural scene at 25 fps with a 640 × 640 declaration. Real-time dataset includes 1500 objects of six person face emotion including neutral caused by 25 Indian female objects. Every objects has been priced on six emotions modules by 1500 Indian subjects.
Figure 5 shows the real-time database images considered for facial expression recognition. The proposed training set is organized of 900 objects (the every set organizes 15 persons and every person includes 60 objects). And remaining procedure, the proposed test set includes 600 objects that are evaluated of random choosing 10 objects from each and every emotions.
5.2 Performance Evaluation
This research work provides a methodical investigation of multiclass system. Standard estimation calculation include precision (P) = True Positive∕(True Positive + False Positive), recall (R) = True Positive∕(True Positive + False Negative), specificity (S) = True Negative∕(True Negative + False Positive), and F-measure = 2 Precision Recall∕(Precision + Recall) that are utilized to estimate the presentation of the projected facial emotion classification system. These types of calculations provide the excellent perception on the recognition and presentation for facial emotion system.
5.3 Results Obtained Using SVM and Tree-Based Classifiers
This section presents human facial emotion classification system using support vector machine technique and using WEKA Tool [10]. For this reason, human facial emotion datasets shown in Fig. 4 were used. The presentation result of the decision trees (DT) and random forests (RF) classifiers were calculated using tenfold cross-validation model and support vector machine using multiclass classification problem. The complete outputs for the average of recall, precision, and specificity examination for the experimental results are obtained in Fig. 5. Table 1 gives the confusion matrix for proposed facial emotion recognition using SVM (RBF) multiclass classification with proposed feature. Tables 2, 3, and 4 correspondingly tabulate the confusion matrix of the tree-based classifiers using the proposed facial features on six emotions by using tenfold cross-validation for emotion recognition.
The accuracy results obtained with proposed feature are by using SVM RBF kernel and tree-based classifier for each individual human facial emotion recognition, and they are presented in Table 5. Figure 6 shows the comparison of the precision, recall, and specificity value of proposed facial features by using support vector machine [SVM] with radial basis function, random forests, decision trees (DT), and Naïve Bayes techniques. The projected approach provides superior performance value of proposed feature by using SVM RBF when compared to tree-based classifiers. Computer-aided detection and diagnosis of techniques is proposed in Balaji et al. [25].
6 Conclusion and Future Work
The proposed research has shown the accurate performance calculation by using SVM and tree-based classifier for human facial emotion classification approach. The research result gives significant attention to four steps, such as preprocessing, face identification, facial feature identification, and finally human facial emotion recognition. SVM RBF kernel gives higher accuracy compared to tree-based classifier. In forthcoming research, the suppleness of this proposed research is developed by using deep learning (DL) to classify facial emotions by using a variety of methods in a complicated environment.
References
B. Dixit, A. Gaikwad, Facial features based emotion recognition. IOSR J. Eng. (IOSRJEN) 8(8), 2250–3021 (2018). ISSN (p): 2278-8719
L. Zhang, M. Kamlesh, C.N. Siew, P.L. Chee, Intelligent facial emotion recognition using moth-firefly optimization, Elsevier Publ. Int. J. Knowl. Based Syst. 111, 248–267 (2016)
B.-F. Wu, C.-H. Lin, Adaptive feature mapping for customizing deep learning based facial expression recognition model. IEEE Access 6, 12451–12461 (2018)
Y. Ding, Q. Zhao, B. Li, X. Yuan, Facial expression recognition from image sequence based on LBP and Taylor expansion. IEEE Access Spec. Sect. Sequential Data Model. Emerg. Appl. 5, 19409–19419 (2017)
M.M. Daisy, P. Kannan, Investigation of rotated local Gabor features in face recognition using fusion techniques. J. Ambient Intell. Human Comput. 12, 5895–5908 (2021)
Y. Zhao, D. Chen, A Facial Expression Recognition Method Using Improved Capsule Network Model (Hindawi Scientific Programming, 2020). Article ID 8845176
V.G.V. Mahesh, C. Chen, V. Rajangam, A.N. Joseph Raj, P.T. Krıshnan, Shape and texture aware facial expression recognition using partial pyramid Zernike moments and Law’s textures feature set. IEEE Trans. J. 4 (2021). Article in IEEE Access
L. Zhang, D. Tjondronegoro, Facial expression recognition using facial movement features. IEEE Trans. Affect. Comput. 2(4), 219–228 (2011)
A. Poursaberi, H. Ahmadi, S.N. Yanushkevich, M. Gavrilova, Gauss–Laguerre wavelet textural feature fusion with geometrical information for facial expression identification. EURASIP J. Image Video Process. Springer Open (2012)
I.H. Witten, E. Frank, Datamining: Practical Machine Learning Tools and Techniques with Java Implementations (1999)
T. Vijayakumar, R. Vinothkanna, M. Duraipandian, Fuzzy logic based aeration control system for contaminated water. J. Electron. 2(1), 10–17 (2020)
A. Sungheetha, R. Sharma, Fuzzy chaos whale optimization and BAT integrated algorithm for parameter estimation in sewage treatment. J. Soft Comput. Paradigm (JSCP) 3(01), 10–18 (2021)
S.R. Mugunthan, Wireless rechargeable sensor network fault modeling and stability analysis. J. Soft Comput. Paradigm (JSCP) 3(01), 47–54 (2021)
S.O. Adeshina, H. Ibrahim, S.S. Teoh, S.C. Hoo, Custom face classification model for classroom using Haar-like and LBP features with their performance comparisons. Electronics 10(102) (2021). https://doi.org/10.3390/electronics10020102
J. Barreto, P. Menezes, J. Dias, Human-robot interaction based on Haar-like features and eigenfaces, in International Conference on Robotics and Automation (2004)
S.A. Ramola, A.K. Shakya, D.V. Pham, Study of Statistical Methods for Texture Analysis and Their Modern Evolutions (Wiley, 2020). https://doi.org/10.1002/eng2.12149
S. Theodoridis, A. Pikrakis, K. Koutroumbas, D. Cavouras, Introduction to Pattern Recognition: A Matlab Approach (Academic Press, 2010).
J. Liu, F.-P. An, Image Classification Algorithm Based on Deep Learning-Kernel Function (Hindawi Scientific Programming, 2020). Article ID 7607612. https://doi.org/10.1155/2020/7607612
V.N. Vapnik, V. Vapnik, Statistical Learning Theory, vol. 1 (Wiley, New York, 1998)
C.-C. Chang, C.-J. Lin, Libsvm: a library for support vector machines. ACM Trans. Intel. Syst. Technol. (TIST) 2(3), 27 (2011)
R. Sathya, M. Kalaiselvi Geetha, Framework for traffic personnel gesture recognition, in International Conference on Information and Communication Technologies (ICICT 2014), Procedia Computer Science, vol. 46, pp. 1700–1707 (2015)
T. Mitchell, Machine Learning (McGraw-Hill Computer Science Series, 1997)
J.R. Quinlan, C4.5: Programs for Machine Learning (Morgan Kaufmann, San Mateo, California, 1993)
C.-C. Chang, C.J. Lin, LIBSVM: a library for support vector machines. ACM Trans. Intel. Syst. Technol. 2, 1–27 (2011)
G.N. Balaji, T.S. Subashini, P. Madhavi, C.H. Bhavani, A. Manikandarajan, Computer-aided detection and diagnosis of diaphyseal femur fracture, in Smart Intelligent Computing and Applications (Springer, Singapore, 2020), pp. 549–559
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Sathya, R., Manivannan, R., Vaidehi, K. (2022). Vision-Based Personal Face Emotional Recognition Approach Using Machine Learning and Tree-Based Classifier. In: Smys, S., Balas, V.E., Palanisamy, R. (eds) Inventive Computation and Information Technologies. Lecture Notes in Networks and Systems, vol 336. Springer, Singapore. https://doi.org/10.1007/978-981-16-6723-7_42
Download citation
DOI: https://doi.org/10.1007/978-981-16-6723-7_42
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-16-6722-0
Online ISBN: 978-981-16-6723-7
eBook Packages: EngineeringEngineering (R0)