Abstract
Electroencephalography is widely used to study the dynamics of neural information processing in the brain and to diagnose brain disorder and cognitive processes. In this paper, we proposed EEG based emotion recognition system using Discrete Wavelet Transformation. A set of highly significant features based on wavelets coefficients has been extracted which also includes modified wavelet energy features. In order to minimize redundancy and maximize relevancy among features, mRMR algorithm is significantly applied for feature selection. Multi class Support Vector Machine is used to perform classification of four classes of human emotions. EEG recordings of “DEAP” database are used in this experiment. The proposed approach shows significant performance compared to existing algorithms.
Access provided by CONRICYT-eBooks. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Automatic detection and recognition of different emotional states is a salient topic in the fast growing research field of affective computing. Emotions are complex states of minds comprised of numerous psychophysiological components, such as bodily changes, cognitive reactions and thoughts. Numerous computational models and algorithms for automatic recognition of emotions have been provided by Affective computing, which integrates knowledge of computer science, physiology, artificial intelligence and biomedical engineering.
Emotions affect all aspects of our daily life and has significant influence on our health. State of depression, anxiety and anger disrupt human immune system and thus associated with many chronic diseases. Bringing into play the current advances in IOT and sensor networks [1], smart healthcare systems should be introduce to improve overall quality of life. The development of automatic emotion recognition system can be very useful in regulating self-emotions and would revolutionize applications in education, entertainment and security.
Philosophers and psychologists presented various theories of emotions. Based on these theories, numerous methods have been developed to detect and recognize different emotions using facial images [3], speech signals [4], gestures and physiological signals. Physiological signals including EEG signals are considered to be the most useful signals for human emotion recognition due to its strong correlation with emotions and independence of people’s will. Facial images based emotion recognition have a major flaw of suppressing and intentional control of emotions while physiological signal originate from Autonomous Nervous System activity, cannot be controlled intentionally. Experimental evidence shows that physiological/bio signals can be influence by the activity of Autonomic Nervous System (ANS) and can convey information regarding human emotion [5, 6]. In this paper, we proposed human emotion recognition system using wavelet energy feature along with statistical and modified wavelet energy feature in order to improve the performance of emotion recognitions systems.
This paper is organized as follows: Sect. 2 provides a review on different factor of emotion recognition system. Section 3, illustrates the detailed methodology including dataset description, procedure for feature extraction, feature selection and classification. Performance of the proposed method and concluding remarks are given in Sects. 4 and 5 respectively.
2 Related Work
EEG based emotion recognition has gain a lot of interest and different emotion classification system have been proposed by the researchers. The results of these systems highly depends on five basic factors which includes number of participants, stimulus, emotion modeling, feature extraction and classifier. Different techniques have been employed by the researchers to investigate these factors and thus these emotion classification systems cannot be compared. However a short review of the recent work done is presented in this section.
Emotion modeling: Emotion is a mental state or feeling that arises involuntarily and comprised of different components such as feelings, bodily changes, behavior and thoughts. In literature numerous emotion models have been proposed. However the two most utilized categories are Discrete Emotional Models (DEM) and Affective dimensional model (ADM). DEM deals with six basic universal categories of emotions: happiness, surprise, anger, disgust, sadness and fear [2]. ADM deals with the description of emotions in some coordinate system. It characterize emotion into two affective parameters, Arousal and Valence. The most commonly used dimensional model is Circumplex Model of Affects (CMA) [7] as given in Fig. 1.
Emotion elicitation: There are numerous methods of emotion elicitation. The most widely used methods for emotion induction includes images, video clips and sound clips etc. The most popular existing databases are: IAPS—the International Affective Picture System and IADS—the International Affective Digitized Sound System facilitate the task of emotion recognition. IAPS [8] and IADS [9] provides a collection of stimuli publically to researchers in the study of emotion.
Feature extraction and classification: Different characteristics of EEG signals can be captured and used as features for classification of emotions. These features can be placed in one of two domain, time domain and frequency domain. Time domain feature can be extracted from raw EEG signal and includes mean, standard deviation etc. [10]. Frequency domain features include the power of different frequency bands of EEG signals. In addition to these, several other features extraction techniques have been presented in the literature which includes High Order Crossings [10], Discrete Wavelet Transformation [11], fractal dimensions [12] and Independent Component Analysis [13]. For classification task, several machine learning algorithms have been used as classifiers. These classifiers includes Support Vector machine [14], Neural Networks [15] and Quadratic Discriminant Analysis [16].
3 Methodology
3.1 Dataset
Recent advances in emotion recognition have increased the interest of many researchers and encouraged them to create databases containing visual, speech and physiological emotion data. These databases includes MAHNOB-HCI [17] and DECAF [18] etc. In this study, we used a public available database called DEAP proposed by Koelstra et al. [19]. 32 healthy participants (16 male and 16 female), aged between 19 and 32, took part in the experiment. The EEG and peripheral physiological signals which includes electrocardiogram, galvanic skin response, respiration, skin temperature, blood volume, electromyograms (EMG) and electrooculogram (EOG) were recorded from these subjects while watching 40 different music videos. Biosemi Active Two system was used to record EEG signals over the scalp from 32 electrodes according to the international 10–20 system as shown in Fig. 2. Preprocessing of EEG signals was performed in order to denoise the heavily distorted EEG signals from motion artifacts, EOG artifacts due to eye blinking and power supply noise. Initially EEG signals were recorded with 512 Hz sampling frequency which were down sampled to 128 Hz during preprocessing. A bandpass frequency filter from 4.0–45.0 Hz was applied to filter EEG signals. The elimination of EOG artifacts was performed using blind source separation technique. In the last of experiment every subject performed self-assessment to evaluate and rate the emotional state caused by each video using Self-Assessment Manikins. Rating was performed by participant for each music video in term of levels of arousal, valence, like/dislike, dominance and familiarity. In this paper, we used the preprocessed data released by Koelstra et al. [19].
3.2 Feature Extraction
Statistical-Based Features
Due to the nonlinear nature of EEG signals and brain complexity, nonlinear feature like high order crossing and fractal dimension are widely employed by researchers in recent publication. However, simple features like mean, standard deviation and band power are still considered beneficial for emotion recognition system. For 32 channel EEG data provided in DEAP dataset, we extracted statistical features in combination with wavelet based feature to improve emotion recognition accuracy. These features include
-
Mean of the raw signal
-
The standard deviation of the raw signal
-
The mean of the absolute values of the first difference of the raw signal
-
The mean of the absolute values of the first signal of the standardized signal
-
The mean of the absolute values of the second difference of the raw signal
-
The mean of the absolute values of the second difference of the standardized signal
Wavelet-Based Features
Discrete Wavelet Transform is a powerful analytical tool for non-stationary signals and is widely used for time-frequency analysis of EEG signals due to its non-stationary nature. DWT decomposes EEG signal into different frequency bands with successive high pass and low pass filters. The high pass filter gives detail coefficients while low pass filter gives approximation coefficients. In this paper, we used Daubechies Wavelet Transform (db4) coefficients which is considered best for multiresolution analysis of EEG signals and EEG signal with 128 Hz sampling rate. Five levels Daubechies wavelet of order 4 is applied for decomposition of EEG signals into five frequency bands, delta, theta, alpha, beta and gamma as given in the Table 1.
After Discrete Wavelet Transformation, we estimated wavelet energy and wavelet entropy according to Eqs. 1 and 2 respectively.
Where C X (l, n) Wavelet coefficients associated with all five sub-bands are used to estimate wavelet energy and wavelet entropy. Another energy based feature set proposed by Murugappan et al. [11] is used in this paper. These features includes Recoursing Energy efficiency (REE), Logarithmic Recoursing Energy Efficiency (LREE) and Absolute Logarithmic Recoursing Energy Efficiency (ALREE). These features are estimated for gamma band as follows.
3.3 Feature Selection and Classification
Feature selection is performed in order to mitigate the high dimensionality feature space problem. In this step, the most suitable subset of all derived features is selected which not only solve the problem of dimensionality but also increase the classification accuracy due to the reduction of noise caused by irrelevant features. In this paper, successfully applied maximum relevancy and minimum redundancy algorithm (mRMR) for feature selection. After selection of most relevant features, classification is performed using machine learning classifier. For this purpose a multi class Support Vector Machine is used with radial basis function.
4 Experimental Results
The EEG recordings of 32 subjects of DEAP database have been used to classify four main classes. The arousal and valence scores on the scale from 0 to 9 is mapped into two levels, high and low. The resulting four classes are, high arousal/high valence (HAHV), high arousal/low valence (HALV), low arousal/high valence (LAHV) and low arousal/low valence (LALV). For performance evaluation of the system, EEG data is divided into two portions, training data and test data. 70% of the total data is used for training purpose while the remaining 30% was used for testing. For classification, two machine learning algorithms, Support Vector Machine and Quadratic discriminant analysis are used to classify EEG data into four classes. Grid search approach was adopted for parameter optimization. In this experiment, best performance is given by SVM with overall accuracy of 49.7% using all channels data.
In order to implement a less complex and user friendly emotion recognition method, we reduced the numbers of EEG channels as much as possible. For this purpose, we selected a group of 15 EEG channels namely Fp1, Fp2, AF3, F3, F4, F7, F8, P7, O1, O2, P8, CP3, CP4, C4 and C3, that belongs to all four major lobs of the brain. Research shows that, left frontal lobe and right frontal lobe exhibit certain activity when a negative or positive emotion is experienced by a person [10]. Furthermore, the related research also reveals that different sub bands (Delta, theta, Alpha, Beta and Gamma) are activated by different emotional states in specific brain regions [20]. In this paper, we also investigated the activity of all five bands of EEG signals in selected channels for the aforementioned four classes of emotion. The classification of emotions is performed using each frequency band separately for combination of different sets of channels. EEG signal acquired from frontal lobe (FP1, FP2, F3 and F4) and temporal lobe (T7, T8) showed best performance with features extracted from Gamma band and achieved an overall classification accuracy of 48.8% for four classes of emotions which is close to the prior accuracy gained using all channels data of EGG signals (Table 2).
5 Conclusion
In this paper, we presented emotion recognition system using the most significant features set extracted from coefficients of Discrete Wavelet Transformation. A public available database called DEAP has been in this work. The EEG recordings of 32 participants have been utilized to extract statistical based feature and wavelet based feature. A feature selection algorithm was adopted to select the most significant and relevant features in order to mitigate the problem of dimensionality, irrelevancy and redundancy.
The proposed approach can significantly classify four classes of emotions using Support Vector Machine from which following can be concluded. First, feature extracted using Discrete Wavelet Transformation effectively represent emotional state of the users. Second, Gamma band holds rich information of all four classes of emotions. Third, we found that there is a strong correlation in frontal brain region related to Gamma band for all four classes of emotions which validates the role of frontal lobe in emotion recognition. In future, research will be conducted on fusion of EEG signals with others physiological signals for high performance.
References
Bilal, M., Kang, S.-G.: An authentication protocol for future sensor networks. Sensors 17(5), 979 (2017)
Ekman, P.: Emotions Revealed. Times Books (2003)
Zhang, L., Tjondronegoro, D.: Facial expression recognition using facial movement features. IEEE Trans. Aff. Comput. 2, 219–229 (2011)
Wang, K., Ning, A., Li, B.N., Zhang, Y.: Speech emotion recognition using Fourier parameters. IEEE Trans. Aff. Comput. 6, 69–75 (2015)
Anttonen. J., Surakka, V.: Emotions and heart rate while sitting on a chair. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 491–499 (2005)
Jones, C.M., Troen, T.: Biometric valence and arousal recognition. In: Proceedings of the 19th Australasian Conference on Computer-Human Interaction, pp. 191–194 (2007)
Posner et al.: The circumplex model of affect: an integrative approach to affective neuroscience, cognitive development, and psychopathology. Dev. Psychopathol. 17 (2005)
Lang, P.J., Bradley, M. M., Cuthbert, B.N.: International affective picture system (IAPS): affective ratings of pictures and instruction manual. Technical repory A-8 (2008)
Bradley, M.M., Lang, P.J.: The international affective digitized sounds (iads-2): Affective ratings of sounds and instruction manual. University of Florida, Gainesville, FL, USA, Technical report B-3 (2007)
Petrantonakis et al.: Emotion recognition from EEG using higher order crossings. IEEE Trans. Inf. Technol. Biomed. 14(2) 186–197 (2010)
Murugappan et al.: Classification of human emotion from EEG using discrete wavelet transform. J. Biomed. Sci. Eng. 3(4), 390 (2010)
Sourina, O., Yisi, L.: A fractal-based algorithm of emotion recognition from EEG using arousal-valence model. BIOSIGNALS (2011)
Lan, T., et al.: Estimating cognitive state using EEG signals. In: 2005 13th European IEEE Signal Processing Conference (2005)
Koelstra, S., et al.: Single trial classification of EEG and peripheral physiological signals for recognition of emotions induced by music videos. In: Proceeding of the International Conference on Brain Informatics, BI 2010, Toronto, Canada, pp. 89–100 (2010)
Wijeratne, U., et al.: Intelligent emotion recognition system using electroencephalography and active shape models. In: Proceedings of the IEEE EMBS Conference on Biomedical Engineering and Sciences, IECBES 2012, pp. 636–641 (2012)
Khalili, Z., Moradi, M. H.: Emotion recognition system using brain and peripheral signals: using correlation dimension to improve the results of EEG. In: Proceedings of the International Joint Conference on Neural Networks, IJCNN 2009, Atlanta, pp. 1571–1575 (2009)
Soleymani, M., et al.: A multimodal database for affect recognition and implicit tagging. IEEE Trans. Aff. Comput. 3(1), 42–55 (2012)
Abadi, M., et al.: DECAF: MEG-based multimodal database for decoding affective physiological responses. IEEE Trans. Aff. Comput. 6(3), 209–222 (2015)
Koelstra, S., et al.: Deap: a database for emotion analysis; using physiological signals. IEEE Trans. Aff. Comput. 3(1), 18–31 (2012)
Daimi, S.N., Saha, G.: Classification of emotions induced by music videos and correlation with participants rating. Expert Syst. Appl. 41(13), 6057–6065 (2014)
Acknowledgement
This work was supported by Institute for Information & communications Technology Promotion (IITP) grant funded by the Korea government(MSIP) (Development of SW fused Wearable Device Module and Flexible SW Application Platform for the integrated Management of Human Activity).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zubair, M., Yoon, C. (2018). EEG Based Classification of Human Emotions Using Discrete Wavelet Transform. In: Kim, K., Kim, H., Baek, N. (eds) IT Convergence and Security 2017. Lecture Notes in Electrical Engineering, vol 450. Springer, Singapore. https://doi.org/10.1007/978-981-10-6454-8_3
Download citation
DOI: https://doi.org/10.1007/978-981-10-6454-8_3
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-6453-1
Online ISBN: 978-981-10-6454-8
eBook Packages: EngineeringEngineering (R0)