scholarly journals Emotion classification from EEG signals using wearable sensors: pilot test

2020 ◽  
pp. 1-9
Author(s):  
Alejandro JARILLO-SILVA ◽  
Víctor A. GOMEZ-PEREZ ◽  
Eduardo A. ESCOTTO-CÓRDOVA ◽  
Omar A. DOMÍNGUEZ-RAMÍREZ

The objective of this work is to present a procedure for the classification of basic emotions based on the analysis of EEG signals (electroencephalogram). For this case, 25 subjects were stimulated, of whom 17 were men and 9 women between 20 and 35 years of age. The stimulus to induce positive, negative and neutral emotions with a certain level of excitation (activation) was a set of video clips previously evaluated. The processed and analyzed signals belong to the gamma and beta frequency bands of the F3, F4, P7, P8, T7, T8, O1 and O2 electrodes. The characteristic variables with the best result are the entropy of each band of each electrode. The cross validation algorithms are applied and later the main component analysis algorithm. Finally, four classifier algorithms are used: classifier trees, Support- Vector-Machine (SVM), Linear-Discriminant-Analysis (LDA) and k-Nearest-Neighbors (KNN). The results confirm that by carrying out the proposed procedure, the EEG signals contain enough information to allow the recognition of basic emotions.

2021 ◽  
Vol 14 ◽  
Author(s):  
Yinfeng Fang ◽  
Haiyang Yang ◽  
Xuguang Zhang ◽  
Han Liu ◽  
Bo Tao

Due to the rapid development of human–computer interaction, affective computing has attracted more and more attention in recent years. In emotion recognition, Electroencephalogram (EEG) signals are easier to be recorded than other physiological experiments and are not easily camouflaged. Because of the high dimensional nature of EEG data and the diversity of human emotions, it is difficult to extract effective EEG features and recognize the emotion patterns. This paper proposes a multi-feature deep forest (MFDF) model to identify human emotions. The EEG signals are firstly divided into several EEG frequency bands and then extract the power spectral density (PSD) and differential entropy (DE) from each frequency band and the original signal as features. A five-class emotion model is used to mark five emotions, including neutral, angry, sad, happy, and pleasant. With either original features or dimension reduced features as input, the deep forest is constructed to classify the five emotions. These experiments are conducted on a public dataset for emotion analysis using physiological signals (DEAP). The experimental results are compared with traditional classifiers, including K Nearest Neighbors (KNN), Random Forest (RF), and Support Vector Machine (SVM). The MFDF achieves the average recognition accuracy of 71.05%, which is 3.40%, 8.54%, and 19.53% higher than RF, KNN, and SVM, respectively. Besides, the accuracies with the input of features after dimension reduction and raw EEG signal are only 51.30 and 26.71%, respectively. The result of this study shows that the method can effectively contribute to EEG-based emotion classification tasks.


Sensors ◽  
2021 ◽  
Vol 21 (15) ◽  
pp. 5135
Author(s):  
Ngoc-Dau Mai ◽  
Boon-Giin Lee ◽  
Wan-Young Chung

In this research, we develop an affective computing method based on machine learning for emotion recognition using a wireless protocol and a wearable electroencephalography (EEG) custom-designed device. The system collects EEG signals using an eight-electrode placement on the scalp; two of these electrodes were placed in the frontal lobe, and the other six electrodes were placed in the temporal lobe. We performed experiments on eight subjects while they watched emotive videos. Six entropy measures were employed for extracting suitable features from the EEG signals. Next, we evaluated our proposed models using three popular classifiers: a support vector machine (SVM), multi-layer perceptron (MLP), and one-dimensional convolutional neural network (1D-CNN) for emotion classification; both subject-dependent and subject-independent strategies were used. Our experiment results showed that the highest average accuracies achieved in the subject-dependent and subject-independent cases were 85.81% and 78.52%, respectively; these accuracies were achieved using a combination of the sample entropy measure and 1D-CNN. Moreover, our study investigates the T8 position (above the right ear) in the temporal lobe as the most critical channel among the proposed measurement positions for emotion classification through electrode selection. Our results prove the feasibility and efficiency of our proposed EEG-based affective computing method for emotion recognition in real-world applications.


Author(s):  
Negin Manshouri ◽  
Mesut Melek ◽  
Temel Kayikcioglu

Despite the long and extensive history of 3D technology, it has recently attracted the attention of researchers. This technology has become the center of interest of young people because of the real feelings and sensations it creates. People see their environment as 3D because of their eye structure. In this study, it is hypothesized that people lose their perception of depth during sleepy moments and that there is a sudden transition from 3D vision to 2D vision. Regarding these transitions, the EEG signal analysis method was used for deep and comprehensive analysis of 2D and 3D brain signals. In this study, a single-stream anaglyph video of random 2D and 3D segments was prepared. After watching this single video, the obtained EEG recordings were considered for two different analyses: the part involving the critical transition (transition-state) and the state analysis of only the 2D versus 3D or 3D versus 2D parts (steady-state). The main objective of this study is to see the behavioral changes of brain signals in 2D and 3D transitions. To clarify the impacts of the human brain’s power spectral density (PSD) in 2D-to-3D (2D_3D) and 3D-to-2D (3D_2D) transitions of anaglyph video, 9 visual healthy individuals were prepared for testing in this pioneering study. Spectrogram graphs based on Short Time Fourier transform (STFT) were considered to evaluate the power spectrum analysis in each EEG channel of transition or steady-state. Thus, in 2D and 3D transition scenarios, important channels representing EEG frequency bands and brain lobes will be identified. To classify the 2D and 3D transitions, the dominant bands and time intervals representing the maximum difference of PSD were selected. Afterward, effective features were selected by applying statistical methods such as standard deviation (SD), maximum (max), and Hjorth parameters to epochs indicating transition intervals. Ultimately, k-Nearest Neighbors (k-NN), Support Vector Machine (SVM), and Linear Discriminant Analysis (LDA) algorithms were applied to classify 2D_3D and 3D_2D transitions. The frontal, temporal, and partially parietal lobes show 2D_3D and 3D_2D transitions with a good classification success rate. Overall, it was found that Hjorth parameters and LDA algorithms have 71.11% and 77.78% classification success rates for transition and steady-state, respectively.


2021 ◽  
Vol 11 (23) ◽  
pp. 11252
Author(s):  
Ayana Mussabayeva ◽  
Prashant Kumar Jamwal ◽  
Muhammad Tahir Akhtar

Classification of brain signal features is a crucial process for any brain–computer interface (BCI) device, including speller systems. The positive P300 component of visual event-related potentials (ERPs) used in BCI spellers has individual variations of amplitude and latency that further changse with brain abnormalities such as amyotrophic lateral sclerosis (ALS). This leads to the necessity for the users to train the speller themselves, which is a very time-consuming procedure. To achieve subject-independence in a P300 speller, ensemble classifiers are proposed based on classical machine learning models, such as the support vector machine (SVM), linear discriminant analysis (LDA), k-nearest neighbors (kNN), and the convolutional neural network (CNN). The proposed voters were trained on healthy subjects’ data using a generic training approach. Different combinations of electroencephalography (EEG) channels were used for the experiments presented, resulting in single-channel, four-channel, and eight-channel classification. ALS patients’ data represented robust results, achieving more than 90% accuracy when using an ensemble of LDA, kNN, and SVM on four active EEG channels data in the occipital area of the brain. The results provided by the proposed ensemble voting models were on average about 5% more accurate than the results provided by the standalone classifiers. The proposed ensemble models could also outperform boosting algorithms in terms of computational complexity or accuracy. The proposed methodology shows the ability to be subject-independent, which means that the system trained on healthy subjects can be efficiently used for ALS patients. Applying this methodology for online speller systems removes the necessity to retrain the P300 speller.


Sensors ◽  
2020 ◽  
Vol 20 (9) ◽  
pp. 2694
Author(s):  
Sang-Yeong Jo ◽  
Jin-Woo Jeong

Visual memorability is a method to measure how easily media contents can be memorized. Predicting the visual memorability of media contents has recently become more important because it can affect the design principles of multimedia visualization, advertisement, etc. Previous studies on the prediction of the visual memorability of images generally exploited visual features (e.g., color intensity and contrast) or semantic information (e.g., class labels) that can be extracted from images. Some other works tried to exploit electroencephalography (EEG) signals of human subjects to predict the memorability of text (e.g., word pairs). Compared to previous works, we focus on predicting the visual memorability of images based on human biological feedback (i.e., EEG signals). For this, we design a visual memory task where each subject is asked to answer whether they correctly remember a particular image 30 min after glancing at a set of images sampled from the LaMemdataset. During the visual memory task, EEG signals are recorded from subjects as human biological feedback. The collected EEG signals are then used to train various classification models for prediction of image memorability. Finally, we evaluate and compare the performance of classification models, including deep convolutional neural networks and classical methods, such as support vector machines, decision trees, and k-nearest neighbors. The experimental results validate that the EEG-based prediction of memorability is still challenging, but a promising approach with various opportunities and potentials.


Sensors ◽  
2020 ◽  
Vol 20 (4) ◽  
pp. 1235
Author(s):  
Giuseppe Varone ◽  
Sara Gasparini ◽  
Edoardo Ferlazzo ◽  
Michele Ascoli ◽  
Giovanbattista Gaspare Tripodi ◽  
...  

The diagnosis of psychogenic nonepileptic seizures (PNES) by means of electroencephalography (EEG) is not a trivial task during clinical practice for neurologists. No clear PNES electrophysiological biomarker has yet been found, and the only tool available for diagnosis is video EEG monitoring with recording of a typical episode and clinical history of the subject. In this paper, a data-driven machine learning (ML) pipeline for classifying EEG segments (i.e., epochs) of PNES and healthy controls (CNT) is introduced. This software pipeline consists of a semiautomatic signal processing technique and a supervised ML classifier to aid clinical discriminative diagnosis of PNES by means of an EEG time series. In our ML pipeline, statistical features like the mean, standard deviation, kurtosis, and skewness are extracted in a power spectral density (PSD) map split up in five conventional EEG rhythms (delta, theta, alpha, beta, and the whole band, i.e., 1–32 Hz). Then, the feature vector is fed into three different supervised ML algorithms, namely, the support vector machine (SVM), linear discriminant analysis (LDA), and Bayesian network (BN), to perform EEG segment classification tasks for CNT vs. PNES. The performance of the pipeline algorithm was evaluated on a dataset of 20 EEG signals (10 PNES and 10 CNT) that was recorded in eyes-closed resting condition at the Regional Epilepsy Centre, Great Metropolitan Hospital of Reggio Calabria, University of Catanzaro, Italy. The experimental results showed that PNES vs. CNT discrimination tasks performed via the ML algorithm and validated with random split (RS) achieved an average accuracy of 0.97 ± 0.013 (RS-SVM), 0.99 ± 0.02 (RS-LDA), and 0.82 ± 0.109 (RS-BN). Meanwhile, with leave-one-out (LOO) validation, an average accuracy of 0.98 ± 0.0233 (LOO-SVM), 0.98 ± 0.124 (LOO-LDA), and 0.81 ± 0.109 (LOO-BN) was achieved. Our findings showed that BN was outperformed by SVM and LDA. The promising results of the proposed software pipeline suggest that it may be a valuable tool to support existing clinical diagnosis.


2020 ◽  
Author(s):  
Negin Manshouri ◽  
Mesut Melek ◽  
Temel Kayıkcıoglu

Abstract Despite the long and extensive history of 3D technology, it has recently attracted the attention of researchers. This technology has become the center of interest of young people because of the real feelings and sensations it creates. People see their environment as 3D because of their eye structure. In this study, it is hypothesized that people lose their perception of depth during sleepy moments and that there is a sudden transition from 3D vision to 2D vision. Regarding these transitions, the EEG signal analysis method was used for deep and comprehensive analysis of 2D and 3D brain signals. In this study, a single-stream anaglyph video of random 2D and 3D segments was prepared. After watching this single video, the obtained EEG recordings were considered for two different analyses: the part involving the critical transition (transition state) and the state analysis of only the 2D versus 3D or 3D versus 2D parts (steady state). The main objective of this study is to see the behavioral changes of brain signals in 2D and 3D transitions. To clarify the impacts of the human brain’s power spectral density (PSD) in 2D-to-3D (2D_3D) and 3D-to-2D (3D_2D) transitions of anaglyph video, nine visual healthy individuals were prepared for testing in this pioneering study. Spectrogram graphs based on short time Fourier transform (STFT) were considered to evaluate the power spectrum analysis in each EEG channel of transition or steady state. Thus, in 2D and 3D transition scenarios, important channels representing EEG frequency bands and brain lobes will be identified. To classify the 2D and 3D transitions, the dominant bands and time intervals representing the maximum difference of PSD were selected. Afterward, effective features were selected by applying statistical methods such as standard deviation, maximum (max) and Hjorth parameters to epochs indicating transition intervals. Ultimately, k-nearest neighbors, support vector machine and linear discriminant analysis (LDA) algorithms were applied to classify 2D_3D and 3D_2D transitions. The frontal, temporal and partially parietal lobes show 2D_3D and 3D_2D transitions with a good classification success rate. Overall, it was found that Hjorth parameters and LDA algorithms have 71.11% and 77.78% classification success rates for transition and steady state, respectively.


Author(s):  
Jingxia Chen ◽  
Dongmei Jiang ◽  
Yanning Zhang ◽  
◽  

To effectively reduce the day-to-day fluctuations and differences in subjects’ brain electroencephalogram (EEG) signals and improve the accuracy and stability of EEG emotion classification, a new EEG feature extraction method based on common spatial pattern (CSP) and wavelet packet decomposition (WPD) is proposed. For the five-day emotion related EEG data of 12 subjects, the CSP algorithm is firstly used to project the raw EEG data into an optimal subspace to extract the discriminative features by maximizing the Kullback-Leibler (KL) divergences between the two categories of EEG data. Then the WPD algorithm is used to decompose the EEG signals into the related features in time-frequency domain. Finally, four state-of-the-art classifiers including Bagging tree, SVM, linear discriminant analysis and Bayesian linear discriminant analysis are used to make binary emotion classification. The experimental results show that with CSP spatial filtering, the emotion classification on the WPD features extracted with bior3.3 wavelet base gets the best accuracy of 0.862, which is 29.3% higher than that of the power spectral density (PSD) feature without CSP preprocessing, is 23% higher than that of the PSD feature with CSP preprocessing, is 1.9% higher than that of the WPD feature extracted with bior3.3 wavelet base without CSP preprocessing, and is 3.2% higher than that of the WPD feature extracted with the rbio6.8 wavelet base without CSP preprocessing. Our proposed method can effectively reduce the variance and non-stationary of the cross-day EEG signals, extract the emotion related features and improve the accuracy and stability of the cross-day EEG emotion classification. It is valuable for the development of robust emotional brain-computer interface applications.


2021 ◽  
Vol 7 ◽  
pp. e766
Author(s):  
Ammar Amjad ◽  
Lal Khan ◽  
Hsien-Tsung Chang

Speech emotion recognition (SER) is a challenging issue because it is not clear which features are effective for classification. Emotionally related features are always extracted from speech signals for emotional classification. Handcrafted features are mainly used for emotional identification from audio signals. However, these features are not sufficient to correctly identify the emotional state of the speaker. The advantages of a deep convolutional neural network (DCNN) are investigated in the proposed work. A pretrained framework is used to extract the features from speech emotion databases. In this work, we adopt the feature selection (FS) approach to find the discriminative and most important features for SER. Many algorithms are used for the emotion classification problem. We use the random forest (RF), decision tree (DT), support vector machine (SVM), multilayer perceptron classifier (MLP), and k-nearest neighbors (KNN) to classify seven emotions. All experiments are performed by utilizing four different publicly accessible databases. Our method obtains accuracies of 92.02%, 88.77%, 93.61%, and 77.23% for Emo-DB, SAVEE, RAVDESS, and IEMOCAP, respectively, for speaker-dependent (SD) recognition with the feature selection method. Furthermore, compared to current handcrafted feature-based SER methods, the proposed method shows the best results for speaker-independent SER. For EMO-DB, all classifiers attain an accuracy of more than 80% with or without the feature selection technique.


Author(s):  
Caroline Dakoure ◽  
Mohamed Sahbi Benlamine ◽  
Claude Frasson

It is of great importance to detect users’ confusion in a variety of situations such as orientation, reasoning, learning, and memorization. Confusion affects our ability to make decisions and can lower our cognitive ability. This study examines whether a confusion recognition model based on EEG features, recorded on cognitive ability tests, can be used to detect three levels (low, medium, high) of confusion. This study also addresses the extraction of additional features relevant to classification. We compare the performance of the K-nearest neighbors (KNN), support vector memory (SVM), and long short-term memory (LSTM) models. Results suggest that confusion can be efficiently recognized with EEG signals (78.6% accuracy in detecting a confused/unconfused state and 68.0% accuracy in predicting the level of confusion). Implications for educational situations are discussed.


Sign in / Sign up

Export Citation Format

Share Document