Brainsci 14 00271
Brainsci 14 00271
Brainsci 14 00271
sciences
Article
Subject-Independent Emotion Recognition Based on EEG
Frequency Band Features and Self-Adaptive Graph Construction
Jinhao Zhang, Yanrong Hao, Xin Wen, Chenchen Zhang, Haojie Deng, Juanjuan Zhao and Rui Cao *
Abstract: Emotion is one of the most important higher cognitive functions of the human brain and
plays an important role in transaction processing and decisions. In traditional emotion recognition
studies, the frequency band features in EEG signals have been shown to have a high correlation
with emotion production. However, traditional emotion recognition methods cannot satisfactorily
solve the problem of individual differences in subjects and data heterogeneity in EEG, and subject-
independent emotion recognition based on EEG signals has attracted extensive attention from
researchers. In this paper, we propose a subject-independent emotion recognition model based
on adaptive extraction of layer structure based on frequency bands (BFE-Net), which is adaptive
in extracting EEG map features through the multi-graphic layer construction module to obtain a
frequency band-based multi-graphic layer emotion representation. To evaluate the performance of
the model in subject-independent emotion recognition studies, extensive experiments are conducted
on two public datasets including SEED and SEED-IV. The experimental results show that in most
experimental settings, our model has a more advanced performance than the existing studies of
the same type. In addition, the visualization of brain connectivity patterns reveals that some of the
findings are consistent with previous neuroscientific validations, further validating the model in
subject-independent emotion recognition studies.
Citation: Zhang, J.; Hao, Y.; Wen, X.; Keywords: electroencephalogram; frequency band; subject-independent; adjacency matrix; SEED
Zhang, C.; Deng, H.; Zhao, J.; Cao, R. dataset; deep learning
Subject-Independent Emotion
Recognition Based on EEG Frequency
Band Features and Self-Adaptive
Graph Construction. Brain Sci. 2024, 1. Introduction
14, 271. https://doi.org/10.3390/
brainsci14030271
Emotion is one of the most important high-level cognitive functions of the human
brain, appearing in all aspects of human life. Different emotions can affect the physical
Academic Editor: Ateke Goshvarpour health of human beings and, at the same time, emotion also plays an important role
Received: 31 January 2024
in the processing of affairs and rational decision-making [1]. Recognition of emotion is
Revised: 3 March 2024 an advanced embodiment of artificial intelligence, and related research has become a
Accepted: 5 March 2024 hotspot of cross-research in many interdisciplinary fields, such as computer, neuroscience,
Published: 12 March 2024 psychology, brain science, biomedical engineering, and robotics, aiming to analyze, explain,
and recognize human emotions and provide results that will promote our understanding
of the cognitive mechanism of emotion [2]. The human brain is an extremely complex
system in nature, and its complexity is not only manifested in the hundreds of billions of
Copyright: © 2024 by the authors. neurons and trillions of connections but also in the diversity of connection patterns, i.e.,
Licensee MDPI, Basel, Switzerland.
in the different patterns of connections manifested in cognition, thought, sensation, and
This article is an open access article
behavior. Furthermore, the connection patterns of the brains of different individuals are
distributed under the terms and
not the same [3].
conditions of the Creative Commons
With the advantages of noninvasiveness, high temporal resolution, and easy acqui-
Attribution (CC BY) license (https://
sition, EEG signals are widely used in brain science research. In the past decades, many
creativecommons.org/licenses/by/
4.0/).
neuropsychological studies have revealed the correlation between cortical regions and
human emotions. The different emotions generated in the hypothalamus, amygdala, hip-
pocampus, deep limbic system, and anterior band gyrus are aroused and relayed through
the brainstem reticular formation, which is modulated and integrated by the frontal and
temporal lobes of the cerebral cortex [4–6]. The contribution of EEG signals from different
regions of the cerebral lobe is different during emotional cognition [7].
In the field of EEG emotion recognition, various studies have been devoted to exploring
effective methods for extracting nonlinear and complex EEG features to recognize emotions.
It has been shown that introducing handcrafted feature extraction into classification models
can improve emotion recognition performance to some extent. Various hand-crafted
features with superior ability to enhance different emotional features were employed. For
example, higher-order cross-features [8] and Hjorth features [9], which are time-domain
features, are capable of extracting temporal information from a signal. Power Spectral
Density (PSD) [10], Wavelet Transform [11,12], and Discrete Wavelet Transform [13], which
have the ability to capture local features in the frequency domain, have been widely
used in the field of processing EEG emotion recognition. In addition, building brain
networks by exploring the relationships among EEG channels has also been used as a
feature extraction method. For example, some works have constructed brain networks
based on Pearson’s correlation coefficient [14], mutual information [15], etc., which are used
in brain network modeling. In addition, many works have used entropy measurements
to extract discriminative features by measuring the complexity of EEG signals, such as
Shannon entropy (ShEn) [16], sample entropy (SampEn) [17], and differential entropy
(DE) [18]. Differential entropy has superior robustness and feature extraction ability and
has been widely used in EEG signal analysis, especially in emotion recognition based on
EEG signals [19,20]
In terms of EEG feature selection, the more common traditional EEG features are
time domain features, frequency domain features, and time–frequency domain features.
The main features in the time domain such as statistical features [21], fractal dimension
features [22], etc., can be used for emotion characterization. In contrast to time domain
features, frequency domain analysis methods can reveal the frequency components of
a signal [23]. In the frequency domain, power spectral density [20] and approximate
entropy [24] are often used in research. With the deepening of the study, researchers have
found that in order to more comprehensively respond to the characteristic information
of EEG signals, combining time and frequency domain features to form time–frequency
domain features should be used for the comprehensive analysis of EEG signals. For
example, wavelet transform entropy [25] and discrete wavelet transform features [26] are
widely used in the field of EEG emotion recognition and have achieved good performance.
Most of the above traditional EEG features are extracted and studied based on a single
channel, and according to previous studies, it has been shown that the huge number of
neurons and brain regions in the brain are interconnected to form a complex network [27].
In order to further understand the specific emotional states of the brain and consider the
correlations and interactions between channels, the research method of constructing inter-
channel adjacency matrices using spatial distances of EEG channels [28] and functional
connectivity metrics has also been gradually adopted by most researchers. Functional
connectivity metrics are usually selected such as the phase-locked value PLV [29], Pearson
correlation coefficient PCC [30], etc. However, in most of the previous studies, the extraction
of the adjacency matrix was based on a priori knowledge, which did not fully consider the
correlation relationship between channels embedded in the EEG signals. In contrast, the
method proposed in this study is more capable of utilizing neural networks to adaptively
explore the inter-channel relationships in EEG signals to obtain brain connectivity patterns
with more emotional representations.
With the continuous development of artificial intelligence, machine learning and
deep learning methods are being gradually applied in the research field of EEG emotion
recognition, and the more representative ones mainly include support vector machine
SVM [31], CNN [32], RNN [33], and other methods, which have achieved good results. It
Brain Sci. 2024, 14, 271 3 of 19
has been shown that traditional neural networks cannot directly deal with non-Euclidean
data. EEG signals are discrete and discontinuous in the spatial domain and thus, it is more
favorable to construct the structure of EEG effective graphs based on the knowledge of
graph theory and use graph neural networks to deal with the information in the graph
domain to better characterize the intrinsic relationship between the channels [20]. A graph
neural network is a type of neural network that processes data in the graph domain,
such as molecular structures, social networks, and knowledge graphs [34]. A graph
convolutional neural network is a neural network with faster localized convolutional
operations, where the convolutional layers can be stacked K times to efficiently convolve
the K-order neighborhood of a node [35]. Therefore, in the field of EEG emotion recognition,
graph representations have also achieved satisfactory performance in processing EEG
signals, e.g., a Dynamic Graph Convolutional Neural Network (DGCNN) for emotion
recognition was proposed, whose graph structure is defined by a dynamic adjacency matrix
reflecting the intrinsic relationship between different EEG electrodes [36]. To capture
local and global inter-channel relationships, a regularized graph neural network (RGNN)
was proposed, which achieved state-of-the-art performance on the SEED and SEED-IV
datasets [37]. Since the attention mechanism [38] was proposed, it has been a focal point
in the field of deep learning. Thus, Transformer has been widely used in the fields of
translation, imaging, etc. The use of the attention mechanism in Transformer helps to
capture the long-term relevance of the data and improves the interpretability. Indeed,
it is used in the field of Brain–Computer Interfaces to capture deep features from the
EEG signals [39]. Therefore, in this paper, inspired by using graph representations with
Transformer in the field of brain science, a new adaptive multi-graph layer research method
based on Transformer components is proposed.
The issue of subject independence in the field of EEG emotion recognition research
has been of great interest. Subject independence refers to the use of different individuals
as subjects in experiments to verify the scalability and robustness of emotion recognition
models. EEG signals are usually weak and prone to be adulterated with noise, as well as
non-stationary properties, resulting in large differences in EEG signals between subjects, or
even between the same subject over longer time spans. In a previous study, a multi-source
domain adaptive approach was proposed that considered both domain-invariant and
domain-specific features, and a one-to-one domain adaptation method was used to extract
domain features to reduce the impact of EEG variability on emotion recognition studies [40].
Therefore, in this study, the subject-independent experimental division method is used as
the only index to evaluate the model, aiming to explore the performance of the proposed
method in the EEG heterogeneity problem, and the results show that the proposed method
achieves more stable performance.
Meanwhile, among the existing research methods for EEG-based emotion recognition,
both machine learning and deep learning algorithms have achieved better results. These
algorithms are more appropriate for dealing with complex problems by virtue of their
strong learning ability and have been introduced into EEG signal-based emotion recognition.
However, there are still some shortcomings in the current research including the following:
First, frequency band features should not be neglected because it has been shown
that EEG signals of different frequency bands have different relevance in emotion recogni-
tion [20]. Second, multichannel EEG signals have structural characteristics of biological
topography in non-Euclidean domains [36]. Directly applying deep learning methods to
EEG-based feature recognition does not allow for better characterization of emotions, as
these methods are designed for computer vision and natural language processing tasks.
Previous research methods have manually extracted connectivity metrics between different
brain regions through a priori knowledge. However, due to the non-smoothness and
specificity of EEG, in the field of emotion recognition research, the use of self-adaptive
methods to measure brain connectivity patterns has become a new perspective to explore
connectivity patterns, which dynamically determines connectivity metrics between dif-
ferent brain regions from input EEG signals or features and provides better access to the
Brain Sci. 2024, 14, 271 4 of 19
connectivity metrics between different emotional states of the channels and the connectivity
between brain regions [41]. In addition, EEG signals vary significantly between individuals,
which makes subject-independent emotion recognition studies a challenge.
To solve the above three problems in EEG-based emotion recognition research, we
propose a neural network model based on extracted frequency band layer features (BFE-
Net), which is a new frequency band-based self-adaptive graph construction emotion
recognition model, and fully consider the significance of subject-independent research in
our experiments.
(1) In order to research the contribution of EEG frequency bands to emotion recognition,
we use DE features extracted from five frequency bands as inputs to the model.
(2) To explore the spatial topological information embedded in EEG signals, we use CNNs
and Transformer models to adaptively extract the frequency band graph layer structure.
(3) We use a Graphical Convolutional Neural Network (GCN) to aggregate features to
obtain a single-band representation of emotion and recognize emotion by fusing
features from the five bands.
Figure 1. Schematic
Figure 1. Schematic diagram
diagram of
of the
the 62
62 electrodes
electrodes in
in the
the EEG
EEG caps
caps used
used for
for the
the SEED
SEED and
and SEED-IV
SEED-IV
datasets. The figure shows the approximate location of each electrode in the brain.
datasets. The figure shows the approximate location of each electrode in the brain.
In the SEED
During dataset, W varies
data preprocessing, fromdata
the EEG 185 were
to 265, while in the to
down-sampled SEED-IV dataset,
200 Hz and W
filtered
varies from 12 to 64. To standardize the data lengths, the length of the SEED dataset was
using a 0–75 Hz band-pass filter to divide the EEG data into 4 s non-overlapping time
standardized to 265, the length of the SEED-IV dataset was standardized to 64, and the
windows of data. To make a fair comparison with existing studies, the differential entropy
features with a short temporal window were used with zeros as padding. Therefore, the
(DE) feature provided by the dataset, smoothed by a linear dynamical system (LDS), was
data format of each feature sample of the SEED dataset was 62 × 265 × 5 and that of the
used directly in this experiment. Differential entropy extends the concept of Shannon’s
SEED-IV dataset was 62 × 64 × 5.
entropy and measures the complexity of continuous random variables, and according to
This study evaluates and compares different emotion recognition models based on the
previous studies, DE features are more effective for emotion recognition compared with
publicly available datasets SEED and SEED-IV.
other features. For each EEG signal per second in each EEG channel, DE features on five
frequency
2.2. Proposed bands (delta, theta, alpha, beta, and gamma) were extracted. Therefore, the data
Methodology
format of DE features for each subject in one experiment was the following: 62 × W × 5,
An overview of the general model proposed in this study is shown in Figure 2. To
where 62 denotes the EEG channel, 5 denotes the five frequency bands mentioned above,
fully consider the contribution of frequency band information in EEG signals to emotion
and W denotes the number of time windows in each trial, with different trials having
recognition, a frequency band-based EEG emotion classification network is designed in
different
this study.WThe values
modelbecause the video durations
implementation are not exactly
will be publicly available equal in different trials.
at https://github.com/
Doubleb0424/BFEnet (accessed on 2 March 2024). The input of the model dataset,
In the SEED dataset, W varies from 185 to 265, while in the SEED-IV W varies
is represented as
from
XϵR N ×12
T ×to
C , where N denotes the number of EEG channels, T denotes the sample lengthwas
64. To standardize the data lengths, the length of the SEED dataset of
astandardized
single band DE to 265, the and
feature, length of the SEED-IV
C Indicates the numberdataset was standardized
of frequency to 64, and the
bands characterized by
features
DE, i.e., Cwith
= 5,awhere
short temporal window
the five bands were used
are delta, theta,with zeros
alpha, beta,asand
padding.
gammaTherefore, the
respectively.
data format
After that, it of eachinto
is fed feature sample of band
the frequency the SEED dataset
feature was 62network
extraction × 265 × to
5 and that of the
aggregate the
SEED-IV dataset was 62 × 64 × 5.
features, and the obtained features are spliced to obtain the full band fusion features. Finally,
This study
the emotion evaluates
recognition and compares
is realized different
after the emotion layer
fully connected recognition modelslayer.
and SoftMax based on
the publicly available datasets SEED and SEED-IV.
In a single band feature extraction module, a Band Feature Extraction Neural Network
(BFE-Net) is proposed, as shown in Figure 3. The BFE-Net consists of three main modules,
2.2. Proposed
namely, Methodology neural network layer (CNN layer), the multi-graphic layer
the convolutional
construction module,
An overview of and
the the graph
general convolution
model andinfeature
proposed fusion
this study is layer
shown(GCN layer).2.The
in Figure To
DE features of each frequency band are fed into the BFE-Net, which first
fully consider the contribution of frequency band information in EEG signals to emotiongoes through
the convolutional
recognition, layer toband-based
a frequency extract theEEG
deepemotion
features,classification
is then fed network
into the multi-graphic
is designed in
layer
this construction
study. The module to construct
model the graphic layer
implementation will features, and, finally,
be publicly is fed into
available at
the graph convolutional neural network to(accessed
https://github.com/Doubleb0424/BFEnet. aggregate onthe 2 features and obtain
March 2024). the feature
The input of the
representation of a single
model is represented 𝑋𝜖ℝ𝑁×𝑇×𝐶 , where 𝑁 denotes the number of EEG channels, 𝑇
as band.
denotes the sample length of a single band DE feature, and 𝐶 Indicates the number of
frequency bands characterized by DE, i.e., 𝐶 = 5, where the five bands are delta, theta,
alpha, beta, and gamma respectively. After that, it is fed into the frequency band feature
Brain Sci. 2024, 14, x FOR PEER REVIEW 6 of 20
extraction network to aggregate the features, and the obtained features are spliced to
Brain Sci. 2024, 14, 271 obtain the full band fusion features. Finally, the emotion recognition is realized after
6 of the
19
fully connected layer and SoftMax layer.
Figure2.2.Schematic
Figure Schematicdiagram
diagramofofthe
thearchitecture
architectureofofan anemotional
emotionalnetwork
networkbased
basedononEEG
EEGfrequency
frequency
bands.The
bands. Theinput
inputofofthe
thenetwork
networkisisthe
thefeature
featureofoffive
fivefrequency
frequencybands.
bands.After
Afterlearning
learningby
bythe
theBand
Band
FeatureExtraction
Brain Sci. 2024, 14, x FOR PEER REVIEW
Feature ExtractionNeural
NeuralNetwork,
Network,thethefused
fusedfeatures
featuresare areused
usedfor
foremotion
emotionrecognition,
recognition,and
7and the
ofthe
20
output of the network is emotion labels.
output of the network is emotion labels.
Figure
Figure 3.
3. The
The proposed
proposed BFE-Net
BFE-Net model,
model, whose
whose inputs
inputs are
are single
single band features. The
band features. The model
model uses
uses aa
CNN layer and layer construction with a GCN layer to obtain a single-band sentiment
CNN layer and layer construction with a GCN layer to obtain a single-band sentiment representation.
representation.
2.2.1. CNN Layer
2.2.1.The
CNN Layer
input DE features are fed into the convolutional layer, which plays an important
Thethe
role in input DE features
model. are fed model
The BFE-Net into thecan
convolutional
extract EEGlayer, which
features atplays an important
different levels of
role in the using
abstraction model.theThe BFE-Net model
convolutional can extract
layer. Taking EEG dataset
the SEED featuresasat
andifferent
example,levels of
its input
abstraction
size is 62 × using the convolutional
265 (number layer. Taking
of EEG channels the SEED
× DE feature dataset as which
dimension), an example, its input
represents the
size is 62 × 265 (number of EEG channels × DE feature dimension), which represents the
DE feature data of a single frequency band of a subject. To make the input DE features
more capable of characterizing emotion, and considering the small dimension of DE
feature data, a CNN is chosen to be used to aggregate the features. Based on the non-
Euclidean characteristics of EEG signals and the need to use single-channel features as the
Brain Sci. 2024, 14, 271 7 of 19
DE feature data of a single frequency band of a subject. To make the input DE features
more capable of characterizing emotion, and considering the small dimension of DE feature
data, a CNN is chosen to be used to aggregate the features. Based on the non-Euclidean
characteristics of EEG signals and the need to use single-channel features as the node
features in the graph structure when constructing the graph structure, as well as to ensure
that the features between EEG channels do not mix with each other and then extract the
single-channel EEG features, we chose to use a one-dimensional convolutional neural
network to extract the single-channel features of the EEG. The module consists of three
consecutive convolutional layers, each of which consists of a one-dimensional convolutional
kernel, a dropout, and a maximum pooling layer. The convolutional kernel sizes were
all selected as 1 × 5, and it was proposed to extract the EEG single-channel features at
different abstraction levels with 64,128,256 convolutional kernels in the three convolutional
layers, respectively. The dropout was set to 0.1 to prevent the overfitting phenomenon from
occurring, while the maximum pooling layer was applied to down-sample the features
afterward. In the first and second layers of the network structure, as shown in Figure 3, the
outputs of the CNNs are fed into the connected multi-layer construction module and the
next layer of CNNs, respectively, aiming to deepen the depth of the network while carrying
out deep feature extraction of the original DE features and obtaining the single-channel
features with a greater characterization capability.
Brain Sci. 2024, 14, x FOR PEER REVIEW 8 of 20
2.2.2. Multi-Graphic Layer Construction
This module uses the CNN feature as input and proposes a new method of adaptively
learning the adjacency matrix (i.e., characterizing the correlations between different EEG
Transformer
channels) usingmodel
neuraltonetworks.
adaptively extract
It uses the adjacency
the Encoder matrix,
component and
in the then latermodel
Transformer uses this
adjacency
to adaptivelymatrix
extractwith the CNN matrix,
the adjacency featureand
constructed
then later as a this
uses graph structure
adjacency as the
matrix layer
with
structure
the of this constructed
CNN feature network hierarchy.
as a graphThe way theasgraph
structure is constructed
the layer structure of for
thiseach layer is
network
shown in The
hierarchy. Figure
way4.the graph is constructed for each layer is shown in Figure 4.
The distribution positions of the electrodes of the EEG cap are defined by a number
of standards, such as the International 10/20 System. The distribution positions of the
electrodes are fixed and regular, and thus, EEG signals can be considered as classical non-
Euclidean structured data, which are well suited for graphical data representation.
In addition, inspired by the successful application of Transformer in the field of NLP,
researchers introduced it into the field of computer vision and proposed the VIT-
Brain Sci. 2024, 14, 271 8 of 19
The distribution positions of the electrodes of the EEG cap are defined by a number
of standards, such as the International 10/20 System. The distribution positions of the
electrodes are fixed and regular, and thus, EEG signals can be considered as classical
non-Euclidean structured data, which are well suited for graphical data representation.
In addition, inspired by the successful application of Transformer in the field of
NLP, researchers introduced it into the field of computer vision and proposed the VIT-
Transformer model, which compensates for the shortcomings of Transformer in processing
sequential data and can utilize the mechanism of self-attention to better capture spatial
and temporal information. Therefore, in this study, the Encoder part of VIT-Transformer is
introduced and improved, which is utilized to adaptively extract the intrinsic correlation
relationship between different EEG channels and then obtain the edge features in the graph
structure. In each layer of the network structure, the feature data outputted from the CNN
are fed into the Encoder. Let the input feature data be Xi :
where WQ , WK , WV is the weight matrix for learning. The query (Qi ), key (Ki ), and value
(Vi ) vectors are split into multiple headers, respectively. In Vit-Transformer, it is usually
split into the h attention header, a hyperparameter, which can be derived from previous
studies and usually takes values of a constant 8 or 12. In the experiments, a smaller number
of attention heads can make the model computationally more efficient, and a larger number
of attention heads can provide richer expressive capabilities. In this study, the data volume
is small, and to ensure that the proposed method is more practical, the number of model
parameters cannot be too large, so it is set to h = 8.
For the self-attention sublayer in the Encoder section, for each Xi , its attention output
is calculated as: !
Qi KiT
Attention( Qi , Ki , Vi ) = so f tmax √ Vi (3)
dk
" !#
Qi KiT
A = Dropout so f tmax √ (4)
dk
√
where dk denotes the scaling factor and dk is the dimension of Qi and Ki in the attention
header. In this study, the transpose matrix of the query value Qi and the key value Ki
is subjected to a matrix product operation, which is used to characterize the correlation
relationship between the EEG channels embedded in the EEG features after the so f tmax
and Dropout operations. It is represented by using A in Equation (4), i.e., the adjacency
matrix in Figure 4.
non-linearly. G denotes the output of the encoder part of the method, i.e., the G-feature in
Figure 4, which contains the global and local feature information in the EEG.
S = GG T (7)
E = so f tmax ( A + S) (8)
The multiplication operation of G with the self-transpose vector is performed to obtain
a new matrix S, denoting the self-attention matrix in Figure 4. Then, the self-attention matrix
is added with the output adjacency matrix in the multi-head attention, and the so f tmax
activation function is used to obtain the bounded and positive new adjacency matrix,
denoted by E in Equation (8), to represent the edge features in the graph structure, i.e., the
edge feature in Figure 4. The significance of this is that the EEG features, after being partially
learned by Encoder, contain more global and local emotional representations, which are
expressed after transposing and multiplying them, and the global and local connections
embedded in the EEG channels are better presented. In each network depth, the edge
features of inter-channel connections and graph structures are dynamically determined by
the corresponding input features.
In general, the computational cost of sparse graphs is much lower than that of com-
plete graphs. In this study, to construct the sparse graph structure, we introduce the
top-k technique, i.e., the first k largest weights in the adjacency matrix are retained while
the remaining small connection weights are set to zero. The top-k operation is applied
as follows:
f or i = 1, 2, · · · N
index
h = argtopk
i ( E[i, :]) (9)
E i, index = 0
where argtopk(·) is a function to obtain the indexes of the first k largest values of each vector
E[i, :] in the adjacency matrix E. index denotes the range of indexes that do not belong to
the first k values in E[i, :]. In the generated adjacency matrix, after passing through the
top-k technique, only the first k maxima in each row vector of the adjacency matrix are
retained, while the remaining values are assigned as 0. In fact, the top-k technique can be
considered as an improved maximum pooling layer.
G = (V , ε) (10)
V = {Vi |i = 1, 2, · · · , N } (11)
ε = ε ij |Vi , V j ϵV (12)
E = eij (13)
Single-channel EEG signals collected by the EEG cap can be considered as nodes of a
graph. Therefore, we consider a multichannel EEG signal as a graph structure. G denotes
the graph, V denotes the set of vertices in graph G, and ε denotes the set of edges in graph G.
N is the number of EEG channels in the EEG signal. In the graph structure representation,
node Vi is typically used to represent a single-channel EEG signal, while edge ε ij denotes
the connectivity between node Vi and V j . E denotes the adjacency matrix of graph G. eij
denotes the strength of the associated connectivity between nodes Vi and V j . The set of
edges in the multi-graphic layer construction obtained through the above modeling is
Brain Sci. 2024, 14, 271 10 of 19
where H ′ is the node feature representation of the output, σ denotes the Relu activation
function, and D̂ is the diagonal matrix with diagonal elements Dii = ∑ j Âij . The core idea
is to utilize the spectral information of the graph to perform a convolution operation by
feature decomposition of the Laplace matrix. This enables spectral graph convolution to
perform effective feature propagation to nodes while preserving the graph structure.
The output of the graph convolution layer is expanded and concatenated into feature
vectors in each BFE-Net to characterize the feature output of that frequency band. The
feature vectors of the five frequency bands are concatenated and fed into a fully connected
layer with a SoftMax activation function to predict emotional categories. The band emo-
tional recognition model in this study can be trained by minimizing the cross-entropy error
between its predicted and true values.
average area under the curve (AUC) of the model was monitored from all emotional cat-
egories by applying a random dropout operation with a dropout of 0.1 during training.
If the average AUC reached 0.99 during the training process, the training process was
stopped, and the last saved model weights were used to categorize the subjects for emotion
recognition. For SEED and SEED-IV with 15 subjects, each round of experiments was
conducted in 15 sessions, and the average validation accuracy was considered as the final
performance of the model so that it could be compared and evaluated with other emotion
recognition studies.
3. Results
This section focuses on the metric performance of the frequency band and self-adaptive
graph convolution-based BFE-Net proposed in this paper on the SEED and SEED-IV. It
discusses the extent to which the network hierarchy and parameter tuning affect the results,
as well as the ability to discriminate among different emotional categories.
In the experimental results for the full frequency band, BFE-Net obtained an average
accuracy of 92.29% in SEED and achieved an average accuracy of 79.81% in SEED-IV, where
both values were higher than the other methods using the same feature input. Meanwhile,
BFE-Net obtained 4.65% and 4.11% standard deviation in experiments with different
subjects, which were lower than the other research methods in the table. This indicates that
the BFE-Net proposed in this study has a more stable performance in subject-independent
emotion recognition studies.
Brain Sci. 2024, 14, 271 12 of 19
Table 2. Impact of network depth on the model performance of BFE-Net (i.e., mean/standard
deviation). N − layer represents the number of network layers in the BFE-Net, N = 1, 2, 3.
Dataset
SEED SEED-IV
Layers
One-layer 85.40/05.18 73.98/05.42
Two-layer 88.53/05.01 75.09/04.95
Three-layer 92.29/04.65 79.81/04.11
Note: Bolded numbers indicate the highest accuracy with the lowest standard deviation.
In the multi-graphic layer construction module, this study uses the Encoder component
of Transformer to dynamically obtain an adjacency matrix suitable for the EEG signals
of different subjects. After TOP-K filtering, sparse adjacency matrices are obtained so as
to construct the corresponding graph structures for obtaining affective representations
independently from other subjects. As shown in Table 3, four different K-value sizes, i.e.,
(K = 5, 10, 15, 20), were selected for this experiment to discuss the effect of the K-value on
the performance of emotion recognition, and the inputs to the same model were full-band
features. From the table, it can be seen that different classification accuracies are achieved
as the value of k is varied, and the best performance of the model is achieved when K = 10.
Table 3. Effect of K-value size on the model performance of BFE-Net (i.e., mean/standard deviation).
Dataset
SEED SEED-IV
K
K=5 90.51/04.40 75.18/05.43
K = 10 92.29/04.65 79.81/04.11
K = 15 91.11/05.79 74.16/03.71
K = 20 89.92/05.18 75.00/04.75
Note: Bolded numbers indicate the highest accuracy with the lowest standard deviation.
This is due to the fact that both SEED and SEED-IV used in this study have 62 EEG
channels, and in the adjacency matrix, selecting a smaller K-value will discard more feature
information, which makes the model’s ability to learn to generalize decrease. On the
contrary, when a larger K-value is selected, the adjacency matrix automatically acquired
by the model will contain more emotional brain connectivity patterns, resulting in an
increase in accuracy. However, at the same time, as the K-value increases, the number of
model parameters becomes larger, the feature information of the adjacency matrix becomes
redundant, and the effective features are less likely to be captured by the model, which
reduces the recognition performance of the model. Therefore, the results reported in Table 1
are the performance at the time K = 10 was chosen.
Brain Sci. 2024, 14, 271 13 of 19
Dataset
SEED SEED-IV
Model
BFE-Net 92.29/04.65 79.81/04.11
w/Distance 80.52/12.35 65.31/09.23
w/PLV 87.56/10.89 70.99/11.78
w/o Self-Matrix 86.15/09.78 75.16/08.65
Note: Bolded numbers indicate the highest accuracy with the lowest standard deviation.
From Table 4, we can see that if the spatial distance is used as the adjacency matrix, the
average accuracy of the model decreases by 11.77% in SEED and 14.5% in SEED-IV; when
PLV is used as the adjacency matrix, the average accuracy of the model decreases by 4.73%
in SEED and 8.82% in SEED-IV; and when the self-attention matrix is removed, the average
accuracy of the model decreases by 6.14% in SEED and 4.65% in SEED-IV. The results
show that the method of extracting the adjacency matrix adaptively in BFE-Net is superior
to the method of extracting the adjacency matrix using spatial distances and functional
connectivity. In addition, it is able to effectively capture the specific emotional features of
different subjects, thus improving the performance of the model. The introduction of the
self-attention matrix can effectively enhance the EEG inter-channel connectivity features of
the self-adaptive adjacency matrix.
3.5. Visualization
To analyze the inter-channel relationships learned by the proposed model, we choose
SEED to visualize the adjacency matrix generated adaptively by the model, as shown
in Figure 6. To explore how brain regions are connected and activated under different
emotional states, firstly, the single-band adjacency matrix is subjected to an average nor-
malization operation, and then the diagonal elements in the adjacency matrix generated by
the three emotions are extracted and converted into a topographic map of the brain.
represents the predicted labels of the model, and the vertical axis represents the ground
truth labels. In the SEED dataset, the labels of the three categories are positive, neutral,
and negative from left to right and top to bottom, and neutral, sad, fear, and happy in the
SEED-IV, respectively. The BFE-Net model shows better performance in identifying
Brain Sci. 2024, 14, 271 negative, neutral, and positive emotions in SEED in terms of the classification results.
14 The
of 19
model recognition results were 87%, 95%, and 94%, respectively.
Figure
Figure 5.5. Confusion
Confusionmatrices
matrices for
for the
the SEED
SEED and
and SEED-IV
SEED-IV datasets,
datasets, denoted
denoted by
by (a,b).
(a,b). The
The horizontal
horizontal
Brain Sci. 2024, 14, x FOR PEER REVIEW 1
coordinates
coordinates inin the
the figure
figure indicate
indicate the
the predicted
predicted labels
labels and
and the
the vertical
vertical coordinates
coordinates indicate
indicate the
the true
true
labels.
labels. The
Thenumbers
numbers are are used
used to
to indicate
indicate the
the number
number of
of predictions
predictions obtained
obtained from
from the
the experiments.
experiments.
3.5. Visualization
To analyze the inter-channel relationships learned by the proposed model, we choose
SEED to visualize the adjacency matrix generated adaptively by the model, as shown in
Figure 6. To explore how brain regions are connected and activated under different
emotional states, firstly, the single-band adjacency matrix is subjected to an average
normalization operation, and then the diagonal elements in the adjacency matrix generated
by the three emotions are extracted and converted into a topographic map of the brain.
4. Discussion
4. Discussion
To further analyze the validity and feasibility of the methodology proposed in this
To further analyze the validity and feasibility of the methodology proposed
paper, the experimental results are discussed in detail in this section.
paper, the experimental results are discussed in detail in this section.
We conducted a series of experiments on two datasets, as shown in Table 1. Most of
We conducted a series of experiments on two datasets, as shown in Table 1. M
these methods focused on using DE features and frequency band features. In addition,
these methods focused on using DE features and frequency band features. In additi
all of them used subject-independent data partitioning to further discuss the feasibility of
of them used subject-independent data partitioning to further discuss the feasibility
the proposed methods. BFE-Net obtained higher accuracy with lower standard deviation
proposed methods. BFE-Net obtained higher accuracy with lower standard dev
compared with the same type of studies, which indicates the stability and validity of the
proposed method compared with the same type
in subject-independent of studies,
EEG-based which
emotion indicates the
recognition. Westability
analyzed and
in validity
proposed method in subject-independent EEG-based
detail the differences between BFE-Net and the existing studies as follows: emotion recognition. We ana
in detail the differences between BFE-Net and the existing studies as follows:
(1) SVM [37]: Support vector machine, abbreviated as SVM, is a classical supervised
(1) SVMalgorithm.
machine learning [37]: Support vector many
It exhibits machine, abbreviated
unique advantagesas in
SVM, is a small
solving classical supe
machine learning algorithm. It exhibits many
samples and nonlinear and high-dimensional pattern recognition. unique advantages in solving
samples and nonlinear and high-dimensional pattern
(2) SA [42]: A new domain adaptive algorithm. The source and target domains are recognition.
(2) bySA
represented [42]: A described
subspaces new domain adaptive
by feature algorithm. The source and target domai
vectors.
represented by subspaces described by feature vectors.
(3) DGCNN [36]: Multi-channel EEG-based emotion classification method bas
DGCNNs that initializes the adjacency matrix and trains the adjacency
dynamically through backpropagation.
(4) TANN [43]: A transferable attention neural network for EEG emotion recog
which learns the emotional discriminative information by highlightin
Brain Sci. 2024, 14, 271 15 of 19
(3) DGCNN [36]: Multi-channel EEG-based emotion classification method based on DGC-
NNs that initializes the adjacency matrix and trains the adjacency matrix dynamically
through backpropagation.
(4) TANN [43]: A transferable attention neural network for EEG emotion recognition,
which learns the emotional discriminative information by highlighting the trans-
ferable EEG brain region data and samples adaptively through local and global
attention mechanisms.
(5) BIDANN [44]: A neural network that maps left hemisphere and right hemisphere
EEG data into discriminative feature spaces separately, and the feature data are
later categorized.
(6) BIDANN-S [45]: A deep learning method for EEG-based emotion classification that
uses raw EEG features extracted from each cerebral hemisphere to extract discretized
deep features and a domain discriminator to mitigate domain differences between
source and target domains.
(7) BIHDM [46]: A bi-hemispheric discrepancy model that learns asymmetrical differ-
ences between two hemispheres using four recurrent neural networks to capture infor-
mation from EEG electrodes in each hemisphere from horizontal and vertical streams.
(8) RGNN [37]: A regularized GNN, which mainly contains node-domain adversarial
training and emotion-aware distribution algorithms to achieve emotion recognition
(9) SOGNN [47]: A Self-Organizing GNN for EEG Cross-Subject Emotion Classification,
which builds graph structures based on input EEG features self-using and uses GNN
learning features for emotion recognition.
The above methods involve algorithms such as machine learning, CNN, GNN, RNN,
and domain adaptation, which are used to research subject-independent EEG emotion
recognition from different perspectives. In previous machine learning-based studies, such
as SVM, feature extraction of EEG was performed based on a priori knowledge. The feature
information embedded in the raw EEG data was discarded. This challenge can be solved
by utilizing data-driven methods, which use neural networks to learn features for better
recognition performance. In previous studies based on domain adaptation, such as SA, the
feature space of source and target domains was utilized for emotion recognition, but the
spatial structure information embedded in EEG could not be well expressed. BIDANN and
BIDANN-S, mentioned in the above methods, integrate the characteristics of the brain’s left
and right hemispheric differences with domain adaptation for research. In addition to the
above two methods, BIHDM can also effectively assign the contribution of the brain’s left
and right hemisphere lateralization to emotion recognition. Brain networks represent the
complex connection patterns of the brain, and the introduction of GNN can effectively learn
the complex spatial topology in EEG, such as DGCNN, RGNN, and SOGNN. Different
graph structures can be constructed by using different adjacency matrices and EEG features.
Constructing the graph structure based on a priori knowledge has some limitations and
cannot correctly simulate the dynamic EEG of different subjects in different emotional states.
DGCNN can dynamically adjust the adjacency matrix, and SOGNN can adaptively extract
the adjacency matrix based on different feature inputs. These two methods can dynamically
build the graph structure according to the EEG of different subjects. In addition, TANN
introduces the attention mechanism and combines brain regions for emotion recognition.
The BFE-Net model proposed in this paper fully considers the advantages and dis-
advantages of the above methods. BFE-Net takes the frequency band features of EEG as
model inputs and utilizes a neural network to fully learn the features. A one-dimensional
convolutional kernel is used in the CNN layer to aggregate the EEG single-channel features,
which prevents the features of certain channels from being neglected and more realistically
restores the EEG generation process. The Transformer model is introduced to adaptively
acquire the adjacency matrix, and its output self-attention matrix is added to generate the
self-attention adjacency matrix. Because the Transformer model is suitable for processing
time series data, the temporal characteristics of the EEG obtain a better representation, and
the adjacency matrix contains global information, as shown in the third row of Table 4. In
Brain Sci. 2024, 14, 271 16 of 19
the multi-graphic layer construction module, three graph layers with different network
depths are constructed. As the network layers deepen, the graph structure of each layer is
dynamically constructed and independent of each other. After feeding the graph structure
into the GNN for further aggregation and learning, comprehensive and accurate senti-
ment features for a single frequency band can be obtained, as shown in Table 2. Finally,
the five single-band features are fused for emotion recognition. BFE-Net can construct
multi-graphical layers of emotion features independent of each other based on feature
inputs from different subjects, which makes it reliable in subject-independent EEG-based
emotion recognition.
In the single-band experimental results in Table 1, we compare the performance of
each model in different frequency bands. In general, most models, including ours, achieve
better recognition performance in the beta and gamma bands than the delta, theta, and
alpha bands. SOGNN achieves the best performance in the theta band, followed by the
beta and gamma bands. Because the higher frequency bands of EEG signals have better
emotional capabilities, the beta and gamma bands obtain a more reliable accuracy, thus
improving the accuracy, which is similar to the results of previous studies [37,42,45,47].
In Figure 5, we find that neutral and positive emotions are more easily recognized in
the SEED dataset, and negative emotions are more easily recognized as positive emotions.
This is because positive and negative emotions are more likely to cause large fluctuations
in EEGs, while neutral emotions cause smoother EEGs. In the SEED-IV dataset, sad,
happy, and fear emotions are more easily recognized, and neutral emotions are the worst
recognized. This is because neutral emotions produce smoother EEGs than the other three
emotions, resulting in lower recognition accuracy.
In Figure 6, we plot the pattern of EEG channel connectivity of BFE-Net under different
emotions. According to the topographic map, we derive three electrode locations with
maximum weights, namely, T7, FC1, and CP5, and show the five electrodes with maximum
weights connected to these three channels. In the three emotional states, the channels with
the largest weights have a larger weight between the channels with the largest weights and
the EEG channels that are mostly from the prefrontal lobe of the brain, i.e., the pattern of
brain region connections is mostly concentrated in the prefrontal lobe region of the brain.
According to previous studies [48], the activation of frontal lobe regions of the brain is
associated with positive and negative emotions. Also, the activation patterns between
the left and right hemispheres of the brain are more pronounced according to the planar
presentation, and it has been shown in previous studies that the asymmetry of the activity
of the EEG between the left and right hemispheres is crucial for emotion recognition [37].
In addition, asymmetric frontal and parietal regional connectivity patterns may reflect the
process of emotional change [49]. As demonstrated by previous studies, our findings are
consistent with the existing findings.
5. Conclusions
In this paper, a new subject-independent emotion recognition model named BFE-Net
is proposed. The following are the main contributions of this paper:
(1) A new neural network-based emotion recognition method for EEG frequency bands
(BFE-Net) is proposed, which constructs EEG multi-graphic layer emotion features by
adaptively acquiring the adjacency matrix.
(2) Based on the frequency band features of EEG, BFE-Net investigates the effect of
different frequency bands on emotion recognition. In the process of single-band
feature extraction, a new Transformer Encoder-based adaptive extraction of adjacency
matrix is proposed to extract the multilayer structure used to characterize emotion
according to the network depth. It can also be used to explore emotional activities
among different functional brain regions.
(3) BFE-Net obtains optimal performance in subject-independent emotion recognition,
with an average accuracy of 92.29% in SEED, as well as 79.81% in SEED-IV. Extensive
Brain Sci. 2024, 14, 271 17 of 19
ablation and comparison experiments are also conducted to validate the performance
of the proposed BFE-Net.
Based on the experimental results we can see that BFE-Net achieves advanced per-
formance on two publicly available EEG emotion datasets under the same experimental
setup. This indicates that the EEG frequency band features with multi-graph layer map
structure are effective in subject-independent emotion recognition. BFE-Net analyzes the
EEG features of different subjects independently, which effectively reduces the effect of
EEG data heterogeneity on subject-independent emotion recognition. We hope to provide
new perspectives for other EEG-based studies.
Admittedly, there are some limitations in this study. The validity of EEG time–
frequency domain features was demonstrated in previous studies based on EEG frequency
bands. Spectral analysis and wavelet transform can convert EEG to graphs for process-
ing [50]. Wavelet transform can detect and analyze time-varying signals [51]. Fast Fourier
Transform (FFT) can convert temporal signals from the time domain to the frequency do-
main [52]. It has been suggested that empirical mode decomposition (EMD) can decompose
an EEG into several intrinsic mode functions (IMFs), which can later be used to compute
features using autoregressive models for emotion recognition [53]. Furthermore, in terms
of neural network architectures, previous studies have shown that domain adaptation can
effectively reduce the effect of variability in EEG data [37,45,54]. Multimodal data have
also been applied in previous studies. The fusion of eye movement data with EEG data can
improve the performance of emotion recognition [55].
In our future research, we will complete the following:
(1) Explore more effective EEG features and introduce EEG spectral analysis, wavelet
transform, empirical pattern decomposition, and other feature analysis methods.
Build complex brain networks to study brain-specific connectivity patterns in emo-
tional states and effectively decode high-level cognitive behaviors.
(2) Explore other neural network architectures such as recurrent networks using LSTM
cells and autoencoders in the study of EEG frequency bands and brain networks.
(3) Improve the performance of emotion recognition and further explore human physi-
ological states in emotional states. Integrate multimodal physiological data such
as eye movement data, EMG signals, and facial expressions into emotion recogni-
tion research.
Author Contributions: Conceptualization, J.Z. (Jinhao Zhang); methodology, J.Z. (Jinhao Zhang), Y.H.
and X.W.; data curation, C.Z.; visualization, H.D.; writing—original draft preparation, J.Z. (Jinhao
Zhang) and C.Z.; writing—review and editing, J.Z. (Jinhao Zhang) and R.C.; supervision, J.Z. (Jinhao
Zhang), J.Z. (Juanjuan Zhao), and R.C. All authors have read and agreed to the published version of
the manuscript.
Funding: This research was funded by the National Natural Science Foundation of China (62206196),
the Natural Science Foundation of Shanxi (202103021223035, 202303021221001), and the Science and
Technology Innovation Program for Higher Education Institutions in Shanxi Province (RD2300004062).
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The open-access datasets SEED and SEED-IV were used in our study.
The links are as follows: SEED: https://bcmi.sjtu.edu.cn/~seed/seed.html (accessed on 13 March
2021). SEED-IV: https://bcmi.sjtu.edu.cn/~seed/seed-iv.html (accessed on 13 March 2021).
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Liu, H.; Zhang, Y.; Li, Y.; Kong, X. Review on emotion recognition based on electroencephalography. Front. Comput. Neurosci.
2021, 15, 84. [CrossRef] [PubMed]
2. Gabriels, K. Response to “uncertainty in emotion recognition”. J. Inf. Commun. Ethics Soc. 2019, 17, 295–298. [CrossRef]
Brain Sci. 2024, 14, 271 18 of 19
3. Boccaletti, S.; Latora, V.; Moreno, Y.; Chavez, M.; Hwang, D.-U. Complex networks: Structure and dynamics. Phys. Rep. 2006, 424,
175–308. [CrossRef]
4. Seo, H.; Oemisch, M. Decoding emotion: The amygdala–prefrontal cortex pathway for emotion regulation of children. Biol.
Psychiatry 2020, 88, 517–519. [CrossRef]
5. Rolls, E.T. The cingulate cortex and limbic systems for emotion, action, and memory. Brain Struct. Funct. 2019, 224, 3001–3018.
[CrossRef]
6. Yang, Z.; Wildschut, T.; Izuma, K.; Gu, R.; Luo, Y.L.; Cai, H.; Sedikides, C. Patterns of brain activity associated with nostalgia: A
social-cognitive neuroscience perspective. Soc. Cogn. Affect. Neurosci. 2022, 17, 1131–1144. [CrossRef]
7. Rolls, E.T. The neuroscience of emotional disorders. Handb. Clin. Neurol. 2021, 183, 1–26.
8. Petrantonakis, P.C.; Hadjileontiadis, L.J. Emotion recognition from EEG using higher order crossings. IEEE Trans. Inf. Technol.
Biomed. 2009, 14, 186–197. [CrossRef]
9. Hjorth, B. EEG analysis based on time domain properties. Electroencephalogr. Clin. Neurophysiol. 1970, 29, 306–310. [CrossRef]
10. Frantzidis, C.A.; Bratsas, C.; Papadelis, C.L.; Konstantinidis, E.; Pappas, C.; Bamidis, P.D. Toward Emotion Aware Computing: An
Integrated Approach Using Multichannel Neurophysiological Recordings and Affective Visual Stimuli. IEEE Trans. Inf. Technol.
Biomed. 2010, 14, 589–597. [CrossRef]
11. Subasi, A. EEG signal classification using wavelet feature extraction and a mixture of expert model. Expert Syst. Appl. 2007, 32,
1084–1093. [CrossRef]
12. Rosso, O.; Martin, M.; Plastino, A. Brain electrical activity analysis using wavelet-based informational tools. Phys. A Stat. Mech.
Its Appl. 2002, 313, 587–608. [CrossRef]
13. Orhan, U.; Hekim, M.; Ozer, M. EEG signals classification using the K-means clustering and a multilayer perceptron neural
network model. Expert Syst. Appl. 2011, 38, 13475–13481. [CrossRef]
14. Chen, H.; Song, Y.; Li, X. A deep learning framework for identifying children with ADHD using an EEG-based brain network.
Neurocomputing 2019, 356, 83–96. [CrossRef]
15. Li, P.; Liu, H.; Si, Y.; Li, C.; Li, F.; Zhu, X.; Huang, X.; Zeng, Y.; Yao, D.; Zhang, Y. EEG based emotion recognition by combining
functional connectivity network and local activations. IEEE Trans. Biomed. Eng. 2019, 66, 2869–2881. [CrossRef]
16. Vijayan, A.E.; Sen, D.; Sudheer, A. EEG-based emotion recognition using statistical measures and auto-regressive modeling. In
Proceedings of the 2015 IEEE International Conference on Computational Intelligence & Communication Technology, Ghaziabad,
India, 13–14 February 2015; pp. 587–591.
17. Hu, J.; Min, J. Automated detection of driver fatigue based on EEG signals using gradient boosting decision tree model. Cogn.
Neurodyn. 2018, 12, 431–440. [CrossRef] [PubMed]
18. Duan, R.-N.; Zhu, J.-Y.; Lu, B.-L. Differential entropy feature for EEG-based emotion classification. In Proceedings of the 2013 6th
International IEEE/EMBS Conference on Neural Engineering (NER), San Diego, CA, USA, 6–8 November 2013; pp. 81–84.
19. Zheng, W.-L.; Lu, B.-L. Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural
networks. IEEE Trans. Auton. Ment. Dev. 2015, 7, 162–175. [CrossRef]
20. Xing, B.; Zhang, H.; Zhang, K.; Zhang, L.; Wu, X.; Shi, X.; Yu, S.; Zhang, S. Exploiting EEG signals and audiovisual feature fusion
for video emotion recognition. IEEE Access 2019, 7, 59844–59861. [CrossRef]
21. Zhang, Q.; Lee, M. A hierarchical positive and negative emotion understanding system based on integrated analysis of visual
and brain signals. Neurocomputing 2010, 73, 3264–3272. [CrossRef]
22. Liang, Z.; Oba, S.; Ishii, S. An unsupervised EEG decoding system for human emotion recognition. Neural Netw. 2019, 116,
257–268. [CrossRef] [PubMed]
23. Zhang, Q.; Lee, M. Emotion recognition in natural scene images based on brain activity and gist. In Proceedings of the 2008 IEEE
International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China,
1–8 June 2008; pp. 3050–3057.
24. Namazi, H.; Aghasian, E.; Ala, T.S. Complexity-based classification of EEG signal in normal subjects and patients with epilepsy.
Technol. Health Care 2020, 28, 57–66. [CrossRef]
25. Gao, Q.; Wang, C.-H.; Wang, Z.; Song, X.-L.; Dong, E.-Z.; Song, Y. EEG based emotion recognition using fusion feature extraction
method. Multimed. Tools Appl. 2020, 79, 27057–27074. [CrossRef]
26. Shahnaz, C.; Shoaib Bin, M.; Hasan, S.M.S. Emotion recognition based on wavelet analysis of Empirical Mode Decomposed
EEG signals responsive to music videos. In Proceedings of the 2016 IEEE Region 10 Conference (TENCON), Singapore, 22–25
November 2016; pp. 424–427.
27. Fraschini, M.; Demuru, M.; Hillebrand, A.; Cuccu, L.; Porcu, S.; Di Stefano, F.; Puligheddu, M.; Floris, G.; Borghero, G.; Marrosu,
F. EEG functional network topology is associated with disability in patients with amyotrophic lateral sclerosis. Sci. Rep. 2016,
6, 38653. [CrossRef]
28. Moon, S.-E.; Chen, C.-J.; Hsieh, C.-J.; Wang, J.-L.; Lee, J.-S. Emotional EEG classification using connectivity features and
convolutional neural networks. Neural Netw. 2020, 132, 96–107. [CrossRef]
29. Wang, Z.; Tong, Y.; Heng, X. Phase-locking value based graph convolutional neural networks for emotion recognition. IEEE
Access 2019, 7, 93711–93722. [CrossRef]
30. Bi, J.; Wang, F.; Yan, X.; Ping, J.; Wen, Y. Multi-domain fusion deep graph convolution neural network for EEG emotion recognition.
Neural Comput. Appl. 2022, 34, 22241–22255. [CrossRef]
Brain Sci. 2024, 14, 271 19 of 19
31. George, F.P.; Shaikat, I.M.; Ferdawoos, P.S.; Parvez, M.Z.; Uddin, J. Recognition of emotional states using EEG signals based on
time-frequency analysis and SVM classifier. Int. J. Electr. Comput. Eng. 2019, 9, 1012–1020. [CrossRef]
32. Iyer, A.; Das, S.S.; Teotia, R.; Maheshwari, S.; Sharma, R.R. CNN and LSTM based ensemble learning for human emotion
recognition using EEG recordings. Multimed. Tools Appl. 2023, 82, 4883–4896. [CrossRef]
33. Delvigne, V.; Facchini, A.; Wannous, H.; Dutoit, T.; Ris, L.; Vandeborre, J.-P. A Saliency based Feature Fusion Model for EEG
Emotion Estimation. In Proceedings of the 2022 44th Annual International Conference of the IEEE Engineering in Medicine &
Biology Society (EMBC), Glasgow, UK, 11–15 July 2022; pp. 3170–3174.
34. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural
Netw. Learn. Syst. 2020, 32, 4–24. [CrossRef] [PubMed]
35. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907.
36. Song, T.; Zheng, W.; Song, P.; Cui, Z. EEG emotion recognition using dynamical graph convolutional neural networks. IEEE Trans.
Affect. Comput. 2018, 11, 532–541. [CrossRef]
37. Zhong, P.; Wang, D.; Miao, C. EEG-based emotion recognition using regularized graph neural networks. IEEE Trans. Affect.
Comput. 2020, 13, 1290–1301. [CrossRef]
38. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473.
39. Chen, P.; Gao, Z.; Yin, M.; Wu, J.; Ma, K.; Grebogi, C. Multiattention adaptation network for motor imagery recognition. IEEE
Trans. Syst. Man Cybern. Syst. 2021, 52, 5127–5139. [CrossRef]
40. Chen, H.; Jin, M.; Li, Z.; Fan, C.; Li, J.; He, H. MS-MDA: Multisource marginal distribution adaptation for cross-subject and
cross-session EEG emotion recognition. Front. Neurosci. 2021, 15, 778488. [CrossRef] [PubMed]
41. Gao, Y.; Fu, X.; Ouyang, T.; Wang, Y. EEG-GCN: Spatio-temporal and self-adaptive graph convolutional networks for single and
multi-view EEG-based emotion recognition. IEEE Signal Process. Lett. 2022, 29, 1574–1578. [CrossRef]
42. Fernando, B.; Habrard, A.; Sebban, M.; Tuytelaars, T. Unsupervised visual domain adaptation using subspace alignment. In
Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2960–2967.
43. Li, Y.; Fu, B.; Li, F.; Shi, G.; Zheng, W. A novel transferability attention neural network model for EEG emotion recognition.
Neurocomputing 2021, 447, 92–101. [CrossRef]
44. Li, Y.; Zheng, W.; Cui, Z.; Zhang, T.; Zong, Y. A Novel Neural Network Model based on Cerebral Hemispheric Asymmetry for
EEG Emotion Recognition. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI,
Stockholm, Sweden, 13–19 July 2018; pp. 1561–1567.
45. Li, Y.; Zheng, W.; Zong, Y.; Cui, Z.; Zhang, T.; Zhou, X. A bi-hemisphere domain adversarial neural network model for EEG
emotion recognition. IEEE Trans. Affect. Comput. 2018, 12, 494–504. [CrossRef]
46. Li, Y.; Wang, L.; Zheng, W.; Zong, Y.; Qi, L.; Cui, Z.; Zhang, T.; Song, T. A novel bi-hemispheric discrepancy model for EEG
emotion recognition. IEEE Trans. Cogn. Dev. Syst. 2020, 13, 354–367. [CrossRef]
47. Li, J.; Li, S.; Pan, J.; Wang, F. Cross-subject EEG emotion recognition with self-organized graph neural network. Front. Neurosci.
2021, 15, 611653. [CrossRef]
48. Pan, J.; Zhan, L.; Hu, C.; Yang, J.; Wang, C.; Gu, L.; Zhong, S.; Huang, Y.; Wu, Q.; Xie, X. Emotion regulation and complex brain
networks: Association between expressive suppression and efficiency in the fronto-parietal network and default-mode network.
Front. Hum. Neurosci. 2018, 12, 70. [CrossRef] [PubMed]
49. Lin, Y.-P.; Wang, C.-H.; Jung, T.-P.; Wu, T.-L.; Jeng, S.-K.; Duann, J.-R.; Chen, J.-H. EEG-based emotion recognition in music
listening. IEEE Trans. Biomed. Eng. 2010, 57, 1798–1806. [PubMed]
50. Almanza-Conejo, O.; Almanza-Ojeda, D.L.; Contreras-Hernandez, J.L.; Ibarra-Manzano, M.A. Emotion recognition in EEG signals
using the continuous wavelet transform and CNNs. Neural Comput. Appl. 2023, 35, 1409–1422. [CrossRef]
51. Greco, A.; Costantino, D.; Morabito, F.; Versaci, M. A Morlet wavelet classification technique for ICA filtered sEMG experimental
data. In Proceedings of the International Joint Conference on Neural Networks, Portland, OR, USA, 20–24 July 2003; pp. 166–171.
52. Uddin, M.Z.; Nilsson, E.G. Emotion recognition using speech and neural structured learning to facilitate edge intelligence. Eng.
Appl. Artif. Intell. 2020, 94, 103775. [CrossRef]
53. Zhang, Y.; Zhang, S.; Ji, X. EEG-based classification of emotions using empirical mode decomposition and autoregressive model.
Multimed. Tools Appl. 2018, 77, 26697–26710. [CrossRef]
54. Cai, S.; Li, H.; Wu, Q.; Liu, J.; Zhang, Y. Motor imagery decoding in the presence of distraction using graph sequence neural
networks. IEEE Trans. Neural Syst. Rehabil. Eng. 2022, 30, 1716–1726. [CrossRef]
55. Zheng, W.L.; Dong, B.N.; Lu, B.L. Multimodal Emotion Recognition using EEG and Eye Tracking Data. In Proceedings of the 36th
Annual International Conference of the IEEE-Engineering-in-Medicine-and-Biology-Society (EMBC), Chicago, IL, USA, 26–30
August 2014; pp. 5040–5043.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.