Skip to main content
Cognitive Neurodynamics logoLink to Cognitive Neurodynamics
. 2022 Oct 10;17(5):1271–1281. doi: 10.1007/s11571-022-09890-3

Feature hypergraph representation learning on spatial-temporal correlations for EEG emotion recognition

Menghang Li 1,2, Min Qiu 1,2, Li Zhu 1,2, Wanzeng Kong 1,2,
PMCID: PMC10542078  PMID: 37786664

Abstract

Electroencephalogram(EEG) becomes popular in emotion recognition for its capability of selectively reflecting the real emotional states. Existing graph-based methods have made primary progress in representing pairwise spatial relationships, but leaving higher-order relationships among EEG channels and higher-order relationships inside EEG series. Constructing a hypergraph is a general way of representing higher-order relations. In this paper, we propose a spatial-temporal hypergraph convolutional network(STHGCN) to capture higher-order relationships that existed in EEG recordings. STHGCN is a two-block hypergraph convolutional network, in which feature hypergraphs are constructed over the spectrum, space, and time domains, to explore spatial and temporal correlations under specific emotional states, namely the correlations of EEG channels and the dynamic relationships of temporal stamps. What’s more, a self-attention mechanism is combined with the hypergraph convolutional network to initialize and update the relationships of EEG series. The experimental results demonstrate that constructed feature hypergraphs can effectively capture the correlations among valuable EEG channels and the correlations inside valuable EEG series, leading to the best emotion recognition accuracy among the graph methods. In addition, compared with other competitive methods, the proposed method achieves state-of-art results on SEED and SEED-IV datasets.

Keywords: EEG, Emotion recognition, Hypergraph learning, Self-attention mechanism

Introduction

EEG-based connectivity patterns can be used for classifying different mental states. Graph-based models have been proposed to describe pairwise relationships within multichannel EEG recordings. For example, Zhang et al. (2019) proposed a graph-based hierarchical model to classify movement intention based on the relationships among EEG signals and their spatial information. However, in real scenarios, the state of a channel being activated may be related to multiple channels, which means that there is a higher-order relationship among the channels. Current graph-based methods neglect the complex higher-order correlations existed in EEG recordings, and are incapable of depicting such set-like relations.

It is worth noting that EEG is a non-stationary signal, and its statistics related to brain regions vary with time. However, the relationship between EEG timing samples is still under investigation. In order to tackle the above challenges, we model the higher-order interdependence of EEG features as hypergraphs to extract two different types of EEG relations. To verify the effective of the idea, we apply the model to EEG-based emotion recognition.

Emotion recognition is a technique for capturing human emotional states, which is vital in human-computer interaction. Typically, two types of signals can be utilized for emotion recognition. One is non-physiological signals, which are easily controlled by humans and are biased when people try to hide them. By contrast, physiological signals are difficult to be disguised and are helpful to obtaining real emotions, like electroencephalogram(EEG) signals, electrocardiogram(ECG) signals, electromyography(EMG), etc. This discovery inspired researchers to use physiological signals for emotion recognition. Among them, EEG signals are a direct response to brain activity belonging to the central nervous system so that real emotions can be captured more objectively and reliably. Also, their acquisition costs are low, making them the most commonly used physiological signals in emotion recognition.

Traditional machine learning methods, such as SVM (Wang et al. 2011; Zheng and Lu 2015), KNN (Bahari and Janghorbani 2013), etc., are used for emotion recognition at the beginning. However, traditional machine learning can not fully extract emotional features existed in EEG recordings. Methods based on convolutional neural networks(CNN)(Lotfi and Akbarzadeh-T 2014; Li et al. 2020; Deng et al. 2021) can make up for the shortcomings of traditional machine learning methods, improve the accuracy of emotion recognition. With the development of graph neural networks, graph-based methods (Feng et al. 2019; Zhong et al. 2020; Zhonget al. 2020) are also applied to EEG emotion recognition. Since these methods consider relationships between EEG channels, they have achieved more accurate recognition results. Till now, higher-order relationships of EEG recordings are not extracted and utilized in obtaining more accurate emotion recognition results.

To fill this gap, we propose a spatial-temporal hypergraph convolutional network named STHGCN, a hypergraph-based architecture for extracting higher-order EEG relations related to emotion recognition in spatial and temporal domains. The STHGCN consists of a spatial block and a temporal block. Each block is composed of a Feature Extraction Layer and a Hypergraph convolutional Layer(3.2). A Fusion layer to fuse the spatial and temporal dependencies in EEG features is added after two blocks 3.3. Through experiments on EEG recordings in two datasets, SEED, SEED-IV 4.1, we show that STHGCN outperforms most state-of-the-art methods in terms of both accuracy and standard variance 5.1. Lastly, we perform an hypergraph visualization and ablation study (5.2, 5.3) to contextualize each component’s effectiveness and pave the future directions towards combining hypergraphs and EEG correlations extraction 6 . The code will be available on GitHub(https://github.com/lmenghang/STHGNN.git)

The main contributions of the paper are as follows:

We use feature hypergraph representation learning for EEG emotion recognition for the first time, propose a parallel spatial-temporal hypergraph convolution structure, and capture the higher-order correlations of EEG signals for better feature extraction.

To capture the higher-order correlations among EEG channels, we construct the hypergraph of EEG channels by combining two kinds of correlations of EEG recordings in spatial and spectral domains, and updated the hypergraph automatically via STHGCN.

To capture the higher-order correlations inside the EEG series, we construct and update the hypergraph of the EEG series via designing a self-attention modal with learnable parameters.

Extensive experiments are conducted on two benchmark datasets. The experimental results show that our STHGCN consistently outperforms all state-of-the-art models.

Related work

Hypergraph learning

Hypergraph learning has made progress in problems where relations among data points extend beyond pairwise interactions, owing to its ability to extract patterns from higher-order relationships, and reduce information loss. With the popularity of deep learning, many methods of combining hypergraphs with neural networks have emerged. The earliest ones are the hypergraph neural network proposed by Feng et al. (2019) and the hypergraph convolutional network proposed by Yadati et al. (2019). Dynamic hypergraph convolutional neural network (Jiang et al. 2019) uses KNN and K-Means to dynamically update the hypergraph structure, improving the ability to capture data relationships and can extract local and global relationships of data. In order to better apply the graph learning strategy to hypergraphs, Huang and Yang (2011) proposed UniGNN, which not only solved the over-smoothing problem of deep hypergraph convolutions but also proved that hypergraphs have powerful representation capabilities. The addition of the attention mechanism also greatly improves the representation ability of the hypergraph. For example, Bai et al. (2021) used the attention mechanism to dynamically update the weights of hyperedges in a hypergraph.

There are a few studies combining hypergraph learning with prediction tasks, such as session-based recommendation (Xia et al. 2021), classification problems across biological networks (Lugo-Martinez and Radivojac 2017), stock selection (Sawhney et al. 2021), etc. However, there is no research bridging hypergraph neural networks and EEG-based emotion recognization, and we are the first to fill this gap.

Contemporary methods in EEG-based emotion recognition

EEG recordings consists of emotional features in different domains, like the time domain, space domain, and frequency domain. Paper (Deng et al. 2021; Ding et al. 2021; Wang et al. 2022) considered both temporal and spatial features and use them for EEG-based emotion recognition. Deng et al. (2021) linearly combined the capsule network with LSTM. TSception (Ding et al. 2021) used multi-scale 1D convolutional kernels to dynamically learn temporal features of EEG signals, and multiple 1D convolutions to globally learn spatial features of brain hemispheres. Wang et al. (2022) used Transformer to extract emotional features from EEG signals, which is based on the spatio-temporal self-attention mechanism. Recently, some methods consider emotional features in the spectral domain. LEDPatNet19 (Tuncer et al. 2021) proposed a multilevel hand-crafted feature generation network to better extract emotion-related EEG features. SST-EmotionNet (Jia et al. 2020) combines spectral, temporal, and topological features to construct two sets of 3D feature maps, which are input into the proposed network for extracting temporal-spatial features and spatial-spectral features. The idea of the method proposed by Shen et al. (2020) is to combine frequency, time, and topological features to form a 4D emotion feature representation. Xiao et al. (2022) used the attention mechanism to further utilize emotion-related EEG features. The above studies verify that EEG signals possess complementarity among the spectral-spatial-temporal emotional features.

Neurological studies have shown that emotional response is closely related to the cerebral cortex of the human brain, which is a 3D spatial structure and is not completely symmetrical between the left and right hemispheres (Lotfi and Akbarzadeh-T 2014). Thus, many studies have focused on capturing the relationships between channels to improve the accuracy of recognizing emotion. Due to the discrepancy of the cerebral cortex, Li et al. (2020) employed four directed recurrent neural networks(RNN) to keep the intrinsic spatial dependence, and captured the discrepancy information between two hemispheres for final classification. And Bagherzadeh et al. (2022) analyze effective brain connectivity to extract relationships among EEG channels. Graph-based methods are more capable of modeling inter-channel relationships, which help learn the representation of spatial relationships in EEG recordings. Feng et al. (2019) adopted a dynamical graph convolutional neural network (DGCNN) to extract more discriminative spatial features for emotion recognition. RGNN (Zhong et al. 2020) considered the biological topology among different brain regions to capture both local and global relations among different EEG channels. LGGNet (Zhonget al. 2020) also considered the local and global relationships between EEG channels, and obtains the global relationship by integrating the local relationships of the four functional brain regions. Utilizing the symmetry of the left and right hemispheres of the brain, and the symmetry of the upper and lower hemispheres of the brain, SFE-Net (Deng et al. 2021) performed data increments based on this, achieving a higher classification accuracy.

Since the cerebral cortex has a 3D structure, and the EEG signal is essentially a time sequence, the relationship of EEG recordings exists not only in the spatial domain, but also in the temporal domain. Similarly, emotion-related EEG features also have relationships in both temporal and spatial domains. Only considering spatial relationships is not enough to accurately represent EEG emotional features. However, it is still challenging to make full use of features in different domains to generate discriminative relationship representations for different emotions.

Method

In this section, we first introduce the notions and definitions used throughout this paper, and then we show how EEG recordings is modeled as a hypergraph. After that, we present our spatial-temporal hypergraph convolutional network for EEG-based Emotion Recognition. Finally, we show the way to optimize STHGCN.

Notations and definitions

Let XT=(S1,S2,,ST)RC×T denotes an EEG signal sample containing T timesteps, where C is the number of electrods. Each sample at a timestep t is represented as a set St=(st1,st2,,stE,)RE and ste(1eE) represents the EEG signal of an anonymous electrodes collected at the timestep t. St of all timesteps are used to construct incidence matrix MTRT×N. MT and XT are then transformed into temporal hypergraph GT, where N is the number of hyperedges in GT. Similarly, Let XE=(T1,T2,,TC)RT×C denotes an EEG signal sample containing C electrods, where T is the number of timesteps, XE is the transpose of XT. Each sample at electrods c is represented as a set Tc=(tc1,tc2,,tcE,)RE and tct(1tT) represents the EEG signal of electrode c collected at an anonymous time. Tc of all electrodes are used to construct incidence matrix MERC×N. ME and XE are then transformed into channelwise hypergraph GE, where N is the number of hyperedges in GE. The task of EEG-based emotion recognition is to predict the emotion label Y for any given hypergraph sets, GT and GE.

Definition 1

(Hypergraph) Let G=(V,E) denotes a hypergraph, where V is a set containing C unique vertices and E is a set containing N hyperedges. Each hyperedge ϵE contains two or more vertices and is assigned a positive weight Wϵϵ, all the weights formulate a diagonal matrix WRN×N. The hypergraph can be represented by an incidence matrix HRC×N where Hiϵ=1 if the hyperedge ϵE contains a vertex viV, otherwise Hiϵ=0. For each vertex and hyperedge, their degree Dii and Bϵϵ are respectively defined as Dii=ϵ=1NWϵϵHiϵ, Bϵϵ=i=1CHiϵ. D and B are diagonal matrics.

Definition 2

(Hypergraph Convolution) Let XlRC×F(l), where F is the demension of vertex features. Hypergraph laplacian is defined as =I-B-1/2HWHTD-1B-1/2. To learn the interdependence among vertices, the hypergraph convolution update rule, an information transfer way from vertex to edge, and then from edge to vertex, is defined as,

Xl+1=σ(B-1/2HWHTD-1B-1/2XlWl)

where σ(·) is an active function. WlRF(l)×F(l+1) is a learnable matrix. D and B are degree matrixes for normalization.

Spatial-temporal hypergraph

The overall stucture of the spatial-temporal hypergraph convolutional network (STHGCN) is illustrated in Fig. 1. The proposed STHGCN is used to decode EEG signals and extract spatial relationships of EEG channels and dynamic relationships of EEG series. It consists of a spatial block and a temporal block, which are independent but have a similiar network structure. The input of spatial block is original EEG recordings with constructed hypergraph containing spatial and spectral relations of EEG channels. The input of temporal block is rotated EEG recordings. The hypergraph containing temporal relations of EEG series was constructed via a self-attention mechanism in temporal block. In the end, the captured spatial and temporal features are fused for classification.

Fig. 1.

Fig. 1

Architecture of STHGCN for emotion recognition. STHGCN is composed of a spatial block and a temporal block which share a similar structure. C denotes EEG channels. n and e denote nodes and edges of hypergraph

Hypergraph construction

Spatial Hypergraph Construction To capture the beyond pairwise relations between channels in EEG-based emotion recognition, we adopt a hypergraph GE=(V,E) to represent each EEG channel cC as a node vV and the subset of their relations as a hyperedge ϵE. We inject domain knowledge by initializing spatial and spectral hyperedges among channels based on two types of relations: spatial correlations and spectral coherences.

Spatial Hyperedges: Electrods adjacent in space, usually produce a time series with similar variation trends. To leverage this finding, we define relations between EEG channels using Pearson correlation coefficient. Formally, we construct a hyperedge epsilon1E1 that connects EEG signals having the same or similar shapes and increasing/decreasing simultaneously and proportionally.

Spectral Hyperedges: Similarly, we construct a hyperedge epsilon2E2 that connects EEG signals whose trends can be described using same or similar linear systems, using spectral coherence coefficient. We combine these two kind of relations among channels of EEG recordings as hyperedges E=E1E2 to construct the hypergraph GE, equivalently denoted by an incidence matrix McRV×E, with entries h(v,ϵ) defined as:

h(v,ϵ)=1,vϵ0,vϵ

The degree of each vertex v is obtained using a function d(v), and stored in a diagonal matrix DvRV×V as:

d(v)=ϵEw(ϵ)h(v,ϵ)

w(ϵ) represents the weight of hyperedge ϵ which is set to 1 at initialization. The degree of each hyperedge ϵ is obtained using a function d(ϵ), and stored in a diagonal matrix DERE×E as:

d(ϵ)=vVh(v,ϵ)

Temporal hypergraph construction To capture the beyond pairwise relations between timesteps in EEG-based emotion recognition, we adopt a hypergraph GT=(V,E) to represent each timestep tT as a node vV and the subset of their relations rR as a hyperedge ϵE. We inject time domain knowledge by constructing hyperedges among timesteps based on relations captured by self-attention mechanism. Time Hyperedges: The importance of different timesteps inside an EEG sequence for identifying emotions are different. Relations among timesteps in an EEG sequence is exploring the intra-correlation of EEG recordings, which could help marking significant period with more information of an EEG sequence. The self-attention mechanism has the ability to screen important information and can give more attention to the time period that is more important in emotion recognition. Herein is used to construct the incidence matrix HRE×E of hypergraph .

The temporal hypergraph construction flowchart are shown in Fig. 2. As illustrated, the EEG recordings is organized as 3D sequences, containing spatial, spectral and temporal information. We first let each node represent a timestep, containing spatial and spectral features. Incidence matrix is initialized with zero. Then, self-attention mechanism is adopted to model the relationship between timesteps. And incidence matrix is constructed according to this. Finally node features and incidence matrix are combined to construct the temporal hypergraph.

Fig. 2.

Fig. 2

The temporal hypergraph construction flowchart. n and e denote nodes and edges of the constructed hypergraph

The differences of emotion are presented in the degree of activation of different frequency bands, timesteps, and brain regions. There are some stable and discriminative patterns of higher-order relationships inside time series and among brain regions for each emotion state. Therefore, it is important to extract relations of EEG recordings in these specific patterns for emotion recognition.

Spatial hypergraph convolution

First, to learn the interdependence among EEG channels, we use a hypergraph convolution on the hypergraph GE. We first define a single hypergraph convolution SConv(·), where the input to the hypergraph convolution layer is a matrix of EEG emotional features XE(l)RC×T. The hypergraph convolution layer updates XE(l) to new features XE(l+1), using their relationships represented in hypergraph GE. Following (Jiang et al. 2019, we define the hypergraph Laplacian as =I-DV-1/2HWDe-1HTDV-1/2 resulting in the hypergraph convolution update rule as:

XE(l+1)=SConv(X(l),H,Θ)=ELU(DV-1/2HWDe-1HTDV-1/2XE(l)Θ) 1

ΘRT×T is learnable parameters, ELU(·) is exponential linear unit activation. We set W=I for initialization. Then, a 1D spatial CNN is used to better represent spatial features of EEG signals for feature fusion.

Temporal hypergraph convolution

To learn the interdependence between EEG channels, we use a hypergraph convolution on the hypergraph GT, We first define a single hypergraph convolution TConv(·), where the input to the hypergraph convolution layer is a matrix of EEG emotional features XT(l)RT×C. The hypergraph convolution layer updates XT(l) to new features XT(l+1) using the same strategy with SConv(·). Then, a 1D temporal CNN is used to better represent temporal features of EEG signals for feature fusion.

Learning to predict and network optimization

We employ spatial and temporal hypergraph convolution concurrently to obtain spatial and temporal representations of EEG recordings respectively. Then, these two representations are concatenated. And label prediction is implemented through two fully-connected layer with softmax activation.

I^=softmax(Wf2×(Wf1×[SConv(·),TConv(·)]+b1)+b2) 2

where Wf1, Wf2, b1 and b2 are learnable parameters. In this paper, the classification cross entropy is used as loss function, which is defined as follow:

L=-i=1Nk=1Myiklog(p(yik^))

where yik is a binary indicator meaning the label of sample i is k, p(yik^) represents the probability that label prediction of sample i is right.

Experiment setup

Datasets

We validate our model on SEED (Zheng and Lu 2015) SEED-IV (Zheng et al. 2018) datasets.

SEED contains three different states of emotions, namely positive, negative, and neutral. EEG recordings were collected from fifteen participants while they were watching stimulus videos. The participants were asked to give feedback immediately after each experiment. Experiments are designed for each participant on three sessions with each session using 15 video clips repeatedly. The EEG series of 62 channels were recorded by an EEG cap and down-sampled into 200Hz. Emotion-related features are pre-computed over different frequency bands for each EEG sample in each channel.

SEED-IV contains four different states of emotions, namely happy, sad, fear, and neutral. Similiar to SEED, the experiment consists of 15 participants and the EEG recordings contains 62 channels. Three experiments are designed for each participant on three sessions, and 24 stimuli video clips in each session are different. DE feature is pre-computed over five frequency bands for each EEG sample in each channel.

Training step

We implement the STHGCN on NVIDIA RTX 3080 GPU. In the training process, we use the RAdam optimizer with learning rate 1e-3 with multi-step decay to 1e-7. The training epoch is set to 600 and the batch size is 32. For each experiment, we randomly shuffle the samples and use 5-fold cross validation.

Evaluation metrics

To show the performance of the proposed method, we compare it with several latest methods on SEED and SEED-IV datasets, including graph methods, like DGCNN, RGNN, and some other SOTA models. We use the recognition accuracy to show the performance of the proposed model. Also, the standard deviation, reflecting the degree of dispersion of a dataset, is adopted to verify the robustness of the proposed model.

Results and analysis

Comparison with baselines

We compare our model with baseline models on SEED and SEED-IV datasets. The left part of Table 1 presents the average recognition accuracy and the standard deviation of these models for EEG based emotion recognition on SEED dataset. The proposed STHGCN achieves the state-of-the-art performance on the SEED dataset. Obviously, deep learning models outperform the traditional SVM. DGCNN only considers the spatial relationship of EEG channels, and employs graph convolution to extract spatial information. BiHDM uses two directional RNNs to extract spatial information of EEG signals, reaching 93.12% on classification accuracy. R2G-STNN extracts spatial information by hierarchically learning improtant information from local to global.

Table 1.

The performance comparison of the state-of-the-art models on the SEED and SEED-IV dataset

Model SEED SEED-IV
ACC(%) STD(%) ACC(%) STD(%)
SVMWang et al. (2011) 83.99 9.72 56.61 20.05
DBNZheng and Lu (2015) 86.08 8.34 66.77 7.38
DGCNNSong et al. (2018) 90.40 8.49 69.88 16.29
BiDANNLi et al. (2018) 92.38 7.04 70.29 12.63
BiHDMLi et al. (2020) 93.12 6.06 74.35 14.09
R2G-STNNLi et al. (2019) 93.38 5.96 68.51 12.37
RGNNZhong et al. (2020) 94.24 5.95 79.37 10.54
4D-CRNNShen et al. (2020) 94.74 2.32 81.27 11.62
SST-EmotionNetJia et al. (2020) 96.02 2.17 84.92 6.66
EeTWang et al. (2022) 96.28 4.39 83.27 8.37
STHGCN 97.51 2.01 85.99 7.33

4D-CRNN represents EEG recordings in spatial, spectral and temporal domain achieving 94.74%. SST-EmotionNet takes the complementarity of spatial, spectral and temporal information into consideration, achieves better performance with an accuracy of 96.02%. Adopting transformer structure, EeT achieves the accuracy of 96.28%, demonstrating the effectiveness of updated transformer with ‘joint attention’. Different from the fomer methods, STHGCN adopts feature hypergraph to construct the dynamic relationship inside EEG series and the spatial relationship among EEG channels, which enables our model to comprehensively capture valuable features from EEG signals for emotion recognition. Compared with the baseline models, the accuracy of our model is further improved. In addition, the left map of Fig. 3 shows the confusion graph of our model on SEED dataset. The results indicate that positive emotion is slightly easier to recognize for STHGCN than negative and neutral emotion.

Fig. 3.

Fig. 3

Confusion map of STHGCN. The left graph is on SEED and the right graph is on SEED-IV

The right part of Table 1 presents the performance of all models on the SEED-IV dataset. The proposed STHGCN achieves the state-of-the-art performance on the SEED-IV dataset. For the four-category classification task, DBN achieved an accurracy of 66.77%, and the graph-based network DGCNN and RGNN further improve the accuracy to 69.88% and 79.37%, respectively. BiDANN and BiHDM fully utilize the bi-hemispheric discrepancy of EEG recordings to achieve 70.29% and 74.35% repectively. SST-EmotionNet and EeT add attention mechanism to learn emotion features from different domains, achieving 84.92% and 83.27% respectively. STHGCN focuses on modeling channelwise and dynamical relationship of EEG recordings. Compared with the baseline models, the accuracy of our model is further improved to 85.99%. In addition, the confusion graph is shown in the right of Fig. 3. The confusion graph presents that STHGCN does well in the recognition on happy and neutral emotion. Sad and fear emotion may be more confused than happy and neutral emotion. The recognition accuracy of fear emotion was 68%, 15% easily confused with sad emotion, and 11% neutral emotion. We infer that the extracted relation patterns among EEG channels for fear emotions are not so unique and share some similarities with sadness and neutral emotions.

We can deduce from the above experiments that deep learning models outperform traditional ones, considering different domain information of EEG signals can improve the performance on emotion recognition, and fear emotion is more confused comparing to happy, sad and neutral.

Higher-order relationships in EEG recordings

The left hypergraph in Fig. 4 illustrates correlations among EEG channels with different dot colors representing different brain regions and different curve colors representing different hyperedges. The size of dots illustrates the importance of EEG channels. Conceptually, if a channel appears in more hyperedges, it is relative with more channels, which means that the channel is more important. And a hyperedge with a higher weight illustrates a stronger connection among channels linked by it. It is clear from Fig. 4 that important channels are most on the frontal lobe and parietal lobe, indicating that these regions are strongly related to the emotion processing in the brain. As we can see, the relationships extracted by hypergraph indicating that local connections and global connections are both valuable for emotion recognition. Through the experiment, we note that modeling inter-channel dependence through spatial correlation as hyperedges drastically supports valuable channels selection. And more importantly, hypergraphs effectively capture higher-order relations among EEG channels.

Fig. 4.

Fig. 4

Five strongest higher-order channel correlations and temporal correlations captured by Hypergraph. The left graph is obtained by spatial hypergraph convolutional network. The right graph is obtained by temporal hypergraph convolutional network

The right hypergraph in Fig. 4 illustrates correlations inside EEG series with different dot colors representing different timesteps and different curve colors representing different hyperedges. The size of dots illustrates the importance of timesteps. We have tested different lengths of EEG series and adopted 32 as the best choice. We depict that periodicity maybe is existed in EEG series. Conceptually, if a stamp appears in more hyperedges, it is relative with more timesteps, which means that the stamp is more important. And a hyperedge with a higher weight illustrates a stronger connection of timesteps linked by it. It is clear from Fig. 4 that important samples which contribute more on emotion recognition are selected to give higher weight. Through the experiment, we note that modeling dependence inside EEG series through temporal correlation pays more attention to valuable samples. And hypergraph effectively capture higher-order relations inside EEG series.

Model component ablation study

Ablation studies on two blocks

We conduct ablation studies on two blocks: spatial block and temporal block. Table 2 presents that the two-block structure outperforms the single-block structure(only with spatial block or temporal block). For example, the accuracy of two-block structure increases by 15.75% and 2.40% than the model only with spatial block or temporal block. The results show that two-block structure effectively uses the spatial-spectral features to construct channelwise hypergraph, uses temporal features to construct temporal hypergraph, and finally improves the classification accuracy. What’s more, the variant model only with spatial block has a worse performance than the one only with temporal block. It indicates the importance of finding dynamic correlations of EEG signals.

Table 2.

The performance comparison of variant models on the SEED dataset

Model ACC(%) STD(%)
SHGCN 81.76 5.58
THGCN 95.11 4.88
STHGCN 97.51 2.01

On the effectiveness of hypergraphs

Most existing graph-based models ignore higher-order relationships, only focus on pairwise patterns of EEG channel correlations. And graph methods left dynamical relationship inside EEG series behind. These limit the accuracy of emotion recognition. To study the effectiveness of hypergraphs, we use identity matrix to replace incidence matrix of each hypergraph, without modeling correlations of spatial relationships, temporal relationships of EEG recordings. That is, no hyperedge exists. As shown in Table 3, hyperedges play an important role in emotion recognition. We observe that the accuracy decreases as we remove hyperedges representing correlations of EEG channels and relationships of EEG series, and performs the worst after all hyperedges are removed. Especially when hyperedges representing temporal relationships are removed, the accuracy drops from 97.51% to 94.69%.

Table 3.

The performance comparison of variant hypergraphs on the SEED dataset

Model ACC(%) STD(%)
STHGCN 97.51 2.01
-Spatial 95.68 3.08
-Temporal 94.69 4.73
-Spatial, -Temporal 94.36 4.74

Conclusion and future

In this paper, we propose the STHGCN model for EEG emotion recognition. STHGCN first explores dependencies among EEG channels via channel hypergraph and captures dynamic correlations via temporal hypergraph. Previous studies based on the graph method exist a problem that they could only extract dependencies between two EEG channels, ignoring the higher-order dependencies among multiple EEG channels, which can be solved by proposed spatial hypergraph convolution. Then, temporal hypergraph convolution with a self-attention modal is proposed to capture dynamic correlations of EEG recordings. Moreover, to learn spatial and temporal relationships of EEG recordings accurately, the hypergraph is updated automatically through hypergraph convolution. Through ablative and qualitative experiments on SEED and SEED-IV, we probe the effectiveness of STHGCN, having better performance than all baselines.

Discussion

Recently, more graph-based methods in EEG recordings representation are proposed for their ability to extract relations between pairwise EEG channels. However, the higher-order relationship among EEG channels and the higher-order relationship among EEG series have not been explored. To fill this gap, our study constructs feature hypergraphs for the first time to extract higher-order relationships in temporal and spatial for EEG-based emotion recognition. Similarly, Different higher-order relation patterns may exist in EEG recordings in different cognitive functions. For example, Zhang et al. (2019) have used the graph method to extract spatial relations for movement intention detection. Thus, extracting spatial and temporal relations could help classify different mental states for different downstream tasks based on EEG.

The results of the proposed model show that higher-order relationships in space and time exist in EEG recordings. And our analysis shows that the valuable channels are distributed mostly in the frontal and parietal lobes. This result is consistent with empirical evidence that the frontal and parietal lobes are related to emotion. The former study did not involve the relation of EEG time series. And our analysis shows that it is worth paying more attention to.

Our model mainly focuses on the spatial relationship among channels, which may be too fine-grained for emotion recognition and prone to high variance between subjects. In the future, we aim to explore the relationship among different brain regions using a hypergraph to confirm our conjecture and reduce the impact of individual differences. Our work is only tested on datasets based on a discrete emotion model, which limits the extendency of our model and the explanation of the result. In the future, we aim to test the effectiveness of hypergraphs on datasets based on emotion dimensional models.

Acknowledgements

This work was supported by National Key R &D Program of China for Intergovernmental International Science and Technology Innovation Cooperation Project (2017YFE0116800), National Natural Science Foundation of China (U20B2074, U1909202), supported by Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province (2020E10010), Key R &D Project of Zhejiang Province(2021C03003).

Data Availibility Statement

The datasets analysed during the current study are available in the SEED repository,https://bcmi.sjtu.edu.cn/~seed/index.html.

Declarations

Conflicts of interest

All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  1. Bagherzadeh S, Maghooli K, Shalbaf A, Maghsoudi A (2022) Emotion recognition using effective connectivity and pre-trained convolutional neural networks in EEG signals. Cogn Neurodyn 1–20 [DOI] [PMC free article] [PubMed]
  2. Bahari F, Janghorbani A (2013) EEG-based emotion recognition using recurrence plot analysis and k nearest neighbor classififier, Paper presented at 20th Iranian conference on biomedical engineering (ICBME). 10.1109/ICBME.2013.6782224
  3. Bai S, Zhang F, Torr PH. Hypergraph convolution and hypergraph attention. Pattern Recognit. 2021;110:107637. doi: 10.1016/j.patcog.2020.107637. [DOI] [Google Scholar]
  4. Deng L, Wang X, Jiang F, Doss R. EEG-based emotion recognition via capsule network with channel-wise attention and lstm models. CCF Trans Pervasive Comput Interact. 2021;3(4):425–435. doi: 10.1007/s42486-021-00078-y. [DOI] [Google Scholar]
  5. Deng X, Zhu J, Yang S (2021) Sfe-net: EEG-based emotion recognition with symmetrical spatial feature extraction. In: Proceedings of the 29th ACM international conference on multimedia, pp. 2391–2400
  6. Ding Y, Robinson N, Zhang S, Zeng Q, Guan C (2021) Tsception: capturing temporal dynamics and spatial asymmetry from EEG for emotion recognition. arXiv preprint arXiv:2104.02935
  7. Feng Y, You, H, Zhang, Z (2019) Hypergraph neural networks. Biomedical engineering, Paper presented at the Proceedings of the AAAI conference on artificial intelligence, 7(3), 162–175
  8. Huang J, Yang J (2021) Unignn: a unified framework for graph and hypergraph neural networks. arXiv preprint arXiv:2105.00956
  9. Jia Z, Lin Y, Cai X, Chen H, Gou H, Wang J (2020) Sst-emotionnet: spatial-spectral-temporal based attention 3d dense network for EEG emotion recognition. In: Proceedings of the 28th ACM international conference on multimedia, pp. 2909–2917
  10. Jiang J, Wei Y, Feng Y, Cao J, Gao Y (2019) Dynamic hypergraph neural networks. In: IJCAI, pp. 2635–2641
  11. Li Y, Wang L, Zheng W, Zong Y, Qi L, Cui Z, Zhang T, Song T. A novel bi-hemispheric discrepancy model for EEG emotion recognition. IEEE Trans Cogn Dev Syst. 2020;13(2):354–367. doi: 10.1109/TCDS.2020.2999337. [DOI] [Google Scholar]
  12. Li Y, Zheng W, Cui Z, Zhang T, Zong Y (2018) A novel neural network model based on cerebral hemispheric asymmetry for EEG emotion recognition. In: IJCAI, pp. 1561–1567
  13. Li Y, Zheng W, Wang L, Zong Y, Cui Z (2019) From regional to global brain: a novel hierarchical spatial-temporal neural network model for EEG emotion recognition. IEEE Trans Affect Comput
  14. Lotfi E, Akbarzadeh-T M-R. Practical emotional neural networks. Neural Netw. 2014;59:61–72. doi: 10.1016/j.neunet.2014.06.012. [DOI] [PubMed] [Google Scholar]
  15. Lugo-Martinez J, Radivojac P (2017) Classification in biological networks with hypergraphlet kernels. arXiv preprint arXiv:1703.04823 [DOI] [PMC free article] [PubMed]
  16. Sawhney R, Agarwal S, Wadhwa A, Derr T, Shah RR (2021) Stock selection via spatiotemporal hypergraph attention network: a learning to rank approach. In: Proceedding of AAAI, 497–504
  17. Shen F, Dai G, Lin G, Zhang J, Kong W, Zeng H. EEG-based emotion recognition using 4D convolutional recurrent neural network. Cogn Neurodyn. 2020;14(6):815–828. doi: 10.1007/s11571-020-09634-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Song T, Zheng W, Song P, Cui Z. EEG emotion recognition using dynamical graph convolutional neural networks. IEEE Trans Affect Comput. 2018;11(3):532–541. doi: 10.1109/TAFFC.2018.2817622. [DOI] [Google Scholar]
  19. Tuncer T, Dogan S, Subasi A (2021) Ledpatnet19: automated emotion recognition model based on nonlinear led pattern feature extraction function using EEG signals. Cogn Neurodyn 1–12 [DOI] [PMC free article] [PubMed]
  20. Wang XW, Nie D, Lu BL (2011) Eeg-based emotion recognition using frequency domain features and support vector machines (2011). Paper presented at international conference on neural information processing
  21. Wang Z, Wang Y, Hu C, Yin Z, Song Y (2022) Transformers for EEG-based emotion recognition: a hierarchical spatial information learning model. IEEE Sens J
  22. Xiao G, Shi M, Ye M, Xu B, Chen Z, Ren Q (2022) 4D attention-based neural network for EEG emotion recognition. Cogn Neurodyn 1–14 [DOI] [PMC free article] [PubMed]
  23. Xia X, Yin H, Yu J, Wang Q, Cui L, Zhang X (2021) Self-supervised hypergraph convolutional networks for session-based recommendation. In: Proceedings of the AAAI conference on artificial intelligence, vol. 35, pp. 4503–4511
  24. Yadati N, Nimishakavi M, Yadav P (2019) Hypergcn: a new method for training graph convolutional networks on hypergraphs. Adv Neural Inf process syst 32
  25. Zhang D, Yao L, Chen K, Wang S, Haghighi PD, Sullivan C. A graph-based hierarchical attention model for movement intention detection from EEG signals. IEEE Trans Neural Syst Rehabil Eng. 2019;27(11):2247–2253. doi: 10.1109/TNSRE.2019.2943362. [DOI] [PubMed] [Google Scholar]
  26. Zheng W-L, Lu B-L. Investigating critical frequency bands and channels for eeg-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Dev. 2015;7(3):162–175. doi: 10.1109/TAMD.2015.2431497. [DOI] [Google Scholar]
  27. Zheng W-L, Liu W, Lu Y, Lu B-L, Cichocki A. Emotionmeter: a multimodal framework for recognizing human emotions. IEEE Trans Cybern. 2018;49(3):1110–1122. doi: 10.1109/TCYB.2018.2797176. [DOI] [PubMed] [Google Scholar]
  28. Zhong P, Wang, D, Miao C (2020) EEG-based emotion recognition using regularized graph neural networks. IEEE Transact Affect Comput
  29. Zhong P, Wang D, Miao C (2020) EEG-based emotion recognition using regularized graph neural networks. IEEE Trans Affect Comput

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The datasets analysed during the current study are available in the SEED repository,https://bcmi.sjtu.edu.cn/~seed/index.html.


Articles from Cognitive Neurodynamics are provided here courtesy of Springer Science+Business Media B.V.

RESOURCES