Skip to main content
PLOS One logoLink to PLOS One
. 2025 Jun 26;20(6):e0321942. doi: 10.1371/journal.pone.0321942

A novel dual-branch network for comprehensive spatiotemporal information integration for EEG-based epileptic seizure detection

Xiaobing Deng 1,*
Editor: Yuvaraj Rajamanickam2
PMCID: PMC12200712  PMID: 40570079

Abstract

Epilepsy is a neurological disorder characterized by recurrent seizures caused by abnormal brain activity, which can severely affects people’s normal lives. To improve the lives of these patients, it is necessary to develop accurate methods to predict seizures. Electroencephalography (EEG), as a non-invasive and real-time technique, is crucial for the early diagnosis of epileptic seizures by monitoring abnormal brain activity associated with seizures. Deep learning EEG-based detection methods have made significant progress, but still face challenges such as the underutilization of spatial relationships, inter-individual physiological variability, and sequence intricacies. To tackle these challenges, we introduce the Dual-Branch Deepwalk-Transformer Spatiotemporal Fusion Network (Deepwalk-TS), which effectively integrates spatiotemporal information from EEG signals to enable accurate and reliable epilepsy diagnosis. Specifically, the Spatio-branch introduces an adaptive multi-channel deepwalk-based graph framework for capturing intricate relationships within EEG channels. Furthermore, we develop a Guided-CNN Transformer branch to optimize the utilization of temporal sequence features. The novel dual-branch networks co-optimize features and achieve mutual gains through fusion strategies. The results of extensive experiments demonstrate that our method achieves the state-of-the-art performance in multiple datasets, such as achieving 99.54% accuracy, 99.07% sensitivity and 98.87% specificity. This shows that the Deepwalk-TS model achieved accurate epilepsy detection while analyzing the spatiotemporal relationship between EEG and seizures. The method further offers an optimized solution for addressing health issues related to seizure diagnosis.

Introduction

Epilepsy, a chronic neurological disorder that affects approximately 50 million people worldwide, ranks second as the second largest neurosystemic disease globally. Epilepsy seizures are marked by sudden, transient, and repetitive, often accompanied by uncontrolled muscletwitches and short-term loss of consciousness [13]. Therefore, early diagnosis of seizures is important for timely intervention, reducing the risk of irreversible brain damage, and improving the overall quality of life for patients [46]. Advanced medical imaging techniques typically detect epilepsy by identifying lesion information, but cannot capture ongoing seizures in the time domain. Therefore, to facilitate timely epilepsy diagnosis, Electroencephalogram (EEG), as a reliable method to capture brain electrical activity, has become a standard diagnostic technique in current clinical practices for epilepsy [7,8].

The computer-aided diagnosis of epilepsy based on EEG signals (EEGs) involves the automated screening of epileptic signals, representing a crucial step in achieving an accurate and efficient diagnosis. Expertise-dependent visual observations suffer from subjectivity and time-intensive analysis due to high-frequency acquisition of EEGs, leading to potential inaccuracies, misdiagnoses, and missed diagnoses [7,9]. Therefore, accurately decoding EEGs is crucial for assisting medical diagnosis, helping neurologists with treatment, and reducing risks in engineering.

The field of deep learning-based EEG seizure detection has attracted extensive research. However, existing methods do not fully consider the inherent characteristics of EEGs, such as multi-channel and temporal sequencing, resulting in insufficient feature extraction and low identification accuracy. The main challenges are as follows. (1) Underutilization of spatial relationships among EEGs across channels, impeding the conversion of raw EEG data into a structured graph. (2) The significant inter-individual physiological variability poses obstacles to deploying automated seizure detection, particularly when applied to unfamiliar patients. (3) The high complexity of EEGs encompasses aspects such as long duration, information redundancy, and a low signal-to-noise ratio due to various noise sources. (4) The extraction and fusion of features present challenges in handling long time series and integrating features in the frequency and time domain.

Therefore, to tackle the challenges, previous studies have explored the extensive applications of traditional feature-based machine learning algorithms (ML) [10] and deep learning algorithms (DL) [11,12] for automatic epilepsy detection. Initially, ML dominated epilepsy detection, but relying on manually designed features made it challenging to capture discriminative features. Recently, DL technology has been widely applied, automatically extracting meaningful features from EEGs in an end-to-end manner and enhancing accuracy. DL-based models are always categorized into Convolutional Neural Networks (CNN)-based [11], Long Short-Term Memory networks (LSTM)-based, Graph Convolutional Networks (GCN)-based [13], and hybrid models [14]. Researchers have proposed alternatives such as 3D CNNs [15] and multi-channel CNN networks [16]. Modified CNNs [17] exhibit discriminative feature learning in EEG classification. However, the CNN-based method struggles with long-term dependencies in time series analysis. LSTM-based methods have been introduced to capture temporal features [18], but always overlook graph structure information in EEGs. Hybrid models, such as CNN+LSTM, have been introduced in various studies to combine the strengths of different architectures and overcome the limitations of individual models.

It is well known that brain structural connectivity reflects the natural links between different regions, making it a key aspect of EEG analysis. The correlation between EEG data and epilepsy detection tasks can involve brain topology or nodal features. Therefore, GCN-based methods are widely used for modeling brain connectivity, as they leverage node characteristics and graph structure to capture complex relationships [19,20]. However, their effectiveness can be affected by poor graph connectivity and high node complexity. DeepWalk is a graph-based machine learning algorithm that learns latent representations of graph vertices by performing random walks [21]. Compared to GCN methods, it offers greater scalability, simplicity, and robustness to the quality of the graph structure.

To address these challenges, we introduce an innovative approach to seizure detection that incorporates both the temporal and spatial aspects of EEGs for the first time. The method, named Dual-Branch Deepwalk-Transformer Spatiotemporal Fusion Network (Deepwalk-TS), consists of two key components. Firstly, to fully extract spatial information between brain channels, we introduce a novel DeepWalk method to reconstruct the node features and topological structures of a brain channel graph. Then, we utilize GCN to extract features and enhance the model’s ability to learn complex relationships. This approach allows for dynamically adjustable random walks, which is well-suited for multi-channel EEG seizure detection. Secondly, we propose a Guided-CNN Transformer branch to address the underutilization of temporal information in EEGs. It adopts CNN and a self-attention mechanism as a guide to enhance Transformer architectures, maximizing the extraction of crucial temporal features. The proposed dual-branch structure collaborates and enhances each other to achieve end-to-end spatiotemporal feature extraction of EEGs, thereby optimizing the performance and efficiency of epilepsy detection. Our contributions are as follows.

  • Based on the characteristics of brain EEGs, we designed an end-to-end dual-branch network (Deepwalk-TS) specifically for spatiotemporal EEGs extraction and epileptic seizure detection.

  • We propose a novel spatial adaptive multi-channel graph construction framework based on deepwalk. DeepWalk fully leverages the stochastic nature of random walks to efficiently extract relevant features between brain channels.

  • We further introduce the Guided-CNN transformer branch for learning attention weights to improve the long-time consistency of EEGs. The outputs of these two branches are then fused using a joint optimization strategy.

  • We conducted extensive experiments across multiple datasets and provided a detailed analysis of each branch. The results show that the Deepwalk-TS method achieves state-of-the-art performance, significantly improving the accuracy of epilepsy detection.

The rest of this paper is organized as follows: The Related Work reviews existing research on epileptic seizure detection, focusing on various machine learning and deep learning approaches. Methodology Section describes the proposed Deepwalk-TS model, explaining its architecture, key components, and spatiotemporal features. The Experiments present the datasets used, performance metrics, and comparison with recent studies. The Discussion section analyzes the implications of the results, limitations, and future directions. Finally, the Conclusion section summarizes the main findings of the study and outlines the potential impact of the proposed approach in real-world epilepsy detection applications.

Related work

Machine-learning techniques for epileptic seizure detection

Traditional epileptic seizure detection methods primarily rely on manual feature extraction and machine learning models, such as Support Vector Machines (SVM) and k-Nearest Neighbors (KNN) [22]. These methods often depend on fine preprocessing and feature engineering of EEGs. For instance, Sun used Discrete Wavelet Transform (DWT) and uniform one-dimensional binary patterns to extract texture features from EEG recordings [23]. Al-Hadeethi et al. achieved good results with the AB-LS-SVM model by reducing EEG dimensionality, extracting statistical features, and selecting key features [24]. Vicnesh et al. classified various types of epilepsy by analyzing nonlinear EEG features and organizing them into a decision tree [25]. However, the transition of EEGs from non-ictal to ictal states is complex and dynamic, with significant differences in EEGs between patients and even between seizures in the same patient. Thus, accurately extracting seizure-related features remains a significant challenge.

Deep-learning methods for epileptic seizure detection

ML has shown significant efficacy in detecting and recognizing abnormal behaviors in individuals, thus the need for automated detection and identification of these behaviors. Recently, the application of DL models for seizure detection has increased, especially with the use of CNN [11,26], GCN [13], and LSTM [18]. Emami et al. segmented filtered EEGs into one-second segments, converted them into images, and classified each image using CNN to distinguish between ictal and non-ictal states [27]. Hu et al. introduced an innovative approach using a deep bidirectional LSTM (BiLSTM) network for seizure detection [18]. Roy developed a multi-scale feature CNN with adaptive transfer learning for EEG motor imagery classification [28]. It improves accuracy by addressing inter-subject variability and combining convolution scales with transfer learning.

Recent advances in epileptic seizure detection

While CNNs have demonstrated high performance in various studies, they often struggle with generalizing to new data or dealing with shifts in datasets. This limitation has prompted recent studies to explore transformer-based networks. For instance, Pan et al. proposed a transformer-based model for epilepsy detection, addressing sample imbalance by oversampling the minority class [29]. Shu et al. introduced EpilepsyNet [30], a transformer-based network that capitalizes on the strengths of transformers to enhance seizure detection accuracy.

In parallel, the success of GCNs in computer vision tasks, such as visual detection [31], disease prediction [32], and image segmentation, has inspired researchers to apply GCNs to epilepsy detection. GCNs excel at capturing complex relationships and structured data, making them highly effective for modeling the spatiotemporal and long-term dependencies in EEGs. Moreover, recent research has explored combining GCNs with transformers, further improving their performance in seizure detection tasks. Hybrid architectures that merge GCNs and transformers, like the GCN-Transformer models, have also shown significant promise in handling multimodal information.

Materials and methods

In this section, we will provide a detailed introduction to the proposed Deepwalk-TS method for epilepsy detection. The method comprises two key branches, including the spatial graph reconstruction and the temporal sequence feature extraction. The former introduces a random Deepwalk-based GCN network to enhance the topological structures and node features. The latter Transformer is designed for temporal sequence extraction from the original EEGs. The detection model utilizes a joint loss optimization strategy to ensure mutual enhancement between the two branches. Fig 1 illustrates the overall Deepwalk-TS model framework, including the input of the initial graph structure and EEGs, the network details of the dual-branch spatiotemporal model, and the seizure prediction module and results. The model takes the raw EEGs as input, collected from the International 10-20 EEG electrode system. The main part of the model consists of the upper and lower sections of Fig 1. The upper figure shows the DeepWalk-based GCN used for reconstructing the channel graph module, while the lower figure displays the Guided-CNN Transformer module used for sequence feature extraction. The prediction module employs FC and Softmax to obtain the prediction results. A detailed description is in the following section.

Fig 1. An overview of proposed Dual-Branch Deepwalk-Transformer Spatiotemporal Fusion Network (Deepwalk-TS) for seizure detection.

Fig 1

The upper shows the spatial branch of Deepwalk-based GCN used to reconstruct the channel graph. The lower shows the temporal branch of Guided-CNN Transformer for sequence feature extraction.

Graph structure learning with DeepWalk

Firstly, this branch performs a random walk on the original graph to construct a channel-channel similarity topological space, moving between nodes from one or a series of starting points following specific rules. We give a graph G=(V,E,X,Y) representation. E is the edge of GCN, and V is defined as {v1,v2,,v|V|}. XR(N×T) is the feature matrix, N is the number of nodes and T is the dimension. YR(N×L) is the labeling matrix (L is the dimension). The initial mode is vi1, Wvi1={(vi1,vi2),(vi2,vi3),,(vin1,vin)} is a random walk and a length n rooted at vi1, there is an edge connection between vij and vij+1. With a probability p(0<p1), the next step moves to a neighboring node. With a probability of 1–p, it jumps to any node in the graph, with jump probability continuously guiding the random walk for the next step. For each sequence of random walks, it maximizes the co-occurrence probability of vertices in the window w,

Pr({vjw,,vj+w}/vjϕ(vj))=i=jw,ijj+wPr(viϕ(vj)) (1)

Fig 2 shows a process of reconstructed graph using random wandering. The order of the walks is marked with a red line and extracted in the next walk graph. After that, we can use the obtained "walks" to reconstruct the graph [57]. The image on the right shows a random walk of length 11 nodes (marked in red). The weights αk,(k=1,2,3,...,11) are used to weight the original k-hopping nodes away from the node c. The random walk generator takes a random vertex vj in the graph G as the root of the random walk wvj. The topology between brain channels is then reconstructed using the randomness of the Deepwalk. The skipgram algorithm iterates over all possible matches for a random wandering sequence in the window w, to obtain the representation matrix φ. Then, we maximize the probability that this vertex is a neighbor in the walking sequence. But it is infeasible to compute Pr(viϕ(vj)). Therefore, we introduce the Softmax layer to approximate Pr, which is expressed as,

Fig 2. An illustration of the Deepwalk algorithm operation involves a random walk on the graph, resulting in a “walk” sequence, which is then utilized to construct a graph.

Fig 2

Pr(viϕ(vj))=k=1[log|V|]11+eϕ(vj)·φ(bk) (2)

where φ(bk)Rd represents the parent node of tree node bk. The path to vj is identified by the sequence of tree nodes (b0,b1,,log|v|), bk = root and log |v|=vj. We utilize the Huffman tree to assign shorter paths to frequently occurring vertices in deepwalk to speed up training time.

Subsequently, we perform graph convolution on the graph reconstructed by DeepWalk to obtain node vectors and graph embeddings. The next step involves using a 12-layer GCN to extract spatial structural features between brain channels. Specifically, the reconstructed graph G is fed into a multi-layer GCN with a parameter sharing strategy to obtain embeddings 𝐙ct(l) that are shared in different channel spaces. The topology graph (Gt,X) as follows,

𝐙ct(l)=ReLU(𝐃~t12𝐆~t𝐃~t12𝐙ct(l1)𝐖c(l)), (3)

where 𝐖c(l) is the lth layer weight matrix of GCN and 𝐙ct(l1) is the node embedding in the (l−1)th layer and 𝐙ct(0)=𝐗. D is a diagonal matrix and represents the characteristic dimension of the node. Dt~ represents the degree matrix of the vertex. The feature graph (Gf,X) is extract the shared information, the weight matrix 𝐖c(l) for every layer of GCN as follows,

𝐙cf(l)=ReLU(𝐃~f12𝐆~f𝐃~f12𝐙cf(l1)𝐖c(l)), (4)

where 𝐙cf(l) is the l-layer output embedding. The shared weight matrix can filter out the shared features from two spaces. Then, the multi-layer GCNs are employed to reorganize the information of the target node and its neighbors.

hvk=σ(av~u·huk1·Wk) (5)

where hvk is the hidden representation of node v in the k-th layer. av~u is the numbers on row v and column u of the reconstructed graph matrix G, indicating the proximity between nodes v and u. Wk is the parameter matrix. σ() denotes the ReLU function. In short, the branch can combine the graph of the nodes of the harvester at multiple levels to obtain more informative representations. Finally, the obtained representations are fed into the vectors to the downstream classifiers to accomplish the seizure classification task.

Guided-CNN transformer feature extraction

For the second temporal branch, we introduced a Guided-CNN Transformer method to comprehensively extract both local and global features from EEGs. The branch is composed of three components: a CNN-based module, a Transformer module, and an interactive module, as depicted at the bottom of Fig 1. Firstly, the raw EEGs are input into the CNN feature extractor, which comprises multiple stacked convolutional layers (1×1 Conv, 3×3 Conv, 1×1 Conv) to extract local series features. Secondly, each Transformer block contains a multi-head self-attention mechanism (MHA) and a multi-layer perception block. The dot product between Q and K calculates global location dependencies, which are dotted with V to incorporate these long-range dependencies into the feature. The original self-attention can be calculated,

Att(Qi,Ki,Vi)=Softmax(QiKiTdk)Vheadi=Attention(Qi,Ki,Vi)MHA(Q,K,V)=Concat[headi]Wo (6)

where i=1,,6, QI, KI, and VI represent the query, key, and value matrix, respectively. Att (Qi,Ki,Vi) is a weighted. dk represents Ki dimension. Wo represents the linear transformation of the final output and headi represents the head of MHA.

Then, we add learnable bias to each self-attention module, which is to enhance the location information,

MHA(Q,K,V)=softmax(QKTdk+B)V (7)

where K=Conv(K) and V=Conv(V). Additionally, each head of the MHA module contains a squeezed self-attention layer. The output sequence has the shape of n×dh and these h are then concatenated into a n×d sequence. Then, in the transformer module, the feedforward network (FFN) is a fully connected (FC) layer, expressed as,

FFN=max(0,XW1+b1)W2+b2 (8)

We have further designed an RFFN to replace the traditional FFN, enhancing both computational efficiency and nonlinear expression capability. The RFFN utilizes a structure similar to the inverted residual block, consisting of two 1×1 conv and a 3×3 depth-wise conv.

RFFN=Conv(DWConv(Conv(X)))+X (9)
Xi=LIL(Xi1) (10)
Xi=MHA(LN(Xi))+Xi (11)
Xi=RFFN(LN(Xi))+Xi (12)

where Xi and Xi refer to the output of the LIL layer and the SMHSA for block i, respectively. LN denotes the layer normalization.

The designed interactive module (inter) integrates the CNN and transformer fusion process. The module includes a transfer feature layer (TF) for keeping invariance, a down/up sample layer for global feature extraction, and an RFFN for enhancing nonlinear expressiveness. The feature X is fed into multi-layer,

inter(X)=Conv(X)+Up/down(X) (13)

We further design a fusion and joint optimization network using dot-multiplication and summation for the proposed spatial-temporal dual-branch method. The final representation vector zv of node v is obtained and can be sent to the downstream classifier to predict the category vectors y^v. A Softmax classifier is then employed for classification,

Softmax(z)i=exp(zi)j=1dexp(zj) (14)

where Softmax(z)i is the i-th component of the vector and d is the dimension of z.

Finally, guided by labeled data, we optimize the proposed model via back propagation and learn the embedding of nodes for classification. We use cross-entropy as the loss function to train the model,

loss=v(yvlogy^v+(1yv)log(1y^v)) (15)

Experimental

We employ a systematic experimental evaluation, including pre-processing steps, compared with various existing approaches and ablation studies. This analysis dissects the contributions of each branch, validating the effectiveness of their holistic spatiotemporal information integration. To further elucidate the impact of parameters, we present detailed comparative experiments. Additionally, we introduce various visualization techniques to enhance interpretability.

Datasets and processing

We perform extensive experiments to evaluate the proposed Deepwalk-TS on two widely used EEG datasets: the pediatric patient data from Boston Children’s Hospital (CHB-MIT, https://physionet.org/content/chbmit/1.0.0/) and the neurological Siena dataset (Siena). These datasets have distinct origins, enhancing the universality of the model assessment. The CHB-MIT comprises multi-channel EEG recordings from long-term monitoring of pediatric epilepsy patients, featuring various sub-datasets with different time spans and diverse seizure types. The dataset includes 22 subjects (5 males and 17 females) with 23 case records. The Siena primarily records EEGs from 14 patients, including 9 males (aged 25-71) and 5 females (aged 20-58). This dataset was recorded at a sampling rate of 512 Hz, and includes multiple channels and diverse patients.

Then, we present the data preprocessing workflow in Fig 3. We applied bandpass filtering to the raw EEGs in the range of 1-40 Hz and standardized the processed multi-channel EEGs (0-1). We employed 1-second or 0.5-second overlapping windows to segment samples of seizure and non-seizure events. For the experiments, we selected the raw EEGs from 16 channels in the CHB-MIT to construct the graph, including channels such as "FP1-F7", "F7-T7", "P7-O1", etc. Subsequently, we combined the EEGs of channels with their corresponding spatial adjacency matrix. We initially formed the graph structure for each case, establishing training and testing datasets.

Fig 3. An illustration of the data preprocessing workflow.

Fig 3

Experimental setup and evaluation

To rigorously assess the performance of our proposed Deepwalk-TS model, we conducted a series of comprehensive experiments with carefully structured setups and evaluations. In our experiments, we maintain a training ratio of 1: 9 for positive (epileptic) to negative (non-epileptic) samples. For the patient, we select 400 seconds of positive samples and 3600 seconds of negative samples (16×256 dimensionality). We employ a 5-fold cross-validation, and the results are presented as the mean  + − standard deviation. The model is trained with a batch size of 128,150 epochs and moment 0.9. For the Deepwalk-based GCN module, we use a two-layer GCN as a graph extractor with identical hidden and output dimensions. where nhid1{512,768} and nhid2{32,128,256}. In our experiments, the tour length t of the random tour is fixed. The traversal sequence randomly selects the neighbors of the last visited node until it reaches the maximum length t. L2 norm penalties are applied to the conv kernel parameters in the hierarchical GCN, and L1 is applied to the adjacency matrix. The dropout rate is set to 0.5 during training, while the retention probability is set to 1 during testing. The learning rate varies from 0.0001 to 0.0005, and the coefficients for graph consistency constraints are {0.01,0.001,0.0001}. The experiments are conducted on a 3080Ti GPU with Cuda 10.2, using the PyTorch 1.9.0 framework and updating parameters with the SGD optimizer.

In order to evaluate the performance of our proposed Deepwalk-TS model for epilepsy detection from multiple dimensions, containing Accuracy, Specificity, Sensitivity, Area Under Curve (AUC), ROC curve and F1-score (F1).

Accuracy(Acc.)=TN+TPFP+TN+FN+TP (16)
Specificity(Spe.)=TNFP+TN (17)
Sensitivity(Sen.)=TPFN+TP (18)
Precision(Pre.)=TPFP+TP (19)
F1=2Sensitivity×PrecisionSensitivity+Precision (20)

where TP means that both the true label and the predicted label are positive, and FP means that the actual label is negative and the predicted label is positive. TN means that both the true label and the predicted label are negative. FN means that the true label is positive and the predicted label is negative.

Experimental results

Single-patient seizure detection results.

We first evaluated the seizure detection results for individual patients using the CHB-MIT dataset, which includes data from all 23 cases. As shown in Table 1, our proposed model dem onstrates outstanding performance, achieving an average accuracy of 99.54%, sensitivity of 99.07%, specificity of 98.87%, F1-Score of 98.84%, and AUC of 99.06%. Notably, accuracy exceeded 98% for 21 patients. However, some variability in performance was observed among patients. Patients such as chb02, chb05, chb08, chb11, chb16, chb18, chb21, chb22 consistently scored perfectly across all evaluation metrics. In contrast, chb12 exhibited slightly lower accuracy at 98.50%, possibly due to less apparent epileptic activity in the selected nine channels. Additionally, Chb14 achieved perfect accuracy and sensitivity but showed lower specificity at 96.83%, likely due to the shorter duration of seizure episodes. Furthermore, notable AUC and F1 were observed for each patient. For instance, chb01 achieved an AUC of 98.99% and an F1-Score of 97.91%, while chb13 demonstrated an even higher AUC of 99.42% and an F1 of 98.86%. The model demonstrated strong stability, with an average specificity of 98.87%, and most patients exhibiting specificity above 98%, indicating its high accuracy in distinguishing non-epileptic states.

Table 1. Results of CHB-MIT single patient experiments.
Case Acc. Sen. Spe. F1 AUC
chb01 99.79 99.58 99.11 97.91 98.99
chb02 100.00 100.00 100.00 100.00 100.00
chb03 98.61 100.00 96.42 98.18 97.32
chb04 99.33 99.27 98.81 97.77 98.27
chb05 100.00 100.00 100.00 100.00 100.00
chb06 99.52 100.00 99.58 98.04 98.12
chb07 98.91 97.55 100.00 99.25 99.51
chb08 100.00 100.00 100.00 100.00 100.00
chb09 98.81 97.24 99.29 98.23 98.93
chb10 98.69 100.00 95.74 98.94 97.56
chb11 100.00 100.00 100.00 100.00 100.00
chb12 98.50 95.64 99.39 97.82 98.69
chb13 100.00 97.75 100.00 98.86 99.42
chb14 100.00 100.00 96.83 95.96 96.72
chb15 100.00 100.00 100.00 100.00 100.00
chb16 100.00 100.00 100.00 100.00 100.00
chb17 98.74 100.00 95.71 96.34 99.46
chb18 100.00 100.00 100.00 100.00 100.00
chb19 98.89 96.85 100.00 98.18 99.11
chb20 99.91 99.11 100.00 99.55 99.79
chb21 100.00 100.00 100.00 100.00 100.00
chb22 100.00 100.00 100.00 100.00 100.00
chb23 99.63 95.56 93.10 98.34 96.45
Mean 99.54 99.07 98.87 98.84 99.06

We also include this model’s additional computational complexity to provide a more comprehensive analysis. Deepwalk-based GCN complexity can be approximated as O(n×k×d), where n is the number of nodes (n = 16), k is the number of random walks per node (k = 9), and d is the dimensionality of the node embeddings. The random walk incurs higher costs with more nodes, walks, and embedding dimensions. Two-layer GCNs typically involve O(v×F×F1), where v is the number of nodes, F and F1 are the original and output feature dimensions. The module’s complexity is O(n×k×d)+O(v×F×F1)+O(E×F1), E is the edges. Through extensive experiments, we have determined that the optimal number of EEG channels is 11, which helps manage the computational overhead. Guided-CNN Transformer complexity is focused on the self-attention mechanisms of Transformer. Q, K, V matrix shape is n×d, Q×Kt=(n×d×n), scores×V=softmax(Q·KTd)×V=O(n2d), so the complexity is O(n2d+nd)=O(n2), the quadratic complexity comes from the attention calculation between all input token pairs. Then, the two-branch fuse mainly involves element-by-element operations or connections, which does not incur significant computational overhead. The overall complexity is controllable up to O(n2).

We also conducted single-patient evaluations on the Siena dataset. Table 2 displays the experimental results, showing an average accuracy of 99.18% across all participants. The evaluation results indicate that the majority of cases exhibit a sensitivity above 98%, specificity above 99%, an F1 score with a minimum and average of 98.98%, and AUC values above 98.94%. It is noteworthy that all metrics for PN00, PN11, and PN13 reach 100%. This indicates that the model’s predictions are entirely correct, accurately detecting both epileptic seizures and non-epileptic states. Furthermore, the high sampling rate of 512Hz suggests that Deepwalk-TS maintains good performance at different sampling rates.

Table 2. Results of single-patient experiments on the Sina dataset.
Case Acc. Sen. Spe. F1 AUC
PN00 100 100 100 100 100
PN01 99.48 99.68 99.92 99.45 98.98
PN03 97.89 99.77 100 99.52 99.40
PN05 100 100 99.00 98.39 99.32
PN06 98.12 97.85 98.94 97.91 97.18
PN07 98.60 98.65 100 99.69 99.82
PN09 97.71 98.47 98.26 98.97 97.88
PN10 99.20 99.50 98.75 98.62 99.01
PN11 100 100 100 100 100
PN12 99.23 99.32 100 98.87 99.56
PN13 98.56 99.20 98.40 98.98 99.12
PN14 99.76 100 97.61 97.36 98.60
PN16 98.93 100 96.75 98.45 97.78
PN17 99.84 99.92 99.76 99.88 99.90
Mean 99.18 98.88 99.02 98.98 98.94

Comparison with existing methods.

In Table 3, we conducted extensive experiments to evaluate results compared with recent studies. We categorized methods into four types: machine learning, CNN-based, LSTM-based, and GCN-based. We first compared Deepwalk-TS with several traditional machine learning approaches, outperforming GMM, Wavelet, PCA-Hybrid, and Random Forest across all metrics, showing a remarkable 12.61% improvement in accuracy over GMM’s 86.93%. Then, we compared Deepwalk-TS with CNN-based methods. The CNN+Transformer achieved 98.76% accuracy, 97.70% sensitivity, and 97.60% specificity, respectively, highlighting the effectiveness of our method in capturing intricate patterns in EEG data.

Table 3. Comparison with existing methods on the CHB-MIT dataset.
Types Author Method Accuracy Sensitivity Specificity
Traditional Gill et al. (2015) [41] GMM hybrid 86.93 86.26 87.58
Janjarasjitt et al. (2017) [42] Single Wavelet 96.87 72.99 98.13
Selvakumari et al. (2019) [43] PCA-Hybrid 95.63 95.70 96.55
Chakraborty et al. (2021) [44] Multiscale+RF 95.06 98.12 99.17
Li et al. (2021) [45] WT, EMD+SVM 97.49 97.34 97.50
Chakrabarti et al. (2022) [33] Random Forster 91.90 94.10 89.70
Li et al. (2023) [46] SSTFT + FKNN 98.81 98.53 99.27
Li et al. (2023) [46] Deepwalk + SSTFT 99.01 98.95 98.20
Amiri et al. (2023) [34] FSST+LDA - 98.44 99.19
CNN-based Yuan et al. (2018) [16] Spec-CNN 93.21 89.54 84.32
Hossain et al. (2019) [47] CNN 98.05 90.00 91.65
Fujita et al. (2019) [47] GAN+1DCNN 96.15 93.53 99.05
Omar et al. (2020) [48] SOC-CNN 96.74 82.35 100.00
Wang et al. (2021) [49] S-1D-CNN 85 90.09 99.81
Cimr et al. (2023) [9] normal-CNN 96.99 97.06 96.89
Xu et al. (2023) [35] 3D-CNN - 95.00 -
Jiang et al. (2023) [50] PCA-SVM 96.67 97.72 95.62
Xiao et al. (2024) [36] SLAM 97.07 96.68 97.42
Zhao et al. (2023) [37] CNN+Transf 98.76 97.70 97.60
LSTM-based Shahbazi et al. (2018) [51] CNN + LSTM 95.51 95.14 94.86
Huang et al. (2019) [52] CNN-BiRNA - 92.88 93.94
Hu et al. (2020) [18] BiLSTM - 93.61 91.85
Yao et al. (2021) [38] LSTM 88.63 87.00 88.63
Asma et al. (2023) [39] LSTM+att 96.48 96.28 96.88
GCN-based Chen et al. (2020) [53] EGCN 98.35 - -
He et al. (2020) [54] GCN+LSTM 98.52 97.75 94.34
Zhao et al. (2021) [13] A-GRN+fl 98.70 99.16 98.66
Jibon et al. (2023) [56] LGCN-Den 98.00 97.84 98.33
Zheng et al. (2021) [55] HGCN 99.40 99.53 88.87
Zheng et al. (2021) [55] HGCN + TS 99.44 98.51 92.36
Ours Deepwalk-TS 99.54 99.07 98.87

Given the sequential feature of EEGs, LSTM-based detection methods have been widely studied. We further compare our approach with classical methods such as CNN+LSTM and LSTM+attention. While LSTM achieved an accuracy of 88.63%, Deepwalk-TS surpassed it with an impressive 10.91% improvement. This improvement is because LSTM primarily focuses on single sequences, making it challenging to integrate information directly. In contrast, Deepwalk-TS integrates the graph structure between channels established by Deepwalk theory. We also compared Deepwalk-TS with GCN-based methods, which demonstrates significant improvements. For instance, against A-GRN+focal loss with an accuracy of 98.70%, Deepwalk-TS achieves an improvement of 0.84%. This indicates that our multi-channel modeling effectively captures spatial relationships.

In experiments exploring the effectiveness of existing feature extraction methods for our module, we combined the optimal SSTFT features with our Deepwalk, achieving an accuracy of 99.01%, sensitivity of 98.95%, and specificity of 98.20%. This demonstrates that both our graph-based feature extraction and existing features contribute to improved epilepsy detection. Similarly, when we combined the representative HGCN method with our TS, we achieved an accuracy of 99.44%, marking a significant improvement over the original method. This proves that incorporating the described seizure-related features into their models enhances their performance.

To further validate the performance, we compared the AUC of our method with the top four performing methods. To present epilepsy detection results intuitively, we further used the ROC graph visualization for assessment. The Fig 4 depicted the average ROC curve of the Deepwalk-TS, with an average AUC of 98.87% across 23 patients, illustrating the strong capability in real-world detection scenarios. In clinical epilepsy monitoring, Accuracy more than 95% is sufficient to alert doctors or instruments to make timely diagnostic interventions. Our proposed Deepwalk-TS model has achieved a very satisfactory 99.54% Accuracy in the test phase. While 100% accuracy is rare in real-world tasks like seizure detection due to noise and variability in the data. The proposed Deepwalk-TS model’s sensitivity of 99.07% ensures timely seizure detection, while specificity of 98.97% effectively distinguishes seizure from non-seizure events, minimizing false positives.

Fig 4. ROC curve comparing our method with advanced performance methods.

Fig 4

Qualitative analysis.

First, we delve into graph structure analysis and visualization. We reveal the graph structure changes of the spatial graph reconstruction branch and the stochastic generation capability of the Deepwalk-TS model. To showcase the significantly diverse graph structures generated by different patients, we randomly selected and displayed structures from two patients (ch00, ch05) with distinct random walks in Fig 5. Calculating similarity by averaging Manhattan distances demonstrates the ability to generate dynamic graphs at different points in time randomly. The last column represents the final weight matrix of channel attention. There is a noticeable difference in the relationships between channels for each sample.

Fig 5. Example of the process of constructing a map of multichannel EEGs.

Fig 5

Additionally, we visualized the epilepsy test results using t-SNE technology to highlight the feature embedding capability of the Deepwalk-TS model, as shown in Fig 6. t-SNE is a nonlinear dimensionality reduction method that effectively reduces high-dimensional feature data to a two-dimensional space. We displayed the raw EEG signals and the embedded data output from the model’s fully connected (FC) layer for all samples. Different colors represent normal (blue) and epileptic seizure (orange) signals.

Fig 6. t-SNE visualization of the original and model-embedded distributions of the test data.

Fig 6

Different blue and orange represent the characteristic sample points of normal and epileptic seizures, respectively. (a) Original data distribution. (b) Distribution of feature embeddings for model prediction results.

In Fig 6(a), the original feature space shows little distinction between normal and epileptic brain signals. However, Fig 6(b) presents the t-SNE plot of the data embeddings obtained after training Deepwalk-TS, derived from the model’s FC layer output. This visualization clearly separates the two classes in the two-dimensional space, improving the interpretability of the model’s performance in detecting epileptic seizures.

Finally, to address the reviewer’s suggestion, we have added the learning curves, specifically the accuracy and loss curves, of the DeepWalk-TS model in the experimental section. These curves provide a clear visualization of the training dynamics, model convergence, and overall performance during the training and validation phases. As shown in Fig 7, the model exhibits stable convergence, as evidenced by the consistent decrease in the loss function over time, indicating effective optimization. The smooth decline in both training and validation loss suggests efficient learning. Typically, a large gap between training and validation loss signals overfitting, while a minimal gap indicates better generalization. In our results, the validation loss closely follows the training loss, confirming that overfitting is minimal and does not compromise the model’s overall performance.

Fig 7. Accuracy and loss curves for DeepWalk-TS model during the training and validation process.

Fig 7

Ablation studies

In this section, we conducted ablation studies on Deepwalk-TS to explore the relationships between spatial channels. We systematically analyzed the impact of the spatial and temporal branches and the influence of each component on overall performance. Firstly, we performed experiments by removing the Guided-CNN transformer, indicating a decline in performance and highlighting its significant positive impact on model performance. Secondly, we conducted experiments by removing the graph embedding Deepwalk-based GCN. The results showed a weakened ability of the model to distinguish between normal and epileptic signals upon removing graph embedding. The specific results of the ablation experiments are detailed in the table below, outlining performance metrics under different model configurations.

As shown in Fig 8, Deepwalk-TS demonstrates the best performance across all evaluation metrics. As a baseline model, GCN achieved an accuracy of 97.89%, sensitivity of 97.32%, specificity of 98.56%, F1-Score of 97.98%, and AUC of 98.10%. However, GCN slightly lags behind other models in terms of accuracy and AUC, suggesting further room for improvement in exploring the spatial relationships of EEG channels. The designed Deepwalk submodule shows a significant improvement over GCN with 98.76% accuracy, 98.42% sensitivity and 99.12% specificity. It shows the effectiveness of the node embedding in learning channel relationships.

Fig 8. Visual histogram of ablation test performance.

Fig 8

Furthermore, the Transformer temporal branches achieve an accuracy of 98.45%, a significant improvement over traditional LSTM. Its self-attention mechanisms provide a significant advantage, delivering outstanding sensitivity of 99.18% and specificity of 98.76%. Compared to Deepwalk, our Deepwalk-TS demonstrates an improvement of 0.78% in accuracy, 0.26% in F1-Score, and 0.86% in AUC. The Transformer module positively influences overall performance, particularly in enhancing temporal information capture and significantly improving AUC. The model components work synergistically, demonstrating a well-designed and effective overall architecture. Their combined strengths enable Deepwalk-TS to excel in epilepsy detection, with Deepwalk enhancing spatial relationships between EEG channels and the Transformer improving temporal information capture.

Discussion

Our study on epileptic seizure detection using the Deepwalk-TS method has yielded several notable findings. This section provides an in-depth discussion of the primary results, real-world performance, strengths and limitations, and potential directions for future research.

Our results demonstrate that the Deepwalk-TS model significantly outperforms both traditional and contemporary models in epileptic seizures detection across multiple datasets. The dual-branch architecture, integrating the Deepwalk-based GCN for spatial analysis and the Guided-CNN Transformer for temporal feature extraction, has shown to be highly effective. Compared to machine learning approaches like GMM and PCA-Hybrid models, Deepwalk-TS shows remarkable improvements in accuracy, sensitivity, and specificity. Unlike those methods [33,34], Deepwalk-TS leverages advanced deep learning techniques to automatically extract meaningful features. CNN-based methods, while effective in capturing local features from EEG signals, often struggle to capture long-term dependencies and spatial relationships [3537]. Similarly, LSTM-based methods, which are adept at handling sequential data, often neglect the spatial structure of EEG signals [38,39].

The Deepwalk-TS has several notable strengths. First, its dual-branch architecture effectively integrates spatial and temporal features, addressing key challenges in EEGs analysis. The Deepwalk-based spatial branch dynamically captures high-order relationships among EEG channels through adaptive graph construction. Second, the Guided-CNN Transformer temporal branch achieves spatial feature extraction and global attention dependencies effectively. This combination enables the model to extract meaningful sequential patterns, significantly enhancing its sensitivity to seizure-related signals. Third, extensive experiments demonstrate that Deepwalk-TS achieves performance on multiple datasets, underscoring its generalizability and effectiveness across diverse patient populations and data sources. Unlike existing methods that focus on either spatial or temporal dimensions, our dual-branch architecture integrates both, enabling the model to capture complex EEG channel relationships and temporal dependencies.

Despite these promising outcomes, there are several limitations to consider. It is still affected by the inherent variability in individual physiological characteristics across different patients, which impacts the generalizability of the model to unseen data. Furthermore, the complexity of EEG sequences and the high dimensionality of the substantial labeled data pose challenges to computational efficiency and scalability. While the visual interpretability of graph structures and channel attention is a strength, the framework lacks detailed mechanisms for clinical feedback loops. Finally, while our method performs well in offline scenarios, real-time detection may present computational challenges.

The Deepwalk-TS method shows great application potential in online epilepsy detection and real-world deployment. For real-time detection, the spatiotemporal information integration enables the algorithm to process continuous EEGs effectively, achieving highly accurate seizure detection. The framework can be optimized for low-latency detection through techniques such as model pruning, lightweight adaptations, and parallel computations. Regarding integration into closed-loop systems, further optimization is required to enhance real-time performance and feedback integration. The method promises seamless integration into clinical workflows, enabling early epilepsy intervention and real-time feedback in closed-loop systems.

Future research should focus on several key areas to further enhance the capabilities and applicability of the Deepwalk-TS model. Firstly, refining the spatial exploration capabilities can lead to even better utilization of the inherent relationships between EEG channels. Enhancing the robustness of the model to handle inter-individual variability via multi-view representation learning. To further enhance the model’s performance across diverse datasets, we will involve exploring advanced augmentation techniques, such as transfer learning, and domain adaptation. Additionally, techniques such as model pruning and unified representation will be employed to reduce computational complexity while maintaining interpretability in clinical settings. Adapting real-time EEG processing for emotion detection offers significant potential in healthcare, human-computer interaction, and education. Extending our work to emotion classification could lead to systems that dynamically respond to emotional states, improving decision-making in real-world scenarios [40]. Focusing on these areas will not only enhance the performance of the model in epileptic seizure detection but also provide a foundation for extending its application to other neurological disorders.

Conclusion

Our investigation into epileptic seizure detection underscores the importance of addressing persistent challenges, such as underutilized spatial relationships among EEG channels, inter-individual physiological variability, and complexities in EEG sequences. To improve the accuracy of epilepsy detection, we introduced the innovative Dual-Branch Deepwalk-Transformer Spatiotemporal Fusion Network (Deepwalk-TS), specifically designed to address these issues in EEGs. This model integrates a Deepwalk-based GCN for spatial analysis and a Guided-CNN Transformer for temporal sequence feature extraction. Our experiments on multiple datasets show that Deepwalk-TS outperforms existing models and significantly improves the accuracy of epilepsy detection. Ablation experiments affirm the rational and effective design of our model, demonstrating the complementary nature of its various components. In summary, our proposed Deepwalk-TS model, with its dual-branch architecture and fusion of spatiotemporal features, represents a promising advancement in epileptic seizure detection. Future research should refine spatial exploration, enhance model robustness to inter-individual variability, and optimize performance across datasets. Reducing computational complexity with pruning will improve the model’s applicability and interpretability, enabling broader use in epilepsy detection and other neurological disorders.

Acknowledgments

We gratefully acknowledge the valuable discussions from our colleagues at Nanchang Institute of Technology, whose expertise contributed to developing seizure detection. We also thank the relevant research teams for providing public datasets and the college for offering experimental facilities. Our appreciation extends to the anonymous reviewers for their constructive feedback on earlier versions of this paper.

Data Availability

The data is publicly available. The data underlying the results presented in the study are available from (https://physionet.org/content/chbmit/1.0.0/).

Funding Statement

This research was funded by the project “Design and Implementation of Urban Economic Intelligent Analysis and Management Platform (GJJ2202706)”. The project host is Xiaobing Deng. The funders participated in a series of work in this study, including algorithm design, experimental validation, and preparation of manuscripts and publication.

References

  • 1.Boada CM, Grossman SN, Grzeskowiak CL, Dumanis S, French JA. Proceedings of the 2020 epilepsy foundation pipeline conference: emerging drugs and devices. Epilepsy Behav. 2021;125:108364. [DOI] [PubMed]
  • 2.Raghu S, Sriraam N, Gommer ED, Hilkman DMW, Temel Y, Rao SV, et al. Cross-database evaluation of EEG based epileptic seizures detection driven by adaptive median feature baseline correction. Clin Neurophysiol. 2020;131(7):1567–78. doi: 10.1016/j.clinph.2020.03.033 [DOI] [PubMed] [Google Scholar]
  • 3.Shoeibi A, Moridian P, Khodatars M, Ghassemi N, Jafari M, Alizadehsani R, et al. An overview of deep learning techniques for epileptic seizures detection and prediction based on neuroimaging modalities: Methods, challenges, and future works. Comput Biol Med. 2022;149:106053. doi: 10.1016/j.compbiomed.2022.106053 [DOI] [PubMed] [Google Scholar]
  • 4.Diab E, Nouboue C, Perin B, Chen S, de Jonckheere J, Gacquer W, et al. N°232 – Detection of epileptic seizures based on electrocardiogram analysis a machine-learning feasibility study. Clin Neurophysiol. 2023;150:e146–7. doi: 10.1016/j.clinph.2023.03.224 [DOI] [Google Scholar]
  • 5.Anita M, Meena Kowshalya A. Automatic epileptic seizure detection using MSA-DCNN and LSTM techniques with EEG signals. Exp Syst Appl. 2024;238:121727. doi: 10.1016/j.eswa.2023.121727 [DOI] [Google Scholar]
  • 6.Jahan S, Nowsheen F, Antik MM, Rahman MdS, Kaiser MS, Hosen ASMS, et al. AI-based epileptic seizure detection and prediction in internet of healthcare things: a systematic review. IEEE Access. 2023;11:30690–725. doi: 10.1109/access.2023.3251105 [DOI] [Google Scholar]
  • 7.Song Y, Zheng Q, Liu B, Gao X. EEG conformer: convolutional transformer for EEG decoding and visualization. IEEE Trans Neural Syst Rehabil Eng. 2023;31:710–9. doi: 10.1109/TNSRE.2022.3230250 [DOI] [PubMed] [Google Scholar]
  • 8.Yan A, Zhou W, Yuan Q, Yuan S, Wu Q, Zhao X, et al. Automatic seizure detection using Stockwell transform and boosting algorithm for long-term EEG. Epilepsy Behav. 2015;45:8–14. doi: 10.1016/j.yebeh.2015.02.012 [DOI] [PubMed] [Google Scholar]
  • 9.Cimr D, Fujita H, Tomaskova H, Cimler R, Selamat A. Automatic seizure detection by convolutional neural networks with computational complexity analysis. Comput Methods Programs Biomed. 2023;229:107277. doi: 10.1016/j.cmpb.2022.107277 [DOI] [PubMed] [Google Scholar]
  • 10.Siddiqui MK, Morales-Menendez R, Huang X, Hussain N. A review of epileptic seizure detection using machine learning classifiers. Brain Inform. 2020;7(1):5. doi: 10.1186/s40708-020-00105-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Zhou M, Tian C, Cao R, Wang B, Niu Y, Hu T, et al. Epileptic seizure detection based on EEG signals and CNN. Front Neuroinform. 2018;12:95. doi: 10.3389/fninf.2018.00095 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Qiu X, Yan F, Liu H. A difference attention ResNet-LSTM network for epileptic seizure detection using EEG signal. Biomed Signal Process Control. 2023;83:104652. doi: 10.1016/j.bspc.2023.104652 [DOI] [Google Scholar]
  • 13.Zhao Y, Dong C, Zhang G, Wang Y, Chen X, Jia W, et al. EEG-based seizure detection using linear graph convolution network with focal loss. Comput Methods Prog Biomed. 2021;208:106277. doi: 10.1016/j.cmpb.2021.106277 [DOI] [PubMed] [Google Scholar]
  • 14.Ahmad I, Wang X, Javeed D, Kumar P, Samuel OW, Chen S. A hybrid deep learning approach for epileptic seizure detection in EEG signals. IEEE J Biomed Health Inform. 2023;10:1–12. [DOI] [PubMed]
  • 15.Choi G, Park C, Kim J, Cho K, Kim TJ, Bae H, Min K, Jung K, Chong J. A novel multi-scale 3D CNN with deep neural network for epileptic seizure detection. In: 2019 IEEE International Conference on Consumer Electronics (ICCE). 2019; 1–2.
  • 16.Yuan Y, Xun G, Jia K, Zhang A. A multi-view deep learning framework for EEG seizure detection. IEEE J Biomed Health Inform. 2019;23(1):83–94. doi: 10.1109/JBHI.2018.2871678 [DOI] [PubMed] [Google Scholar]
  • 17.Nogay HS, Adeli H. Detection of epileptic seizure using pretrained deep convolutional neural network and transfer learning. Eur Neurol. 2020;83(6):602–14. doi: 10.1159/000512985 [DOI] [PubMed] [Google Scholar]
  • 18.Hu X, Yuan S, Xu F, Leng Y, Yuan K, Yuan Q. Scalp EEG classification using deep Bi-LSTM network for seizure detection. Comput Biol Med. 2020;124:103919. doi: 10.1016/j.compbiomed.2020.103919 [DOI] [PubMed] [Google Scholar]
  • 19.Wang Y, Cui W, Yu T, Li X, Liao X, Li Y. Dynamic multi-graph convolution-based channel-weighted transformer feature fusion network for epileptic seizure prediction. IEEE Trans Neural Syst Rehabil Eng. 2023;31:4266–77. doi: 10.1109/TNSRE.2023.3321414 [DOI] [PubMed] [Google Scholar]
  • 20.Li Y, Liu Y, Guo Y-Z, Liao X-F, Hu B, Yu T. Spatio-temporal-spectral hierarchical graph convolutional network with semisupervised active learning for patient-specific seizure prediction. IEEE Trans Cybern. 2022;52(11):12189–204. doi: 10.1109/TCYB.2021.3071860 [DOI] [PubMed] [Google Scholar]
  • 21.Berahmand K, Nasiri E, Rostami M, Forouzandeh S. A modified DeepWalk method for link prediction in attributed social network. Computing. 2021;103(10):2227–49. doi: 10.1007/s00607-021-00982-2 [DOI] [Google Scholar]
  • 22.Akbari H, Sadiq MT. Detection of focal and non-focal EEG signals using non-linear features derived from empirical wavelet transform rhythms. Phys Eng Sci Med. 2021;44(1):157–71. doi: 10.1007/s13246-020-00963-3 [DOI] [PubMed] [Google Scholar]
  • 23.Sun C, Cui H, Zhou W, Nie W, Wang X, Yuan Q. Epileptic seizure detection with EEG textural features and imbalanced classification based on easyensemble learning. Int J Neural Syst. 2019;29(10):1950021. doi: 10.1142/S0129065719500217 [DOI] [PubMed] [Google Scholar]
  • 24.Al-Hadeethi H, Abdulla S, Diykh M, Deo RC, Green JH. Adaptive boost LS-SVM classification approach for time-series signal classification in epileptic seizure diagnosis applications. Exp Syst Appl. 2020;161:113676. doi: 10.1016/j.eswa.2020.113676 [DOI] [Google Scholar]
  • 25.Vicnesh J, Hagiwara Y. Accurate detection of seizure using nonlinear parameters extracted from EEG signals. J. Mech. Med. Biol. 19:1940004; 2019.
  • 26.O’Shea A, Lightbody G, Boylan G, Temko A. Neonatal seizure detection from raw multi-channel EEG using a fully convolutional architecture. Neural Netw. 2020;123:12–25. doi: 10.1016/j.neunet.2019.11.023 [DOI] [PubMed] [Google Scholar]
  • 27.Emami A, Kunii N, Matsuo T, Shinozaki T, Kawai K, Takahashi H. Seizure detection by convolutional neural network-based analysis of scalp electroencephalography plot images. Neuroimage Clin. 2019;22:101684. doi: 10.1016/j.nicl.2019.101684 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Roy AM. Adaptive transfer learning-based multiscale feature fused deep convolutional neural network for EEG MI multiclassification in brain-computer interface. Eng Appl Artif Intell. 2022;116:105347.
  • 29.Pan H, Gong S, Dong F, Jiang L. Epilepsy seizure detection using Transformer. HSET. 2022;1:325–9. doi: 10.54097/hset.v1i.482 [DOI] [Google Scholar]
  • 30.Lih OS, Jahmunah V, Palmer EE, Barua PD, Dogan S, Tuncer T, et al. EpilepsyNet: Novel automated detection of epilepsy using transformer model with EEG signals from 121 patient population. Comput Biol Med. 2023;164:107312. doi: 10.1016/j.compbiomed.2023.107312 [DOI] [PubMed] [Google Scholar]
  • 31.Cao P, Zhu Z, Wang Z, Zhu Y, Niu Q. Applications of graph convolutional networks in computer vision. Neural Comput Appl. 2022;34(16):13387–405. doi: 10.1007/s00521-022-07368-1 [DOI] [Google Scholar]
  • 32.Salim I, Hamza AB. Classification of developmental and brain disorders via graph convolutional aggregation. Cogn Comput. 2023;16(2):701–16. doi: 10.1007/s12559-023-10224-6 [DOI] [Google Scholar]
  • 33.Chakrabarti S, Swetapadma A, Pattnaik PK. An improved method for recognizing pediatric epileptic seizures based on advanced learning and moving window technique. AIS. 2022;14(1):39–59. doi: 10.3233/ais-210042 [DOI] [Google Scholar]
  • 34.Amiri M, Aghaeinia H, Amindavar HR. Automatic epileptic seizure detection in EEG signals using sparse common spatial pattern and adaptive short-time Fourier transform-based synchrosqueezing transform. Biomed Signal Process Control. 2023;79:104022. doi: 10.1016/j.bspc.2022.104022 [DOI] [Google Scholar]
  • 35.Xu Y, Yang J, Ming W, Wang S, Sawan M. Deep learning for short-latency epileptic seizure detection with probabilistic classification. IEEE Trans Affect Comput. 2023;14:382–93.
  • 36.Xiao T, Wang Z, Zhang Y, Lv H, Wang S, Feng H, et al. Self-supervised learning with attention mechanism for EEG-based seizure detection. Biomed Signal Process Control. 2024;87:105464. doi: 10.1016/j.bspc.2023.105464 [DOI] [Google Scholar]
  • 37.Zhao Y, Chu D, He J, Xue M, Jia W, Xu F, et al. Interactive local and global feature coupling for EEG-based epileptic seizure detection. Biomed Signal Process Control. 2023;81:104441. doi: 10.1016/j.bspc.2022.104441 [DOI] [Google Scholar]
  • 38.Yao X, Li X, Ye Q, Huang Y, Cheng Q, Zhang G-Q. A robust deep learning approach for automatic classification of seizures against non-seizures. Biomed Signal Process Control. 2021;64:102215. doi: 10.1016/j.bspc.2020.102215 [DOI] [Google Scholar]
  • 39.Baghdadi A, Fourati R, Aribi Y, Daoud S, Dammak M, Mhiri C, et al. A channel-wise attention-based representation learning method for epileptic seizure detection and type classification. J Ambient Intell Human Comput. 2023;14(7):9403–18. doi: 10.1007/s12652-023-04609-6 [DOI] [Google Scholar]
  • 40.Prakash A, Poulose A. Electroencephalogram-based emotion recognition: a comparative analysis of supervised machine learning algorithms. Data Sci Manag. 2025. doi: 10.1016/j.dsm.2024.12.004 [DOI] [Google Scholar]
  • 41.Gill AF, Fatima SA, Usman Akram M, Khawaja SG, Awan SE. Analysis of EEG signals for detection of epileptic seizure using hybrid feature set. In: Theory and Applications of Applied Electromagnetics: APPEIC 2014. Springer; 2015. p. 49–57. [Google Scholar]
  • 42.Janjarasjitt S. Performance of epileptic single-channel scalp EEG classifications using single wavelet-based features. Australas Phys Eng Sci Med. 2017;40(1):57–67. doi: 10.1007/s13246-016-0520-4 [DOI] [PubMed] [Google Scholar]
  • 43.Selvakumari RS, Mahalakshmi M, Prashalee P. Patient-specific seizure detection method using hybrid classifier with optimized electrodes. J Med Syst. 2019;43(5):121. doi: 10.1007/s10916-019-1234-4 [DOI] [PubMed] [Google Scholar]
  • 44., Chakraborty M, Mitra D. A computationally efficient automated seizure detection method based on the novel idea of multiscale spectral features. Biomed Signal Process Control. 2021;70:102990. doi: 10.1016/j.bspc.2021.102990 [DOI] [Google Scholar]
  • 45.Li C, Zhou W, Liu G, Zhang Y, Geng M, Liu Z, et al. Seizure onset detection using empirical mode decomposition and common spatial pattern. IEEE Trans Neural Syst Rehabil Eng. 2021;29:458–67. doi: 10.1109/TNSRE.2021.3055276 [DOI] [PubMed] [Google Scholar]
  • 46.Li M, Chen W, Xia M. GNMF-based quadratic feature extraction in SSTFT domain for epileptic EEG detection. Biomed Signal Process Control. 2023;80:104274. doi: 10.1016/j.bspc.2022.104274 [DOI] [Google Scholar]
  • 47.Fujita H, Cimr D. Computer aided detection for fibrillations and flutters using deep convolutional neural network. Inf Sci 2019;486:231–9.
  • 48.Kaziha O, Bonny T. A convolutional neural network for seizure detection. In: 2020 Advances in Science and Engineering Technology International Conferences (ASET). 2020. p. 1–5. doi: 10.1109/aset48392.2020.9118362 [DOI] [Google Scholar]
  • 49.Wang X, Wang X, Liu W, Chang Z, Kärkkäinen T, Cong F. One dimensional convolutional neural networks for seizure onset detection using long-term scalp and intracranial EEG. Neurocomputing. 2021;459:212–22. doi: 10.1016/j.neucom.2021.06.048 [DOI] [Google Scholar]
  • 50.Jiang L, He J, Pan H, Wu D, Jiang T, Liu J. Seizure detection algorithm based on improved functional brain network structure feature extraction. Biomed Signal Process Control. 2023;79:104053. doi: 10.1016/j.bspc.2022.104053 [DOI] [Google Scholar]
  • 51.Shahbazi M, Aghajan H. A generalizable model for seizure prediction based on deep learning using CNN-LSTM architecture. In: 2018 IEEE Global Conference on Signal and Information Processing (GlobalSIP). 2018. p. 469–73. doi: 10.1109/globalsip.2018.8646505 [DOI] [Google Scholar]
  • 52.Huang C, Chen W, Cao G. Automatic epileptic seizure detection via attention-based CNN-BiRNN. In: 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). IEEE; 2019. 660–3. [Google Scholar]
  • 53.Chen X, Zheng Y, Niu Y, Li C. Epilepsy classification for mining deeper relationships between EEG channels based on GCN. In: 2020 International Conference on Computer Vision, Image and Deep Learning (CVIDL). IEEE; 2020. p. 701–6. [Google Scholar]
  • 54.He J, Cui J, Zhang G, Xue M, Chu D, Zhao Y. Spatial–temporal seizure detection with graph attention network and bi-directional LSTM architecture. Biomed Signal Process Control. 2022;78:103908. doi: 10.1016/j.bspc.2022.103908 [DOI] [Google Scholar]
  • 55.Zeng D, Huang K, Xu C, Shen H, Chen Z. Hierarchy graph convolution network and tree classification for epileptic detection on electroencephalography signals. IEEE Trans Cogn Dev Syst. 2021;13(4):955–68. doi: 10.1109/tcds.2020.3012278 [DOI] [Google Scholar]
  • 56.Jibon FA, Miraz MH, Khandaker MU, Rashdan M, Salman M, Tasbir A, Nishar NH, Siddiqui FH. Epileptic seizure detection from electroencephalogram (EEG) signals using linear graph convolutional network and DenseNet based hybrid framework. J Radiat Res Appl Sci. 2023;16:100607.
  • 57.Qiu J, Dong Y, Ma H, Li J, Wang K, Tang J. Network embedding as matrix factorization: unifying deepwalk, line, pte, and node2vec. Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining. 2018. p 459–67.

Decision Letter 0

Yuvaraj Rajamanickam

3 Jan 2025

PONE-D-24-53519A Novel Dual-Branch Network for Comprehensive Spatiotemporal Information Integration for EEG-based Epileptic Seizure DetectionPLOS ONE

Dear Dr. Deng,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Feb 17 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Yuvaraj Rajamanickam, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf.

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. Thank you for stating in your Funding Statement:

“This study was supported by the project foundation design and implementation of intelligent analysis and management platform for the urban economy (GJJ2202706), Xiaobing Deng.”

Please provide an amended statement that declares *all* the funding or sources of support (whether external or internal to your organization) received during this study, as detailed online in our guide for authors at http://journals.plos.org/plosone/s/submit-now. Please also include the statement “There was no additional external funding received for this study.” in your updated Funding Statement.

Please include your amended Funding Statement within your cover letter. We will change the online submission form on your behalf.

4. Thank you for stating the following in the Acknowledgments Section of your manuscript:

“This study was supported by the project foundation design and implementation of intelligent analysis and management platform for urban economy (GJJ2202706).”

We note that you have provided additional information within the Acknowledgements Section that is not currently declared in your Funding Statement. Please note that funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

“This study was supported by the project foundation design and implementation of intelligent analysis and management platform for the urban economy (GJJ2202706), Xiaobing Deng.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

Additional Editor Comments (if provided):

Comments:

Please discuss whether the application of the algorithm is possible for online detection (i.e., close loop system).

What are the strengths and weaknesses of the proposed framework? Please include them in the paper.

What about the computational complexity of the proposed method? Some description related to computational complexity is required in the paper.

There is a huge amount of literature regarding EEG-derived fingerprints of seizure detection. Do authors consider that the incorporation of already described seizure-related features into their model could improve their results? Some discussion on that is needed. Without this more rigorous comparison with published work, this work is just another incremental addition.

The authors never state what level of performance would be considered adequate for a “real-world scenario” system that is useful and valid.

How the proposed emotion classification system be useful in the real-world scenario? Need to discuss.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors have performed a study on a deep-learning based model to detect epileptic seizures. The paper is well-written and the proposed methodology is robust.

Some minor points need to be revised.

Introduction:

- page 1: the authors state that “early diagnosis of seizures is important for patients with refractory epilepsy to prevent irreversible damage to the patient’s brain”. This sentence makes too many assumptions in implying that single epileptic seizures can damage the brain irreversibly. Seizure detection is an important goal, but this sentence needs to be reformulated.

Results:

- Page 7: section 4.1 and 4.2 should be moved to Method section, since they do not represent a result of the current research.

Discussion:

- Page 14: the authors have analyzed data from two different databases, one of them being represented by a pediatric population. This may influence the results and should be discussed.

Reviewer #2: 1. Please define SOTA in the abstract.

2. Please mention the details in the abstract.

3. Please add paper organization at the end of the introduction.

4. The related work section is missing.

5. Fig.6, please indicate the blue and orange colours.

6. Limitations of the proposed system should be discussed in the paper.

7. The learning curve (Accuracy and loss) of the DL models is missing from the paper.

Reviewer #3: Comment to author(s):

I have gone through this article. It is based on epileptic seizure detection by employing a DeepWalk algorithm on EEG datasets. The overall study has great potential to add significant value to relevant fields. However, some points should be considered before the final acceptance of this article.

1. The evaluation scores were significantly high with an accuracy of 99.54%. Why does it not reach 100%, and how can overfitting be understood in the proposed work? It is possible to add a description related to these concerns to the Discussion section, or wherever possible in the manuscript, it is feasible.

2. Please mention the acronyms in full form when they are first used in the manuscript, such as "GMM," on page 9.

3. Please check Page no. 5, equation 2 and equation 3, the variables "bk" and "Dt" are defined somewhere or not. Also, check all the typos if they are available in the manuscript, such as in page no. 8, section 4.3.1, "...proposed model dem onstrated..."

Overall, the manuscript has great potential and the proposed work is highly valuable. We look forward to getting this article to the next stage of article processing and acceptance.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Umberto Aguglia

Reviewer #2: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: Comment to author(s).pdf

pone.0321942.s001.pdf (282.8KB, pdf)
PLoS One. 2025 Jun 26;20(6):e0321942. doi: 10.1371/journal.pone.0321942.r002

Author response to Decision Letter 1


3 Feb 2025

Response to Reviewers

Dear Editor and Reviewers:

System Response to Reviewers lacks formula and image modification, for details, please refer to the uploaded file entitled "Response to Reviewers". Thank you for your letter and the reviewers’ comments concerning our manuscript entitled “A Novel Dual-Branch Network for Comprehensive Spatiotemporal Information Integration for EEG-based Epileptic Seizure Detection (No. PONE-D-24-53519)”. We deeply appreciate the constructive comments, helpful critiques, and favorable assessments from all reviewers. Those comments are valuable and have significantly contributed to the revision and improvement of our paper. We have carefully reviewed the comments and made the necessary modifications to the resubmitted manuscript. Below are our detailed point-by-point responses to each comment.

Editor’s Comments:

Thank you for your thorough review of our manuscript, bring your constructive feedback and positive comments on the proposed Deepwalk-TS algorithm for epilepsy detection. Below, we address the issues you raised and outline the changes made to improve the paper’s quality and suitability for publication.

1.Please discuss whether the application of the algorithm is possible for online detection (i.e., close loop system).

We have discussed the online application of Dual-Branch Deepwalk Transformer Spatiotemporal Fusion Network (Deepwalk-TS) in detail in the "Discussion" section. It can be determined that method shows great application potential in online epilepsy detection. The algorithm innovatively integrates the spatial information analysis of Deepwalk-based and the temporal feature extraction of Guided-CNN Transformer, dynamically and efficiently processes EEGs, and has achieved accurate and timely epilepsy detection. Although the real-time performance and feedback integration need to be further optimized, the proposed method provides a solid foundation for advanced clinical systems to support online epilepsy detection in a timely manner.

For real-time detection, the spatiotemporal information integration enables the algorithm to process continuous EEGs effectively, achieving highly accurate seizure detection. The Deepwalk-based graph construction dynamically captures spatial relationships across EEG channels, while the Transformer module efficiently processes time-series data, ensuring real-time detection capability. Although current implementations may require minor delays for feature extraction, the framework can be optimized for low-latency detection through techniques such as model pruning, lightweight architecture adaptations, and parallel computations. The model is well-suited for online streaming data, making it feasible for real-time monitoring.

Regarding integration into closed-loop systems, our study primarily focuses on algorithmic design and localized detection using multiple datasets. The method's robust ability to analyze spatiotemporal EEG features ensures reliable seizure detection, even in complex clinical scenarios. While the method emphasizes accuracy and strong performance in static evaluations, further optimization is required to enhance real-time performance and feedback integration. As part of a feedback loop, the system can process real-time EEG data streams to continuously update the prediction model, leveraging adaptive learning techniques such as incremental training or online fine-tuning. Additionally, incorporating active learning or adaptive model updates based on real-time feedback could further improve the model's ability to adapt to new seizure patterns. The proposed framework promises seamless integration into clinical workflows, enabling early intervention and real-time feedback in closed-loop systems for epilepsy management.

2.What are the strengths and weaknesses of the proposed framework? Please include them in the paper.

We have added a detailed discussion of the strengths and weaknesses of our method in the revised manuscript to enhance the clarity and depth of the paper.

The Deepwalk-TS has several notable strengths: First, its dual-branch architecture effectively integrates spatial and temporal features, addressing key challenges in EEGs analysis. The Deepwalk-based spatial branch dynamically captures high-order relationships among EEG channels through adaptive graph construction, ensuring that spatial dependencies are preserved and utilized efficiently. Second, the Guided-CNN Transformer temporal branch leverages CNN’s spatial feature extraction capabilities and the Transformer’s global attention mechanism to model temporal dependencies effectively. This combination enables the model to extract meaningful sequential patterns, significantly enhancing its sensitivity to seizure-related signals. Third, extensive experiments demonstrate that Deepwalk-TS achieves state-of-the-art performance on multiple datasets, underscoring its generalizability and effectiveness across diverse patient populations and data sources. Additionally, the model’s design facilitates interpretability by allowing visual analysis of learned graph structures and channel attention mechanisms, providing valuable insights for clinical applications.

While Deepwalk-TS exhibits high accuracy and robust feature extraction capabilities, there are several limitations to consider. It is still affected by the inherent variability in individual physiological characteristics across different patients, which impacts the generalizability of the model to unseen data. Another limitation is the reliance on substantial labeled data, as the method's performance with raw, unsegmented signals is untested. Furthermore, while the visual interpretability of graph structures and channel attention is a strength, the framework lacks detailed mechanisms for clinical feedback loops.

Future research should focus on several key areas: Enhancing the robustness of the model to handle inter-individual variability via multi-view representation learning. To further enhance the model's performance across diverse datasets, we will explore advanced techniques such as data augmentation, transfer learning, and domain adaptation. Additionally, techniques such as model pruning and unified representation will be employed to reduce computational complexity while preserving accuracy and interpretability. Focusing on these areas will not only enhance the performance and applicability of the method in epileptic seizure detection but also provide a foundation for researchers to extend its use to other neurological disorders and related fields. We have incorporated these discussions in the revised manuscript to provide a more balanced and comprehensive overview of the framework’s strengths, weaknesses, and directions for future improvements.

3.What about the computational complexity of the proposed method? Some description related to computational complexity is required in the paper.

Thank you for raising the important point regarding the computational complexity of our proposed method. We included additional calculations in Experimental Results subsection to provide a more comprehensive analysis. The computational complexity of this model primarily stems from two key components: the Deepwalk-based GCN and the Guided-CNN Transformer.

Deepwalk-based GCN complexity can be approximated as O(n×k×d), where n is the number of nodes (n = 16), k is the number of random walks per node (k = 9), and d is the dimensionality of the node embeddings. The random walk incurs higher costs with more nodes, walks, and embedding dimensions. Two-layer GCNs typically involve O(v×F×F1), where v is the number of nodes, F and F1 are the original and output feature dimensions. The module's complexity is O(n×k×d)+O(v×F×F1)+ O(E×F1), E is the edges. Through extensive experiments, we have determined that the optimal number of EEG channels is 11, which helps manage the computational overhead. Guided-CNN Transformer complexity is focused on the self-attention mechanisms of Transformer. Q, K, V matrix shape is n×d, Q×Kt = (n×d×n), scores×V = softmax(Q×Kt /d)×V = O(n2d), so the complexity is O(n2d + nd) = O(n2), the quadratic complexity comes from the attention calculation between all input token pairs. Then, the outputs of the GCN space and Transformer time branches are fused, which mainly involves element-by-element operations or connections, which does not incur significant computational overhead. During training, the two branches are jointly optimized through techniques such as balanced loss function and mini-batch gradient descent, and the overall complexity is controllable up to O(n2). We have expanded the discussion of computational complexity in the revised paper, with a clearer breakdown of each component.

4.There is a huge amount of literature regarding EEG-derived fingerprints of seizure detection. Do authors consider that the incorporation of already described seizure-related features into their model could improve their results? Some discussion on that is needed. Without this more rigorous comparison with published work, this work is just another incremental addition.

Our work focuses on developing an end-to-end framework, the Deepwalk-TS, which automatically learns spatiotemporal features directly from raw EEG data. This approach eliminates the dependency on handcrafted features, allowing the model to autonomously capture high-order spatial relationships and temporal dependencies. Below, we address the existing seizure features and works, results comparisons with existing research, and the incremental contributions of our proposed model.

In the revised manuscript, we have added a Related Works section, providing an overview of feature extraction and EEG-based seizure detection methods. In this paper, we focused on a deep learning-based approach that learns both spatial-temporal relationships between EEG channels, which is an area that has not been fully explored in the current works. Then, we revised the manuscript to clarify and expand on the integration of existing seizure-related features in the Experiment section. We conducted additional experiments incorporating mainstream handcrafted-based and GCN-based epilepsy detection methods, fusing features with representative methods like SSTFT and HGCN in Table 3. The results of these experiments have been updated in Table 3 of the revised manuscript.

We combined the optimal SSTFT features with our Deepwalk, achieving an accuracy of 99.01%, sensitivity of 98.95%, and specificity of 98.20%. This demonstrates that both our graph-based feature extraction and existing features contribute to improved epilepsy detection. Similarly, when we combined the representative HGCN method with our TS, we achieved an accuracy of 99.44%, marking a significant improvement over the original method. This proves that incorporating the described seizure-related features into their models enhances their performance. We acknowledge that integrating existing features could enhance the model's generalizability and robustness, providing complementary value and improving its practical efficacy.

While deep learning-based seizure detection is a growing research field, the novelty of Deepwalk-TS lies in its holistic integration of spatiotemporal features. Unlike existing methods that primarily focus on either spatial or temporal dimensions, our dual-branch architecture combines both, allowing the model to capture intricate relationships across EEG channels while effectively modeling temporal dependencies. This unified approach enhances the model's adaptability to complex and heterogeneous datasets, improving its robustness and accuracy in seizure detection. Therefore, our work should not be viewed as merely an incremental addition to existing research, but rather as a significant exploration in epileptic seizure detection.

5.The authors never state what level of performance would be considered adequate for a “real-world scenario” system that is useful and valid.

The reviewer's comment has inspired us to recognize the importance of defining performance thresholds for real-world epilepsy detection, which are crucial for assessing the practical applicability of our method. A system should achieve a minimum threshold of sensitivity (≥ 90%) is vital in clinical applications to reduce the risk of false negatives, which could lead to missed seizures and delayed interventions. In our experiments, the proposed Deepwalk-TS model achieves a sensitivity of 99.07%, which meets the necessity for ensuring timely detection of seizures. Meanwhile, high specificity (≥ 95%) is also important to avoid false alarms leading to unnecessary treatments or disruptions. Our model achieves a specificity of 98.97%, which is suitable for distinguishing between seizure events and non-seizure events, minimizing the occurrence of false positives. In addition, the overall accuracy (≥ 95%) and Area Under the Curve (AUC ≥ 0.98) are also crucial for evaluating the overall performance of the detection system. Our model achieves an accuracy of 99.54% and an AUC of 99.06%, exceeding the necessary thresholds for precise epilepsy detection. In a real-world deployment beyond performance metrics, real-time detection with low-latency processing is essential for timely seizure detection and intervention. Although our model achieves high accuracy and sensitivity, further optimization is required for efficient real-time operation in portable EEG devices or closed-loop systems.

6.How the proposed emotion classification system be useful in the real-world scenario? Need to discuss.

We understand the question asks how our seizure detection work can be applied to emotion classification in real-world scenarios. While our research primarily focuses on EEG-based seizure detection, the core module of Deepwalk-TS can capture both spatial and temporal features from raw EEG signals and can be effectively applied to other domains, including emotion classification.

We have included a discussion on the application of the Deepwalk-TS algorithm for emotion classification in the revised version of the Discussion section. Adapting real-time EEG processing to detect emotional states offers vast potential across healthcare, human-computer interaction, education, and more. Emotion detection using EEGs can help track emotional distress, such as anxiety or depression, aiding clinicians in identifying early signs for timely intervention. It can also enhance customer interactions by detecting emotions like frustration or satisfaction. Additionally, seizure detection methods could extend to entertainment, such as video games, by tailoring user experiences based on emotional states. By extending our work into emotion classification, future research could create systems that respond dynamically to a person’s emotional state, offering better decision-making in various real-world scenarios. We have expanded on the potential applications of emotion classification using similar techniques in the revised manuscript.

Reviewer #1: The authors have performed a study on a deep-learning based model to detect epileptic seizures. The paper is well-written and the proposed methodology is robust. Some minor points need to be revised.

Thank you very much for your positive feedback and constructive comments. We greatly appreciate your recognition of the value of our work and the robustness of methodology. We have carefully considered your comments and incorporated the necessary revisions to further improve the clarity and depth of our paper.

1.Introduction: - page 1 the authors state that “early diagnosis of seizures is important for patients with refractory epilepsy to prevent irreversible damage to the patient’s brain”. This sentence makes too many assumptions in implying that single epileptic seizures can damage the brain irreversibly. Seizure detection is an important goal, but this sentence needs to be reformulated.

We have revised the sentence and refined the logic of the introduction to ensure clarity and precision. The revised sentence now reads: "Early detection of epilepsy is crucial for timely intervention, reducing the risk of irreversible brain damage, and improving the overall quality of life for patients."

In this revised version, we emphasize the importance of early diagnosis in the high prevale

Attachment

Submitted filename: Response to Reviewers.pdf

pone.0321942.s002.pdf (262.2KB, pdf)

Decision Letter 1

Yuvaraj Rajamanickam

28 Feb 2025

PONE-D-24-53519R1A Novel Dual-Branch Network for Comprehensive Spatiotemporal Information Integration for EEG-based Epileptic Seizure DetectionPLOS ONE

Dear Dr. Deng,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Apr 14 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Yuvaraj Rajamanickam, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Thank you for addressing my previous comments and revising the manuscript accordingly. I appreciate the efforts made to improve the clarity and quality of the paper. The modifications have significantly enhanced the manuscript, and I find the current version to be well-structured and scientifically sound. I have no further major concerns at this stage.

Reviewer #2: Please find the attached file for the comments. The paper needs minor revisions before the next submission.

Reviewer #3: The author revised the manuscript where, all comments are addressed and responses are satisfactory. I do not have any further queries.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Umberto Aguglia

Reviewer #2: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: Comments R2.docx

pone.0321942.s003.docx (14.8KB, docx)
PLoS One. 2025 Jun 26;20(6):e0321942. doi: 10.1371/journal.pone.0321942.r004

Author response to Decision Letter 2


11 Mar 2025

Response to Reviewers

Dear Editor and Reviewers:

We sincerely appreciate your letter and the reviewers’ valuable feedback on our manuscript titled "A Novel Dual-Branch Network for Comprehensive Spatiotemporal Information Integration for EEG-based Epileptic Seizure Detection" (PONE-D-24-53519R1). We are grateful for the reviewers’ positive comments, which have helped us further refine and improve our work. We have carefully reviewed all the suggestions and have made the necessary revisions in the Minor Revision resubmission. Below are our detailed point-by-point responses to each comment.

Editor Comments:

1.Please review your reference list to ensure that it is complete and correct.

We have carefully reviewed our reference list and thoroughly checked each citation to ensure completeness and accuracy. Missing fields, such as page numbers and other relevant details, have been corrected and supplemented where necessary. The revised manuscript now includes a fully verified and properly formatted reference list that adheres to the journal’s guidelines.

2.While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool.

Thank you for your valuable feedback, which has helped us improve the overall quality of our submission. We have carefully reviewed all figure files to ensure they meet the journal's formatting requirements. Please let us know if any further modifications are needed. We appreciate your guidance and support throughout the revision process.

Reviewer #1: The modifications have significantly enhanced the manuscript, and I find the current version to be well-structured and scientifically sound. I have no further major concerns at this stage.

Thank you for your positive feedback and recognition of our efforts in improving the manuscript. We sincerely appreciate your thoughtful review and valuable suggestions, which have helped us enhance the clarity, structure, and scientific rigor of our work. Your support and encouragement are greatly appreciated!

Reviewer #2: Please find the attached file for the comments. The paper needs minor revisions before the next submission.

1.The Fig.6 results show the overfitting issue. Please check it. The validation accuracy is higher than the training accuracy in the case of the SEED dataset.

We sincerely thank the reviewer for his insightful comments on the potential overfitting issue in the loss curve. Generally, a larger gap between the training and validation curves indicates strong overfitting, while a minimal gap indicates less severe overfitting. After carefully re-evaluating the training and validation process, we found that the original validation curve was not fully plotted, which could have led to a misinterpretation of the results. We have corrected this issue in the revised figure and observe that the validation accuracy remains consistent with the training accuracy, indicating that the overfitting is small and does not significantly affect the overall model performance.

2.Please consider the following reference for your discussion

We sincerely appreciate the reviewer’s suggestion to incorporate this reference into our discussion. The recommended paper offers valuable insights into EEG-based emotion recognition and the comparative performance of various supervised machine learning algorithms. We have now cited this reference in the revised manuscript, and we greatly appreciate the reviewer’s insightful recommendation.

Reviewer #3: The author revised the manuscript where, all comments are addressed and responses are satisfactory. I do not have any further queries.

Thank you for your thoughtful review and positive feedback. We sincerely appreciate your time and effort in evaluating our manuscript. Your valuable comments have helped us improve the quality of our work, and we are grateful for your support!

We greatly appreciate the positive feedback regarding the potential and value of our work. We are committed to addressing these concerns and will revise the manuscript to improve its clarity and completeness. We are excited to submit the revised version for you to look over. If you have any queries, please don’t hesitate to contact me at the e-mail below.

Sincerely,

Xiaobing Deng

guoguo10201@163.com

Attachment

Submitted filename: Response_to_Reviewers_auresp_2.pdf

pone.0321942.s004.pdf (84.6KB, pdf)

Decision Letter 2

Yuvaraj Rajamanickam

14 Mar 2025

A Novel Dual-Branch Network for Comprehensive Spatiotemporal Information Integration for EEG-based Epileptic Seizure Detection

PONE-D-24-53519R2

Dear Dr. Deng,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Yuvaraj Rajamanickam, Ph.D

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Yuvaraj Rajamanickam

PONE-D-24-53519R2

PLOS ONE

Dear Dr. Deng,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Yuvaraj Rajamanickam

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Comment to author(s).pdf

    pone.0321942.s001.pdf (282.8KB, pdf)
    Attachment

    Submitted filename: Response to Reviewers.pdf

    pone.0321942.s002.pdf (262.2KB, pdf)
    Attachment

    Submitted filename: Comments R2.docx

    pone.0321942.s003.docx (14.8KB, docx)
    Attachment

    Submitted filename: Response_to_Reviewers_auresp_2.pdf

    pone.0321942.s004.pdf (84.6KB, pdf)

    Data Availability Statement

    The data is publicly available. The data underlying the results presented in the study are available from (https://physionet.org/content/chbmit/1.0.0/).


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES