Skip to main content
PLOS One logoLink to PLOS One
. 2026 Apr 17;21(4):e0345204. doi: 10.1371/journal.pone.0345204

Enhancing generalizability in classification of peripheral neural recordings with graph neural network

Rui Qi Ji 1,*, Mehdy Dousty 1,2,, Ryan G L Koh 3,4,, Ervin Sejdić 1,5
Editor: Luca Citi6
PMCID: PMC13089695  PMID: 41996395

Abstract

The peripheral nervous system plays a crucial role in facilitating communication between biological systems. However, decoding neural signals from peripheral nerve recordings remains a challenge due to their complex spatiotemporal patterns. In this study, we propose a graph-based learning approach to more effectively capture temporal and spatial information for classifying neural signal patterns. Unlike previous work, our method incorporates the physical geometry of the nerve cuff, addressing the underrepresented relationships between electrodes. We used a publicly available dataset consisting of neural recordings from eight Long-Evans rats, obtained using a 56-channel nerve cuff electrode. We constructed graphs where each node represents the time series recorded from an electrode, and edges correspond to the distances between electrodes along the surface of the nerve cuff (e.g., geodesic distance). We employed a leave-one-out strategy to evaluate the generalizability of the approach. We further evaluated the within-rat performance of the model by training on two folds and testing on the remaining fold of each rat’s data. In generalizability evaluation, we achieved a mean F1 score of 65.03%, representing a 17.74% improvement over the previous study, and in within-rat testing, we achieved a mean F1 score of 77.50%, representing a 3.14% increase. These findings highlight the value of incorporating the recording geometry into model design, particularly in this small dataset setting, where explicit spatial priors help compensate for limited training examples and improve decoding performance.

Introduction

The nervous system is a fundamental component of biological function, serving as the primary communication network within the body [1]. It is responsible for transmitting electrical and chemical signals that regulate movement, sensation, and autonomic processes [2,3]. The peripheral nervous system facilitates bidirectional communication between the brain and the rest of the body [4,5]. As such, peripheral neural signals encode crucial information about sensory inputs and motor commands, making their accurate interpretation essential for various biomedical engineering applications, and the diagnosis of neurologic disorders [68].

Nevertheless, obtaining reliable and selective recordings in the peripheral nervous system is challenging [9]. Recent studies have utilized multi-contact electrodes [1012] to achieve encoding both temporal and spatial information of neural recordings in Long-Evan rats. Convolutional neural networks (CNNs) have demonstrated significant capability in analyzing physiological data [1316]. Koh et al. proposed a CNN-based approach to classify three afferent activities in peripheral neural recordings from Long-Evans rats: dorsiflexion, plantarflexion, and pricking [12]. These movements are fundamental to locomotion, balance, and pain perception [17]. Thus, classifying these three types of neural activity is critical for understanding motor control mechanisms. Such classification has direct implications in neuroprosthetics, rehabilitation, and sensory feedback systems for amputees or patients with neuromuscular disorders [17,18].

CNNs operate through receptive fields and therefore capture the structural layout of multichannel neural recordings, enabling them to model local spatial dependencies by leveraging weight sharing [19]. Transformers, based on self-attention, have been shown to lack the strong architectural inductive biases of convolutional networks, such as limited receptive fields and translational invariance, that can be beneficial for spatially structured inputs [20]. Although Transformers can model long-range dependencies [21], they require substantially more data to learn useful representations in the absence of such biases, and do not inherently encode the geometric relationships present in nerve cuff recordings. Similarly, other neural network architectures commonly applied in the literature for peripheral neural signal analysis or classic machine learning classifiers do not encode the true inter-electrode distances or circumferential arrangement around the nerve, meaning that important physiological dimensions of electrode placement, such as the true inter-electrode distances and the way neural signals propagate across adjacent contacts, remain unmodeled. In contrast, graph neural networks (GNNs) offer greater flexibility in representing irregular topologies, as they enable a more adaptive encoding of both temporal and spatial relationships in neural data [22]. Unlike CNNs, which require large amounts of data to learn spatial filters on fixed grid structures, GNNs incorporate structural priors through graph connectivity, enabling more effective biosignal analysis and improved generalization in data-limited settings [2327]. Previous work has also highlighted the importance of how combining both temporal and spatial information yields the highest performance [28], thus, the main motivation for using GNNs in our context is their ability to simultaneously encode the data’s temporal and structural relationships. For instance, electrodes located on opposite sides of a nerve cuff may record similar physiological activity, yet this relationship cannot be effectively captured using conventional CNN or Transformer architectures. By modeling these electrode connections as edges in a graph, GNNs naturally integrate this spatial context alongside temporal dynamics.

In this paper, we propose a graph-based learning approach for classifying peripheral neural signals. Our key contributions are

  • 1) Geometry-aware graph construction: We model nerve cuff electrodes as graph nodes and define edges based on geodesic distances to encode both the physical arrangement of electrodes and functional similarities between neural signals. We examine how varying graph connectivity affects model performance and conduct ablation studies to verify that the proposed graph-based approach is the primary contributor to the performance gains. In the ablation studies, we also compare the feature extraction from neural recordings using a LSTM module versus a 1D CNN module.

  • 2) Improved generalizability in small-data regimes: Our approach achieves higher classification accuracy and better generalization, outperforming the CNN baseline from previous studies and highlighting the importance of encoding information with graphs in smaller datasets.

Materials and methods

Data description & preprocessing

Neural recordings were previously collected from nine Long-Evan rats from the sciatic nerve using a 56-multi-contact nerve cuff electrode were used [11]. In this study, we excluded one rat due to an issue with the degradation of the plantarflexion signal, resulting in a dataset comprising of eight rats. Neural recordings were collected from a 56 channel nerve cuff electrode comprised of 7 rings of 8 contact, evenly distributed over the length of the electrode. The recordings were acquired at a sampling frequency of 30 kHz with a neural data acquisition board (RHD2000, Intan Technologies, USA) [11]. Three afferent activities, dorsiflexion, plantarflexion, and pricking, were manually performed to evoke neural activity in the recording. Naturally evoked compound action potentials (nCAPs), produced by proprioceptive or mechano-sensory afferent activity in response to physiological limb movements or mechanical stimulation, were detected, and used to construct spatiotemporal signatures for each activity. Each spatiotemporal signature is a matrix in which rows represent neural activity of individual channels over time, and columns capture signals at specific time points across channels, resulting in a matrix of size 56 by 100 (number of contacts by time samples) [11,12,29]. These spatiotemporal signatures were then constructed into graph structures, which were fed into our proposed model for classification of the three activities. Representing the data as graphs allows us to explicitly encode spatial relationships between electrodes, which may capture physiologically meaningful similarities that conventional matrix-based approaches overlook. The number of samples for each rat is presented in Table 1.

Table 1. Number of samples from each Long-Evan rat, and number of samples for each class.

Rat # Number of Samples Dorsiflexion Plantarflexion Pricking
Rat 2 34,238 4,044 13,272 16,922
Rat 4 31,346 8,602 12,248 10,496
Rat 5 37,610 10,088 9,948 17,574
Rat 6 29,878 11,744 3,704 14,430
Rat 7 22,506 4,438 9,640 8,428
Rat 8 23,532 5,208 11,872 6,452
Rat 9 21,396 9,358 3,306 8,732
Rat 10 11,174 2,600 3,004 5570

Graphs & adjacency matrices

A graph can be represented mathematically as an order pair G=(V,E), where V represents the nodes, and E represents the edges, connecting pairs of nodes together [30]. The graph can be mathematically expressed through an adjacency matrix A, where Aij indicates the presence of an edge between the nodes. If only the existence of edges is considered, the graph is unweighted, where Aij=1 if the edge exists between i and j, and Aij=0 otherwise. In weighted graphs, non-zero values of Aij also represent the strength or significance of the connection [31]. GNNs are deep learning architectures designed to operate directly on graph-structured data. Unlike conventional models that assume regularly structured 1D, 2D, or 3D inputs, GNNs leverage both node features and the connectivity defined by edges to learn representations that capture the underlying topology [22]. Through iterative message passing, each node aggregates information from its neighbors, enabling the model to integrate both local and global structural patterns in the data.

In this work, we represent the electrode contact points (i.e., channels) as nodes and define the edges based on the geodesic distance relationships between them. The nerve cuff consists of multiple electrodes arranged circumferentially around the nerve, so electrodes positioned on opposite sides may capture similar neural activity due to their proximity to the same underlying fibers. The geodesic distance measures the shortest path along a curved surface rather than the straight-line Euclidean distance [32]. This distinction is particularly important for nerve cuff electrodes, which are positioned on a cylindrical surface around the nerve. While Euclidean distance may treat two electrodes on opposite sides of the cuff as far apart, geodesic distance captures their true proximity along the nerve’s surface, providing a more physiologically meaningful representation of spatial relationships.

Given the structure of the nerve cuff electrode, this distance is computed using Equation 1 below, in which x and y represent the row and column indices (e.g., electrode coordinates) of the electrode channels in the nerve cuff array, respectively. The term |xixj| captures the vertical distance between electrodes, while min(|yiyj|,8|yiyj|) accounts for the wrap-around nature of the circular arrangement in the horizontal direction, as shown in Fig 1.

Fig 1. Geodesic distance configuration in graph construction, where x and y represent the row and column indices (e.g., electrode coordinates).

Fig 1

dgeodesic(i,j)=|xixj|+min(|yiyj|,8|yiyj|) (1)

By utilizing geodesic distance, electrodes that are closer to each other on the cuff are more likely to be directly connected in the graph, representing stronger spatial relationships. This connectivity structure enables the model to learn signal propagation patterns along the electrodes, rather than being limited to local information as in other methods, thereby better capturing spatial correlations and neural dynamics within the recordings. With this approach, we constructed a weighted adjacency matrix of size 56 by 56, where the weights are computed based on the distance between electrode positions, measured by their geodesic distance, as shown in Equation 2 [33].

weight(i,j)=exp(dgeodesic(i,j)22σ2) (2)

To refine the graph topology, we use the distance-scaling parameter σ and the number of nearest neighbours, k, as tunable hyperparameters. Specifically, for each node, edges are first formed only with its k nearest neighboring electrodes based on spatial distance, and the corresponding edge weights are assigned using a distance-based weighting function parametrized by σ. With lower values of σ, only electrodes that are immediately adjacent on the nerve cuff will have substantial edge weights, leading to sharper decay in edge weights and emphasizing local interactions. Larger σ values result in high weights even for electrodes located further apart along the cuff, allowing the model to integrate global information from distant regions of the nerve. By restricting connectivity to the k nearest neighbors, the graph remains sparse and focuses message passing on the most relevant spatial relationships, while σ controls the strength of these connections. This combined-tuning mechanism enables flexible tuning of the graph structure, balancing the trade-off between local specificity and global connectivity, ultimately improving the model’s ability to capture informative patterns from the neural data.

Model architecture

The model architecture is implemented with a hybrid approach, incorporating both sequential and graph-based learning paradigms. The full model architecture is shown in Fig 2. Following the graph formulation introduced above, each neural recording is represented as a graph G = (V,E), where each node viV corresponds to an electrode contact and each edge eijE encodes the spatial relationship between electrodes. The graph structure is represented by a weighted adjacency matrix A, where each non-zero entry Aij denotes the strength of the connection between nodes i and j, as defined by the geodesic distance-based graph construction procedure.

Fig 2. Proposed model architecture.

Fig 2

We first extract temporal representation from the neural signals using a Long Short-Term Memory (LSTM) layer with 256 units. The resulting hidden representations serve as node feature vectors hi for each viV. These node features, together with the weighted adjacency matrix A, are then processed by an edge convolution layer followed by a general graph convolution layer.

The edge convolution layer follows a message-passing formulation in which node features are updated by aggregating information from neighboring nodes j𝒩(i), explicitly incorporating edge information through the corresponding adjacency weights Aij. In this work, edge features are defined directly by these weighted adjacency values, which encode spatial proximity between electrode contacts. This allows adaptive feature learning that captures spatial dependencies expressed in adjacency matrices [34]. An edge convolution update can be represented by Equation 3, where hi(l) denotes the feature vector of node i at layer l, and hjl denotes the feature vector of a neighboring node j𝒩(i). The term Aij corresponds to the weighted adjacency value between nodes i and j, which represents edge features and encodes the spatial relationship between electrode contacts based on the distance-based graph construction. The function ϕ(·) is a learnable mapping implemented as a multilayer perceptron that combines node features and edge information to generate messages from neighboring nodes.

General graph convolutions then aggregate node features based on neighborhood information, enabling the model to learn feature dependencies across the graph structure, as shown in Equation 4 [35]. Here, H(l) represents the matrix of node features at layer l, 𝐀~=𝐀+𝐈 is the adjacency matrix with added self-loops, and 𝐃~ is the corresponding degree matrix. The matrix W(l) contains learnable weights, and σ(·) denotes a nonlinear activation function. This formulation enables the model to propagate and integrate information across the graph while preserving its underlying structure. Together, these convolutional layers ensure that both local and global structural properties of the graph input are effectively processed.

𝐡i(l+1)=j𝒩(i)ϕ(𝐡i(l),𝐡j(l),Aij), (3)
𝐇(l+1)=σ(𝐃~12𝐀~𝐃~12𝐇(l)𝐖(l)), (4)

In this model, an edge convolution layer with 32 units and a general graph covolution with 128 units were used, with an L2 regularizer of magnitude 5 × 10−3 to improve generalizability. The final feature representation undergoes global average pooling before being passed through fully connected layers with rectified linear unit (ReLU) activations. Finally, we use a softmax layer to output the classification probabilities. During training, a batch size of 1024 and a learning rate of 0.001 were used. A 20% dropout rate is applied to the dense layers to enhance regularization and reduce overfitting.

Prior to training, we applied data augmentation in the form of low-amplitude Gaussian noise, which serves as a physiologically meaningful perturbation for neural recordings. Gaussian noise injection acts as a regularizer by simulating naturally occurring variability in peripheral nerve signals—such as background neural activity, electrode impedance fluctuations, and minor recording noise—while preserving the underlying spatiotemporal structure of the compound action potentials. This approach helps improve model robustness and generalizability without altering the temporal dynamics or spatial relationships that are essential for accurate decoding.

Training & evaluation

Hyperparameter tuning.

We tuned the model hyperparameters using data from two randomly selected rats (e.g., Rats 2 and 10), which served as a validation set for hyperparameter optimization. The training hyperparameters, including the learning rate (1e-4 –1e-2), batch size (128–1024), and L2 regularization coefficient (1e-5 – 5e-3), were systematically tuned based on validation performance. In addition, architectural parameters such as the number of hidden units (128–512), dropout rates (0.1–0.8), and overall model layer configurations were also selected through the same validation-based tuning procedure to balance model capacity and generalization. Once the optimal hyperparameters were determined, these rats were reincorporated into the training and evaluation process to maximize data utilization. In addition, we systematically varied key graph construction parameters, such as the number of neighbors and the distance decay parameter (σ) in the adjacency matrix, to analyze their impact on model performance. Model development and training were carried out in Python using the TensorFlow deep learning framework.

Evaluation.

Model performance was evaluated using both cross-subject and within-subject strategies. For generalizability across rats, we employed an eight-fold leave-one-out cross-validation approach, where data from seven rats were used for training and the remaining rat was held out for testing. This design allowed us to rigorously assess across-subject generalization, addressing a limitation of prior CNN-based studies that focused only on within-subject performance [12]. Performance was quantified using test accuracy and the macro-averaged F1 score, which is more reliable for imbalanced datasets. To ensure comparability with previous work, we also conducted within-subject evaluations by performing a cross-validation in which two folds were used for training and the remaining fold was used for testing. This process was repeated such that each fold served as the test set once.

Ablation studies

To evaluate the contribution of specific architectural and structural design choices in our proposed framework, we also conducted three sets of ablation studies. These studies were aimed at disentangling the effects of (1) temporal modeling via LSTM and (2) graph construction based on geodesic distances. To ensure that the changes in performance were not solely attributable to the use of LSTM, we replaced the it with a 1D CNN module. This design choice also aligns more closely with the architecture used by the previous study [11,12], thereby enabling a more direct comparison with prior work. In addition to the proposed geodesic graph, we constructed graphs based on Euclidean distances between electrode contacts to evaluate whether preserving the true surface geometry of the nerve cuff provides advantages over simpler spatial proximity measures. Euclidean distances were computed directly from the two-dimensional spatial coordinates of the electrode contacts, without accounting for the curved surface geometry of the nerve cuff. Furthermore, we constructed a random graph baseline in which edges were assigned randomly between nodes. This setup preserved the graph’s sparsity while removing the physiological prior embedded in the geodesic topology, allowing us to observe the performance of the model under different spatial constraints.

Results

Generalizability performance

Table 2 presents the classification accuracies (%) and macro-averaged F1 score for each individual rat when used as a test set, as well as the mean and standard deviation across all rats. The first row shows the performance of the baseline CNN model reimplemented from the previous study [12] and the subsequent rows show the graph-based model proposed in this work. For the graph-based model, we compared how varying the number of neighbouring nodes changes the performance at a fixed σ of 2. This allows us to better examine the effects of neighbouring nodes in classification performances. For the graph-based models, we have grayed out the performance scores for Rats 2 and 10, as these subjects were used as a validation set during hyperparameter tuning. Accordingly, their results were excluded from the calculation of the mean ± standard deviation. The scores are nonetheless reported to provide additional context regarding their individual performance. Overall, the graph-based approaches resulted in an improved mean classification accuracy compared to the CNN model, which achieved 54.00 ± 5.21% (* beside the mean ± standard deviation indicates significant difference, p < 0.05 from the baseline model performance, computed using the t-test). The graph-based approach with connectivity defined by four or five neighbors showed a statistically significant improvement from the CNN model in terms of F1-score, and among the graph learning models evaluated, the graphs constructed with five neighbors was the highest-performing model, outperforming the baseline CNN by 14.32% in mean classification accuracy and 17.74% in F1 score. Fig 3 illustrates graph connectivity with five neighbors, which is the optimal connectivity found.

Table 2. Generalization performance (accuracy and macro-averaged F1 score) of baseline model and graph-based models with varying connectivity (Rats 2 and 10 are excluded from final mean ± standard deviation evaluation as they were used in the validation set for hyperparameter tuning).

Metrics Model Rat 2 Rat 4 Rat 5 Rat 6 Rat 7 Rat 8 Rat 9 Rat 10 Mean ± Standard Deviation
Accuracies CNN 48.00 53.57 51.57 47.70 58.41 61.82 50.93 43.96 54.00 ± 5.21
2 neighbors 54.63 57.32 54.07 77.88 52.53 60.81 43.22 51.38 57.64 ± 11.55
3 neighbors 53.49 58.71 67.49 69.78 62.51 58.92 54.14 50.45 61.93 ± 5.88
4 neighbors 49.43 60.91 68.25 73.14 60.48 63.92 58.95 51.02 64.28 ± 5.45*
5 neighbors 51.71 62.89 64.98 72.90 68.65 72.71 67.78 52.17 68.32 ± 4.03*
6 neighbors 52.30 60.08 62.88 68.43 65.05 56.73 68.49 53.92 63.61 ± 4.68
7 neighbors 50.58 58.99 66.42 64.33 56.30 69.27 37.59 49.77 58.82 ± 11.44
8 neighbors 50.58 40.09 53.66 59.55 48.28 69.27 43.74 49.84 52.43 ± 10.78
9 neighbors 38.75 55.33 55.64 66.06 30.90 56.57 41.28 49.87 50.96 ± 12.62
F1 Score CNN 35.99 44.67 47.22 46.55 47.09 46.44 51.76 33.34 47.29 ± 2.37
2 neighbors 40.38 54.45 51.77 65.43 40.27 54.50 29.29 29.69 49.29 ± 12.66
3 neighbors 36.08 58.62 61.40 59.82 60.80 55.52 50.77 26.91 57.82 ± 4.03*
4 neighbors 22.05 60.13 56.50 66.46 58.88 61.24 55.10 27.56 59.72 ± 4.01*
5 neighbors 38.18 60.27 62.54 63.98 65.56 72.64 65.19 33.63 65.03 ± 4.20*
6 neighbors 46.57 60.19 54.80 60.80 59.86 59.43 62.86 46.00 59.66 ± 2.67*
7 neighbors 40.00 49.30 61.33 57.10 48.99 60.83 27.13 23.62 50.78 ± 12.78
8 neighbors 41.16 35.74 53.08 53.86 36.10 70.46 27.67 24.57 46.15 ± 15.81
9 neighbors 39.05 55.53 55.07 49.66 29.95 56.86 23.06 24.90 45.02 ± 14.71

Fig 3. Graph connectivity with five neighbors, defined by geodesic distance.

Fig 3

To systematically analyze the effect of graph construction hyperparameters, we present a comprehensive heatmap of averaged F1-scores across all rats while varying the number of neighbors and σ (Fig 4). This visualization highlights how model performance changes with different parameter combinations and demonstrates that the chosen hyperparameters (σ of 2 and number of neighbors of 5) achieve the highest classification accuracy. We then performed ablation studies with these chosen hyperparameters to evaluate model generalizability, with results summarized in Table 3. Fig 5a and 5b further illustrate the graph connectivity constructed using Euclidean distance and random assignments, respectively. Compared with the geodesic-distance-based connectivity shown in Fig 3, these alternative graph structures exhibit remarkably different topologies.

Fig 4. Heatmap of F1-scores across different distance decay parameters and number of neighbors.

Fig 4

Table 3. Generalization performance (accuracy and macro-averaged F1 score) of ablation studies (Rats 2 and 10 are excluded from final mean ± standard deviation evaluation as they were used in the validation set for hyperparameter tuning).

Metrics Model Rat 2 Rat 4 Rat 5 Rat 6 Rat 7 Rat 8 Rat 9 Rat 10 Mean ± Standard Deviation
Accuracies Graph - 1D CNN 52.74 59.32 60.04 75.02 71.70 61.33 54.32 60.81 63.62 ± 7.98
Graph – Random 52.20 59.05 55.45 50.78 52.36 50.45 42.38 49.66 51.74 ± 5.62
Graph – Euclidean Distance 54.21 56.54 59.49 68.29 60.57 67.41 65.09 51.02 62.90 ± 4.72
F1 Score Graph - 1D CNN 45.87 57.10 51.81 68.67 67.01 60.11 49.41 58.51 59.02 ± 7.83
Graph – Random 35.14 56.26 54.87 26.03 37.22 22.33 24.66 22.85 36.89 ± 15.35
Graph – Euclidean Distance 41.36 56.87 56.00 51.69 50.36 64.32 58.9 31.67 56.36 ± 5.06

Fig 5. Comparison of graph connectivity definitions in ablation studies.

Fig 5

(a) Euclidean distance–based graph. (b) Randomly constructed graph.

Substituting the LSTM branch with a CNN did not provide significant performance gains, and our model still outperformed prior CNN-based work. Interestingly, when using a random graph in place of the correct adjacency matrix, the accuracy remained close to that of the CNN baseline. This observation likely reflects the limited generalization capability of the CNN baseline in the across-subject setting. CNNs rely on fixed, grid-based receptive fields and tend to learn subject-specific spatial patterns that do not transfer well across rats. As a result, both the CNN and the random-graph model lack an explicit inductive bias that enforces physiologically meaningful spatial relationships between electrodes, leading to similar generalization performance. In contrast, incorporating anatomically informed graph connectivity enables the GNN to leverage consistent spatial organization across subjects, resulting in substantially improved performance.

Within-rat performance

We then evaluated the within-rat performance to establish a direct comparison with the previous study (note that the results are slightly different as one rat is removed in this study) [12]. The accuracies and F1-scores from both the model proposed in the previous study, the model proposed in this study, as well as the ablation studies performed, are summarized in Table 4. The graph-based approach achieved a 1.92% improvement in accuracy and a 3.14% improvement in F1-score compared to the CNN baseline. Paired t-test analysis revealed that, although the geodesic-distance-based graph model consistently outperformed the CNN baseline, the observed improvements were not statistically significant (p > 0.05). In contrast, both the random graph and Euclidean-distance-based graph ablations resulted in significantly lower performance compared to the geodesic graph, indicating that preserving the physiologically meaningful geodesic structure is critical and represents the most effective choice for graph construction.

Table 4. Within-rat performance (accuracy and macro-averaged F1 score) of baseline model, graph-based models with optimal connectivity, as well as ablation studies.

Metrics Model Rat 2 Rat 4 Rat 5 Rat 6 Rat 7 Rat 8 Rat 9 Rat 10 Mean ± Standard Deviation
Accuracies CNN 59.89 76.31 76.14 84.38 87.00 86.82 92.17 64.40 78.76 ± 10.41
Graph 72.69 80.15 80.98 83.41 81.57 81.82 89.89 74.95 80.68 ± 5.66
Graph - 1D CNN 72.54 78.34 82.21 84.74 81.82 81.60 91.76 72.46 80.68 ± 6.36
Graph – Random 63.14 68.04 72.01 81.07 78.16 75.26 87.67 62.10 73.43 ± 8.88*
Graph – Euclidean Distance 65.29 70.58 75.33 81.87 72.28 73.31 87.29 66.41 74.04 ± 7.45*
F1 Score CNN 48.73 76.69 75.32 75.77 84.99 85.42 88.70 59.25 74.36 ± 13.53
Graph 72.39 80.06 80.29 72.86 76.13 79.96 86.33 72.21 77.50 ± 5.13
Graph - 1D CNN 71.96 76.50 81.52 76.26 76.97 79.97 87.86 65.34 77.04 ± 6.65
Graph – Random 44.94 67.84 72.28 59.51 69.40 68.44 83.23 54.11 64.97 ± 11.81*
Graph – Euclidean Distance 49.17 70.24 73.94 68.65 54.21 73.03 77.57 57.36 65.52 ± 10.46*

Discussion

The proposed approach in this work showed the significance and effectiveness of using graphs to encode information from neural recordings.

The results demonstrated that the proposed graph-based learning approach outperforms conventional CNNs in classifying peripheral neural signals, both in across-subject generalization and within-subject evaluations. Compared to the reimplemented CNN baseline from [12], the graph-based model achieved substantial improvements in mean accuracy and F1 score (Tables 2 and 4), highlighting the advantage of explicitly incorporating spatial relationships between electrodes into the learning process. These gains were consistent across most test subjects, indicating robustness to subject, specific variability, a key challenge in neural decoding. It should be noted that the reported performance differences also reflect the exclusion of one rat from the analysis, which accounts for slight deviations from previously reported values [12].

Ablation studies (Tables 3 and 4) confirm that the performance advantage is not solely attributable to the LSTM module, as replacing it with a 1D CNN yielded comparable results. In contrast, removing the geodesic-distance-based graph construction and replacing it with random connectivity led to a sharp performance drop, underscoring the importance of physiologically meaningful graph structures. Using Euclidean-distance-based graphs also resulted in a noticeable reduction in performance compared to geodesic connectivity, although it consistently outperformed the random graph baseline. This pattern suggests that incorporating spatial proximity alone is beneficial, but Euclidean distances do not reflect the meaningful geometry of the nerve cuff, as they ignore the circumferential arrangement of electrodes and the way neural signals propagate along the nerve surface. In contrast, geodesic distances capture both the physical layout and the physiologically relevant pathways through which activity spreads across adjacent contacts, leading to superior performance. By modeling electrode positions using geodesic distances, the graph-based approach leverages both local and global spatial dependencies, enabling more informative spatiotemporal feature extraction than CNNs, which primarily capture local spatial patterns.

Our analysis of connectivity hyperparameters reveals that graphs constructed with more than three neighbors outperformed those with fewer, with peak performance achieved at five neighbors and a σ of 2. This configuration appears to optimally balance local specificity and global context. The distance decay parameter plays a key role in modulating the model’s sensitivity to spatial distance: higher σ reduces the decay of edge weights with distance, allowing for more global integration when many neighbors are retained, while lower σ enforces stronger locality. This interaction suggests that careful tuning of both parameters can maximize the model’s ability to capture meaningful spatial relationships while avoiding overfitting to noise.

From a physiological perspective, the geodesic-based connectivity preserves the true spatial organization of electrodes on the nerve cuff, reflecting how neural signals propagate through the peripheral nervous system. Electrodes positioned close together are more likely to record correlated activity, while distant contacts often provide complementary, non-redundant information. Capturing these relationships is crucial for improving model performance and interpretability.

Overall, graph-based approaches not only outperformed traditional methods such as CNNs in classification performance but also have great potential in clinical explainability. One of the advantages of GNNs is their ability to provide deeper understanding into the model’s decision-making process. For instance, future studies may examine the learned weights in the GNN layers to identify the most influential nodes (e.g., electrodes) and edges (e.g., connections) that contribute to the final classification decision. By investigating node importance, we can determine which specific electrode channels contributed more in distinguishing different neural patterns, providing valuable information about the distribution of neural activity. Similarly, analyzing edge importance can reveal how different electrodes interact and contribute to temporal dynamics, helping us understand how neural signals propagate across the nerve. Additionally, the graph-based approach has proven to be more generalizable, allowing for the incorporation of data from multiple subjects rather than relying solely on a single animal model for training. This ensures consistent model performance across different subjects and enhances real-world applicability.

Future studies should explore different strategies for defining graph connectivity, such as adaptive thresholding based on statistical dependencies between channels or dynamic graphs that incorporate learnable edges. Subgraph-based approaches could be investigated to focus on the most informative nodes and edges, potentially improving computational efficiency and model interpretability [36]. Additionally, self-supervised learning techniques, such as contrastive learning or graph autoencoders, could be explored to enhance the ability of the model to extract meaningful representations without relying on a large amount of labeled data. Prior studies have shown that graph-based approaches are particularly effective in low-data regimes, as they leverage relational inductive biases to learn richer representations compared to grid-based methods [37,38]. Our findings align with this evidence, suggesting that integrating self-supervised objectives could further strengthen performance under limited data availability. A notable limitation of the present study is the relatively small dataset, consisting of recordings from only eight rats. Although statistical testing demonstrated significant performance differences between the graph-based models and the CNN baseline, the small sample size inherently limits the strength of these conclusions. This constraint reflects the practical challenges of collecting nCAPs in vivo, an experimental process that is resource-intensive and time-consuming. As larger datasets become available, future work should evaluate graph-based methods on more extensive cohorts to further validate generalizability and robustness.

While this study focuses on rat peripheral nerve recordings, the proposed graph-based framework is well positioned for translation to human applications. Human peripheral neural signals are characterized by greater anatomical variability, differences in nerve size, and increased signal heterogeneity arising from subject-specific physiology, electrode placement, and clinical noise sources. By explicitly modeling inter-electrode relationships rather than relying on fixed grid-based assumptions, the graph formulation provides a flexible representation that can naturally adapt to these sources of variability.

Conclusion

We explored a graph-based approach for classifying three afferent activities using neural recordings obtained from Long-Evan rats with a 56-channel nerve cuff electrode. The GNN model effectively captured temporal patterns through nodal features, and by constructing weighted graph adjacency matrices with geodesic distance, the model was able to extract more informative and temporal and spatial features, outperforming the CNN model. Under the leave-one-out strategy, the GNN model outperformed the CNN model by 14.32% in mean classification accuracy, and by 17.74% in macro-averaged F1 score, and in the within-rat evaluation, the proposed model in this study achieved an improvement of 1.92% in accuracy and 3.14% in F1 score. In summary, our findings highlight the potential of graph-based models for decoding neural signals by effectively utilizing both temporal and spatial relationships. A particularly promising future direction is the interpretability of learned graph representations, which may provide insight into how individual electrodes and their interactions contribute to classification decisions. Analyzing node- and edge-level importance within the learned graphs could help reveal physiologically meaningful patterns of neural activity and improve transparency for neuroscience and clinical applications. Future work in this area can further refine graph-based approaches to enhance classification accuracy and interpretability, and expand the applicability of these models to other neural decoding tasks.

Data Availability

The data used in this study were from a publicly available dataset available at Borealis, U of T Dataverse: https://doi.org/10.5683/SP3/JRZDDR.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Vizi ES, Kiss JP, Lendvai B. Nonsynaptic communication in the central nervous system. Neurochem Int. 2004;45(4):443–51. doi: 10.1016/j.neuint.2003.11.016 [DOI] [PubMed] [Google Scholar]
  • 2.Oosting PH. Signal transmission in the nervous system. Rep Prog Phys. 1979;42(9):1479–532. doi: 10.1088/0034-4885/42/9/001 [DOI] [Google Scholar]
  • 3.Hildebrand JG. Analysis of chemical signals by nervous systems. Proc Natl Acad Sci U S A. 1995;92(1):67–74. doi: 10.1073/pnas.92.1.67 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Kamimura D, Tanaka Y, Hasebe R, Murakami M. Bidirectional communication between neural and immune systems. Int Immunol. 2020;32(11):693–701. doi: 10.1093/intimm/dxz083 [DOI] [PubMed] [Google Scholar]
  • 5.Townsend KL. One Nervous System: Critical Links Between Central and Peripheral Nervous System Health and Implications for Obesity and Diabetes. Diabetes. 2024;73(12):1967–75. doi: 10.2337/dbi24-0004 [DOI] [PubMed] [Google Scholar]
  • 6.Thakor NV, Wang Q, Greenwald E. Bidirectional peripheral nerve interface and applications. Annu Int Conf IEEE Eng Med Biol Soc. 2016;2016:6327–30. doi: 10.1109/EMBC.2016.7592175 [DOI] [PubMed] [Google Scholar]
  • 7.Petrini FM. Interfacing the peripheral nervous system: towards the development of a bidirectional neural communication. 2015.
  • 8.Varho T, Jääskeläinen S, Tolonen U, Sonninen P, Vainionpää L, Aula P, et al. Central and peripheral nervous system dysfunction in the clinical variation of Salla disease. Neurology. 2000;55(1):99–104. doi: 10.1212/wnl.55.1.99 [DOI] [PubMed] [Google Scholar]
  • 9.Koh RGL, Zariffa J, Jabban L, Yen S-C, Donaldson N, Metcalfe BW. Tutorial: a guide to techniques for analysing recordings from the peripheral nervous system. J Neural Eng. 2022;19(4):10.1088/1741-2552/ac7d74. doi: 10.1088/1741-2552/ac7d74 [DOI] [PubMed] [Google Scholar]
  • 10.Larson CE, Meng E. A review for the peripheral nerve interface designer. J Neurosci Methods. 2020;332:108523. doi: 10.1016/j.jneumeth.2019.108523 [DOI] [PubMed] [Google Scholar]
  • 11.Koh RGL, Nachman AI, Zariffa J. Classification of naturally evoked compound action potentials in peripheral nerve spatiotemporal recordings. Sci Rep. 2019;9(1):11145. doi: 10.1038/s41598-019-47450-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Koh RGL, Balas M, Nachman AI, Zariffa J. Selective peripheral nerve recordings from nerve cuff electrodes using convolutional neural networks. J Neural Eng. 2020;17(1):016042. doi: 10.1088/1741-2552/ab4ac4 [DOI] [PubMed] [Google Scholar]
  • 13.Dousty M, Fleet DJ, Zariffa J. Personalized Video-Based Hand Taxonomy Using Egocentric Video in the Wild. IEEE J Biomed Health Inform. 2025;29(9):6214–25. doi: 10.1109/JBHI.2024.3495699 [DOI] [PubMed] [Google Scholar]
  • 14.Dousty M, Fleet DJ, Zariffa J. Hand Grasp Classification in Egocentric Video After Cervical Spinal Cord Injury. IEEE J Biomed Health Inform. 2024;28(2):645–54. doi: 10.1109/JBHI.2023.3269692 [DOI] [PubMed] [Google Scholar]
  • 15.Riek NT, Akcakaya M, Bouzid Z, Gokhale T, Helman SM, Kraevsky K, et al. ECG-SMART-NET: A Deep Learning Architecture for Precise ECG Diagnosis of Occlusion Myocardial Infarction. IEEE Trans Biomed Eng. 2025;72(12):3613–20. doi: 10.1109/TBME.2025.3573581 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Somani S, Russak AJ, Richter F, Zhao S, Vaid A, Chaudhry F, et al. Deep learning and the electrocardiogram: review of the current state-of-the-art. Europace. 2021;23(8):1179–91. doi: 10.1093/europace/euaa377 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Mueller MJ, Minor SD, Schaaf JA, Strube MJ, Sahrmann SA. Relationship of plantar-flexor peak torque and dorsiflexion range of motion to kinetic variables during walking. Phys Ther. 1995;75(8):684–93. doi: 10.1093/ptj/75.8.684 [DOI] [PubMed] [Google Scholar]
  • 18.Micera S, Navarro X. Bidirectional interfaces with the peripheral nervous system. Int Rev Neurobiol. 2009;86:23–38. doi: 10.1016/S0074-7742(09)86002-9 [DOI] [PubMed] [Google Scholar]
  • 19.Romero DW, Knigge DM, Gu A, Bekkers EJ, Gavves E, Tomczak JM. Towards a general purpose CNN for long range dependencies in N D. 2022. https://doi.org/arXiv:220603398
  • 20.Nerella S, Bandyopadhyay S, Zhang J, Contreras M, Siegel S, Bumin A, et al. Transformers and large language models in healthcare: A review. Artif Intell Med. 2024;154:102900. doi: 10.1016/j.artmed.2024.102900 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Lin T, Wang Y, Liu X, Qiu X. A survey of transformers. AI Open. 2022;3:111–32. doi: 10.1016/j.aiopen.2022.10.001 [DOI] [Google Scholar]
  • 22.Zhou J, Cui G, Hu S, Zhang Z, Yang C, Liu Z, et al. Graph neural networks: A review of methods and applications. AI Open. 2020;1:57–81. doi: 10.1016/j.aiopen.2021.01.001 [DOI] [Google Scholar]
  • 23.Demir A, Koike-Akino T, Wang Y, Haruna M, Erdogmus D. EEG-GNN: Graph Neural Networks for Classification of Electroencephalogram (EEG) Signals. Annu Int Conf IEEE Eng Med Biol Soc. 2021;2021:1061–7. doi: 10.1109/EMBC46164.2021.9630194 [DOI] [PubMed] [Google Scholar]
  • 24.Tang S, Dunnmon JA, Saab K, Zhang X, Huang Q, Dubost F. Self-supervised graph neural networks for improved electroencephalographic seizure analysis. arXiv preprint. 2021. 10.48550/arXiv.2104.08336 [DOI]
  • 25.Li R, Yuan X, Radfar M, Marendy P, Ni W, O’Brien TJ, et al. Graph Signal Processing, Graph Neural Network and Graph Learning on Biological Data: A Systematic Review. IEEE Rev Biomed Eng. 2023;16:109–35. doi: 10.1109/RBME.2021.3122522 [DOI] [PubMed] [Google Scholar]
  • 26.Atoar Rahman SM, Ibrahim Khalil M, Zhou H, Guo Y, Ding Z, Gao X, et al. Advancement in Graph Neural Networks for EEG Signal Analysis and Application: A Review. IEEE Access. 2025;13:50167–87. doi: 10.1109/access.2025.3549120 [DOI] [Google Scholar]
  • 27.Tang S, Dunnmon JA, Liangqiong Q, Saab KK, Baykaner T, Lee-Messer C. Modeling multivariate biosignals with graph neural networks and structured state space models. In: Conference on health, inference, and learning, 2023. 50–71. [Google Scholar]
  • 28.Wang Z, Wang Y, Zhang J, Hu C, Yin Z, Song Y. Spatial-Temporal Feature Fusion Neural Network for EEG-Based Emotion Recognition. IEEE Trans Instrum Meas. 2022;71:1–12. doi: 10.1109/tim.2022.3165280 [DOI] [Google Scholar]
  • 29.Koh RGL, Nachman AI, Zariffa J. Use of spatiotemporal templates for pathway discrimination in peripheral nerve recordings: a simulation study. J Neural Eng. 2017;14(1):016013. doi: 10.1088/1741-2552/14/1/016013 [DOI] [PubMed] [Google Scholar]
  • 30.Hamilton WL. Graph representation learning. Morgan & Claypool Publishers. 2020. [Google Scholar]
  • 31.Kovalenko A, Pozdnyakov V, Makarov I. Graph Neural Networks With Trainable Adjacency Matrices for Fault Diagnosis on Multivariate Sensor Data. IEEE Access. 2024;12:152860–72. doi: 10.1109/access.2024.3481331 [DOI] [Google Scholar]
  • 32.Whiteley N, Gray A, Rubin-Delanchy P. Matrix factorisation and the interpretation of geodesic distance. Advances in Neural Information Processing Systems. 2021;34:24–38. [Google Scholar]
  • 33.Agarwal R, Aziz A, Krishnan AS, Challa A, Danda S. ESW Edge Weights: Ensemble Stochastic Watershed Edge Weights for Hyperspectral Image Classification. IEEE Geosci Remote Sensing Lett. 2022;19:1–5. doi: 10.1109/lgrs.2022.3173793 [DOI] [Google Scholar]
  • 34.Coupeau P, Fasquel J-B, Dinomais M. On the relevance of edge-conditioned convolution for GNN-based semantic image segmentation using spatial relationships. In: 2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA), 2022. 1–6. 10.1109/ipta54936.2022.9784143 [DOI] [Google Scholar]
  • 35.Gama F, Marques AG, Leus G, Ribeiro A. Convolutional Graph Neural Networks. In: 2019 53rd Asilomar Conference on Signals, Systems, and Computers, 2019. 452–6. 10.1109/ieeeconf44664.2019.9048767 [DOI] [Google Scholar]
  • 36.Zhao L, Jin W, Akoglu L, Shah N. From stars to subgraphs: uplifting any GNN with local structure awareness. In: 2021. 10.48550/arXiv.2110.03753 [DOI]
  • 37.Pappu A, Paige B. Making graph neural networks worth it for low-data molecular machine learning. arXiv preprint. 2020. 10.48550/arXiv.2011.12203 [DOI]
  • 38.Agarwal S, Dubey T, Gupta S, Bedathur S. A transfer framework for enhancing temporal graph learning in data-scarce settings. In: 2025. https://arxiv.org/abs/250300852

Decision Letter 0

Luca Citi

30 Nov 2025

Dear Dr. Ji,

  • Data leakage in hyperparameter tuning. The authors report that hyperparameter optimisation (e.g., σ, k) was performed on data later used for evaluation, risking inflated performance claims. Possibly due to a lack of critical details (see next point), it is hard to gauge to what extent this poses a risk of data leakage. On one hand a clear separation between training, model selection / parameter tuning and test datasets is a pillar of any proper model evaluation procedure. On the other hand, the authors provide a heatmap of hyperparameter performance (Figure 4) and the initial tuning step might have been used only to identify a reasonable parameter values from which to explore other combinations. It is essential that the authors clarify this point unequivocally and convincingly. Please describe the training/evaluation split and procedure step-by-step.

  • Insufficient method details. Critical details (e.g., edge feature definitions, exact GNN formulations, code availability) are omitted, hindering reproducibility. Clarify role of "k". Equations and pseudocode should be added where needed. Consider sharing code unless not doing so is justifiable.

  • Inappropriate data augmentation. The listed augmentations are intended for image data and not meaningful for spatio-temporal neural signals. Clarify the rationale for using them. Also confirm unequivocally that any data augmentation is performed within CV, on the training data only, and not on the pooled data prior to splitting.

  • Weak ablation design. No baseline that uses inter‑channel correlation matrices is included (e.g. a baseline graph whose edge weights are Euclidean distances). Thus the ablation does not fully isolate the benefit of the geodesic geometry.

  • Statistical significance. Are these differences observed statistically significant? If so, provide details of test used. The overlapping standard deviations and large variances make it difficult to claim a statistically meaningful difference.

  • Generalisation and Clinical Relevance. Discuss how the method generalises from rat to human recordings, considering signal variability and electrode configurations (e.g., larger cuffs).

Please submit your revised manuscript by Jan 14 2026 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols....

We look forward to receiving your revised manuscript.

Kind regards,

Luca Citi, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS One has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. Please update your submission to use the PLOS LaTeX template. The template and more information on our requirements for LaTeX submissions can be found at http://journals.plos.org/plosone/s/latex.

4. We note you have included a table to which you do not refer in the text of your manuscript. Please ensure that you refer to Table 1 in your text; if accepted, production will need this reference to link the reader to the Table.

5. If the reviewer comments include a recommendation to cite specific previously published works, please review and evaluate these publications to determine whether they are relevant and should be cited. There is no requirement to cite these works unless the editor has indicated otherwise.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

Reviewer #1: Yes

Reviewer #2: No

Reviewer #3: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously? -->?>

Reviewer #1: Yes

Reviewer #2: No

Reviewer #3: No

**********

3. Have the authors made all data underlying the findings in their manuscript fully available??>

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.-->

Reviewer #1: Yes

Reviewer #2: No

Reviewer #3: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

Reviewer #1: The paper presents a solid study that applies Graph Neural Networks to classify peripheral neural recordings, integrating the geometry of the cuff electrode into the graph construction. Below, I provide some observations to improve the quality of the manuscript, listed in order of importance.

Strong:

The state of the art should be further developed. A good comparison is made with CNN, based on the results of Koh et al., but Transformers are only mentioned without a direct comparison or concrete results. Moreover, it would be helpful to include other models often used in the literature for peripheral neural recordings analysis, such as LSTM, Inception Time, or simpler approaches like SVM, for a more complete comparison and to better contextualize the results.

In the final discussion, it would be useful to add an in-depth analysis of the generalization of the approach from rat to human, considering the differences in peripheral neural signals. The analysis could explore how the methodology can be adapted to clinical settings, where signal variability is greater. Furthermore, it would be interesting to discuss how the graph structure may need to be modified based on different electrode configurations, such as cuffs with more channels or devices for larger nerve sections, to assess the robustness of the model in real clinical scenarios.

It would be helpful to add a discussion on the interpretability of the model, to help neuroscience experts better understand how the model makes decisions. A possible future development (if not already covered) could be the analysis of the connection weights in the graphs to explore the interactions between electrodes, or how the importance of individual connections contributes to the classification results. If not addressed in the paper, it might be interesting to mention this in the conclusions as a future direction.

Minor:

It would be useful to add context on the nerve interface and include an explanatory figure of the cuff electrode to facilitate understanding, especially for less experienced readers.

The results are presented in tables, but it would be more elegant and immediate to use graphs, which could make the differences between the models and various experiments more evident.

I suggest split/shortening some long sentences to make reading easier. Some punctuation errors (for example, missing puntis “.” at the end of the figures), a general check would be usefull.

Reviewer #2: In the manuscript entitled “Enhancing Generalizability in Classification of Peripheral Neural Recordings with Graph Neural Network,” the authors propose a graph-based learning framework for classifying peripheral nerve recordings. The topic is timely and relevant to the application of graph neural networks in analysis of neural signals. Nevertheless, I have several major concerns regarding the methodological design, analysis, and presentation that limit the strength of the conclusions. My comments are organised below into major and minor points:

Major

1.Data leakage in hyperparameter tuning

Two randomly selected rats are used for hyperparameter optimisation and then reincorporated into training and evaluation. This compromises the independence of test data and likely inflates reported results. These subjects should be excluded from final evaluation, or the authors should employ nested cross-validation to avoid overlap. Furthermore, it is unclear which parameters were actually tuned, as architectural settings were fixed while only σ and k were “systematically varied.”

2. Questionable performance claims

The claim that the proposed method “outperforms the CNN” is not well supported. Table 2 shows that only a subset of configurations exceed CNN accuracy or F1-score, and significance is claimed without clear justification. Reported improvements may reflect overfitting from hyperparameter sweeps rather than genuine generalisation gains. Statistical testing on n = 8 subjects is underpowered and should not be used to claim significance.

3.Within-subject results not statistically meaningful

In Table 4, within-rat results exhibit large standard deviations and overlapping confidence intervals across methods. Differences are likely random variation and should not be interpreted as real improvements.

4. Weak ablation design

The ablation comparing geodesic graphs to random graphs is insufficient to support claims about the value of geometric priors. Stronger baselines—e.g., graphs built from inter-channel correlation matrices—are needed to determine whether improvements truly stem from spatial geometry or merely from introducing structured connectivity.

5. Inappropriate data augmentation

The listed augmentations (brightness, contrast, saturation, hue perturbation) are intended for image data and not meaningful for spatio-temporal neural signals. Their use could distort physiological relationships. The authors should justify these operations or remove them entirely.

6.Unclear architectural and training hyperparameters

Hidden-unit sizes, L2 regularisation, batch size, learning rate, and dropout rate are fixed without explanation. The basis for these choices must be stated, or the parameters should be included in the tuning procedure.

7.Insufficient methodological clarity

The paper must specify the exact formulations of the edge convolution and general graph convolution layers, including equations and citations of the specific variants used. The authors should explicitly state what constitutes edge features (e.g., edge weights or other quantities).

Minor

- Clarify dataset composition and class balance; report number of samples per class

- The description of “naturally evoked compound action potentials (nCAPs)” should be expanded for a general audience. In Equation (1), variables x and y should be explicitly defined as electrode coordinates on the cuff.

- Figures and table captions should be more detailed to allow easy understanding of what’s exactly being shown

- Specify what software was used for reproducibility

Reviewer #3: Main comments:

This manuscript is generally well written and the idea of incorporating electrode geometric distances on the nerve cuff into decoding via a GNN is sensible. However, critical methodological details are missing, and some claims are not convincingly supported because of flaws in the experimental design.

1. Insufficient method details. Figure 2 describes the proposed method at a high level, but crucial details are omitted. Although EdgeConv and GeneralConv exist in literature, the authors should provide mathematical formulations (at least in the supplement) and ideally share code so readers can reproduce the work. Without these specifics, it is difficult to determine the source of the reported improvements for the GNN relative to other methods.

2. Unclear role of k. The authors state, “To refine the graph topology, we use the σ value and number of nearest neighbours k as tunable hyperparameters.” Since the adjacency matrix A is determined by σ, the role of k is ambiguous. Given that k strongly affects results, please explain precisely how k is used to construct A (pseudocode or equations in the supplement would help).

3. Inappropriate baseline in ablation. In the ablation studies (Sec. 2.5) the authors use a random graph as a baseline. To support the claim that GNNs improve over CNNs, a more appropriate baseline is a graph whose edge weights are the Euclidean distances between electrodes in the 2D plane (instead in the cylinder). This better approximates the local receptive-field scanning performed by CNNs. Please add this comparison.

4. Unexpected random-graph result requires explanation. The authors report that “Interestingly, when using a random graph in place of the correct adjacency matrix, the accuracy remained close to that of the CNN baseline. This suggests…”. This requires a deeper explanation, since it implies that CNN weights may be unimportant—counterintuitive to expectations (random < CNN < GNN). Please analyze and discuss potential reasons.

Minor comments:

1. Abstract claim on small-data regimes. The abstract claims benefits “particularly in small-data regimes where models cannot reliably learn such relationships solely from the data.” I do not see experiments testing varying dataset sizes. If this claim is retained, provide supporting experiments or tone down the claim.

2. Figure 3 / neighborhood k. In Figure 3 the reported neighborhood size k = 5 is not visually consistent (e.g., electrode 20 appears connected to only four neighbors). Please correct the figure or clarify the visualization.

3. Within-subject cross-validation details. In Sec. 2.4.2 you state “two folds for training” for within-subject evaluations. Please describe the training/evaluation split and procedure step-by-step in the supplement (fold definitions, random seeds, how validation/test sets were selected, etc.).

4. Inconsistent CNN accuracy values. In Sec. 3.1 CNN accuracy is reported as 52.00 ± 5.17%, while it’s 52.00 ± 5.86% in Table 3. Please reconcile and explain any differences.

5. Misleading axis label “temperature.” In Figure 4 the x-axis labeled “temperature” appears to represent σ. Rename the axis to σ or give it a meaningful term consistent with the graph-construction procedure. Replace “temperature” throughout the manuscript.

6. Statistical significance. In Sec. 3.2 the authors report “a 1.92% improvement in accuracy and a 3.14% improvement in F1-score compared to the CNN baseline.” Are these differences statistically significant? Are the differences in Table 4 significant? Please report appropriate statistical tests (e.g., paired tests across folds/datasets) and include p-values or confidence intervals.

**********

what does this mean?). If published, this will include your full peer review and any attached files.). If published, this will include your full peer review and any attached files.). If published, this will include your full peer review and any attached files.). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our For information about this choice, including consent withdrawal, please see our For information about this choice, including consent withdrawal, please see our For information about this choice, including consent withdrawal, please see our Privacy Policy..-->

Reviewer #1: No

Reviewer #2: Yes: Dominik KleplDominik KleplDominik KleplDominik Klepl

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

To ensure your figures meet our technical requirements, please review our figure guidelines: https://journals.plos.org/plosone/s/figures

You may also use PLOS’s free figure tool, NAAS, to help you prepare publication quality figures: https://journals.plos.org/plosone/s/figures#loc-tools-for-figure-preparation.

NAAS will assess whether your figures meet our technical requirements by comparing each figure against our figure specifications.

PLoS One. 2026 Apr 17;21(4):e0345204. doi: 10.1371/journal.pone.0345204.r002

Author response to Decision Letter 1


15 Jan 2026

We sincerely thank the editor for the opportunity to revise our manuscript and for recognizing the importance of our work, and greatly appreciate the constructive feedback provided. We have considered each comment and have made revisions to the manuscript to address all concerns raised. Responses to each reviewer comment are provided in blue text throughout this letter.

Editor’s comments:

1. Data leakage in hyperparameter tuning. The authors report that hyperparameter optimisation (e.g., σ, k) was performed on data later used for evaluation, risking inflated performance claims. Possibly due to a lack of critical details (see next point), it is hard to gauge to what extent this poses a risk of data leakage. On one hand a clear separation between training, model selection / parameter tuning and test datasets is a pillar of any proper model evaluation procedure. On the other hand, the authors provide a heatmap of hyperparameter performance (Figure 4) and the initial tuning step might have been used only to identify a reasonable parameter values from which to explore other combinations. It is essential that the authors clarify this point unequivocally and convincingly. Please describe the training/evaluation split and procedure step-by-step.

We thank the editor for raising this point. In the revised manuscript, we have fully addressed the concern regarding hyperparameter tuning and data leakage. Specifically, the two rats used for hyperparameter optimization (Rats 2 and 10) are now excluded from all final performance evaluations in Tables 2 and 3, ensuring a strict separation between validation and test data. We also clarify exactly which hyperparameters were tuned, training parameters (learning rate, batch size, L2 penalty), architectural parameters (hidden units, dropout), as well as graph construction parameters (number of neighbours and distance-scaling parameter) in Methods, subsection Training & Evaluation, Hyperparameter Tuning (pages 6-7). To clarify further, the heatmap in Figure 4 is provided specifically to illustrate how different graph topologies influence performance. The purpose of this analysis is to demonstrate the importance of graph topology in problems where physiologically informed data structure matters, and the variation across σ and k confirms the critical role of incorporating meaningful nerve geometry.

For a more detailed training/evaluation procedure, we have shared our graph construction and model training code on GitHub: https://github.com/ruiqi1124/graph_nerve_recording.git

2. Insufficient method details. Critical details (e.g., edge feature definitions, exact GNN formulations, code availability) are omitted, hindering reproducibility. Clarify role of "k". Equations and pseudocode should be added where needed. Consider sharing code unless not doing so is justifiable.

We thank the editor for raising this point, we have now expanded our Methods, subsection Model Architecture (pages 5 – 6) to elaborate on the mathematical formulations of the GNN layers.

3. Inappropriate data augmentation. The listed augmentations are intended for image data and not meaningful for spatio-temporal neural signals. Clarify the rationale for using them. Also confirm unequivocally that any data augmentation is performed within CV, on the training data only, and not on the pooled data prior to splitting.

We have now clarified this in the manuscript, in Methods, subsection Model Architecture (last paragraph, page 6). We also confirm that any data augmentation was only performed during model training, on the training set.

4. Weak ablation design. No baseline that uses inter channel correlation matrices is included (e.g. a baseline graph whose edge weights are Euclidean distances). Thus the ablation does not fully isolate the benefit of the geodesic geometry.

We have included a third ablation study, using graphs constructed with Euclidean distances. The description has been added to Methods, subsection Ablation Studies (pages 7-8), and the results are presented in Tables 3 and 4. More relevant discussion has also been added in the Discussion section (page 11).

5. Statistical significance. Are these differences observed statistically significant? If so, provide details of test used. The overlapping standard deviations and large variances make it difficult to claim a statistically meaningful difference.

Paired t-tests were used in comparing results in Tables 2 and 3, and any statistical significance is commented on in the Results section (subsection Generalizability Performance, page 8 for Table 2, and subsection Within-Rat Performance, page 10 for Table 3).

6. Generalisation and Clinical Relevance. Discuss how the method generalises from rat to human recordings, considering signal variability and electrode configurations (e.g., larger cuffs).

We have added a paragraph on this in the Discussion section (last paragraph, page 12).

Reviewers’ Comments

We sincerely thank the reviewers for their time, thoughtful feedback, and careful evaluation of the initial version of our manuscript. We have carefully considered the comments and have revised the manuscript accordingly and added further details and clarifications to address the concerns raised.

Reviewer #1: The paper presents a solid study that applies Graph Neural Networks to classify peripheral neural recordings, integrating the geometry of the cuff electrode into the graph construction. Below, I provide some observations to improve the quality of the manuscript, listed in order of importance.

We thank the reviewer for taking the time to provide us with constructive feedback. Additional explanations were added to relevant sections as emphasized below.

Strong:

1. The state of the art should be further developed. A good comparison is made with CNN, based on the results of Koh et al., but Transformers are only mentioned without a direct comparison or concrete results. Moreover, it would be helpful to include other models often used in the literature for peripheral neural recordings analysis, such as LSTM, Inception Time, or simpler approaches like SVM, for a more complete comparison and to better contextualize the results.

We revised the Introduction section to provide a broader overview of existing methods (third paragraph, page 2). We expanded the comparison to explicitly discuss other deep learning techniques and machine learning classifiers for peripheral nerve decoding. We still believe that CNN is the most meaningful comparison with graph-based models, as no other models can encode the spatial relationship, and the geometric topology of the nerve cuff. We have highlighted this point even more in the introduction, and added a new citation 20 to support our point. We also added rationale for why these approaches cannot encode cuff geometry.

Additionally, we have clarified in the Introduction section that CNN and LSTM are indeed compared, in ablation studies.

2. In the final discussion, it would be useful to add an in-depth analysis of the generalization of the approach from rat to human, considering the differences in peripheral neural signals. The analysis could explore how the methodology can be adapted to clinical settings, where signal variability is greater. Furthermore, it would be interesting to discuss how the graph structure may need to be modified based on different electrode configurations, such as cuffs with more channels or devices for larger nerve sections, to assess the robustness of the model in real clinical scenarios.

We added a new paragraph in the Discussion section explicitly analyzing translation from rat to human recordings (last paragraph, page 12). The revised text discusses differences in fascicle size, nerve diameter, and signal variability, and addresses how graph structure may change for cuffs with different channel densities or geometries. We also outline how the approach can adapt to clinical nerve-interface designs.

3. It would be helpful to add a discussion on the interpretability of the model, to help neuroscience experts better understand how the model makes decisions. A possible future development (if not already covered) could be the analysis of the connection weights in the graphs to explore the interactions between electrodes, or how the importance of individual connections contributes to the classification results. If not addressed in the paper, it might be interesting to mention this in the conclusions as a future direction.

We thank the reviewer for mentioning this important direction for future work. We have elaborated on this in the third last paragraph in the Discussion section (page 11).

Minor:

It would be useful to add context on the nerve interface and include an explanatory figure of the cuff electrode to facilitate understanding, especially for less experienced readers.

The results are presented in tables, but it would be more elegant and immediate to use graphs, which could make the differences between the models and various experiments more evident.

In response to the points above, we have added figures illustration graph connectivity using geodesic distance, Euclidean distance, and random graph construction (Figures 3, page 7, and 5, page 10). These provide more context for the models described in the tables. The geodesic distance-based graph is the main proposed model of this study, while the other ones are part of ablation studies.

I suggest split/shortening some long sentences to make reading easier. Some punctuation errors (for example, missing puntis “.” at the end of the figures), a general check would be useful.

We thank the reviewers for pointing this out. We have revised this accordingly.

Reviewer #2: In the manuscript entitled “Enhancing Generalizability in Classification of Peripheral Neural Recordings with Graph Neural Network,” the authors propose a graph-based learning framework for classifying peripheral nerve recordings. The topic is timely and relevant to the application of graph neural networks in analysis of neural signals. Nevertheless, I have several major concerns regarding the methodological design, analysis, and presentation that limit the strength of the conclusions. My comments are organised below into major and minor points:

We thank the reviewer for taking the time to provide us with constructive feedback. Additional explanations were added to relevant sections as emphasized below.

Major

1.Data leakage in hyperparameter tuning

Two randomly selected rats are used for hyperparameter optimisation and then reincorporated into training and evaluation. This compromises the independence of test data and likely inflates reported results. These subjects should be excluded from final evaluation, or the authors should employ nested cross-validation to avoid overlap. Furthermore, it is unclear which parameters were actually tuned, as architectural settings were fixed while only σ and k were “systematically varied.”

We thank the reviewer for requesting this clarification and agree that it is essential for transparency and reproducibility. In the revised manuscript, Rats 2 and 10, which were used for hyperparameter tuning, have been excluded from all final performance metrics (mean ± standard deviation) in Tables 2 and 3 to prevent data leakage in the generalizability analysis. Their individual results remain in the tables solely for contextual reference. Statistical analyses (paired t-tests) are now conducted only on the remaining rats, and we have clearly specified the hyperparameters tuned in Methods, subsection Training & Evaluation, Hyperparameter Tuning (pages 6 – 7).

2. Questionable performance claims

The claim that the proposed method “outperforms the CNN” is not well supported. Table 2 shows that only a subset of configurations exceed CNN accuracy or F1-score, and significance is claimed without clear justification. Reported improvements may reflect overfitting from hyperparameter sweeps rather than genuine generalisation gains. Statistical testing on n = 8 subjects is underpowered and should not be used to claim significance.

We acknowledge that the sample size is small and agree that this is an inherent limitation of the study. Unfortunately, in vivo peripheral nerve recording experiments are technically demanding and resource-intensive, and datasets of this modality are scarce in the literature. We now explicitly describe this limitation in the second-to-last paragraph of the Discussion (page 12). Importantly, after addressing the first major revision point and excluding the two rats used for hyperparameter tuning, the graph-based approach with 4 or 5 neighbors still significantly outperforms the CNN baseline in generalizability (p < 0.05, paired t-test). These results have been updated and clarified in Table 2 and in subsection Generalizability Performance of the Results (page 8).

3.Within-subject results not statistically meaningful

In Table 4, within-rat results exhibit large standard deviations and overlapping confidence intervals across methods. Differences are likely random variation and should not be interpreted as real improvements.

We have now clarified this in Results, subsection Within-Rat Performance (page 10), expanding more on the performance results in Table 4.

4. Weak ablation design

The ablation comparing geodesic graphs to random graphs is insufficient to support claims about the value of geometric priors. Stronger baselines—e.g., graphs built from inter-channel correlation matrices—are needed to determine whether improvements truly stem from spatial geometry or merely from introducing structured connectivity.

We added a Euclidean-distance graph baseline, generated using inter-electrode 2D planar distances. Results are presented in updated Tables 3 and 4. We discuss how Euclidean graphs perform better than random graphs but worse than geodesic graphs, supporting the claim that true anatomical geometry provides additional benefit (third paragraph of Discussion section, page 11).

5. Inappropriate data augmentation

The listed augmentations (brightness, contrast, saturation, hue perturbation) are intended for image data and not meaningful for spatio-temporal neural signals. Their use could distort physiological relationships. The authors should justify these operations or remove them entirely.

We have now clarified this and removed any inappropriate augmentations described. The final augmentation pipeline uses only low-amplitude Gaussian noise, which is physiologically meaningful for neural signals. This is described clearly in the revised Methods, subsection Model Architecture (last paragraph, page 6).

6.Unclear architectural and training hyperparameters

Hidden-unit sizes, L2 regularisation, batch size, learning rate, and dropout rate are fixed without explanation. The basis for these choices must be stated, or the parameters should be included in the tuning procedure.

We have now clarified this in subsection Training & Evaluation, Hyperparameter Tuning of Methods (page 7).

7.Insufficient methodological clarity

The paper must specify the exact formulations of the edge convolution and general graph convolution layers, including equations and citations of the specific variants used. The authors should explicitly state what constitutes edge features (e.g., edge weights or other quantities).

We thank the reviewer for asking for clarification on formulations of the convolutional layers. These have now been incorporated into Methods, subsection Model Architecture (pages 5 – 6).

Minor

- Clarify dataset composition and class balance; report number of samples per class

- The description of “naturally evoked compound action potentials (nCAPs)” should be expanded for a general audience. In Equation (1), variables x and y should be explicitly defined as electrode coordinates on the cuff.

- Figures and table captions should be more detailed to allow easy understanding of what’s exactly being shown

- Specify what software was used for reproducibility

We thank the reviewer for pointing these out, and we have addressed each item in the revised manuscript. Dataset composition has been updated and is now summarized in Table 1. The description of naturally evoked CAPs has been clarified in Methods, subsection Data Des

Attachment

Submitted filename: Response to Reviewers.pdf

pone.0345204.s001.pdf (99.2KB, pdf)

Decision Letter 1

Luca Citi

24 Feb 2026

Dear Dr. Ji,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that the paper is close to meeting PLOS ONE’s publication criteria but we would like you to consider some final suggestions. Therefore, we invite you to submit a revised version of the manuscript that considers the points raised during the review process.

Please submit your revised manuscript by Apr 10 2026 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

  • A letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols....

We look forward to receiving your revised manuscript.

Kind regards,

Luca Citi, PhD

Academic Editor

PLOS One

Journal Requirements:

1. If the reviewer comments include a recommendation to cite specific previously published works, please review and evaluate these publications to determine whether they are relevant and should be cited. There is no requirement to cite these works unless the editor has indicated otherwise.

2. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

Reviewer #1: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions??>

Reviewer #1: Yes

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously? -->?>

Reviewer #1: Yes

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available??>

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.-->

Reviewer #1: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #1: Yes

Reviewer #3: Yes

**********

Reviewer #1: I would like to thank the authors for their comprehensive responses and the revisions made to the manuscript. The clarifications provided are exhaustive and satisfactorily address my previous concerns regarding the state-of-the-art comparison and the methodological details.

I have just one minor note regarding the presentation of the results, which is a suggestion and not a mandatory requirement for acceptance. While I appreciate the inclusion of the network topology figures, I believe that the quantitative results currently presented in Tables 2, 3, and 4 would be much more effective if visualized as charts (e.g., grouped bar plots with error bars). Visualizing the data in this way would allow readers to immediately grasp the performance trends and the stability of the model across subjects, whereas large tables can be harder to interpret at a glance. The detailed numerical tables could then be moved to the Supplementary Material.

However, I leave this decision entirely to the authors' discretion. I am satisfied with the scientific content and the improvements made, and I support the publication of the manuscript in its current form.

Reviewer #3: All my comments have been addressed properly. I don't have further questions. The comparison between geodesic-based and the Euclidean graph looks reasonable, which I think is the key of this paper.

**********

what does this mean?). If published, this will include your full peer review and any attached files.). If published, this will include your full peer review and any attached files.). If published, this will include your full peer review and any attached files.). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our For information about this choice, including consent withdrawal, please see our For information about this choice, including consent withdrawal, please see our For information about this choice, including consent withdrawal, please see our Privacy Policy..-->

Reviewer #1: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

To ensure your figures meet our technical requirements, please review our figure guidelines: https://journals.plos.org/plosone/s/figures

You may also use PLOS’s free figure tool, NAAS, to help you prepare publication quality figures: https://journals.plos.org/plosone/s/figures#loc-tools-for-figure-preparation.

PLoS One. 2026 Apr 17;21(4):e0345204. doi: 10.1371/journal.pone.0345204.r004

Author response to Decision Letter 2


2 Mar 2026

We sincerely thank the editor and reviewers for their time in reviewing the revised manuscript, and greatly appreciate the constructive feedback provided. We have considered each comment. Responses to each reviewer comment are provided in blue text throughout this letter.

Reviewer #1: I would like to thank the authors for their comprehensive responses and the revisions made to the manuscript. The clarifications provided are exhaustive and satisfactorily address my previous concerns regarding the state-of-the-art comparison and the methodological details.

I have just one minor note regarding the presentation of the results, which is a suggestion and not a mandatory requirement for acceptance. While I appreciate the inclusion of the network topology figures, I believe that the quantitative results currently presented in Tables 2, 3, and 4 would be much more effective if visualized as charts (e.g., grouped bar plots with error bars). Visualizing the data in this way would allow readers to immediately grasp the performance trends and the stability of the model across subjects, whereas large tables can be harder to interpret at a glance. The detailed numerical tables could then be moved to the Supplementary Material.

However, I leave this decision entirely to the authors' discretion. I am satisfied with the scientific content and the improvements made, and I support the publication of the manuscript in its current form.

We sincerely thank the reviewer for their kind remarks and thoughtful feedback on the revised manuscript. We appreciate the suggestion to visualize the quantitative results in Tables 2, 3, and 4 using grouped bar plots with error bars. After careful consideration, we chose to retain the tables in the main text because a few of the reported differences between models may be less clearly distinguishable when visualized in plot form, and we believe that presenting the exact numerical values allows for more precise comparison. We are grateful for the reviewer’s support and constructive suggestions that have helped improve the quality and clarity of the manuscript.

Reviewer #3: All my comments have been addressed properly. I don't have further questions. The comparison between geodesic-based and the Euclidean graph looks reasonable, which I think is the key of this paper.

We sincerely thank the reviewer for their time and thoughtful feedback on the revised manuscript. We greatly appreciate the careful evaluation and are pleased that the comparison between geodesic-based and Euclidean graph constructions is now clear and reasonable. The reviewer’s comments have been instrumental in strengthening the manuscript, and we are grateful for their support.

Attachment

Submitted filename: R2_Response_to_Reviewers.pdf

pone.0345204.s002.pdf (38.8KB, pdf)

Decision Letter 2

Luca Citi

3 Mar 2026

Enhancing Generalizability in Classification of Peripheral Neural Recordings with Graph Neural Network

PONE-D-25-46080R2

Dear Dr. Ji,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact  and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact  and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact  and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact billing support....

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Luca Citi, PhD

Academic Editor

PLOS One

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Luca Citi

PONE-D-25-46080R2

PLOS One

Dear Dr. Ji,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS One. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Luca Citi

Academic Editor

PLOS One

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Response to Reviewers.pdf

    pone.0345204.s001.pdf (99.2KB, pdf)
    Attachment

    Submitted filename: R2_Response_to_Reviewers.pdf

    pone.0345204.s002.pdf (38.8KB, pdf)

    Data Availability Statement

    The data used in this study were from a publicly available dataset available at Borealis, U of T Dataverse: https://doi.org/10.5683/SP3/JRZDDR.


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES