Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2021 Feb 10;16(2):e0245589. doi: 10.1371/journal.pone.0245589

Efficient neural spike sorting using data subdivision and unification

Masood Ul Hassan 1,2,*, Rakesh Veerabhadrappa 2, Asim Bhatti 2,*
Editor: Alexandros Iosifidis3
PMCID: PMC7875432  PMID: 33566859

Abstract

Neural spike sorting is prerequisite to deciphering useful information from electrophysiological data recorded from the brain, in vitro and/or in vivo. Significant advancements in nanotechnology and nanofabrication has enabled neuroscientists and engineers to capture the electrophysiological activities of the brain at very high resolution, data rate and fidelity. However, the evolution in spike sorting algorithms to deal with the aforementioned technological advancement and capability to quantify higher density data sets is somewhat limited. Both supervised and unsupervised clustering algorithms do perform well when the data to quantify is small, however, their efficiency degrades with the increase in the data size in terms of processing time and quality of spike clusters being formed. This makes neural spike sorting an inefficient process to deal with large and dense electrophysiological data recorded from brain. The presented work aims to address this challenge by providing a novel data pre-processing framework, which can enhance the efficiency of the conventional spike sorting algorithms significantly. The proposed framework is validated by applying on ten widely used algorithms and six large feature sets. Feature sets are calculated by employing PCA and Haar wavelet features on three widely adopted large electrophysiological datasets for consistency during the clustering process. A MATLAB software of the proposed mechanism is also developed and provided to assist the researchers, active in this domain.

Introduction

Neuro-engineering is an interdisciplinary research domain that provides a collaborative platform for engineers, scientists, neurologists and clinicians to grow a robust and reliable communication network between human brain and computers using advanced engineering procedures, methods, tools and algorithms [13]. It is largely accepted hypothesis that the brain passes information in terms of neurons’ firings i.e. action potential or spikes over specific interval of time, known as neuron firing rate. Neurophysiological study of these hefty action potentials or spikes emanating from the neural network of the brain is essential to reveal the underlying behaviours and properties of neurons. A good understanding of the human brain neuronal network or nervous system is critically important in developing brain machine interfaces (BMIs), neuro-prosthetics and comprehensive brain-computer communication networks [4].

Electrophysiological analysis has attracted paramount importance, in recent years, in deciphering useful information about the underlying functional behaviour of the brain both in spontaneous and stimulated environments [5, 6]. This has paved the way of new discoveries in understanding the impact of external stimuli such as pharmaceuticals [7] and infections on the brain functionality [8]. Researchers have successfully developed the neural decoders from the neurophysiological study of intra neural recordings of human primary motor cortex to drive the artificial prostheses [9]. Electrophysiological studies also find significant importance in treating patients having neurological diseases or mental disorders especially in the case of epileptic disease. In addition, these studies have played vital role in understanding the gamma-protocadherine influences in regulating the neural network endurance and generating new neural synapses [10].

The significance of electrophysiological study of human brain lies in intercepting the neuronal signals with negligible interference in brain’s natural functionality. Numerous electrophysiological methods are found in the literature to monitor the action potentials or spikes from neurons, such as intracellular glass pipette electrodes [11], patch clamp electrodes [12, 13], extracellular single or multi-site electrodes [14], and optical imaging devices [15, 16]. Among all, extracellular recordings using micro fabricated electrode arrays [1719] are largely preferable in research because of its relatively less impact on the normal working behaviour of neurons [20]. Extracellular recordings are further categorised into invasive (in-vivo) and non-invasive (in-vitro) approaches [21]. In in-vivo approach, microelectrodes such as a probe or tetrode (probe with four electrodes) is surgically implanted in the understudy region of the brain. Whereas, in in-vitro approach, neurons are cultured on the separate dishes integrated with microelectrodes [22]. The neurophysiological technology implemented to record neural action potentials is very advanced, but still it is very immature to record the action potentials emanating from a single neuron. Brain consists of closely packed neurons that mostly excites simultaneously to encode information consisting of synchronised and correlated action potentials [23, 24]. Neurons present in the surrounding or neighbourhood of the understudy region, when excited, introduce noise in the neural recordings [25, 26]. Therefore, to study and analyse the behaviour of individual neurons and to group the action potentials having similar features into specific clusters, the concept of ‘Spike Sorting’ is implemented [27, 28].

An overview of in-vivo and in-vitro recordings and complete description of the steps involved in the spike sorting process is illustrated in Fig 1. Spike sorting consists of four main steps. First, raw data is filtered to minimise the effect of noise. The work of Choi et al. in [29] has significance importance in reducing the effect of background noise and detecting useful spikes trains from neural recordings at low signal to noise ratio (SNR) using multi resolution Teager energy operator (MTEO). Paralikar et al. in [30] proposed the virtual referencing (VR) method based on average functional electrode signal and inter-electrode correlation (IEC) method based on correlation coefficient between threshold exceeding spikes segments for common noise reduction. Common noise is generally produced by electromyographic activity, motion artifacts, and electric field pickup, especially in awake/behaving subjects. Pillow et al. in [30] proposed binary pursuit algorithm to significantly reduce the effect of stochastic background component of correlated Gaussian noise from the neural recordings. Takekawa et. al in [31] worked on filtering the biological noise from the neural recording using peak band pass filtering technique. Band pass filtering is a common practice among neural scientists for reducing the effect of background noise. This followed by spike extraction [32]. Abeles and Golstein in [33], elaborated extensively about multi-unit spikes detection. Threshold and inter-spike interval based detection methods are frequent and popular among researchers [34]. However, in the proposed algorithm the focus is on the computation efficiency of spike sorting algorithms rather than the spike estimation. For the proposed research work, spikes are extracted using labels provided with the data to make comparison of performance between different algorithms unbiased due to noise effects. The third step in spike sorting is the feature extraction of detected spikes [35]. The latest feature extraction technique is proposed by Zamani and Demosthenous in [36], however, feature extraction techniques that are largely practised by researchers are Principal Component Analysis (PCA) [3739], Wavelet Transform [4042] and Wavelet Packet Decomposition [43]. The last step in this process is the clustering of spikes into specific action potential groups having similar features [44]. For clustering, scientists have proposed numerous clustering algorithms in the literature [4548] that are mainly classified into two main categories; Supervised [49] and Un-Supervised [50]. In supervised clustering, the number of clusters are predefined and the algorithms forced the spikes to fit into desired number of predefined clusters [51]. Whereas, in unsupervised clustering, algorithms, without having prior clustering information, automatically estimate the total cluster numbers and based on similarity in spike features, label the spikes into their respective groups [52]. The unsupervised clustering is more reliable and useful when there is no prior knowledge about clusters [53]. The spike sorting algorithms are mainly used offline and are implemented for behavioural quantification on pre-recorded neural datasets [54]. However, researchers have developed online spike sorting algorithms that can quantify spike-clusters on live neural recordings [55]. The latest state of art in spike sorting process is presented in [56].

Fig 1. An overview of spike sorting process with in-vivo and in-vitro recordings.

Fig 1

(a) Microscopic image of neural network in the brain. (b) Brains cells cultured on the Micro-Electrode Arrays (MEAs). (c) Implanted probe in the rat brain for in vivo recordings. (d) Data acquisition system to interface with MEAs (e) Computing machine for data processing and spike sorting. (f) Multichannel data acquisition and recording (g) Visualisation of complete spike sorting process. (h) Raw data after sampling and amplification. (i) Noise filtering of data using band pass filters. (j) Spikes detected using the threshold or inter-spike interval methods. (k) Feature extraction of the detected spikes to reduce the dimensionality of the data. (l) Clustered features after applying clustering algorithms extracted spike features. (m) Clustered spikes.

Problem statement

Advancements in nanotechnology and nanofabrication has enabled neuroscientists and engineers to capture the electrophysiological activities of the brain at very high resolution, data rate and fidelity. However, to decipher useful information from these high dense electrode data, performance in terms of computational speed and accuracy of these spike-sorting algorithms, independent of their online and offline nature, plays an important role.

Stevenson and Kording in [57], presented data analysis issues due to progressive technological advancements of neural recordings. Progress in neural recording techniques enabling simultaneous multi channels recording is projected to double every 7 years resulting in high density and large size data. It is estimated that recording from 1000 neurons simultaneously could be achieved by 2025. The most recent automated spike sorting algorithm proposed by Chung et al. in [58] also highlighted the issue of low computational speed of spike sorting algorithms. Although they have proposed an efficient method for spike sorting, it lacks the speed researchers require for optimal results when sorting larger and high dense datasets. Wild et al. in [59] studied the performance evaluation on widely used clustering algorithms. His research outcomes highlighted the dependency of computational speed on data size or number of spikes to be clustered.

Chen and Cai in [60] investigated the issue and proposed that this behaviour is due to complexity of operations involved in the algorithms. They reported, for n size of data, spectral clustering requires O(n2) (second order equation) operations in graph construction and O(n3) (third order equation) operations in Eigen-decomposition. These second order and third order equations prove the non-linear behaviour of spectral clustering. To motivate our analysis, spectral clustering was applied on five datasets of variable length and calculated the corresponding computational time as in Table 1. The plot in Fig 2, clearly depicts the non-linear behaviour in computational time required by spectral clustering to complete its operations with respect to data size.

Table 1. Computational times of five datasets for spectral clustering.

Data Name Data Size # of classes Computational Time
MNIST [61] 70000 10 3654.90
LetterRec [62] 20000 26 195.63
PenDigits [62] 10992 10 60.48
Seismic [63] 98528 3 4328.35
Covtype [62] 581012 7 181006.17

Fig 2. Computational time versus data size plot.

Fig 2

The dependency of speed and computational time on data size in spike-sorting has made it very difficult to efficiently and accurately identify the total number of neurons in large and dense electrophysiological data. Furthermore, based on the work of Napoleon and Pavalakodi on large, dense and high dimensional breast cancer cell data [64], the accuracy of clustering algorithms is also somehow contingent to the data size. With the increase in data size the occurrence of false positives and negatives in spike sorting increases significantly, which reduces the overall efficiency and performance of the algorithms involved in the process.

Despite these challenges, in literature, researchers have developed numerous spike sorting algorithms to address the challenge of handling large and dense electrophysiological data. However, limited work has considered enhancing computational speed and efficiency by changing the way we input data into spike sorting algorithms. The proposed algorithm pre-processes data to significantly reduce computational time and to enhance speed and efficiency of a wide range of existing spike sorting algorithms. The proposed algorithm has great potential to be adopted by parallel computing approaches to further enhance spike sorting algorithms’ efficiency for real-time online spike analysis.

Proposed mechanism

The novelty of the proposed mechanism lies in its capability to operate the existing spike sorting algorithms at their peak efficiency by introducing the optimal length subsets of large electrophysiological data at clustering stage. The overall mechanism consists of three major steps as illustrated in Fig 3. 1) The first step involves subdivision of data into data-subsets of optimal length. The procedure to identify optimal length is discussed in next section. 2) The second step involves clustering spikes in data-subsets using conventional spike sorting algorithms. 3) The last step involves unification of the clustered subsets. The final unified clusters are then used to label the detected spikes representing complete large electrophysiological data into their respective neural classes. The comparison of conventional spike sorting and proposed algorithm is depicted in Fig 4. It is worth mentioning that the proposed mechanism deals with data- subdivision and unification to felicitate and enhance the performance of existing clustering algorithms and does not modify the internal workings of the algorithms employed in this study. A recently developed clustering algorithm “Mountainsort” by Chung et al. [58] uses a density based approach to cluster spikes can also be used with this mechanism for efficient spike sorting.

Fig 3. Illustration of complete proposed mechanism.

Fig 3

The first step is to divide the large electrophysiological data into smaller groups. Second step involves the clustering of data-subsets using the conventional spike sorting algorithms. Last step involves the unification or merging of clustered data-subsets to get optimal clustering of complete large electrophysiological data.

Fig 4. Comparison of conventional and proposed spike sorting process.

Fig 4

A similar approach of data subdivision is used by Pachitariu et al. in [65] for KiloSort algorithm. The algorithm divides the high dense neural data into small batches and uses them for mean-time processing of data filtering in the GPU that reduces the overall time of the spike sorting process. However, clustering of spikes is still deployed at complete large neural datasets, which resulted into the slower computational speed of spike sorting at clustering stage. In addition, as opposed to proposed mechanism, the data-subdivision mechanism is limited to KiloSort and may not be applicable for other spike sorting algorithms. Furthermore, this algorithm failed to introduce the concept of optimal length for data-subdivision which is an important parameter to consider in enhancing the computational speed and operational cost of spike sorting process.

The detailed description of the steps involved in the proposed mechanism is provided in the following sections:

Data subdivision

Subdividing large electrophysiological data into optimal length subsets is the most critical component of proposed mechanism. To form data subsets, let D represents the electrophysiological data recorded at a single acquisition channel. The total number N of optimal subdivisions is estimated as in Eq (1)

N=LOL (1)

where L is the length of data D and OL is the optimal length for data-subsets. The procedure to calculate OL is presented in the next section.

The data-subsets are then estimated as in Eq (2).

Sd(n)={D(1+(n-1)*N):D(n*N)n*N<DtD(1+(n-1):D(Dt)n*NDt}n=1,2,3,4N (2)

where Sd(n) represents n number of subdivided data-subsets of the large data D.

Identification of optimal length (OL) for data-subsets

OL is the range of values from which if the data size is selected to perform clustering, the clustering quality and computational efficiency of the conventional algorithms improve significantly. OL parameter is dependent on the algorithm type rather than on the data dynamics. Therefore it needs to be estimated only once for each algorithm. The OL parameter for ten commonly used clustering algorithms employed in this study, is estimated and shown in Fig 5b.

Fig 5. Identification of optimal length OL.

Fig 5

(a) Illustrates the description of steps involved in identifying the OL for spike sorting algorithms. Computational time versus data size plot. The X-axis shows the length of the data increasing from zero to 2000 while the Y-axis shows the corresponding time taken by the clustering algorithm to perform clustering process, in milliseconds. The computational time is the processing time after movmean filter(20 datapoints length) filtered the unwanted ripples in the plot and returned smooth curves. Detected abrupt changes in the plot taking 0.1 of the maximum rate of change in computational time as threshold (d) Identified optimal length OL of data subsets used for data subdivision. b) Optimal Length (OL) for ten commonly used clustering algorithms. The average value over ten repetitive analyses is given as robustness of the measure in optimal length for data subdivision.

To understand the computational time vs data size behaviour, clustering is performed in an incremental manner. At every increment, the size or length of the data increases and the computational time is plotted with respect to data size as shown in Fig 5a. The size of data for which the clustering algorithm shows smoother behaviour is termed as OL, that needs to be estimated for optimal clustering results.

In this research work OL is estimated by employing the work of Killick [66]. A threshold of 0.1 of the maximum rate of change of the computational time is used. The first change in computational time above the threshold is estimated to be the optimal length of the data-subset.

The procedure proposed in this research work to calculate OL is implemented on ten aforementioned commonly used clustering algorithms. The procedure is repeated hundred times to get an average OL value as an efficient measure for robustness in results. The calculated OLs are depicted in Fig 5b. It is observed that the performance of clustering algorithms is independent of the data dynamics and feature extraction techniques. OL for all the algorithms adopted in this study, lies approximately in the same range for all three data and six feature sets, employed. Therefore, the computational performance of the algorithms depends on the length of the data set and not on the data dynamics.

Deviation of (OL) from the estimated optimal point could lead to inefficient spike sorting performance. Data subdivision using optimal length is a compromise between computations involved in clustering process and unification process. (OL) forms a direct relationship with clustering computations and an inverse relationship with computations involved in the unification process.

Clustering of data

Data subdivision is followed by clustering of data-subsets employing conventional spike sorting algorithms. Ten algorithms, as illustrated in Fig 5, are employed in this study due to their wide adaptability in spike sorting research. The algorithm proposed is independent of the clustering procedure; therefore, any other clustering technique could be adopted in this mechanism.

Unification of subclusters

After the clustering is performed on each data-subset, the unification of the sub clusters is performed. Sub-clusters are unified by identifying the overlap between the bounded regions of sub-clusters. The bounded region (BR) is a ‘m’ dimensional set which consist of minimum and maximum variations of ‘m’ dimensional spike feature waveforms in each dimension for a corresponding sub cluster. The bounded region for jth sub cluster is given by relationship in Eq (3).

BRj,i={[minmax]j,1[minmax]j,2[minmax]j,3[minmax]j,m} (3)

Where BRj,i is the ‘m’ dimensional bounded region for jth sub cluster with j ∈ [1, 2, 3,.., k] and ‘k’ is the total number of sub clusters participated in the unification process. [minmax]j,i are the minimum and maximum variations of spike feature waveforms for jth sub cluster and in ith dimension and i ∈ [1, 2, 3, …, m].

In this study, since 10 PCA or 10 Wavelet features are used to transform the spike waveform into spike feature waveform. So ‘m’ is 10 in this particular case and BR is a 10 dimensional set with minimum and maximum values providing variation of spike feature waveforms in each dimension for a particular sub cluster.

The bounded region is calculated for all ‘k’ sub clusters participated in the unification process. The sub clusters, having overlapping bounded regions in all dimensions, are unified together. The unification process for a 2 dimensional sub clusters is shown in Fig 6. In the Fig 6, it is also illustrated how sub clusters unify in three different scenarios i.e. 1) no overlapping region between sub clusters 2) overlap between two distinct sub clusters and 3) multiple overlapping sub clusters.

Fig 6. Mechanisim to unify or merge clusters.

Fig 6

To eliminate the impact of outliers in deciding the bounded region for unification process, the spike feature waveforms are filtered in each sub clusters. The filter proposed in this study is based on Euclidean distance. A sub cluster having ‘m’ dimensional spike feature waveforms, should have an ‘m’ dimensional centroid ‘C’. It is important to note that, the complete ‘m’ dimensional spike feature waveform is considered as a single point in ‘m’ dimensional space in calculating the Euclidean distance. Therefore, for each spike feature waveform, an Euclidean distance from spike feature waveform to its sub cluster centroid is calculated. The relationship to calculate the Euclidean distances is given in Eq (4).

EDl(Ci,Sl,i)=(C1-Sl,1)2+(C2-Sl,2)2+(C3-Sl,3)2++(Cm-Sl,m)2 (4)

EDl(Ci, Sl,i) is the Euclidean distance calculated for the lth spike feature waveform Sl,i and sub cluster centroid Ci. l ∈ [1, 2, 3, 4 …, n] and i ∈ [1, 2, 3, …, m] where ‘n’ is the total number of spikes in a sub cluster and ‘m’ is the spike feature waveform dimension.

Since, the Euclidean distance is calculated based on Eq (4), a n × 1 Euclidean distance matrix (EDM) is generated, as in Eq (5).

EDM=[ED1ED2ED3EDn] (5)

This EDM matrix is used to identify outliers in spike feature waveforms. From EDM, a Mean ‘μ’ and Standard Deviation ‘σ’ is calculated by using Eqs (6) and (7).

μ=t=1nEDtn (6)
σ=1nt=1n(EDt-μ)2 (7)

Using the mean and standard deviation of a normal distribution curve, the Euclidean distance values are converted into Z scores using the Eq (8).

Zl=(EDl-μ)/σ (8)

The Z score distribution determined by Eq (8) is then used to identify the data outliers in the EDM matrix given by Eq (5). To this aim, we considered two scenarios; 1) when the Z score distribution of the EDM matrix is normal and 2) when the Z score distribution of the EDM matrix is skewed. There are numerous methods that can determine the normality of the data distribution as in [6769]. However, in this study, the normality of the Z score distribution of EDM matrix is determined using the Interquartile Range IQR method [70].

The quartiles are three points that divide the data set into four equal groups, each group comprising a quarter of the data, for a set of data values which are arranged in either ascending or descending order. Q1, Q2, and Q3 are represent the first, second, and third quartile’s value. The Interquartile Range (IQR) is basically a difference between the first quartile (Q1) and third quartile (Q3). The IQR of the Euclidean distance matrix sorted in ascending order can be determined using relation given in Eq (9).

IQR=Q3-Q1 (9)

Where Q1 is first quartile and it is the median of lower half of the euclidean distances sorted in ascending order and Q3 is the third quartile and it is the median of upper half of the euclidean distances sorted in ascending order.

If the distance of the Q1 and Q3 from the median of the complete dataset containing Euclidean distances is equal, the data is normally distributed and the bell shaped curve is symmetric. If the distance from data mid-point to Q1 is bigger than Q3, the data distribution is skewed towards left, and if Q3 is bigger than Q1, the data distribution is skewed towards right.

For a normal distribution of the data, when bell shaped curve is symmetric, Empirical rule is valid and the outlier filter (OF) is is defined as a range between μ±2σ and its is given by Eq (10).

OF=[minmax]=[μ-2σμ+2σ]=[-2Z2Z] (10)

For a nonsymmetric or left and right skewed distributions, 1.5 Interquartile Range (1.5 IQR) filter is used to identify the sub cluster outliers. The factor 1.5 is empirically derived and being used by novel researchers in statistics for skewed data to identify outliers [71, 72]. Therefore, in this study 1.5 IQR based outlier filter (OF) is designed to remove data outliers in skewed distribution and it is given by Eq (11).

OF=[minmax]=[Q1-1.5×IQRQ3+1.5×IQR] (11)

All the featured spikes, having the Euclidean distance lies within the OF range, are considered in estimating the bounded region in Eq (3) for unification of sub clusters.

A similar approach is adopted by Aksenova et al. in [73] to perform training of online spike sorting algorithm employing phase space. Their algorithm is focused on efficient noise reduction rather than optimisation of computational efficiency.

Performance evaluation of the proposed algorithm

In this research work, the performance of the proposed algorithm is evaluated using two indicators, computational time and clustering quality. A comparative performance of the proposed algorithm with respect to the conventional algorithm is presented in Fig 7(a) and 7(b).

Fig 7. Illustration of improved computational speed and clustering accuracy.

Fig 7

(a) Improved computational speed in percentage of understudy ten algorithms across six large neural feature sets. (b) Improved clustering accuracy in percentage of understudy ten algorithms across six large neural feature sets. The proposed data-subdivision and unification method has shown a positive trend in improving the performance of spike sorting algorithms. The improvement in reducing computational time is significantly high, while due to maturity of spike of sorting algorithms, accuracy improvement is relatively lower in some of spike sorting algorithms. The average results for 10 repetitive analysis has been presented and it is worth noting that proposed mechanism has shown promising improvement results around all data types and spike sorting algorithms.

For validation, ten most widely adopted clustering algorithms are employed in the proposed research work. The algorithms include MeanShift (MS) [74], Density-based spatial clustering of applications with noise (DBSCAN) [75], Kmeans (KM) [76], Kmedoids (KMD) [77], Fuzzy C means (FCM) [78], Variational Bayesian Gaussian Mixture Model (VBGMM) [79], Expectation Maximization Gaussian Mixture Model (EMGMM) [80], Agglomerative Hierarchical Clustering (AHC) [81], Birch (BH) [82] and Ordering Points to Identify the Clustering Structure (OPTICS) [83].

To quantify computational efficiency of the proposed algorithm, three data sets are used, reported by Quiroga [84], because of their wider adoptability and ground truth availability. These datasets includes two (2) simulated Dataset 1 (D1) and Dataset 2 (D2) and one human Dataset 3 (D3). Human data is originated from multiunit recording in the temporal lobe of an epileptic patient from Itzhak Fried’s lab at UCLA [84]. The information regarding spatio-temporally overlapping spikes as a result of multi-unit recordings can be identified using “Matching Persuit” algorithm [85]. However in this study, the multi-unit spikes are already detected and labeled in the ground truth. Labels for three distinguished clusters are provided for each of dataset D1, D2 and D3 in their respective ground truth.

Each spike waveform consists of 64 samples. Haar Wavelets and PCA features are employed to reduce the data dimensionality while preserving the variance of the data and spike information. In case of Haar wavelets transform, optimal wavelet features were selected following the study of Quiroga [79], which implemented the four-level multi resolution decomposition. The 64 wavelet coefficients generated provides unique spike characteristic at different scales and times. As each spike class has different multimodal distribution, the Lilliefors modification of Kolmogorov-Smirnov (KS) test for normality [81] was used to select the optimal wavelet features. The maximum deviation of multimodal distribution features from normality defines the optimal features. We refer the readers to Quiroga [79] for further explanation. In this context, 10 wavelet features with largest deviation of normality is regarded as optimal wavelet features.

Similarly, 10 PCA features were selected in this study to validate the computational and performance efficiency of the proposed vs conventional algorithms. The PCA components are not scaled to match their explained variances. The individual variances of PCA components are accumulated and the optimal number of PCA components that gives at least 85% of cumulative explained variance are chosen for the analysis. 10 PCA features are required to get at least 85% cumulative explained variance of the 64 dimensional spikes data used in this study.

It is important to note that, the accuracy of clustering algorithms may be affected by the data dimensionality and number of optimal feature sets used. However, in this study the same 10-dimensional features are used for all the algorithms to maintain the consistency while validating the performance outcomes.

The research work is carried out on a personal computer (PC) consisting of Intel (R) Pentium (R) CPU G4560 @3.5GHz, 8 GB of RAM and 64 Bit windows 10 operating system.

Performance on computational time or speed

To explore and validate the performance of the proposed algorithm, in terms of computational time, as tabulated in Table 2, is estimated using the expression (12).

Ts(%)=(Ct-Pt)Ct×100 (12)

Table 2. Computational times and time based performance improvement for ten clustering algorithms.

Algorithm Method Computational Time (Seconds)
D1, PCA D1, WAV D2, PCA D2, WAV D3, PCA D3, WAV
Meanshift Proposed 0.3 0.02 0.43 0.13 0.11 0.03
Conventional 0.36 0.04 0.76 0.16 0.15 0.06
Time Saved (%) 17.25 60.59 43.69 17.81 23.88 46.63
DBSCAN Proposed 0.75 1.6 3.26 0.54 8.76 3.84
Conventional 1.82 4.81 8.37 1.3 48.63 21.09
Time Saved (%) 58.6 66.64 61.02 58.71 81.98 81.77
Kmeans Proposed 0.04 0 0.01 0 0.04 0.03
Conventional 0.28 0.03 0.03 0.01 0.09 0.05
Time Saved (%) 84 84.7 73.2 55.14 52.55 32.59
Kmedoids Proposed 0.32 0.1 0.17 0.14 1.37 1.31
Conventional 1.14 0.36 0.33 0.41 1.38 1.85
Time Saved (%) 72.39 72.2 47.67 65.99 1.19 29.26
VBGMM Proposed 0.3 0.25 0.6 0.4 1.91 1.05
Conventional 0.44 0.51 0.75 0.63 3.9 2.57
Time Saved (%) 31.17 50.85 19.46 37.15 51.02 59.35
EMGMM Proposed 0.43 0.32 1.2 0.46 3.39 3.07
Conventional 0.46 0.62 1.76 0.58 3.71 4.29
Time Saved (%) 6.83 48.56 32.06 21.17 8.78 28.5
Agglomerative Proposed 0.18 0.07 0.06 0.06 0.2 0.2
Conventional 0.18 0.15 0.15 0.14 1.14 1.04
Time Saved (%) 2.55 54.23 59.75 58.61 82.92 80.46
OPTICS Proposed 1.11 0.44 0.42 0.43 1.72 1.72
Conventional 1.14 1.08 1.02 1.05 7.27 7.18
Time Saved (%) 2.35 59.15 58.56 59.37 76.34 76.07
BIRCH Proposed 1.25 1.61 1.39 1.94 2.22 5.32
Conventional 1.68 4.11 2.35 3.73 5.9 31.56
Time Saved (%) 25.65 60.77 40.94 47.99 62.4 83.15
FCM Proposed 0.01 0.03 0.01 0.05 0.02 0.41
Conventional 0.08 0.05 0.02 0.06 0.05 0.59
Time Saved (%) 81.92 49.8 54.26 13.51 65.12 29.43

Where Ct and Pt are computational times of clustering using conventional and proposed algorithms, respectively.

Performance on clustering accuracy

The clustered spikes from spike sorting algorithms are generally evaluated using the validation indices [46]. In this work clustering accuracy, as is described in [86] and (13), is adopted as a validation index and is calculated using the confusion matrix [87] as in (14).

A=#ofaccuratelyclusteredspikesTotal#ofSpikes%=SumofConf.MatrixDiagonalsTotal#ofSpikes% (13)
C=[Ce1g1Ce1g2Ce1gqCe2g1Ce2g2Ce2gq.........Cemg1Cemg2Cemgq] (14)

Where A and C are accuracy index and confusion matrix, respectively. m is total number of estimated clusters and q is total number of clusters in ground truth. Cei,gi represents the number of spikes estimated and clustered accurately relative to the labels provided with the spikes data ground truth. Where ei refers to estimated cluster index and gi ground truth. The accuracy index highlights the percentage of spikes accurately labelled to the clusters described in the ground truth. There are two scenarios taken into account while calculating accuracies.

m = q: when number of clusters estimated are equal to number of clusters in the ground truth. This leads to the square confusion matrix of size m|m=q and the sum of confusion matrix diagonals divided by total number of spikes provides the percentage of accuracy as in Eq (13).

mq: when the number of clusters estimated m are not equal to the number of clusters in ground truth q, the confusion matrix is generated by taking only the dominant estimated clusters m equal to the total number of clusters q in the ground truth. In case of estimated clusters less than the ground truth clusters, i.e. m < q, the confusion matrix is zero padded. The accuracy is calculated by using the expression (13).

The percentage of accuracy enhancement is estimated using the accuracy difference between proposed and conventional methods, which is tabulated in Table 3.

Table 3. Clustering accuracy and accuracy based performance improvement for ten clustering algorithms.

Algorithm Method Accuracy (%)
D1, PCA D1, WAV D2, PCA D2, WAV D3, PCA D3, WAV
Meanshift Proposed 89.89 97.81 83.76 94.03 72.18 81.82
Conventional 89.18 97.59 62.36 94 52.72 75.05
Improved Acc. (%) 0.71 0.23 21.4 0.03 19.46 6.76
DBSCAN Proposed 87.85 92.16 34.72 84.02 72.18 72.19
Conventional 61.07 35.29 33.79 62.3 51.55 51.88
Improved Acc. (%) 26.77 56.87 0.93 21.72 20.63 20.32
Kmeans Proposed 66.35 99.38 95.21 92.78 71.63 78.87
Conventional 44.69 63.86 65.46 81.58 36.56 62.21
Improved Acc. (%) 21.66 35.52 29.76 11.19 35.06 16.66
Kmedoids Proposed 98.07 99.38 77.52 93.1 61.31 83.92
Conventional 96.79 99.38 48.46 92.78 43.48 69.14
Improved Acc. (%) 1.28 0 29.06 0.32 17.82 14.78
VBGMM Proposed 88.25 77.31 62.3 84.34 70.08 63.51
Conventional 85.26 66.47 54.52 68.85 50.66 31.93
Improved Acc. (%) 2.98 10.85 7.77 15.49 19.42 31.58
EMGMM Proposed 91.31 90.37 72.97 84.66 74.15 57.21
Conventional 89.15 80.41 56.61 65.84 61.95 38.15
Improved Acc. (%) 2.16 9.97 16.36 18.82 12.2 19.05
Agglomerative Proposed 94.55 99.26 88.46 96 56.15 80.88
Conventional 93.7 99.21 84.86 87.91 43.75 51.4
Improved Acc. (%) 0.85 0.06 3.6 8.09 12.4 29.48
OPTICS Proposed 76.58 29.42 31 26.04 62.33 20.58
Conventional 42.67 17.18 22.71 12.38 12.34 4.28
Improved Acc. (%) 33.9 12.24 8.29 13.66 49.99 16.29
BIRCH Proposed 93.19 99.26 86.83 93.33 54.73 80.88
Conventional 72.43 99.18 84.66 92.89 44.76 55.53
Improved Acc. (%) 20.76 0.09 2.18 0.44 9.96 25.35
FCM Proposed 71.78 99.38 84.98 92.63 60.99 77.15
Conventional 71.72 99.38 49.42 92.6 39.77 63.97
Improved Acc. (%) 0.06 0 35.56 0.03 21.22 13.18

Clustering results

To highlight enhancement in clustering quality, visual representation of clusters estimated using proposed and conventional methods employing OPTICS on dataset 3 with PCA features and DBSCAN on dataset 1 with Wavelet features, gives 49.99 and 56.87 percent accuracy improvement in the clustering results with respect to the ground truth as in Table 3. The illustration of clustering results for aforementioned examples is shown in Figs 8 and 9 respectively. It is clear from the results that proposed methodology generates significantly superior results in contrast to conventional methods.

Fig 8. Comparison of clustering results obtained using conventional and proposed mechanism employing OPTICS with dataset 3 and PCA features.

Fig 8

(a) Clustering results using conventional spike sorting method applied on complete dataset containing 3740 Spikes. (b) Performance indication of clustering results based on computational time/speed and clustering accuracy. (c)-(f) Clustering results using proposed spike sorting mechanism applied on data-subdivision of optimal length i.e 935 for OPTICS. (g) Unification of subdivided cluster subsets. (h) Performance indication of clustering results using proposed spike sorting method.

Fig 9. Comparison of clustering results obtained using conventional and proposed mechanism employing DBSCAN with dataset 1 and Wavelet features.

Fig 9

(a) Clustering results using conventional spike sorting method applied on complete dataset containing 3405 Spikes. (b) Performance indication of clustering results based on computational time/speed and clustering accuracy. (c)-(f) Clustering results using proposed spike sorting mechanism applied on data-subdivision of optimal length i.e 1135 for DBSCAN. (g) Unification of subdivided cluster subsets. (h) Performance indication of clustering results using proposed spike sorting method.

Discussion

It is largely observed from the results and performance evaluation that the proposed algorithm shows continuous improvement around all algorithms and datasets. The accuracy is improved up to 56.87% while computational time is reduced up to 84.7%. Hence, proposed mechanism has significant impact on enhancing the speed and accuracy of the spike sorting process. In term of clustering accuracy, DBSCAN demonstrates high accuracy improvement of 56.87 percent followed by OPTICS at 49.99 percent. In terms of computational time, Kmeans shows highest computational speed enhancement of 84.7 percent followed by BIRCH with computational speed enhancement of 83.15 percent. In terms of parameter tuning complexity, MeanShift, FCM and Gaussian Mixture models require one parameter to tune, DBSCAN and OPTICS require two and BIRCH requires three parameters to tune to perform their operations. All the supervised clustering algorithms including Kmeans, Kmedoids and Agglomerative require single parameter to tune. In terms of robustness, Kmeans, Kmedoids,FCM gives different results at every iteration, however, Meanshift, EMGMM, VMGMM, Agglomerative, DBSCAN, OPTICS, BIRCH converged to same results after each iteration. For simplicity of the presentation, the presented results are averaged over 10 repetitions.

Software implementation

The software for proposed mechanism is implemented using MATLAB as shown in Fig 10. The free access to open source software for academic purpose is provided with detailed user instructions online at: https://github.com/ermasood/Handling-Larger-Data-Sets-for-Clustering. The software yields the clustering labels with high accuracy and in a fast and efficient way. The first graph in the software window shows the clustered spikes and the second graph illustrates the clustered features of the inputted data. MATLAB codes provided are tested on 2019b and 2018b MATLAB versions. Additionally,’Linspecer.m’ file [88] from MathWorks is required to generate attractive colour combinations and shades for beautiful visualisations.

Fig 10. Software for proposed clustering mechanism.

Fig 10

Conclusion

Neural spike sorting is prerequisite to deciphering useful information from electrophysiological data recorded from the brain, in vitro and/or in vivo. Significant advancements in nanotechnology and nano fabrication has enabled neuroscientists and engineers to capture the electrophysiological activities of the brain at very high resolution, data rate and fidelity. However, the evolution in spike sorting algorithms to deal with the aforementioned technological advancement and capability to quantify higher density data sets is somewhat limited. It is observed from the experiments that larger datasets highly effect the computational time required to perform clustering. To address this challenge, a novel clustering mechanism is proposed to handle large datasets efficiently and with higher accuracy. The proposed mechanism resolves the issue of high computational time and reduced accuracy in conventional method. The proposed algorithms has demonstrated up to 84% and 56% improvement in terms of computational time and clustering accuracy, respectively. The proposed framework is validated by applying on ten widely used clustering algorithms and six large data sets. PCA and Haar wavelets features are employed for consistency during the clustering process. A MATLAB software of the proposed mechanism is also developed and provided to assist the researchers, active in this domain.

Supporting information

S1 Data

(ZIP)

Data Availability

All relevant data are within the manuscript and its Supporting information files.

Funding Statement

The research work is fully supported by Neural and Cognitive Systems Lab at Institute for Intelligent Systems Research and Innovation, Deakin University". Although we do not have any explicit external funding grant linked to this work however the work is internally funded by the lab.

References

  • 1. Dominique MD. What is Neural Engineering? Journal of Neural Engineering. 2006;4(4). [Google Scholar]
  • 2. Bhatti A, Lee KH, Garmestani H, Lim CP. Emerging Trends in Neuro Engineering and Neural Computation. Springer; 2017. [Google Scholar]
  • 3. He B. Neural engineering. Springer Science & Business Media; 2007. [Google Scholar]
  • 4. Eliasmith C, Anderson CH. Neural engineering. Massachusetts Institue of Technology; 2003;. [Google Scholar]
  • 5. Gaburro J, Bhatti A, Harper J, Jeanne I, Dearnley M, Green D, et al. Neurotropism and behavioral changes associated with Zika infection in the vector Aedes aegypti. 2018;7(1):1–11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Gaburro J, Duchemin JB, Paradkar PN, Nahavandi S, Bhatti AJSr. Electrophysiological evidence of RML12 mosquito cell line towards neuronal differentiation by 20-hydroxyecdysdone. 2018;8(1):10109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Bari MAU, Gaburro J, Michalczyk A, Ackland ML, Williams C, Bhatti A. In: Mechanism of Docosahexaenoic Acid in the Enhancement of Neuronal Signalling. Springer; 2017. p. 99–117. [Google Scholar]
  • 8. Gaburro J, Bhatti A, Sundaramoorthy V, Dearnley M, Green D, Nahavandi S, et al. Zika virus-induced hyper excitation precedes death of mouse primary neuron. 2018;15(1):79. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Mussa-Ivaldi FA, Miller LE. Brain–machine interfaces: computational demands and clinical needs meet basic neuroscience. TRENDS in Neurosciences. 2003;26(6):329–334. 10.1016/S0166-2236(03)00121-8 [DOI] [PubMed] [Google Scholar]
  • 10. Lefebvre JL, Zhang Y, Meister M, Wang X, Sanes JR. Y-Protocadherins regulate neuronal survival but are dispensable for circuit formation in retina. Development. 2008;135(24):4141–4151. 10.1242/dev.027912 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Lee AK, Manns ID, Sakmann B, Brecht M. Whole-Cell Recordings in Freely Moving Rats. Neuron. 2006;51(4):399–407. 10.1016/j.neuron.2006.07.004 [DOI] [PubMed] [Google Scholar]
  • 12. Spira ME, Hai A. Multi-electrode array technologies for neuroscience and cardiology. Nature nanotechnology. 2013;8(2):83 10.1038/nnano.2012.265 [DOI] [PubMed] [Google Scholar]
  • 13. Stuart G, Dodt H, Sakmann B. Patch-clamp recordings from the soma and dendrites of neurons in brain slices using infrared video microscopy. Pflügers Archiv. 1993;423(5-6):511–518. [DOI] [PubMed] [Google Scholar]
  • 14. Zhang J, Laiwalla F, Kim JA, Urabe H, Van Wagenen R, Song YK, et al. Integrated device for optical stimulation and spatiotemporal electrical recording of neural activity in light-sensitized brain tissue. Journal of neural engineering. 2009;6(5):055007 10.1088/1741-2560/6/5/055007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Cui X, Lee VA, Raphael Y, Wiler JA, Hetke JF, Anderson DJ, et al. Surface modification of neural recording electrodes with conducting polymer/biomolecule blends. Journal of Biomedical Materials Research. 2001;56(2):261–272. [DOI] [PubMed] [Google Scholar]
  • 16. Buzsáki G. Large-scale recording of neuronal ensembles. Nature neuroscience. 2004;7(5):446 10.1038/nn1233 [DOI] [PubMed] [Google Scholar]
  • 17. Wise KD, Najafi K. Microfabrication techniques for integrated sensors and microsystems. Science. 1991;254(5036):1335–1342. 10.1126/science.1962192 [DOI] [PubMed] [Google Scholar]
  • 18. Csicsvari J, Henze DA, Jamieson B, Harris KD, Sirota A, Barthó P, et al. Massively parallel recording of unit and local field potentials with silicon-based electrodes. Journal of neurophysiology. 2003;90(2):1314–1323. 10.1152/jn.00116.2003 [DOI] [PubMed] [Google Scholar]
  • 19. Zhang J, Nguyen T, Cogill S, Bhatti A, Luo L, Yang S, et al. A review on cluster estimation methods and their application to neural spike data. Journal of neural engineering. 2018;15(3). 10.1088/1741-2552/aab385 [DOI] [PubMed] [Google Scholar]
  • 20. Veerabhadrappa R, Lim C, Nguyen T, Berk M, Tye S, Monaghan P, et al. Unified selective sorting approach to analyse multi-electrode extracellular data. Scientific reports. 2016;6:28533 10.1038/srep28533 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Khudhair D, Nahavandi S, Garmestani H, Bhatti A. In: Microelectrode Arrays: Architecture, Challenges and Engineering Solutions. Springer; 2017. p. 41–59. [Google Scholar]
  • 22. Veerabhadrappa R, Bhatti A, Berk M, Tye SJ, Nahavandi S. Hierarchical estimation of neural activity through explicit identification of temporally synchronous spikes. Neurocomputing. 2017;249:299–313. 10.1016/j.neucom.2016.09.135 [DOI] [Google Scholar]
  • 23. Hettiarachchi IT, Lakshmanan S, Bhatti A, Lim C, Prakash M, Balasubramaniam P, et al. Chaotic synchronization of time-delay coupled Hindmarsh–Rose neurons via nonlinear control. Nonlinear dynamics. 2016;86(2):1249–1262. 10.1007/s11071-016-2961-4 [DOI] [Google Scholar]
  • 24. Rey HG, Pedreira C, Quiroga RQ. Past, present and future of spike sorting techniques. Brain research bulletin. 2015;119:106–117. 10.1016/j.brainresbull.2015.04.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Kreuz T, Chicharro D, Houghton C, Andrzejak RG, Mormann F. Monitoring spike train synchrony. Journal of neurophysiology. 2012;109(5):1457–1472. 10.1152/jn.00873.2012 [DOI] [PubMed] [Google Scholar]
  • 26. Brown EN, Kass RE, Mitra PP. Multiple neural spike train data analysis: state-of-the-art and future challenges. Nature neuroscience. 2004;7(5):456 10.1038/nn1228 [DOI] [PubMed] [Google Scholar]
  • 27. Einevoll GT, Franke F, Hagen E, Pouzat C, Harris KD. Towards reliable spike-train recordings from thousands of neurons with multielectrodes. Current opinion in neurobiology. 2012;22(1):11–17. 10.1016/j.conb.2011.10.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Zhou H, Mohamed S, Bhatti A, Lim CP, Gu N, Haggag S, et al. Spike sorting using hidden markov models. In: International Conference on Neural Information Processing. Springer;. p. 553–560.
  • 29. Choi JH, Jung HK, Kim T. A new action potential detector using the MTEO and its effects on spike sorting systems at low signal-to-noise ratios. IEEE Transactions on Biomedical Engineering. 2006;53(4):738–746. 10.1109/TBME.2006.870239 [DOI] [PubMed] [Google Scholar]
  • 30. Paralikar KJ, Rao CR, Clement RS. New approaches to eliminating common-noise artifacts in recordings from intracortical microelectrode arrays: Inter-electrode correlation and virtual referencing. Journal of neuroscience methods. 2009;181(1):27–35. 10.1016/j.jneumeth.2009.04.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31. Takekawa T, Ota K, Murayama M, Fukai T. Spike detection from noisy neural data in linear-probe recordings. European Journal of Neuroscience. 2014;39(11):1943–1950. 10.1111/ejn.12614 [DOI] [PubMed] [Google Scholar]
  • 32. Gibson S, Judy JW, Marković D. Spike sorting: The first step in decoding the brain: The first step in decoding the brain. IEEE Signal processing magazine. 2012;29(1):124–143. 10.1109/MSP.2011.941880 [DOI] [Google Scholar]
  • 33. Abeles M, Goldstein MH. Multispike train analysis. Proceedings of the IEEE. 1977;65(5):762–773. 10.1109/PROC.1977.10559 [DOI] [Google Scholar]
  • 34. Abe S. In: Feature selection and extraction. Springer; 2010. p. 331–341. [Google Scholar]
  • 35. Adamos DA, Kosmidis EK, Theophilidis G. Performance evaluation of PCA-based spike sorting algorithms. Computer methods and programs in biomedicine. 2008;91(3):232–244. 10.1016/j.cmpb.2008.04.011 [DOI] [PubMed] [Google Scholar]
  • 36. Zamani M, Demosthenous A. Feature extraction using extrema sampling of discrete derivatives for spike sorting in implantable upper-limb neural prostheses. IEEE Transactions on Neural Systems and Rehabilitation Engineering. 2014;22(4):716–726. 10.1109/TNSRE.2014.2309678 [DOI] [PubMed] [Google Scholar]
  • 37. Shoham S, Fellows MR, Normann RA. Robust, automatic spike sorting using mixtures of multivariate t-distributions. Journal of neuroscience methods. 2003;127(2):111–122. 10.1016/S0165-0270(03)00120-1 [DOI] [PubMed] [Google Scholar]
  • 38. Lagerlund TD, Sharbrough FW, Busacker NE. Spatial filtering of multichannel electroencephalographic recordings through principal component analysis by singular value decomposition. Journal of clinical neurophysiology. 1997;14(1):73–82. 10.1097/00004691-199701000-00007 [DOI] [PubMed] [Google Scholar]
  • 39. Takekawa T, Isomura Y, Fukai T. Accurate spike sorting for multi-unit recordings. European Journal of Neuroscience. 2010;31(2):263–272. 10.1111/j.1460-9568.2009.07068.x [DOI] [PubMed] [Google Scholar]
  • 40. Özkaramanli H, Bhatti A, Bilgehan B. Multi-wavelets from B-spline super-functions with approximation order. Signal processing. 2002;82(8):1029–1046. 10.1016/S0165-1684(02)00212-8 [DOI] [Google Scholar]
  • 41.Bhatti A, Ozkaramanli H. M-band multi-wavelets from spline super functions with approximation order. In: Acoustics, Speech, and Signal Processing (ICASSP), 2002 IEEE International Conference on. vol. 4. IEEE;. p. IV–4172–IV–4172.
  • 42. Hulata E, Segev R, Ben-Jacob E. A method for spike sorting and detection based on wavelet packets and Shannon’s mutual information. Journal of neuroscience methods. 2002;117(1):1–12. 10.1016/S0165-0270(02)00032-8 [DOI] [PubMed] [Google Scholar]
  • 43. Hulata E, Segev R, Shapira Y, Benveniste M, Ben-Jacob E. Detection and sorting of neural spikes using wavelet packets. Physical review letters. 2000;85(21):4637 10.1103/PhysRevLett.85.4637 [DOI] [PubMed] [Google Scholar]
  • 44.Hartigan JA. Clustering algorithms. 1975;.
  • 45.Steinbach M, Karypis G, Kumar V. A comparison of document clustering techniques. In: KDD workshop on text mining. vol. 400. Boston;. p. 525–526.
  • 46. Lewicki MS. A review of methods for spike sorting: the detection and classification of neural action potentials. Network: Computation in Neural Systems. 1998;9(4):R53–R78. 10.1088/0954-898X_9_4_001 [DOI] [PubMed] [Google Scholar]
  • 47.Wehr M, Pezarisl J, Sahani M. Spike Sorting Algorithms;.
  • 48.Eick CF, Zeidat N, Zhao Z. Supervised clustering-algorithms and benefits. In: Tools with Artificial Intelligence, 2004. ICTAI 2004. 16th IEEE International Conference on. IEEE;. p. 774–776.
  • 49.Jain AK, Dubes RC. Algorithms for clustering data. 1988;.
  • 50.Zhao Z. Evolutionary Computing and Splitting Algorithms for Supervised Clustering [Thesis]; 2004.
  • 51.Gibson S, Judy JW, Markovic D. Comparison of spike-sorting algorithms for future hardware implementation. In: Engineering in Medicine and Biology Society, 2008. EMBS 2008. 30th Annual International Conference of the IEEE. IEEE;. p. 5015–5020. [DOI] [PubMed]
  • 52. Stevenson IH, Kording KPJNn. How advances in neural recording affect data analysis. 2011;14(2):139. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Hassan MU, Veerabhadrappa R, Zhang J, Bhatti A. Robust Optimal Parameter Estimation (OPE) for Unsupervised Clustering of Spikes Using Neural Networks. In: 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC). IEEE; 2020. p. 1286–1291.
  • 54. Veerabhadrappa R, Ul Hassan M, Zhang J, Bhatti A. Compatibility evaluation of clustering algorithms for contemporary extracellular neural spike sorting. Frontiers in systems neuroscience. 2020;14:34 10.3389/fnsys.2020.00034 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55. Wouters J, Kloosterman F, Bertrand A. Towards online spike sorting for high-density neural probes using discriminative template matching with suppression of interfering spikes. Journal of neural engineering. 2018;15(5):056005 10.1088/1741-2552/aace8a [DOI] [PubMed] [Google Scholar]
  • 56.Rakesh Veerabhadrappa JZAB Masood Ul Hassan. Compliance Assessment of Clustering Algorithms for Future Contemporary Extracellular Neural Spike Sorting. Frontiers in Systems Neuroscience. 2020;. [DOI] [PMC free article] [PubMed]
  • 57. Wild J, Prekopcsak Z, Sieger T, Novak D, Jech RJJonm. Performance comparison of extracellular spike sorting algorithms for single-channel recordings. 2012;203(2):369–376. [DOI] [PubMed] [Google Scholar]
  • 58. Chung JE, Magland JF, Barnett AH, Tolosa VM, Tooker AC, Lee KY, et al. A fully automated approach to spike sorting. Neuron. 2017;95(6):1381–1394. 10.1016/j.neuron.2017.08.030 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Chen X, Cai D. Large scale spectral clustering with landmark-based representation. In: Twenty-Fifth AAAI Conference on Artificial Intelligence;.
  • 60.LeCun Y. The MNIST database of handwritten digits. http://yann.lecun.com/exdb/mnist/;.
  • 61. Bache K, Lichman M. UCI Machine Learning Repository [http://archive.ics.uci.edu/ml]. Irvine, CA: University of California; School of information and computer science. 2013;28. [Google Scholar]
  • 62. Duarte MF, Hu YH. Vehicle classification in distributed sensor networks. Journal of Parallel and Distributed Computing. 2004;64(7):826–838. 10.1016/j.jpdc.2004.03.020 [DOI] [Google Scholar]
  • 63. Napoleon D, Pavalakodi SJIJoCA. A new method for dimensionality reduction using k-means clustering algorithm for high dimensional data set. 2011;13(7):41–46. [Google Scholar]
  • 64. Killick R, Fearnhead P, Eckley IAJJotASA. Optimal detection of changepoints with a linear computational cost. 2012;107(500):1590–1598. [Google Scholar]
  • 65. Pachitariu M, Steinmetz N, Kadir S, Carandini M, Harris KD. Kilosort: realtime spike-sorting for extracellular electrophysiology with hundreds of channels. BioRxiv. 2016; p. 061481. [Google Scholar]
  • 66. Dokmanic I, Parhizkar R, Ranieri J, Vetterli MJISPM. Euclidean distance matrices: essential theory, algorithms, and applications. 2015;32(6):12–30. [Google Scholar]
  • 67. Drezner Z, Turel O, Zerom D. A modified Kolmogorov–Smirnov test for normality. Communications in Statistics—Simulation and Computation®. 2010;39(4):693–704. 10.1080/03610911003615816 [DOI] [Google Scholar]
  • 68. Mbah AK, Paothong A. Shapiro–Francia test compared to other normality test using expected p-value. Journal of Statistical Computation and Simulation. 2015;85(15):3002–3016. 10.1080/00949655.2014.947986 [DOI] [Google Scholar]
  • 69. Yazici B, Yolacan S. A comparison of various tests of normality. Journal of Statistical Computation and Simulation. 2007;77(2):175–183. 10.1080/10629360600678310 [DOI] [Google Scholar]
  • 70. Mishra P, Pandey CM, Singh U, Gupta A, Sahu C, Keshri A. Descriptive statistics and normality tests for statistical data. Annals of cardiac anaesthesia. 2019;22(1):67 10.4103/aca.ACA_157_18 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71. Hubert M, Van der Veeken S. Outlier detection for skewed data. Journal of Chemometrics: A Journal of the Chemometrics Society. 2008;22(3-4):235–246. 10.1002/cem.1123 [DOI] [Google Scholar]
  • 72. Rousseeuw PJ, Hubert M. Robust statistics for outlier detection. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery. 2011;1(1):73–79. [Google Scholar]
  • 73. Aksenova TI, Chibirova OK, Dryga OA, Tetko IV, Benabid AL, Villa AE. An unsupervised automatic method for sorting neuronal spike waveforms in awake and freely moving animals. Methods. 2003;30(2):178–187. 10.1016/S1046-2023(03)00079-3 [DOI] [PubMed] [Google Scholar]
  • 74.Ester M, Kriegel HP, Sander J, Xu X. A density-based algorithm for discovering clusters in large spatial databases with noise. In: Kdd. vol. 96;. p. 226–231.
  • 75. Lloyd S. Least squares quantization in PCM. IEEE transactions on information theory. 1982;28(2):129–137. 10.1109/TIT.1982.1056489 [DOI] [Google Scholar]
  • 76. Park HS, Jun CH. A simple and fast algorithm for K-medoids clustering. Expert systems with applications. 2009;36(2):3336–3341. 10.1016/j.eswa.2008.01.039 [DOI] [Google Scholar]
  • 77. Bezdek JC, Ehrlich R, Full W. FCM: The fuzzy c-means clustering algorithm. Computers & Geosciences. 1984;10(2-3):191–203. 10.1016/0098-3004(84)90020-7 [DOI] [Google Scholar]
  • 78. Corduneanu A, Bishop CM. Variational Bayesian model selection for mixture distributions In: Artificial intelligence and Statistics. vol. 2001 Morgan Kaufmann; Waltham, MA;. p. 27–34. [Google Scholar]
  • 79. Law MH, Figueiredo MA, Jain AK. Simultaneous feature selection and clustering using mixture models. IEEE transactions on pattern analysis and machine intelligence. 2004;26(9):1154–1166. 10.1109/TPAMI.2004.71 [DOI] [PubMed] [Google Scholar]
  • 80.Davidson I, Ravi S. Agglomerative hierarchical clustering with constraints: Theoretical and empirical results. In: European Conference on Principles of Data Mining and Knowledge Discovery. Springer;. p. 59–70.
  • 81.Zhang T, Ramakrishnan R, Livny M. BIRCH: an efficient data clustering method for very large databases. In: ACM Sigmod Record. vol. 25. ACM;. p. 103–114.
  • 82.Ankerst M, Breunig MM, Kriegel HP, Sander J. OPTICS: ordering points to identify the clustering structure. In: ACM Sigmod record. vol. 28. ACM;. p. 49–60.
  • 83. Quiroga RQ. Concept cells: the building blocks of declarative memory functions. Nature Reviews Neuroscience. 2012;13(8):587 10.1038/nrn3251 [DOI] [PubMed] [Google Scholar]
  • 84. Story M, Congalton RG. Accuracy assessment: a user’s perspective. Photogrammetric Engineering and remote sensing. 1986;52(3):397–399. [Google Scholar]
  • 85.Do TT, Gan L, Nguyen N, Tran TD. Sparsity adaptive matching pursuit algorithm for practical compressed sensing. In: 2008 42nd Asilomar Conference on Signals, Systems and Computers. IEEE; 2008. p. 581–587.
  • 86. Ben-David A. A lot of randomness is hiding in accuracy. Engineering Applications of Artificial Intelligence. 2007;20(7):875–885. 10.1016/j.engappai.2007.01.001 [DOI] [Google Scholar]
  • 87. Dunham MH. Data mining: Introductory and advanced topics. Pearson Education; India; 2006. [Google Scholar]
  • 88.Lansey JC. Beautiful and distinguishable line colors colormap—File Exchange - MATLAB Central;. Available from: https://au.mathworks.com/matlabcentral/fileexchange/42673-beautiful-and-distinguishable-line-colors-colormap.

Decision Letter 0

Gennady Cymbalyuk

15 Oct 2019

PONE-D-19-23638

Efficient Neural Spike Sorting using Data Subdivision and Unification

PLOS ONE

Dear Dr. Bhatti,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

(1) Abstract.

What do you mean by "big data dynamics"? This term is ambiguous. Please change this sentence and the next one, immediately following. Rephrase in order to avoid reference to other algorithms reported in the literature if the Authors do not cite explicitly which ones. The current reference is too general and inappropriate.

How many datasets did you use?

(2) Introduction.

Please, focus the introduction on the problems addressed and thoroughly review the literature and the current state of the art in the field.

1. The review of the literature is not complete, because it missed one key important paper related to this topic, in particular because that paper has introduced for the first time a series of steps that are very close, if not identical, to the steps of data subdivision, clusters formed for each sub-set, unification process by merging neighbor clusters in feature space, thus achieving unified clusters in the end. This paper is the following:

- Aksenova TI, Chibirova OK, Dryga OA, Tetko IV, Benabid AL, Villa AE. An unsupervised automatic method for sorting neuronal spike waveforms in awake and freely moving animals. Methods. 2003; 30(2):178-187. doi: 10.1016/S1046-2023(03)00079-3 : this is the very first paper (2003) to describe unsupervised neural spike sorting based on a fast implementation suitable for real-time application for high-density neural probes.

With respect to application of spike sorting to online experimental procedures, the Authors should also mention:

- Abeles M, Goldstein MH. Multispike train analysis. Proceedings of the IEEE. 1977; 65(5):762-773. doi:10.1109/PROC.1977.10559 : this is a seminal paper (1977) for detecting and identifying the spikes in multispike trains based on signal detection by template matching.

- Wouters J, Kloosterman F, Bertrand A. Towards online spike sorting for high-density neural probes using discriminative template matching with suppression of interfering spikes. J Neural Eng. 2018; 15(5):056005. doi: 10.1088/1741-2552/aace8a : a fast and computationally cheap method for real-time applications.

Consider  recently developed spike sorting algorithms :

Chung, Jason E., Jeremy F. Magland, Alex H. Barnett, Vanessa M. Tolosa, Angela C. Tooker, Kye Y. Lee, Kedar G. Shah, Sarah H. Felix, Loren M. Frank, and Leslie F. Greengard. "A fully automated approach to spike sorting." Neuron 95, no. 6 (2017): 1381-1394.

A more satisfactory review of the literature should also include:

- Zamani M, Demosthenous A. (2014) Feature extraction using extrema sampling of discrete derivatives for spike sorting in implantable upper-limb neural prostheses. IEEE Trans Neural Syst Rehabil Eng. 2014 Jul;22(4):716-726. doi: 10.1109/TNSRE.2014.2309678.

(3) Materials and Methods.

The Authors mention several times the problem of noisy recordings, but they do not examine which types of noise --and/or artifacts-- are present and the methods to face this problem that have been described in the recent literature.

A better way to compare the methods presented by the Authors in their Table 2 and Table 3 could have been to add several known levels of noise to the same benchmarked data set and see how performances and accuracies allow to discriminate the most robust algorithms.

To this end, the Authors should consider these papers:

- Choi JH, Jung HK, Kim T. (2006) A new action potential detector using the MTEO and its effects on spike sorting systems at low signal-to-noise ratios. IEEE Trans Biomed Eng. 2006 Apr;53(4):738-46. doi: 10.1109/TBME.2006.870239

- Paralikar KJ, Rao CR, Clement RS. (2009) New approaches to eliminating common-noise artifacts in recordings from intracortical microelectrode arrays: inter-electrode correlation and virtual referencing. J Neurosci Methods. 2009 Jun 30;181(1):27-35. doi: 10.1016/j.jneumeth.2009.04.014.

- Pillow JW1, Shlens J, Chichilnisky EJ, Simoncelli EP. (2013) A model-based spike sorting algorithm for removing correlation artifacts in multi-neuron recordings. PLoS One. 2013 May 3;8(5):e62123. doi: 10.1371/journal.pone.0062123.

- Takekawa T, Ota K, Murayama M, Fukai T. (2014) Spike detection from noisy neural data in linear-probe recordings. Eur J Neurosci. 2014 Jun;39(11):1943-50. doi: 10.1111/ejn.12614:  an older reference to Takekawa is provided but it should be replaced by this one .

The Authors discuss Spike sorting accuracy (Subsection 3.5) but false alarm ratio is also an extremely important feature to be considered (and discussed in several papers cited above) for the evaluation of the quality of neural spike sorting.

(4) Results.

The Authors should provide the MATLAB codes, with the description of the MATLAB version and environment, of their algorithms.  They compare many methods developed elsewhere and it is of paramount importance to assess that the Authors' implementation follows exactly the algorithms cited in the literature.

A test against a surrogate data set could also be informative for the readers to be convinced of their superior efficiency in the spike sorting procedure claimed by the Authors.

-Optimal length: describe how relevant it is to have the 'optimal length'.

Please, substantiate: 'OL parameter is dependent on the algorithm type rather than on the data dynamics.' The spiking rates may vary by 2 orders of magnitude, so you may end up with clusters that simply don't have enough spikes?

Clarify labeling in Figure 4.

Unification of subclusters:

Describe in detail how you account for differing variances in different dimensions (i.e. principal components).   Explain what 'the standard distribution and normal distribution curves' are.

In general, describe how this technique is applied to the data. Do you apply it to sequential segments, blocks of segments or pairwise across the recording?

Performance evaluation:

Why do you choose two examples where both the conventional and your method do not work for showing performance improvement?

Figure 6: Why those spline fits? Suggests that the different methods are related, please, explain.

Table 3: Numbers suggest a very high accuracy, and no error estimate is given. How did you achieve such a high precision? K-means for example is known to give very different results in different runs. Are these averages over multiple runs? And does the K-means example involve multiple runs to obtain stable clusters? Which of these algorithms converge to the same result every time they are run? Could part of your accuracy improvement be due to running K-means more often, effectively averaging results?

Figure 7:

Lines/symbols are overlapping to an extent that this figure becomes uninformative. Maybe separate plots or cluster centroids for different segments? Please, provide a plot showing the temporal stationarity of firing rates (for different segments).

Please, clarify description of the algorithm concerning temporal speedup. What is the advantage of independent clustering? How does your method compare to a density based approach?

clustering accuracy:

The measure you are using puts a higher weight on large clusters with a lot of spikes. In many datasets, these are multiunit clusters that are hard to separate. It would be nice to have some measure of temporal stationarity.

We would appreciate receiving your revised manuscript by Nov 29 2019 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Gennady Cymbalyuk, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1) Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2)  Thank you for stating the following in the Acknowledgments Section of your manuscript:

[The research work is fully supported by Neural and Cognitive Systems Lab at Institute

for Intelligent Systems Research and Innovation, Deakin University.]

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

 [The author(s) received no specific funding for this work.]

Please include the updated Funding Statement in your cover letter. We will change the online submission form on your behalf.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: No

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: N/A

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors address a relevant problem in spike sorting, namely how to deal with datasets from recordings that become increasingly long due to technological advances in recording techniques.

However, the presentation of their results and statistical analyses do not allow me to make any judgements about the validity of their approach. In fact, I believe that the amount of changes that would be necessary for a revised version of this manuscript would effectively amount to a resubmission of the manuscript.

Specifically, the introduction consists of a rather broad discussion about measuring brain activity and its relevance (not immediately related to the manuscript), but almost completely ignores recently developed spike sorting algorithms (e.g.

Chung, Jason E., Jeremy F. Magland, Alex H. Barnett, Vanessa M. Tolosa, Angela C. Tooker, Kye Y. Lee, Kedar G. Shah, Sarah H. Felix, Loren M. Frank, and Leslie F. Greengard. "A fully automated approach to spike sorting." Neuron 95, no. 6 (2017): 1381-1394.

and the sorting algorithms they use for comparison). It would good to have an introduction more specific to the manuscript and especially describing the current state of the art in the field.

-Optimal length: what I missed here is a discussion of how relevant it is to have the 'optimal length'. Can I be off by a factor of 2 and it doesn't really matter?

Also, I'm not sure how the authors come up with this claim: 'OL parameter is dependent on the algorithm type rather than on the data dynamics.' This may be the case in machine learning examples, but here spiking rates may vary by 2 orders of magnitude, so you may end up with clusters that simply don't have enough spikes?

Labelling in Figure 4 is messy, I don't understand what is plotted.

Unification of subclusters:

I don't understand how you account for differing variances in different dimensions (i.e. principal components). And for distances, in 1D, the 95% claim is valid, but here you're talking about volumes. And I'm completely lost about what 'the standard distribution and normal distribution curves' are.

In general, I'm wondering how this technique is applied to the data. Do you apply it to sequential segments, blocks of segments or pairwise across the recording?

Performance evaluation:

Most strikingly, why do you choose two examples where both the conventional and your method do not work for showing performance improvement? Seems not relevant to the reader.

Figure 6: Why those spline fits? Suggests that the different methods are related, but I do not see how.

Table 3: Numbers suggest a very high accuracy, and no error estimate is given. How did you achieve such a high precision? K-means for example is known to give very different results in different runs. Are these averages over multiple runs? And does the K-means example involve multiple runs to obtain stable clusters? Which of these algorithms converge to the same result every time they are run? I am also concerned that part of your accuracy improvement might be due to running K-means more often, effectively averaging results.

Figure 7:

Lines/symbols are overlapping to an extent that this figure becomes uninformative. Maybe separate plots or cluster centroids for different segments? What I am missing here is also a plot showing the temporal stationarity of firing rates (for different segments).

temporal speedup:

If I understood things correctly (and I'm not sure I did), PCA/Wavelet is run on the whole dataset to obtain low dimensional representations of spikes. Then batches of N spikes are clustered. That sounds similar to what Kilosort does, except that batches are used for optimizing clusters rather than clustering them independently. What is the advantage of independent clustering? Mountainsort on the other hand follows a density based approach, which also seems to scale quite well with recording size. How does your method compare to a density based approach?

clustering accuracy:

The measure you are using puts a higher weight on large clusters with a lot of spikes. In many datasets, these are multiunit clusters that are hard to separate. Also, It would be nice to have some measure of temporal stationarity.

Abstract:

6 or 3 datasets?

Reviewer #2: GENERAL COMMENTS

---------------------------------

The Authors present an interesting manuscript about an efficient method to apply spike sorting on large data sets -- in the order of several hundreds of multiple spike trains recorded simultaneously. This topic is central for any project aimed at real-time decoding of brain activity, in particular for brain-machine interfaces. The paper is well written and reads easily. The main principles and methods are clearly presented and the figures are well done. The recommendation is to accept the paper, but there are few suggested corrections to introduce and the paper may be accepted only after appropriate amendments are introduced.

SPECIFIC COMMENTS

---------------------------------

(1) Abstract.

The Authors use the expression "big data dynamics". What does it mean? This sounds a bit weird because it may assume so many different meanings. Please change this sentence and the next one, immediately following. Rephrase in order to avoid reference to other algorithms reported in the literature if the Authors do not cite explicitely which ones. The current reference is too general and inappropriate.

(2) Introduction.

1. The review of the literature is not complete, because it missed one key important paper related to this topic, in particular because that paper has introduced for the first time a series of steps that are very close, if not identical, to the steps of data subdivision, clusters formed for each sub-set, unification process by merging neighbor clusters in feature space, thus achieving unified clusters in the end. This paper is the following:

- Aksenova TI, Chibirova OK, Dryga OA, Tetko IV, Benabid AL, Villa AE. An unsupervised automatic method for sorting neuronal spike waveforms in awake and freely moving animals. Methods. 2003; 30(2):178-187. doi: 10.1016/S1046-2023(03)00079-3 : this is the very first paper (2003) to describe unsupervised neural spike sorting based on a fast implementation suitable for real-time application for high-density neural probes.

With respect to application of spike sorting to online experimental procedures, the Authors should also mention:

- Abeles M, Goldstein MH. Multispike train analysis. Proceedings of the IEEE. 1977; 65(5):762-773. doi:10.1109/PROC.1977.10559 : this is a seminal paper (1977) for detecting and identifying the spikes in multispike trains based on signal detection by template matching.

- Wouters J, Kloosterman F, Bertrand A. Towards online spike sorting for high-density neural probes using discriminative template matching with suppression of interfering spikes. J Neural Eng. 2018; 15(5):056005. doi: 10.1088/1741-2552/aace8a : a fast and computationally cheap method for real-time applications.

A more satisfactory review of the literature should also include:

- Zamani M, Demosthenous A. (2014) Feature extraction using extrema sampling of discrete derivatives for spike sorting in implantable upper-limb neural prostheses. IEEE Trans Neural Syst Rehabil Eng. 2014 Jul;22(4):716-726. doi: 10.1109/TNSRE.2014.2309678.

(3) Materials and Methods.

The Authors mention several times the problem of noisy recordings, but they do not examine which types of noise --and/or artifacts-- are present and the methods to face this problem that have been described in the recent literature.

A better way to compare the methods presented by the Authors in their Table 2 and Table 3 could have been to add several known levels of noise to the same benchmarked data set and see how performances and accuracies allow to discriminate the most robust algorithms.

To this end, the Authors should consider these papers:

- Choi JH, Jung HK, Kim T. (2006) A new action potential detector using the MTEO and its effects on spike sorting systems at low signal-to-noise ratios. IEEE Trans Biomed Eng. 2006 Apr;53(4):738-46. doi: 10.1109/TBME.2006.870239

- Paralikar KJ, Rao CR, Clement RS. (2009) New approaches to eliminating common-noise artifacts in recordings from intracortical microelectrode arrays: inter-electrode correlation and virtual referencing. J Neurosci Methods. 2009 Jun 30;181(1):27-35. doi: 10.1016/j.jneumeth.2009.04.014.

- Pillow JW1, Shlens J, Chichilnisky EJ, Simoncelli EP. (2013) A model-based spike sorting algorithm for removing correlation artifacts in multi-neuron recordings. PLoS One. 2013 May 3;8(5):e62123. doi: 10.1371/journal.pone.0062123.

- Takekawa T, Ota K, Murayama M, Fukai T. (2014) Spike detection from noisy neural data in linear-probe recordings. Eur J Neurosci. 2014 Jun;39(11):1943-50. doi: 10.1111/ejn.12614: an older reference to Takekawa is provided but it should be replaced by this one .

The Authors discuss Spike sorting accuracy (Subsection 3.5) but false alarm ratio is also an extremely important feature to be considered (and discussed in several papers cited above) for the evaluation of the quality of neural spike sorting.

(4) Results.

The Authors should provide the MATLAB codes, with the description of the MATLAB version and environment, of their algorithms. They compare many methods developed elsewhere and it is of paramount importance to assess that the Authors' implementation follows exactly the algorithms cited in the literature.

A test against a surrogate data set could also be informative for the readers to be convinced of their superior efficiency in the spike sorting procedure claimed by the Authors.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Feb 10;16(2):e0245589. doi: 10.1371/journal.pone.0245589.r002

Author response to Decision Letter 0


21 Nov 2019

Below are the suggested revisions according to valuable comments from the reviewers.

1) Abstract.

1. What do you mean by "big data dynamics"? This term is ambiguous. Please change this sentence and the next one, immediately following. Rephrase in order to avoid reference to other algorithms reported in the literature if the Authors do not cite explicitly which ones. The current reference is too general and inappropriate.

Author Response: The manuscript has been updated according to reviewers comment (Line 16 to 20).

2. How many datasets did you use?

Author Response: The manuscript has been updated according to reviewers comment (Line 24 to 26).

2) Introduction.

Please, focus the introduction on the problems addressed and thoroughly review the literature and the current state of the art in the field.

1. The review of the literature is not complete, because it missed one key important paper related to this topic, in particular because that paper has introduced for the first time a series of steps that are very close, if not identical, to the steps of data subdivision, clusters formed for each sub-set, unification process by merging neighbor clusters in feature space, thus achieving unified clusters in the end. This paper is the following:

Aksenova TI, Chibirova OK, Dryga OA, Tetko IV, Benabid AL, Villa AE. An unsupervised automatic method for sorting neuronal spike waveforms in awake and freely moving animals. Methods. 2003; 30(2):178-187. doi: 10.1016/S1046-2023(03)00079-3 : this is the very first paper (2003) to describe unsupervised neural spike sorting based on a fast implementation suitable for real-time application for high-density neural probes.

Author Response: The manuscript has been updated according to reviewers comment (Line 232 to 234).

2. With respect to application of spike sorting to online experimental procedures, the Authors should also mention:

a) Abeles M, Goldstein MH. Multispike train analysis. Proceedings of the IEEE. 1977; 65(5):762-773. doi:10.1109/PROC.1977.10559 : this is a seminal paper (1977) for detecting and identifying the spikes in multispike trains based on signal detection by template matching.

Author Response: The manuscript has been updated according to reviewers comment (Line 87 to 90).

b) Wouters J, Kloosterman F, Bertrand A. Towards online spike sorting for high-density neural probes using discriminative template matching with suppression of interfering spikes. J Neural Eng. 2018; 15(5):056005. doi: 10.1088/1741-2552/aace8a : a fast and computationally cheap method for real-time applications.

Author Response: The manuscript has been updated according to reviewers comment (Line 109 to 111).

3. Consider recently developed spike sorting algorithms :

Chung, Jason E., Jeremy F. Magland, Alex H. Barnett, Vanessa M. Tolosa, Angela C. Tooker, Kye Y. Lee, Kedar G. Shah, Sarah H. Felix, Loren M. Frank, and Leslie F. Greengard. "A fully automated approach to spike sorting." Neuron 95, no. 6 (2017): 1381-1394.

Author Response: The manuscript has been updated according to reviewers comment (Line 123 to 127).

4. A more satisfactory review of the literature should also include:

Zamani M, Demosthenous A. (2014) Feature extraction using extrema sampling of discrete derivatives for spike sorting in implantable upper-limb neural prostheses. IEEE Trans Neural Syst Rehabil Eng. 2014 Jul;22(4):716-726. doi: 10.1109/TNSRE.2014.2309678.

Author Response: The manuscript has been updated according to reviewers comment (Line 95 to 98).

3) Materials and Methods.

The Authors mention several times the problem of noisy recordings, but they do not examine which types of noise --and/or artifacts-- are present and the methods to face this problem that have been described in the recent literature. A better way to compare the methods presented by the Authors in their Table 2 and Table 3 could have been to add several known levels of noise to the same benchmarked data set and see how performances and accuracies allow to discriminate the most robust algorithms.

1. To this end, the Authors should consider these papers:

Choi JH, Jung HK, Kim T. (2006) A new action potential detector using the MTEO and its effects on spike sorting systems at low signal-to-noise ratios. IEEE Trans Biomed Eng. 2006 Apr;53(4):738-46. doi: 10.1109/TBME.2006.870239

Paralikar KJ, Rao CR, Clement RS. (2009) New approaches to eliminating common-noise artifacts in recordings from intracortical microelectrode arrays: inter-electrode correlation and virtual referencing. J Neurosci Methods. 2009 Jun 30;181(1):27-35. doi: 10.1016/j.jneumeth.2009.04.014.

Pillow JW1, Shlens J, Chichilnisky EJ, Simoncelli EP. (2013) A model-based spike sorting algorithm for removing correlation artifacts in multi-neuron recordings. PLoS One. 2013 May 3;8(5):e62123. doi: 10.1371/journal.pone.0062123.

Takekawa T, Ota K, Murayama M, Fukai T. (2014) Spike detection from noisy neural data in linear-probe recordings. Eur J Neurosci. 2014 Jun;39(11):1943-50. doi: 10.1111/ejn.12614: an older reference to Takekawa is provided but it should be replaced by this one .

Author Response: The manuscript has been updated according to reviewers comment (Line 74 to 87).

2. The Authors discuss Spike sorting accuracy (Subsection 3.5) but false alarm ratio is also an extremely important feature to be considered (and discussed in several papers cited above) for the evaluation of the quality of neural spike sorting.

Author Response: The manuscript has been updated according to reviewers comment (Line 270 to 286).

4) Results.

1. The Authors should provide the MATLAB codes, with the description of the MATLAB version and environment, of their algorithms. They compare many methods developed elsewhere and it is of paramount importance to assess that the Authors' implementation follows exactly the algorithms cited in the literature.

Author Response: The manuscript has been updated according to reviewers comment (Line 316 to 324).

2. A test against a surrogate data set could also be informative for the readers to be convinced of their superior efficiency in the spike sorting procedure claimed by the Authors.

Author Response: The manuscript has been updated according to reviewers comment (Line 250 to 256).

3. Optimal length: describe how relevant it is to have the 'optimal length'. What I missed here is a discussion of how relevant it is to have the 'optimal length'. Can I be off by a factor of 2 and it doesn't really matter?

Author Response: The manuscript has been updated according to reviewers comment (Line 205 to 209).

4. Please, substantiate 'OL parameter is dependent on the algorithm type rather than on the data dynamics.' The spiking rates may vary by 2 orders of magnitude, so you may end up with clusters that simply don't have enough spikes?

Author Response: The manuscript has been updated according to reviewers comment (Line 228 to 231).

5. Clarify labeling in Figure 4., Labelling in Figure 4 is messy, I don't understand what is plotted.

Author Response: The figure has been updated according to reviewers comment (Figure 4).

5) Unification of subclusters:

Describe in detail how you account for differing variances in different dimensions (i.e. principal components). Explain what 'the standard distribution and normal distribution curves' are. In general, describe how this technique is applied to the data. Do you apply it to sequential segments, blocks of segments or pairwise across the recording?

I don't understand how you account for differing variances in different dimensions (i.e. principal components). And for distances, in 1D, the 95% claim is valid, but here you're talking about volumes. And I'm completely lost about what 'the standard distribution and normal distribution curves' are.

In general, I'm wondering how this technique is applied to the data. Do you apply it to sequential segments, blocks of segments or pairwise across the recording?

Author Response: The manuscript has been updated according to reviewers comment (Line 220 to 228).

6) Performance evaluation: Why do you choose two examples where both the conventional and your method do not work for showing performance improvement?

Author Response: The manuscript has been updated according to reviewers comment (Line 288 to 295).

7) Figure 6: Why those spline fits? Suggests that the different methods are related, please, explain.

Author Response: The figure has been updated according to reviewers comment (Figure 6).

8) Table 3: Numbers suggest a very high accuracy, and no error estimate is given. How did you achieve such a high precision? K-means for example is known to give very different results in different runs. Are these averages over multiple runs? And does the K-means example involve multiple runs to obtain stable clusters? Which of these algorithms converge to the same result every time they are run? Could part of your accuracy improvement be due to running K-means more often, effectively averaging results?

Author Response: The manuscript has been updated according to reviewers comment (Line 309 to 313).

9) Figure 7: Lines/symbols are overlapping to an extent that this figure becomes uninformative. Maybe separate plots or cluster centroids for different segments? Please, provide a plot showing the temporal stationarity of firing rates (for different segments).

Author Response: The figures has been updated according to reviewer’s comment (Figure 7 and Figure 8).

10) Temporal speedup: Please, clarify description of the algorithm concerning temporal speedup. What is the advantage of independent clustering? How does your method compare to a density based approach?

If I understood things correctly (and I'm not sure I did), PCA/Wavelet is run on the whole dataset to obtain low dimensional representations of spikes. Then batches of N spikes are clustered. That sounds similar to what Kilosort does, except that batches are used for optimizing clusters rather than clustering them independently. What is the advantage of independent clustering? Mountainsort on the other hand follows a density based approach, which also seems to scale quite well with recording size. How does your method compare to a density based approach?

Author Response: The manuscript has been updated according to reviewers comment (Line 149 to 157).

11) Clustering accuracy: The measure you are using puts a higher weight on large clusters with a lot of spikes. In many datasets, these are multiunit clusters that are hard to separate. It would be nice to have some measure of temporal stationarity.

Author Response: The manuscript has been updated according to reviewers comment (Line 221 to 231).

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1) Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

Author Response: The manuscript is according to the style requirements of PLOS One Journal.

2) Thank you for stating the following in the Acknowledgments Section of your manuscript:

[The research work is fully supported by Neural and Cognitive Systems Lab at Institute

for Intelligent Systems Research and Innovation, Deakin University.]

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

[The author(s) received no specific funding for this work.]

Please include the updated Funding Statement in your cover letter. We will change the online submission form on your behalf.

Author Response: Funding related text is removed from the manuscript. We don’t require any updates in the funding statement.

Attachment

Submitted filename: Response to reviewers.pdf

Decision Letter 1

Gennady Cymbalyuk

23 Dec 2019

PONE-D-19-23638R1

Efficient Neural Spike Sorting using Data Subdivision and Unification

PLOS ONE

Dear Dr. Bhatti,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please, seriously revise the manuscript to to clarify the concerns described  below and  fix  typos.

--Figure 7: Lines/symbols are overlapping to an extent that this figure becomes uninformative.

Maybe separate plots or cluster centroids for different segments? Please, provide a plot

showing the temporal stationarity of firing rates (for different segments).

--10) Temporal speedup: Please, clarify description of the algorithm concerning temporal speedup.

What is the advantage of independent clustering? How does your method compare to a

density based approach?

--11) Clustering accuracy: The measure you are using puts a higher weight on large clusters with a

lot of spikes. In many datasets, these are multiunit clusters that are hard to separate. It would

be nice to have some measure of temporal stationarity.

Other comments:

Please, clarify what readers are supposed to see in the example Figs 7+8 Both the conventional and the proposed method seem to produce identical results, but the sequence of plotting the different lines has changed. The bottom graphs are identical. Are these placeholder figures?

-'The surrounding region between −2SD to 2SD, containing about 95 percent of the

cluster data...' This statement is still wrong.

-What are 'Quirogo datasets'?

We would appreciate receiving your revised manuscript by Feb 06 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Gennady Cymbalyuk, Ph.D.

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: No

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: I Don't Know

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I believe the authors still need more time to polish their manuscript. There are a lot of typos, and a few of my previous comments have not been addressed, specifically:

--Figure 7: Lines/symbols are overlapping to an extent that this figure becomes uninformative.

Maybe separate plots or cluster centroids for different segments? Please, provide a plot

showing the temporal stationarity of firing rates (for different segments).

--10) Temporal speedup: Please, clarify description of the algorithm concerning temporal speedup.

What is the advantage of independent clustering? How does your method compare to a

density based approach?

--11) Clustering accuracy: The measure you are using puts a higher weight on large clusters with a

lot of spikes. In many datasets, these are multiunit clusters that are hard to separate. It would

be nice to have some measure of temporal stationarity.

Other comments:

I'm still not sure what I'm supposed to see in the example Figs 7+8 Both the conventional and the proposed method seem to produce identical results, but the sequence of plotting the different lines has changed. The bottom graphs are identical. Are these placeholder figures?

-'The surrounding region between −2SD to 2SD, containing about 95 percent of the

cluster data...' This statement is still wrong.

-What are 'Quirogo datasets'?

Reviewer #2: The manuscript can be processed for publication as is. All comments have been addressed adequately

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Feb 10;16(2):e0245589. doi: 10.1371/journal.pone.0245589.r004

Author response to Decision Letter 1


3 Feb 2020

PONE-D-19-23638

Efficient Neural Spike Sorting using Data Subdivision and Unification

To the Editor,

Prof Gennady Cymbalyuk

Academic Editor, PLOS ONE

We would like to acknowledge and appreciate the efforts and time of the editor and the reviewers for their invaluable comments and suggestions that has allowed us to enhance the quality of our manuscript.

Below are the revisions according to valuable comments from the reviewers.

• Figure 7: Lines/symbols are overlapping to an extent that this figure becomes uninformative. Maybe separate plots or cluster centroids for different segments? Please, provide a plot showing the temporal stationarity of firing rates (for different segments).

Author Response: New figures (Fig 8 and Fig 09) are introduced clearly highlighting the performance difference between conventional and proposed algorithm in terms of computational efficiency and clustering accuracy. Separate cluster of different data segments are shown. Clustering outcome of intermediate steps is also shown to facilitate readers’ understanding of the proposed algorithm.

• Temporal speedup: Please, clarify description of the algorithm concerning temporal speedup. What is the advantage of independent clustering? How does your method compare to a density based approach?

Author Response: A new figure (Figure 4) is added to address this comment, highlighting the differentiation between conventional spike sorting and proposed spike sorting mechanisms. Please refer to lines 132-140, and Table 1 for explanation on the effect of data size on the computational efficiency and temporal speedup of the clustering algorithms. In addition, please refer to lines 168-186 highlighting the comparison between density based approach and proposed method.

• Clustering accuracy: The measure you are using puts a higher weight on large clusters with a lot of spikes. In many datasets, these are multiunit clusters that are hard to separate. It would be nice to have some measure of temporal stationarity.

Author Response: Please refer to accuracy index “A” highlighted in expression (4) and explanation in lines 290-307 where estimation of clustering accuracy is defined. Accuracy index is defined as the percentage of spikes accurately assigned to the relevant cluster, as per the ground truth, with respect to total number of spikes. This makes the accuracy index “A” independent of the number of spikes or the size of the cluster. As the contributing of the proposed mechanism is data processing rather than the clustering, which is adopted from conventional algorithms, therefore commenting on the temporal stationarity is out of the scope of this work.

• 'The surrounding region between −2SD to 2SD, containing about 95 percent of the cluster data...' This statement is still wrong.

What are 'Quirogo datasets'?

Author Response: The manuscript has been updated according to reviewers comment (Line 24 to 26). The reference of the adopted data set and brief explanation is also provide in lines 267-272.

In addition the statement in reference to 2SD is updated, lines 246-247.

Thanks

Asim Bhatti

Attachment

Submitted filename: Rebuttal Letter_R2.pdf

Decision Letter 2

Alexandros Iosifidis

7 May 2020

PONE-D-19-23638R2

Efficient Neural Spike Sorting using Data Subdivision and Unification

PLOS ONE

Dear Dr. Bhatti,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

We would appreciate receiving your revised manuscript by Jun 21 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Alexandros Iosifidis

Academic Editor

PLOS ONE

Additional Editor Comments (if provided):

Please address the comments of Reviewer 1.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: No

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Apart from a methological issue pointed out below (which needs to be discussed), a few missing details in the Methods and some awkward sentences and typos (I probably missed some and would encourage the authors to do another round of proofreading), this manuscript is now in a good shape.

Main points:

-please fix typos and grammatical errors (see below for a list of suggestions).

-I think I'm still missing some crucial information about the analysis. First I thought, that the performance improvement was somewhat related to nonstationarities in the data and you have shown (great, thanks) that this is clearly not the case. Another thing that I kept pointing out in my reviews and is still somewhat misleading in the presentation of the method is that in a high dimensional multivariate gaussian distribution, the probability for a datapoint to be within a 2 sigma radius from the center is not 95% but rather dependent on the number of dimensions, i.e. at most (95%)^d (for L1 norm), where d is the number of dimensions (or PCA components/ features). I haven't really found the number of dimensions you used in the paper (and you really do need to report it, it is a crucial number), but there is one figure suggesting the use of 10 features/dimensions. This seems high to me (and you may want to discuss such a parameter choice in the Discussion), what would have expected from other work would be 3-4 features. In any case, in the 10 feature case, your 2 sigma radius then accounts for at most 60% of the datapoints, so there are a lot of points outside your cluster boundaries. Does that explain why those widely used algorithms are working so poorly? If so, that's fine, but you want to discuss it in the Discussion section. It is also not clear to me how different dimensions are handled and you should elaborate a bit on that in the Methods. Is each dimension scaled such that variances match? If that is the case you're downweighting the first principal component and effectively explaining noisy, low variance features? Or am I missing something more subtle? You're reporting a performance improvement and I still don't see any reason why this should happen and especially why it would happen so consistently, given that all these algorithms have been used very successfully for years. I'm totally fine with the speed improvement and follow the argument that this should happen. But a general classification performance improvement is very hard to believe, so you need to at least report the specific circumstances under which it happens, i.e. the number of features/ dimensions and make clear that you're potentially inflating tiny differences in principal components with small variances (unless you corrected for that in some way, in that case it should be reported). Ideally, you should have some idea about a mechanism for the the performance improvement and discuss it in the Discussion (is it some kind of regularization effect that would be beneficial for noisy data?).

Specifically, do report the number of features/PCA components used. Do make clear whether the standard deviation was estimated for each component separately, thus enhancing the effect of small components, or whether (and how) you accounted for differences in the variances of features/PCA components. Ideally, specify a typical variation between variances of the features/PCA components (e.g. ratio between largest and smallest) and mention whether the results were sensitive to the number of PCA components.

A thorough analysis of the effect of dimensionality and scaling is certainly beyond the scope of this article, but I'm sure you made observations what happens if you change these parameters. You shall discuss them in the Discussion, and maybe even speculate about a mechanism or a scenario that tends to give performance improvements.

-Figure 7 has errorbars now, so please mention briefly how you obtained them/what they reflect. Further, numbers reported suggest a huge precision in comparison to these errorbars. Please round them, and wherever refered to in the text, add the uncertainty in brackets (e.g. 53+-6 %). You may leave the uncertainty in the table for clarity as it is already shown in Figure 7.

Other remarks

Figure 8+9: markers and labels don't match.

ln65: Brain consists

ln105: automatically estimate

ln119: presented data analysis issues due to progressive technological advancements of neural recordings

ln126: Although they have proposed an

efficient method for spike sorting, it still lacks the speed researchers require

ln130:The larger is the size the slower is the speed and large is the computational time required by spike sorting algorithms. -- rephrase or simply leave out (what, other than the obvious, are you trying to say?)

ln133: They reported, (?)

ln136: These second and third order operations prove the non-linear behaviour of spectral clustering.

ln138: To motivate our analysis,

ln141: The dependency of speed and computational time on data

size in spike-sorting has made it very difficult

ln142: identify the total number of

ln144: breast cancer cell data

ln149: Despite these challenges, ...

ln151: However, limited work has considered enhancing computational

ln153: The proposed algorithm pre-processes data to

ln154: time and to enhance speed and efficiency of a wide range

ln156: by parallel computing approaches to further

ln159: The novelty of the proposed mechanism

ln162:The first step involves subdivision of data into data-subsets of optimal length.

ln164:The second step involves clustering spikes in data-subsets

using conventional spike sorting algorithms.

ln165: The last step involves unification

ln166: clusters are then used to label

ln170:of conventional algorithms but rather performs additional data

ln171: the proposed mechanism very versatile and

ln175: uses a density based

The second step involves clustering

The last step involves unification

ln180: overall time of the spike sorting process.

ln193:The total number N of optimal subdivisions is estimated

ln195, 199: ,where L is the

ln223: of the algorithm depends on the length

ln227: (O L ) forms a direct

ln228: and an inverse relationship

the X-axis

the Y-axis

The computational time is the processing time after a movmean

filter (20 datapoints length) filtered the unwanted ripples in the plot and returned smooth curves. (representing computational time (why twice?))

The average value over ten repetitive analyses

robustness of the measure

ln241:'It is observed that, the

variations in data dimensionality does not have any effect on estimating the bounded

region. Whatever is the dimensionality of data, when the ED is calculated, the result is

always a single entry in one-dimensional space. For all EDs The standard deviation

(SD) is calculated using [66] and normal distribution curves are formed based on [67].' --Not even wrong. If you have a multivariate Gaussian distribution, the density distribution as a function of the radius is not Gaussian. The square of the radius (equals the sum of squares of Gaussian distributed random variables and) follows a Chi-squared distribution (check Wikipedia?) and you can imagine (take the cumulative distribution and rescale the x-axis) what follows for the distribution of the radius itself.

Figure 7: Continues positive trend is observed ???

Errorbars represent...

ln344: To cater for schocastic ??? variations of some of the algorithms

ln335: over 10 repetitons.

Reviewer #2: The latest revision and revised/new figures made the manuscript even more clear. The manuscript can be published as is.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Feb 10;16(2):e0245589. doi: 10.1371/journal.pone.0245589.r006

Author response to Decision Letter 2


21 Jun 2020

PONE-D-19-23638-R3

Efficient Neural Spike Sorting using Data Subdivision and Unification

PLOS ONE

To the Editor,

Prof Alexandros Iosifidis

Academic Editor, PLOS ONE

We would like to acknowledge and appreciate the efforts and time of the editor and the reviewers for their invaluable comments and suggestions that has allowed us to enhance the quality of our manuscript.

Below are the suggested revisions according to valuable comments from the reviewers.

1) I think I'm still missing some crucial information about the analysis. First I thought, that the performance improvement was somewhat related to no stationarities in the data and you have shown (great, thanks) that this is clearly not the case. Another thing that I kept pointing out in my reviews and is still somewhat misleading in the presentation of the method is that in a high dimensional multivariate gaussian distribution, the probability for a datapoint to be within a 2 sigma radius from the center is not 95% but rather dependent on the number of dimensions, i.e. at most (95%)^d (for L1 norm), where d is the number of dimensions (or PCA components/ features).

Author Response: The manuscript is updated with additional information, mathematical expressions and references to address the points raised by the review as well as for the ease of the general readers (Line 237 to 252).

2) I haven't really found the number of dimensions you used in the paper (and you really do need to report it, it is a crucial number), but there is one figure suggesting the use of 10 features/dimensions. This seems high to me (and you may want to discuss such a parameter choice in the Discussion), what would have expected from other work would be 3-4 features.

In any case, in the 10 feature case, your 2 sigma radius then accounts for at most 60% of the datapoints, so there are a lot of points outside your cluster boundaries. Does that explain why those widely used algorithms are working so poorly? If so, that's fine, but you want to discuss it in the Discussion section. It is also not clear to me how different dimensions are handled and you should elaborate a bit on that in the Methods. Is each dimension scaled such that variances match? If that is the case you're down weighting the first principal component and effectively explaining noisy, low variance features? Or am I missing something more subtle? You're reporting a performance improvement and I still don't see any reason why this should happen and especially why it would happen so consistently, given that all these algorithms have been used very successfully for years. I'm totally fine with the speed improvement and follow the argument that this should happen. But a general classification performance improvement is very hard to believe, so you need to at least report the specific circumstances under which it happens, i.e. the number of features/ dimensions and make clear that you're potentially inflating tiny differences in principal components with small variances (unless you corrected for that in some way, in that case it should be reported). Ideally, you should have some idea about a mechanism for the performance improvement and discuss it in the Discussion (is it some kind of regularization effect that would be beneficial for noisy data?). Specifically, do report the number of features/PCA components used.

Do make clear whether the standard deviation was estimated for each component separately, thus enhancing the effect of small components, or whether (and how) you accounted for differences in the variances of features/PCA components. Ideally, specify a typical variation between variances of the features/PCA components (e.g. ratio between largest and smallest) and mention whether the results were sensitive to the number of PCA components. A thorough analysis of the effect of dimensionality and scaling is certainly beyond the scope of this article, but I'm sure you made observations what happens if you change these parameters. You shall discuss them in the Discussion, and maybe even speculate about a mechanism or a scenario that tends to give performance improvements.

Author Response: Author Response: Further explanation is added, please refer to lines 282 to 298.

3) Figure 7 has errorbars now, so please mention briefly how you obtained them/what they reflect. Further, numbers reported suggest a huge precision in comparison to these errorbars. Please round them, and wherever refered to in the text, add the uncertainty in brackets (e.g. 53+-6 %). You may leave the uncertainty in the table for clarity as it is already shown in Figure 7.

Author Response: Taking into account reviewer’s comment, Figure 7 is updated to provide simplified performance comparison. Performance outcomes, averaged over 10 repetitions, are presented for simplicity and ease of understanding.

4) Other remarks

Figure 8+9: markers and labels don't match.

ln65: Brain consists

ln105: automatically estimate

ln119: presented data analysis issues due to progressive technological advancements of neural recordings

ln126: Although they have proposed an

efficient method for spike sorting, it still lacks the speed researchers require

ln130:The larger is the size the slower is the speed and large is the computational time required by spike sorting algorithms. -- rephrase or simply leave out (what, other than the obvious, are you trying to say?)

ln133: They reported, (?)

ln136: These second and third order operations prove the non-linear behaviour of spectral clustering.

ln138: To motivate our analysis,

ln141: The dependency of speed and computational time on data

size in spike-sorting has made it very difficult

ln142: identify the total number of

ln144: breast cancer cell data

ln149: Despite these challenges, ...

ln151: However, limited work has considered enhancing computational

ln153: The proposed algorithm pre-processes data to

ln154: time and to enhance speed and efficiency of a wide range

ln156: by parallel computing approaches to further

ln159: The novelty of the proposed mechanism

ln162:The first step involves subdivision of data into data-subsets of optimal length.

ln164:The second step involves clustering spikes in data-subsets

using conventional spike sorting algorithms.

ln165: The last step involves unification

ln166: clusters are then used to label

ln170:of conventional algorithms but rather performs additional data

ln171: the proposed mechanism very versatile and

ln175: uses a density based

The second step involves clustering

The last step involves unification

ln180: overall time of the spike sorting process.

ln193:The total number N of optimal subdivisions is estimated

ln195, 199: ,where L is the

ln223: of the algorithm depends on the length

ln227: (O L ) forms a direct

ln228: and an inverse relationship

the X-axis

the Y-axis

The computational time is the processing time after a movmean

filter (20 datapoints length) filtered the unwanted ripples in the plot and returned smooth curves. (representing computational time (why twice?))

The average value over ten repetitive analyses

robustness of the measure

ln241:'It is observed that, the

variations in data dimensionality does not have any effect on estimating the bounded

region. Whatever is the dimensionality of data, when the ED is calculated, the result is

always a single entry in one-dimensional space. For all EDs The standard deviation

(SD) is calculated using [66] and normal distribution curves are formed based on [67].' --Not even wrong. If you have a multivariate Gaussian distribution, the density distribution as a function of the radius is not Gaussian. The square of the radius (equals the sum of squares of Gaussian distributed random variables and) follows a Chi-squared distribution (check Wikipedia?) and you can imagine (take the cumulative distribution and rescale the x-axis) what follows for the distribution of the radius itself.

Figure 7: Continues positive trend is observed ???

Errorbars represent...

ln344: To cater for schocastic ??? variations of some of the algorithms

ln335: over 10 repetitons.

Author Response: The manuscript has been thoroughly revised taking into account all the comments by the reviewer.

Thanks

Asim Bhatti

Attachment

Submitted filename: PONE-D-19-23638-R3-Response to Reviewers.pdf

Decision Letter 3

Alexandros Iosifidis

3 Aug 2020

PONE-D-19-23638R3

Efficient Neural Spike Sorting using Data Subdivision and Unification

PLOS ONE

Dear Dr. Bhatti,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Sep 17 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Alexandros Iosifidis

Academic Editor

PLOS ONE

Journal Requirements:

Additional Editor Comments (if provided):

Please address the comments of Reviewer 1, by making sure that all definitions in the paper are precise.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Lines 237-253: Please talk to a statistician (or someone who knows English and statistics) and reframe (and please put references from peer reviewed publications). This is about making your paper understandable to a reader, and I'm not asking for a layman explanation here (in fact, I feel that you're trying to explain a lot of things you don't need to explain, I'm asking for correctness. The standard deviation of a random variable is a well known and defined quantity and your equation does not reflect the standard deviation of euclidean distances.

The euclidean distances are strictly positive numbers, but in Figure 6, you're suggesting that they are Gaussian distributed, and therefore negative values are possible. So I really don't understand what you are doing here. And I would also be very interested whether you scale your principal components in some way, to match their variances, or whether the first principal components have larger weights.

The main issue that I raised in the last revision was that when you're working in a 10 dimensional space, things are a little more complicated. For example, if you have a standard normal distribution in 10 dimensions, then the euclidean distances (ED) follow a Chi-square distribution with 10 degrees of freedom (see https://en.wikipedia.org/wiki/Chi-square_distribution).

other comments:

ln. 288 kolmogorov-Smirnov (KS) test --> Kolmogorov-Smirnov (KS) test

ln. 296 Please make clear what you mean by this sentence: 'It is

observed that 10 PCA features ensures the cumulative explained variance of over 85%

up to 95%, in case of the data sets employed in this study.' Please reframe this sentence.

The matlab function-- Please capitalize MATLAB.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Feb 10;16(2):e0245589. doi: 10.1371/journal.pone.0245589.r008

Author response to Decision Letter 3


23 Sep 2020

To the Editor,

Gennady Cymbalyuk

Academic Editor, PLOS ONE

We would like to acknowledge and appreciate the efforts and time of the editor and the reviewers for their invaluable comments and suggestions that has allowed us to enhance the quality of our manuscript.

Below are the suggested revisions according to valuable comments from the reviewers.

1) Lines 237-253: Please talk to a statistician (or someone who knows English and statistics) and reframe (and please put references from peer reviewed publications). This is about making your paper understandable to a reader, and I'm not asking for a layman explanation here (in fact, I feel that you're trying to explain a lot of things you don't need to explain, I'm asking for correctness. The standard deviation of a random variable is a well-known and defined quantity and your equation does not reflect the standard deviation of Euclidean distances.

Author Response: The equation has been updated with a square term that was previously missing in the equation. (Please refer to Eq. 7.)

2) The Euclidean distances are strictly positive numbers, but in Figure 6, you're suggesting that they are Gaussian distributed, and therefore negative values are possible. So I really don't understand what you are doing here.

Author Response: Detailed explanation of probability distribution of Euclidean distances is discussed at lines 273 to 318.

3) And I would also be very interested whether you scale your principal components in some way, to match their variances, or whether the first principal components have larger weights.

Author Response: The PCA components are not scaled to match their explained variances. The individual variances of PCA components are accumulated and the optimal number of PCA components that gives at least 85% of cumulative explained variance are chosen for the analysis. 10 PCA features are required to get at least 85% cumulative explained variance of the 64 dimensional spikes data. (Lines 353 to 358)

4) The main issue that I raised in the last revision was that when you're working in a 10 dimensional space, things are a little more complicated. For example, if you have a standard normal distribution in 10 dimensions, then the Euclidean distances (ED) follow a Chi-square distribution with 10 degrees of freedom (see https://en.wikipedia.org/wiki/Chi-square_distribution).

Author Response: The manuscript is updated with additional information, mathematical expressions, figures and references to address the points raised by the review as well as for the ease of the general readers. (Fig 6, Fig 7 and Lines 234 to 318).

5) Other comments:

i. ln. 288 kolmogorov-Smirnov (KS) test --> Kolmogorov-Smirnov (KS) test

ii. ln. 296 Please make clear what you mean by this sentence: 'It is observed that 10 PCA features ensures the cumulative explained variance of over 85% up to 95%, in case of the data sets employed in this study.' Please reframe this sentence.

iii. The matlab function-- Please capitalize MATLAB.

Author Response:

The manuscript has been updated according to reviewer comments.

Thanks

Asim Bhatti

Attachment

Submitted filename: Rebuttal Letter_R4.pdf

Decision Letter 4

Alexandros Iosifidis

10 Dec 2020

PONE-D-19-23638R4

Efficient Neural Spike Sorting using Data Subdivision and Unification

PLOS ONE

Dear Dr. Bhatti,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jan 24 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Alexandros Iosifidis

Academic Editor

PLOS ONE

Additional Editor Comments (if provided):

Please address the comments provided by the Reviewer on the current version of the paper, and submit a point-to-point response letter with the revised paper.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: There are a few final edits I'd suggest in the modified part of the manuscript to improve readability, but I think the manuscript is in good shape now, and the methods are presented clearly.

lines 275-286 and Figure 7: I'd suggest to remove this paragraph and Figure 7 since

- I'm not sure how the explanation of a z-score helps in understanding the method.

- you are not plotting normal distributions in any of the Figures anymore, you defined the standard deviation in (7), so Equation (9) is not necessary (you're not using any other property of the normal distribution other than its standard deviation, and you cannot assume that the distribution of z scores is normal).

line 287: A short motivational sentence about what you're planning to do with the z score values would be great, e.g. 'We wanted to determine outliers for each spike cluster. To this aim, we considered two scenarios where the z scores distributions of a given cluster were either consistent with a normal distribution or skewed. There are numerous...'

if you rather want to keep these lines:

line 287: data distribution --> z-score distribution

line 284: Euclidean

line 281, 283: 'used to plot', 'is plotted': I don't find these plots anywhere, so please reformulate.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Feb 10;16(2):e0245589. doi: 10.1371/journal.pone.0245589.r010

Author response to Decision Letter 4


29 Dec 2020

1) lines 275-286 and Figure 7: I'd suggest to remove this paragraph and Figure 7 since

- I'm not sure how the explanation of a z-score helps in understanding the method.

- you are not plotting normal distributions in any of the Figures anymore, you defined the standard deviation in (7), so Equation (9) is not necessary (you're not using any other property of the normal distribution other than its standard deviation, and you cannot assume that the distribution of z scores is normal).

Author Response: As per reviewer’s suggestion paragraph (lines 275-286) and Figure 7 have been removed.

2) line 287: A short motivational sentence about what you're planning to do with the z score values would be great, e.g. 'We wanted to determine outliers for each spike cluster. To this aim, we considered two scenarios where the z scores distributions of a given cluster were either consistent with a normal distribution or skewed. There are numerous...'

Author Response: The manuscript has been updated and a motivational sentence highlighting the use of Z scores is added at the start of the paragraph. (Please refer to lines 275-281).

3) if you rather want to keep these lines:

a. line 287: data distribution --> z-score distribution. line 284: Euclidean

b. line 281, 283: 'used to plot', 'is plotted': I don't find these plots anywhere, so please reformulate.

Author Response: We have removed the suggested text from the manuscript as per reviewer comments (1 and 2).

Attachment

Submitted filename: Rebuttal Letter_R5.pdf

Decision Letter 5

Alexandros Iosifidis

5 Jan 2021

Efficient Neural Spike Sorting using Data Subdivision and Unification

PONE-D-19-23638R5

Dear Dr. Bhatti,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Alexandros Iosifidis

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Authors addressed all Reviewers' comments. Congratulations for the acceptance of your paper.

Reviewers' comments:

Acceptance letter

Alexandros Iosifidis

21 Jan 2021

PONE-D-19-23638R5

Efficient Neural Spike Sorting using Data Subdivision and Unification

Dear Dr. Bhatti:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Prof. Alexandros Iosifidis

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Data

    (ZIP)

    Attachment

    Submitted filename: Response to reviewers.pdf

    Attachment

    Submitted filename: Rebuttal Letter_R2.pdf

    Attachment

    Submitted filename: PONE-D-19-23638-R3-Response to Reviewers.pdf

    Attachment

    Submitted filename: Rebuttal Letter_R4.pdf

    Attachment

    Submitted filename: Rebuttal Letter_R5.pdf

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting information files.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES