Abstract
Development of automated analysis tools for “single ion channel” recording is hampered by the lack of available training data. For machine learning based tools, very large training sets are necessary with sample-by-sample point labelled data (e.g., 1 sample point every 100microsecond). In an experimental context, such data are labelled with human supervision, and whilst this is feasible for simple experimental analysis, it is infeasible to generate the enormous datasets that would be necessary for a big data approach using hand crafting. In this work we aimed to develop methods to generate simulated ion channel data that is free from assumptions and prior knowledge of noise and underlying hidden Markov models. We successfully leverage generative adversarial networks (GANs) to build an end-to-end pipeline for generating an unlimited amount of labelled training data from a small, annotated ion channel “seed” record, and this needs no prior knowledge of theoretical dynamical ion channel properties. Our method utilises 2D CNNs to maintain the synchronised temporal relationship between the raw and idealised record. We demonstrate the applicability of the method with 5 different data sources and show authenticity with t-SNE and UMAP projection comparisons between real and synthetic data. The model would be easily extendable to other time series data requiring parallel labelling, such as labelled ECG signals or raw nanopore sequencing data.
1. Introduction
Ion channels are transmembrane proteins that control the flow of ions from one side of the membrane to the other; they play a fundamental role in the function of key biological processes, such as the generation and propagation of action potentials [1, 2]. Due to their fundamental role in biological processes, dysfunction of ion channels can lead to a number of conditions (known as “channelopathies”), including cystic fibrosis [3, 4]. Beyond being directly responsible for health conditions, ion channels are also targeted for other uses, such as pesticides in agriculture [5] or therapeutic drugs for other conditions such as pain management [6] and a number of neuropsychiatric disorders [7].
The broad roles of ion channels motivates research into their function. The development of the Nobel Prize winning patch-clamp electrophysiological technique [8, 9] allowed for “real time” recording of the function of different ion channels to changes of conditions, such as the application of drugs. In particular, single channel recording has the ability to capture a single ion channel opening and closing in real-time, and provides researchers with the tools necessary to investigate the mechanisms behind the function of these proteins.
Analysis of patch clamp single channel data is notoriously difficult, typically taking skilled researchers significant time to manually supervise. Many different models have been developed to attempt to solve this problem [10–13]–however most of these methods are developed using simulated data generated by simulating a continuous time hidden Markov model (HMM), then applying a layer of filtered Gaussian noise on top [for example, 14–18]. Whilst this method can generate a large amount of data for training, the data itself may have hidden biases (including a user encoded HMM gating schema), since it is entirely dependent on a priori assumptions about underlying model and characteristic noise dynamics.
Generative Adversarial Networks (GANs) [19] have proven to be important for generative tasks, such as image generation [20, 21] or audio synthesis [22]. GANs have been used in electrophysiology to create new unsynchronised datasets [23, 24], with the former example being used to generate data for model training, but continuous labelling of data (millisecond-by-millisecond) has remained an unsolved challenge. GANs have great potential to allow a data driven approach to synthesis, using a small amount of “real”, lab recorded data to generate an unlimited amount of simulated data; the advantage of this method is that noise and artefacts present in the real training dataset also appear in the generated data.
In this work we develop a GAN model to generate realistic, simulated ion channel data from existing patch-clamp recordings from several channel phenotypes, each with a continuous parallel state label. These outputs could then be used to build more accurate ion channel analysis models. We use a further series of analysis tools to evaluate the model’s performance—including T-SNE and UMAP projections to investigate if each channel’s high dimensional characteristics are reflected in the simulated data.
2. Methods
2.1 Experimental workflow summary
(i) Experimental ion channel data (See “source data”, Table 1) was recorded using patch-clamp electrophysiology and then passed to our GAN model network (using a Jupyter notebook) as input data. After training, (ii) the GAN network is set to write simulated data to file. (iii) These simulated data were then passed along with the original raw “seed” partner datasets to a Jupyter notebook that ran dimensionality reduction algorithms to compare the raw and GAN simulated datasets (see “T-SNE and UMAP").
Table 1. Sources of raw data for development and testing of the GAN network.
Code | Channel source | Sample size (segments) | Methods Reference | Fig |
---|---|---|---|---|
Phenotype A | Canine articular chondrocyte | 179 | [28] | 5A, 6A |
Phenotype B | WinEDR simulated | 244 | None, see text below | 5B, 6B |
Phenotype C | Tracheal chondrocyte | 36 | None, see text below | 5C, 6C |
Phenotype D | Paraventricular nucleus of the Hypothalamus | 499 | Feetham, Nunn, Lewis, Dart, & Barrett-Jolley, 2015 | 5D, 6D |
Phenotype E | Equine articular chondrocytes | 286 | [29] | 1–4 |
2.2 Model design
We propose a GAN based model to generate synthetic time-series data that includes realistically similar features to real ion channel molecule currents. Fig 1 describes the complete pipeline of the proposed DeepGANnel model in this work. The architecture of the GAN model introduced in this work is following the regular DC-GAN, but applies the convolutional neural networks that demonstrated efficiently to produce time series datasets in previous works [25, 26]. In the following sections, the utilized neural networks briefly are covered along with the processed data information and evaluation metrics criteria.
2.3 Generative Adversarial Networks (GAN)
A GAN typically consists of two neural networks competing against each other; the generator network tries to convert random noise into observations that seem as if they have been sampled from the original data; while the discriminator network aims to classify whether the sampled data comes from the original dataset or output of the generator network by predicting a class probability. The training process is employed in an adversarial manner between the two networks by updating the parameters of both models based on updates of the discriminator; the generator is looking to maximise the discriminator’s uncertainty in classifying an output as real or fake, and the discriminator is looking to minimise this. A more detailed, mathematical formulation of these loss functions, along with the formulation of the constituent layers of these models can be found in S1 Appendix, and the full model architecture of both generator and discriminator including shapes are included in S2 and S3 Appendices respectively.
2.4 Design of the networks
The DeepGANnel networks are based on the successful DC-GAN model [27], with the input and output shapes of the generator changed to images with dimensions n by m (one channel for the raw signal, one for the idealisation, n and m were typically 2 and 1280, but this numbers can easily be varied depending on context) along with changes to the subsequent hyperparameters along with pre/post processing to facilitate this change.
The architecture of the generator G consists of strided deconvolutional layers that allow the network to undergo spatial learning with its own up-sampling; batch normalisation layers allow for stabilising learning parameters by normalising inputs and Leaky ReLU activation functions for all layers, except for the Tanh function that is used in the output layer. The input to this model is a matrix of latent noise that is transformed into the output signal, and the output of the model is a two-dimensional sample generated from the noise that can be sent to the input of the discriminator for training the model.
The discriminator D, is also a deep convolutional neural network. Similar to the generator, this model combines strided convolution layers to downsample the input data to obtain a binary classification of the input record (real or generated). In our discriminator architecture, the batch normalisation layers are not used, but instead a dropout regularised technique was used at each layer. The last convolution layer in our D network is flattened and passed into a sigmoid function for classification, but otherwise similarly to the generator function; Leaky ReLU activation functions are used for all other layers. Fig 2 shows the architecture of our DeepGANnel model in this work including both G and D networks. The Adam optimiser was used initially with 0.0001 learning rate for initial training process. Manual tuning of the learning rate parameters was needed during training to avoid overfitting from one of the models.
2.5 Data sources
For GAN simulation training, ion channel recordings were obtained from cell-attached patch clamp electrophysiology on a number of different cell types; for detailed methods please see the associated references.
In the case of tracheal chondrocytes (“Phenotype C”), methods similar to (Lewis et al. 2013), except that trachea was isolated, cut into small pieces, and thence treated as for articular chondrocytes. In all cases, animals were previously euthanised for unassociated reasons; no animals were killed or harmed for this study. Typically, patch pipettes were fabricated using 1.5 mm o.d. borosilicate glass capillary tubes (Sutter Instrument, USA, supplied by INTRACEL, UK). They were pulled using a two-step electrode puller (Narishige, Tokyo, Japan) and when filled with recording solutions, had a resistance of approximately 5–10 MΩ. In each case, data was recorded using cell-attached patch clamp with an Axopatch 200/a amplifier (Axon Instruments, USA). Low-pass filtering was set to 1 kHz (Axon’s built in 4-pole Butterworth filter) and data were digitized at 5 kHz with a Digidata 1200A interface or CED 1401 (CED, Cambridge, UK). Recordings were made with WinEDR (John Dempster, University of Strathclyde, UK).
In addition, in order to achieve a diverse population of ion channel records to test our GAN model, we simulated ion channel data using the simulation feature in WinEDR (John Dempster, University of Strathclyde, UK) using default rate constants (“Phenotype B”). Idealisation/annotation of raw data was performed with QuB [17]. Datasets had different record lengths between 1024 to 4096 datapoints per channel. Sample sizes given in Table 1.
For manifold projection data input. The raw data is the same as above (Table 1), and the simulated data are outputs from the GAN, these specific datasets are included in data in the associated repository.
2.6 Evaluation metrics
GANs are considered successful when they implicitly learn the distribution of samples of the real dataset. We assess the efficiency of the proposed DeepGANnel model to simulate single molecule data by comparing real to GAN simulated data using a number of different approaches. The standard metrics for GANs are the so-called generator loss and discriminator loss. These are calculated as the logistic binary cross entropy loss—this is calculated as:
(1) |
Where y is the true label and p the predicted label. For the generator this is calculated once, with the true label being whether the output is fake or not, and the predicted label being whether the discriminator predicted if the output is fake. For the discriminator, this is calculated twice then totalled—once for the fake output as above, and again in the same manner for real outputs.
We also use other methods to measure model success. Although there is currently some debate as to which metrics are most appropriate for measuring GAN performance, we will utilize a two-sample test called maximum mean discrepancy [30], and a classical evaluation metric for time series defined as dynamic time warping [31] to evaluate model performance.
We also use the t-SNE and UMAP dimensionality reduction algorithms as a means to compare and contrast the different ion channels and sources in a more meaningful way.
Maximum mean discrepancy
The maximum mean discrepancy (MMD) measures the dissimilarity between two probability distributions Pr and Pg—one from the real data distribution and one from the GAN respectively, by comparing statistics of the samples. In general, given a kernel K: X x Y → R, and samples , an estimation of MMD is equated as follows:
(2) |
The smaller MMD measurement, the greater similarity between the distributions. A modified Python library’s (Tensorflow) two-sample test [30] was used to determine MMD using the Gaussian kernel for the above calculation.
Dynamic time warping
Dynamic time warping (DTW) is a traditional calculation for measuring the dissimilarity between two groups of time series data. DTW warps the series temporally to minimise the Euclidean distance between points and then calculates the distance between the two warped records. The warping of one signal occurs using the following method. An N by M matrix D is constructed where N is the length of the first signal x and M the length of the second signal y. Starting at D0,0, the matrix is filled out iteratively using the following formula:
(3) |
While i = 1, …., N and j = 1, …., M. This matrix can then be used to find the most appropriate warping for the data, for which then a Euclidian similarity measure can be used. In the present work we used a dedicated library (FastDTW [32]) that approximates DTW.
T-SNE and UMAP
To comp the raw data (Table 1) and the matching GAN outputs; raw data (Table 1) and matching GAN outputs were used as inputs for manifold projection analysis. Dimensionality reduction algorithms allow us to visualise high dimensional data in a low dimensional format, typically within a visualisation such as a scatter plot. Principle component analysis (PCA), for example, reduces the dimension of data by choosing a new orthogonal basis for the data based on the maximal variance of the data. For non-linear datasets, this dimensionality reduction may not accurately convey the shape or pattern of the higher dimensional data, as points that are close together in Euclidian space but far away in the context of the data are brought together during the PCA transformation. The so-called “kernel trick” can solve this problem by providing a non-linear transformation to the space before the PCA algorithm is applied, however this requires that this non-linear transformation to be found algorithmically, which can be time consuming and inaccurate. The t-SNE [33] algorithm focuses on local similarity rather than global similarity by considering only a neighbourhood around each point in high dimensional space. The algorithm first constructs a matrix of probability distributions for each pair of points in the higher dimensional space such that close points have a high probability and far away points have a low probability. The algorithm then constructs a second matrix of probability distributions for the lower dimensional space and minimises the Kullback-Leibler divergence between the two via gradient descent. This method preserves local similarities, but global similarities are lost between far away points. Uniform Manifold Approximation and Projection (UMAP) [34] was developed to attempt to maintain the global similarities by using the same general method as t-distributed stochastic neighbour embedding (T-SNE) (the construction of two matrices and optimisation to fit the lower dimensional one to the higher dimensional one), but uses a topological transformation of the data and a different divergence metric to achieve stronger global similarity. Note that in both cases we trained these projections on the super dataset together (i.e., all raw and GAN records concatenated together). We used the standard Python packages t-SNE from Scikit Learn and UMAP from the umap-learn conda-forge. Full code is in our repository as a Jupyter notebook, but essentially raw and GAN generated datasets were broken into 512 datapoint length “windows” and input to the manifold function with complexity values of 2 (which allows projection onto an x-y plane). To give an objective indication of whether there was a significant difference between manifold clusters between original and GAN simulated data, we tested for statistical difference between each pair with the R package ClusterSignificance [35].
2.7 Computing platform
A Nvidia Titan X GPU with 12 GB of RAM was used for the experiments to train and generate realistic ion channel molecule current records alongside the event classifications of generated records using the DeepGANnel model. The GAN models were built in Tensorflow [36] version 2.x, including Keras, using Jupyter notebooks as a user-friendly interface exploiting the GPU. With this computing platform and software, trained models Once trained, the included code on this equipment generates 0.4 seconds of data (e.g., one 4096x2-datapoint “record” in under 4ms).
3. Results
Single (ion channel) molecule datasets were recorded using the patch-clamp technique and our standard protocols, and approximately 30 seconds was recorded under constant conditions (room temperature etc.). The resulting datasets were annotated (idealised) in QuB to produce a two-dimensional signal (dimension 1 = raw signal, dimension 2 = continuous annotation). Following robust scaling (Scikit learn) and reshaping these data were passed to the DeepGANnel model for approximately 10,000 epochs. Processing within each training epoch included augmentation with an invisibly small amount of Gaussian noise (approximately 0.1% of signal amplitude) applied to each data window. Fig 3A shows the characteristic evolution of discriminator and generator losses. Typical generator losses were 10x or more the discriminator loss. The first 250 epochs of the more sophisticated MMD and DTW losses are shown in Fig 3B. Fig 3C and 3D show examples of raw (real) input data and a representative strip of post train DeepGANnel generated data. The two first analyses that are typically conducted in patch-clamp research are amplitude histograms and kinetic analysis. In addition to the objective metrics, amplitude histograms are shown in Fig 3E (real events) and Fig 3F (GAN simulated events), are also clearly similar in terms of size and shape.
For an in-depth analysis we conducted full kinetic analysis of both raw (real molecular data) and GAN simulated data. These are shown in Fig 4, and there are similarities and differences between the real and GAN events. In terms of closed times, it is apparent that whilst the over-all distribution is similar between real (Fig 4A) and GAN (Fig 4C) there are differences. The real data included some long closed-events that are absent from the GAN simulated equivalent. In terms of open times again the over-all distribution is similar between real (Fig 4B) and GAN simulated (Fig 4D), but there is attenuation of the very long open sojourns in the GAN simulated data.
To further test the ability of the GAN method to simulate a wide range of ion channel-like data we then trained DeepGANnel on 4 further ion channel phenotypes (see Table 1). Each of these is shown in Fig 5, where raw data, including the timeseries label, is shown a above the GAN simulated data. It is clear that each phenotype of channel is well represented by the matching GAN. To investigate more objectively how similar each GAN generated dataset is from its parent raw data (Table 1) we performed t-SNE and UMAP dimension reduction comparisons between each of the 5-phenotypes of data (Fig 6). It is clear that there is a tendency for each GAN manifold cluster (t-SNE or UMAP) to align closely with its parent real data projection. Following this, we systematically analysed the cluster separation between real and GAN clusters, but first, we investigated whether this method could distinguish between each real vs real combination, and each GAN vs GAN combination. We focussed on UMAP projections. Table 2 shows the comparisons between UMAPs of the real data. Each real dataset is significantly different to each other, demonstrating the power of UMAP cluster analysis to objectively classify real world data. Secondly, the equivalent analysis focusses on UMAP comparisons of the 5 different GAN generated datasets is show in Table 3. In this case, all GAN simulated datasets are different from each other, except for the A and C datasets (different types of cartilage ion channels), which are not significantly different from each other. Finally, Table 4 shows that cluster-analysis of UMAP projections is unable to see a statistically significant difference between the canine articular ion channel data (Phenotype A) and its respective GAN, the tracheal ion channel data (Phenotype C) and its respective GAN or the equine cartilage data (Phenotype E) and its derived GAN data. However, the GANs produced from channel phenotypes B and D were significantly distinguishable from their parent datasets (Phenotypes B and D).
Table 2. Statistical separation of UMAP clusters between real datasets.
Phenotype A | Phenotype B | Phenotype C | Phenotype D | Phenotype E | |
Phenotype A | 1 | 0.02 | 0.001 | 0.001 | 0.001 |
Phenotype B | - | 1 | 0.02 | 0.001 | 0.001 |
Phenotype C | - | - | 1 | 0.001 | 0.001 |
Phenotype D | - | - | - | 1 | 0.001 |
Phenotype E | - | - | - | - | 1 |
Table 3. Statistical separation of UMAP clusters between GAN generated datasets.
Phenotype A | Phenotype B | Phenotype C | Phenotype D | Phenotype E | |
Phenotype A | 1 | 0.001 | 0.590 | 0.001 | 0.001 |
Phenotype B | - | 1 | 0.001 | 0.001 | 0.001 |
Phenotype C | - | - | 1 | 0.001 | 0.001 |
Phenotype D | - | - | - | 1 | 0.001 |
Phenotype E | - | - | - | - | 1 |
Table 4. Statistical separation of UMAP clusters between each real dataset and its GAN simulated equivalent.
Phenotype | Source | p-val |
---|---|---|
A | Canine articular chondrocyte | 0.739 |
B | WinEDR simulated | 0.001 |
C | Tracheal chondrocyte | 0.371 |
D | Paraventricular nucleus of the Hypothalamus | 0.001 |
E | Equine articular chondrocytes | 0.993 |
Maximum p-values calculated by the ClusterSignificance permutation package in R [35]. p-values are given after at least 1000 permutations.
4. Discussion
In this work, we have generated synthetic raw single-molecule timeseries data along with continuous synchronised annotation/idealisation using a generative adversarial network (GAN) based on both real ion channel single molecule data from cultured chondrocytes. We demonstrate that the GAN generated raw ion channel data was similar to those obtained by real ion channel data. We assessed success of the GAN by three methods and in each they proved successful, but retain some limitations.
A central problem in single molecule, including “ion channel” research is that analyses of data is laborious and frequently requires a degree of expert hand crafting to complete. The first step in such analysis is idealisation of the record, or in machine learning terms, annotating or labelling. Each time point (of which there will be many million) needs to be annotated as to how many molecule pores are open at that instant. This then becomes, effectively, a two-dimensional representation of the data. Currently, we are working with simple datasets with one type of molecule in the dataset, but in the future such analysis will extend this to more complex datasets. Clearly new analysis methods will also be necessary to get the maximum amount of information from complex single molecule data. A number of tools have been developed to address these issues [10, 17, 37, 38], including our own Deep-Channel deep learning model [13]. For further development of similar or enhanced tools there is a lack of available training data. There are two clear choices for such data; (i) real biological data that has been annotated in some way or (ii) synthetic datasets. Both approaches have biases and severe limitations. Real data cannot be perfectly labelled; the real ground truth is unknowable and so will only be an approximation. In practical terms, the longer the length of the real data, the greater the “ground truth” errors will be. As a result, new machine learning methods will learn the errors of the existing technology. Furthermore, only simple datasets can be annotated and so this sets an upper limit on the complexity of the datasets that could be analysed by potential new tools. However, synthetic datasets also contain many biases and limitations, some of which may be entirely unanticipated or recognised. Therefore, the starting problem that our work addressed here was to create very large datasets of ion channel single molecule activity that could be used to develop single molecule analysis tools. We chose to investigate if GAN technology could provide a useful alternative source of data. In principle this would have the advantages of synthetic data in that one could produce unlimited amounts, but still retain nuance and subtle authenticity missed by mere simulation. We are also hopeful that such synthetic data could be used in development of ion channel modelling software and may allow for a novel type of data inference, extracting critical features in datasets that may be overlooked by traditional analysis. Visual inspection, and analysis of our similarity metrics demonstrates that DeepGANnel can reconstruct faithfully simulate a number of different ion channel phenotypes. Clearly the datasets are not identical (that would be simply a copy), but raises the question of whether the simulation is good enough to be useful. To examine this, we show in Fig 7 an example experiment. We train our Deep-Channel model to analyse/idealise the data from phenotype C (Table 4, Fig 5C). Half of the small length of original raw data (75k datapoints equivalent to approx. 8 seconds of recording) is clearly insufficient to allow Deep-Channel to learn the appropriate features and idealise the remaining half of this dataset (Fig 7A–7C). However, DeepGANnel, once trained can produce any amount of data similar to this original dataset (See Fig 5C). We then trained Deep-Channel on this GAN dataset, 5x the size of the original, and performance is now acceptable (Fig 7D–7F).
Comparisons of GAN methods to traditional synthesis methods
The new GAN synthesis method will have strengths and weaknesses compared to stochastic Markovian model-based processes. (1) Speed: Training a GAN takes a considerable time (hours), but simulation itself takes ms per record. The usual stochastic methods produce records at the ms per record timescale, but do not need training. They still need laborious hand customising to simulate a particular phenotype of channel. Our software only Markovian based simulations, written in Python, take about 8ms to generate one “record” whereas DeepGANnel, on our GPU workstations takes about 4ms (once trained). It should be noted that Stochastic models will take n times longer to simulate data records with n channels within, whereas the GAN method would need retraining to a multi-channel dataset; but still take the same 4ms per record once trained. The method we used to create the training set in our DeepChannel project [13] used stochastic simulation followed by passing these data through a real patch clamp amplifier. This method was the slowest of all, producing data at less than real time (so approximately 1000x slower than other methods). (2) Authenticity: Speed is not the objective of this work however, it is “authenticity”. In the present paper, we provide metrics for authenticity in terms of UMAP/t-SNE cluster similarity for DeepGANnel, however there is no absolute way to do the same with stochastic methods. The more effort the user puts into it, analysing noise and reproducing these, measuring single channel properties and encoding this, the closer it would become, but this entire laborious process would need to be repeated for every type of ion channel and condition to be simulated. With DeepGANnel, the user simply points the script at a new set of seed data. Fundamentally, stochastic data may include unrecognised biases since every feature must be hand crafted, and DeepGANnel would include artefacts found in native data that would likely be omitted using a stochastic approach. The inclusion of these in training data would be important for development of robust analysis software. (3) A priori assumptions: DeepGANnel needs none, but stochastic simulation requires every detail to be estimated. However, stochastically generated data has the advantage that the experimenter could potentially develop software similar to HJCFit [38, 39] and QuB [40], to recover the HMM, and validate accuracy. With a DeepGANnel simulated dataset the underlying HMM would be unknown, and need to be estimated with further software such as HJCFit, in the same way as one would do with real patch clamp data.
Previous use of GAN for time-series data
The literature includes previous examples where timeseries data can be simulated with a GAN [For example, 26], but these lack the synchronous labelling critical in single molecule analysis or many other physiological studies. Generating single-molecule timeseries with a basic GAN (Vanilla-GAN) model should be very effective at producing authentic ion molecule signal patch-clamp signal [41], but this would lack the output of the critical timepoint-by timepoint labelling necessary to meet our goals of creating valid alternative datasets. However, we have shown that using a 2DCNN based model in a shape of (samples x 2 x 1280 x 1) is very effective; with a small amount of carefully annotated seed data generating unlimited synthetic copies. In Figs 3 and 4 we show a direct comparison for a typical electrophysiological work up of both the original and DeepGANnel synthesised data. On the supplementary information and public repository, we include a movie of the training process. The match between synthetic kinetic analysis and the real data is not perfect, but rather close. The notable exception is that the longer states (both open and closed) are missing or have a diminished representation. We attribute this to the necessity to use a finite window (“image” width or “record length”) size. Also, as stated in the methods this was cropped to remove leading and trailing artefacts. Perhaps models using far greater window lengths would be possible, but this does not appear to be a major problem for our purpose (since most single molecule events durations are within this window) and it would increase the model complexity many-fold. This model took 24–48 hours to train on our system and note that performance peaked but would deteriorate if it was left indefinitely. Our code allowed manual adjustment of learning rate as epochs progressed, but we still chose to stop the modelling manually. The ever-increasing GPU power make ever larger window sizes less of an issue in the future. We summarise these strengths and weaknesses in Table 5.
Table 5. Likely strengths and weaknesses of DeepGANnel versus traditional synthesis methods.
Traditional Method | Traditional Method | DeepGANnel | |
---|---|---|---|
A priori assumptions | Stochastic simulation | Everything must be estimated or assumed; channel size, rate constants, open channel noise, thermal noise levels, artefact frequency etc. | None required. |
Authenticity | Stochastic simulation | Depends entirely on the accuracy of a priori-assumptions. Authenticity difficult. | Highly authentic. |
Speed | Stochastic simulation | Moderate speed. | Slow to train, fast to simulate thereafter. |
GPU needed | Stochastic simulation | Typically, these are not used. Future stochastic models may use them. | Realistically these are needed for training, although not for simulation itself. |
Vanilla GAN | yes | - | |
Markov Model | Stochastic simulation | Could include Markovian model structure. | May include Markovian structure, but this is not guaranteed. |
Need for seed data | Stochastic simulation | No. The data can be completely imaginary. | Yes |
Fully Labelled data | Stochastic simulation | Yes | - |
Vanilla GAN | Cannot provide labels in parallel to raw data. | Yes |
This table summarises the pros and cons of DeepGANnel discussed and justified in the text. By definition such comparisons can only be subjective because Traditional Methods vary (entirely dependent on a priori assumptions, that could be simple or complex), as do computing platforms.
Future applications of GAN in electrophysiology
The potential for further exploitation, of GAN technology in electrophysiology beyond the current use for creation of datasets is immense. One future goal will be to simulate far more complex data signals, but we still have the limitation on how to acquire the fully annotated seed data in the first place. Potentially painstaking manual annotation of very short sections of data with known numbers of ion channel molecules by several human experts would be possible. Furthermore, it is possible that single-molecule GAN could be used more directly in electrophysiological modelling. Currently, single molecule behaviour within such models is derived by a set of differential equations based on a set of measured or even estimated parameters [42], but it may be possible in the future to use GAN to generate more realistic stochastic behaviour directly. Additionally, future studies will investigate whether interpretability methods can be used to identify important, defining features within each different dataset that are missed either by eye or by standard single molecule analysis techniques. The DeepGANnel approach will prove especially useful for generation of “raw” data, along with ground truth annotations as analysis methods move to using a big data approach; creating large datasets where underlying models are unknown or too complex to simulate stochastically. Potentially also, our methods could be used to augment data in an analysis pipeline, for example to facilitate secondary analysis where only small samples of data are available.
The architecture we present here, using deep learning to generate physiological timeseries data with continuous annotations, could also be adapted easily for additional usability for equivalent systems in physiology. For example, action potential or electrocardiogram simulation. As proof-of-principle we show here that indeed DeepGANnel can easily synthesise telemetered ECG signal, again fully annotated (Fig 8). In this example the annotation dimension is merely beat (binary state 1) or no beat binary (0), but this could easily be extended to include P-wave (categorical state 2), T-wave (categorical state 3), or abnormal event (categorical state 4) etc with trivial code adaptation. Another example would potentially be Nanopore data which has similar data output to patch-clamp data.
In summary, GANs are increasingly proving a viable method to generate synthetic datasets for biological research, and here we show an implementation that allows simulation of time dependent single molecule (patch clamp ion channel protein) activity along with a continuous state annotation that is extendable for an array of physiological uses.
Supporting information
Data Availability
All relevant data are available on Figshare (DOI: 10.6084/m9.figshare.19236603). All relevant code are available on Github (https://github.com/RichardBJ/DeepGANnel).
Funding Statement
This work was funded by Biotechnology and Biological Sciences Research Council (BBSRC) grant awards to RBJ (as Principal Investigator). Project References BB/R022143/1, BB/S008136/1 and a BBSRC DTP Studentship 2266700. The funder had no influence in the design analyses or publication decisions.
References
- 1.Hodgkin AL, Huxley AF. The components of membrane conductance in the giant axon of Loligo. J Physiol. 1952:473–96. doi: 10.1113/jphysiol.1952.sp004718 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Neher E, Sakmann B. Single-Channel Currents Recorded from Membrane of Denervated Frog Muscle-Fibers. Nature. 1976;260(5554):799–802. doi: 10.1038/260799a0 [DOI] [PubMed] [Google Scholar]
- 3.Quinton PM. Cystic fibrosis: a disease in electrolyte transport; Cystic fibrosis: a disease in electrolyte transport. The FASEB Journal. 1990;4(10):2709–10. doi: 10.1096/fasebj.4.10.2197151 [DOI] [PubMed] [Google Scholar]
- 4.Welsh MJ. Abnormal regulation of ion channels in cystic fibrosis epithelia; Abnormal regulation of ion channels in cystic fibrosis epithelia. The FASEB Journal. 1990;4(10):2718–25. doi: 10.1096/fasebj.4.10.1695593 [DOI] [PubMed] [Google Scholar]
- 5.Silver KS, Du Y, Nomura Y, Oliveira EE, Salgado VL, Zhorov BS, et al. Voltage-Gated Sodium Channels as Insecticide Targets. Elsevier; 2014. p. 389–433. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Skerratt SE, West CW. Ion channel therapeutics for pain. Channels. 2015;9(6):344–51. doi: 10.1080/19336950.2015.1075105 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Imbrici P, Conte Camerino D, Tricarico D. Major channels involved in neuropsychiatric disorders and therapeutic perspectives. Frontiers in Genetics. 2013;4:76. doi: 10.3389/fgene.2013.00076 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Sakmann B, Neher E. Patch Clamp Techniques for Studying Ionic Channels in Excitable Membranes. Annu Rev Physiol. 1984;46(1):455–72. doi: 10.1146/annurev.ph.46.030184.002323 [DOI] [PubMed] [Google Scholar]
- 9.Hamill OP, Marty A, Neher E, Sakmann B, Sigworth FJ. Improved Patch-Clamp Techniques for High-Resolution Current Recording from Cells and Cell-Free Membrane Patches. Pflugers Archiv-European Journal of Physiology. 1981;391(2):85–100. doi: 10.1007/BF00656997 [DOI] [PubMed] [Google Scholar]
- 10.Gnanasambandam R, Nielsen MS, Nicolai C, Sachs F, Hofgaard JP, Dreyer JK. Unsupervised Idealization of Ion Channel Recordings by Minimum Description Length: Application to Human PIEZO1-Channels. Frontiers in Neuroinformatics. 2017;0:31-. doi: 10.3389/fninf.2017.00031 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Qin F. Restoration of Single-Channel Currents Using the Segmental k-Means Method Based on Hidden Markov Modeling. Biophysical Journal. 2004;86(3):1488-. doi: 10.1016/S0006-3495(04)74217-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Hotz T, Schutte OM, Sieling H, Polupanow T, Diederichsen U, Steinem C, et al. Idealizing Ion Channel Recordings by a Jump Segmentation Multiresolution Filter. IEEE Transactions on NanoBioscience. 2013;12(4):376–86. doi: 10.1109/TNB.2013.2284063 [DOI] [PubMed] [Google Scholar]
- 13.Celik N, O’Brien F, Brennan S, Rainbow RD, Dart C, Zheng Y, et al. Deep-Channel uses deep neural networks to detect single-molecule events from patch-clamp data. Commun Biol. 2020;3(1):3. doi: 10.1038/s42003-019-0729-3 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Voldsgaard Clausen M. Obtaining transition rates from single-channel data without initial parameter seeding. Channels (Austin). 2020;14(1):87–97. doi: 10.1080/19336950.2020.1732004 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Colquhoun D, Hawkes AG, Srodzinski K. Joint distributions of apparent open and shut times of single-ion channels and maximum likelihood fitting of mechanisms. Philosophical Transactions of the Royal Society of London Series A: Mathematical, Physical and Engineering Sciences. 1996;354(1718):2555–90. doi: 10.1098/rsta.1996.0115 [DOI] [Google Scholar]
- 16.Anderson DF, Ermentrout B, Thomas PJ. Stochastic representations of ion channel kinetics and exact stochastic simulation of neuronal dynamics. Journal of Computational Neuroscience. 2015;38(1):67–82. doi: 10.1007/s10827-014-0528-2 [DOI] [PubMed] [Google Scholar]
- 17.Nicolai C, Sachs F. Solving ion channel kinetics with the QuB software. Biophysical Reviews and Letters. 2013;8(03n04):191–211. [Google Scholar]
- 18.Gillespie DT. Exact Stochastic Simulation of Coupled Chemical-Reactions. J Phys Chem-Us. 1977;81(25):2340–61. doi: 10.1021/j100540a008 [DOI] [Google Scholar]
- 19.Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative Adversarial Nets. Advances in Neural Information Processing Systems. 2014;27 (NIPS 2014). [Google Scholar]
- 20.Karras T, Laine S, Aila T. A Style-Based Generator Architecture for Generative Adversarial Networks. arXiv. 2019;1812.04948v3. None arxiv:1812.04948v3. [DOI] [PubMed]
- 21.Brock A, Donahue J, Simonyan K. Large Scale GAN Training for High Fidelity Natural Image Synthesis. arXiv. 2019;1809.11096.
- 22.Donahue C, McAuley J, Puckette M. Adversarial Audio Synthesis. arXiv. 2019;1802.04208v3.
- 23.Qin X-H, Wang Z-Y, Yao J-P, Zhou Q, Zhao P-F, Wang Z-Y, et al. Using a one-dimensional convolutional neural network with a conditional generative adversarial network to classify plant electrical signals. Computers and Electronics in Agriculture. 2020;174:105464. doi: 10.1016/j.compag.2020.105464 [DOI] [Google Scholar]
- 24.Truong ND, Kuhlmann L, Bonyadi MR, Querlioz D, Zhou L, Kavehei O. Epileptic seizure forecasting with generative adversarial networks. IEEE Access. 2019;7:143999–4009. [Google Scholar]
- 25.Delaney AM, Brophy E, Ward TE. Synthesis of Realistic ECG using Generative Adversarial Networks. arXiv. 2019;1909.09150.
- 26.Zhu F, Ye F, Fu Y, Liu Q, Shen B. Electrocardiogram generation with a bidirectional LSTM-CNN generative adversarial network. Sci Rep. 2019;9(1):6734. Epub 2019/05/01. doi: 10.1038/s41598-019-42516-z . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Radford A, Metz L, Chintala S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv. 2015;1511.06434.
- 28.Lewis R, Feetham CH, Gentles L, Penny J, Tregilgas L, Tohami W, et al. Benzamil sensitive ion channels contribute to volume regulation in canine chondrocytes. Brit J Pharmacol. 2013;168(7):1584–96. doi: 10.1111/j.1476-5381.2012.02185.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Mobasheri A, Lewis R, Maxwell JEJ, Hill C, Womack M, Barrett-Jolley R. Characterization of a stretch-activated potassium channel in chondrocytes. Journal of Cellular Physiology. 2010;223(2):511–8. doi: 10.1002/jcp.22075 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Gretton A, Borgwardt KM, Rasch MJ, Scholkopf B, Smola A. A Kernel Two-Sample Test. J Mach Learn Res. 2012;13:723–73. [Google Scholar]
- 31.Sakoe H, Chiba S. Dynamic-Programming Algorithm Optimization for Spoken Word Recognition. Ieee T Acoust Speech. 1978;26(1):43–9. doi: 10.1109/Tassp.1978.1163055 [DOI] [Google Scholar]
- 32.Salvador S, Chan P. Toward accurate dynamic time warping in linear time and space. Intell Data Anal. 2007;11(5):561–80. doi: 10.3233/Ida-2007-11508 [DOI] [Google Scholar]
- 33.Van Der Maaten L, Hinton G. Visualizing Data using t-SNE. J Mach Learn Res. 2008;9:2579–605. [Google Scholar]
- 34.McInnes L, Healy J, Melville J. UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction. ArXiv. 2018;1802.03426.
- 35.Serviss JT, Gådin JR, Eriksson P, Folkersen L, Grandér D. ClusterSignificance: a bioconductor package facilitating statistical analysis of class cluster separations in dimensionality reduced data. Bioinformatics. 2017;33(19):3126–8. doi: 10.1093/bioinformatics/btx393 . [DOI] [PubMed] [Google Scholar]
- 36.Abadi M, Agarwal A, Barham P, Brevdo E, Chen Z, Citro C, et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. 2015. [Google Scholar]
- 37.Juette MF, Terry DS, Wasserman MR, Altman RB, Zhou Z, Zhao H, et al. Single-molecule imaging of non-equilibrium molecular ensembles on the millisecond timescale. Nature methods. 2016;13(4):341. doi: 10.1038/nmeth.3769 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Colquhoun D, Hatton CJ, Hawkes AG. The quality of maximum likelihood estimates of ion channel rate constants. Journal of Physiology-London. 2003;547(3):699–728. doi: 10.1113/jphysiol.2002.034165 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Gibb AJ, Ogden KK, McDaniel MJ, Vance KM, Kell SA, Butch C, et al. A structurally derived model of subunit-dependent NMDA receptor function. J Physiol. 2018;596(17):4057–89. Epub 20180801. doi: 10.1113/JP276093 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Qin F, Auerbach A, Sachs F. Estimating single-channel kinetic parameters from idealized patch-clamp data containing missed events. Biophysical Journal. 1996;70(1):264–80. doi: 10.1016/S0006-3495(96)79568-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Celik N, O’Brien F, Zheng Y, Coenens F, Barrett-Jolley R. Automatically detecting single cell ion channel events of neural activity using generative adversarial and recurrent neural networks. COSYNE 2020; Denver, Colorado, USA: 2020. [Google Scholar]
- 42.Feetham CH, Nunn N, Lewis R, Dart C, Barrett-Jolley R. TRPV4 and KCa ion channels functionally couple as osmosensors in the paraventricular nucleus. Br J Pharmacol. 2015;172(7):1753–68. doi: 10.1111/bph.13023 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Nunn N, Feetham CH, Martin J, Barrett-Jolley R, Plagge A. Elevated blood pressure, heart rate and body temperature in mice lacking the XLαs protein of the Gnas locus is due to increased sympathetic tone. Exp Physiol. 2013;98(10):1432–45. doi: 10.1113/expphysiol.2013.073064 [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
All relevant data are available on Figshare (DOI: 10.6084/m9.figshare.19236603). All relevant code are available on Github (https://github.com/RichardBJ/DeepGANnel).