Indexed retention times (iRT), MS1 (the first level of mass analysis) or survey scan charge state distributions, and sequence ion intensities of MSMS (tandom mass spectrometry) spectra were predicted from peptide sequence by use of long-short term memory (LSTM) recurrent neural networks models. Data points on the order of 105 were used to train the iRT and charge state distribution models. An HCD sequence ion prediction model was trained with 2×106 experimental spectra. The models with a simple deep learning architecture can predict those three key LC-MS/MS (Liquid chromatography-tandem mass spectrometry) properties with superior accuracies.
Keywords: Algorithms, Peptides, Statistics, Bioinformatics Software, Bioinformatics, Charge State Distribution, Deep Learning, Indexed Retention Time, Prediction
Graphical Abstract
Highlights
Deep learning models for prediction of LC-MS/MS properties.
Abstract
Deep learning models for prediction of three key LC-MS/MS properties from peptide sequences were developed. The LC-MS/MS properties or behaviors are indexed retention times (iRT), MS1 or survey scan charge state distributions, and sequence ion intensities of HCD spectra. A common core deep supervised learning architecture, bidirectional long-short term memory (LSTM) recurrent neural networks was used to construct the three prediction models. Two featurization schemes were proposed and demonstrated to allow for efficient encoding of modifications. The iRT and charge state distribution models were trained with on order of 105 data points each. An HCD sequence ion prediction model was trained with 2 × 106 experimental spectra. The iRT prediction model and HCD sequence ion prediction model provide improved accuracies over the start-of-the-art models available in literature. The MS1 charge state distribution prediction model offers excellent performance. The prediction models can be used to enhance peptide identification and quantification in data-dependent acquisition and data-independent acquisition (DIA) experiments as well as to assist MRM (multiple reaction monitoring) and PRM (parallel reaction monitoring) experiment design.
The ability to directly predict LC-MS/MS1 behaviors from peptide amino acid residue sequence will power the next generation of proteomics research. Currently, the major behavior or properties of peptides in the context of LC-MS/MS are obtained experimentally, suffering from high cost and inhomogeneous accuracies. This work intends to address those issues directly using a common deep learning model architecture trained with experimental data to predict useful LC-MS/MS properties, such as indexed retention times (iRT), MS1 or survey or precursor scan charge state distributions, and MS/MS sequence ion intensities. Those properties can be directly used for enhancing peptide identification and quantification in both data-dependent acquisition and data-independent acquisition (DIA) experiments as well as for designing MRM and PRM experiments.
The previous research into prediction of those three major classes of LC-MS/MS properties varies significantly. The first use of machine learning algorithms for retention time prediction was with simple dense neural networks (1). Since then, many algorithms have been proposed (2). DeepRT is the most recent method utilizing a deep learning model to predict retention time (3). It is surprising that information of charge state distributions of MS1 scans has not been extensively used to facilitate peptide identification and quantification. Studies for prediction of charge state distributions were carried out with simple models and limited data (4, 5). Prediction of MS/MS spectra is both theoretically interesting and practically useful. Development of the kinetic models by Zhang (6–9) requires deep understanding of ion fragmentation mechanisms. The kinetic model development is a reminiscence of knowledge-based artificial intelligence in proteomics. Like in the other fields, such as image processing and language processing, deep-learning-based approaches in the field of proteomics quickly outperform any knowledge-based approaches. DeepNovo (10), a deep learning model based de novo sequencing method provides an accurate peptide sequencing strategy without the reference to a predetermined protein sequence database and digestion rules. The deep-learning-based de novo sequencing method has been extended for DIA experiments (11). The recently developed pDeep algorithm allows for prediction of MS/MS spectra with peptide sequences without incorporating detailed fragmentation mechanisms into the model (12). Schoenholz et al. (13) developed a similar long-short term memory (LSTM)-model-based search algorithm called “DeepMatch.” Gessulat et al. (14) recently developed a deep-learning-based tool called “Prosit” for prediction of LC retention time and fragment ion intensities from sequences of synthetic peptides. Tiwary et al. (15) reported two deep learning models for MS/MS spectrum prediction.
The core motivation of this work is to identify the key LC-MS/MS properties that are determined solely by peptide sequence. The chosen three properties fit the criteria. The iRT calibration strategy (16) allows establishment of a common scale to calibrate experimental retention times. Solvent gradient, column temperature, sample loading, and other experimental conditions can be incorporated into the iRT versus RT (retention time) or RT versus iRT calibration functions. MS1 charge state distributions may change with many experimental conditions. One of the most important ones is the charge detection method of the mass analyzer (see “Discussion” section). However, given commonly used mass analyzers, MS1 charge state distributions are stable against many other experimental conditions. For MS/MS sequence ion intensity prediction, different ion series may present in an MS/MS spectrum with different fragmentation methods and fragment ion intensities also change with different mass analyzer/detector. Spectral behavior changes also with fragmentation conditions such as activation energy and activation duration. However, MS/MS spectral behavior is remarkably stable under commonly used instrument classes and standard methods. Under the constraints discussed above, we demonstrate in this work that practical deep learning models can be trained to predict those key LC-MS/MS properties with excellent accuracies.
Since the beginning of this decade, neural-network-based deep learning models have made significant impact on image and natural language processing (17). For peptide property prediction, recurrent neural network models, such as LSTM models, are a natural choice because they are able to handle sequential input data with a variable length, such as peptide sequences (in spectrum prediction, charge state is also an input). Deep supervised learning models are capable of classification and regression tasks. Our iRT and spectrum prediction models are classic regression examples, but the charge state prediction model can also be considered as a classification method in which the output is charge state classification probabilities. The three prediction models demonstrated here represent three levels of data in LC-MS/MS: iRT for LC behavior, charge state distribution for MS1, and HCD sequence ion intensities at MSMS level. They also differ in output dimensionality: iRT is a scalar, charge state distribution is a vector, and HCD sequence ion intensities are two-dimensional matrix with one of the dimensions dependent on the peptide length. However, the core learning blocks are the same bidirectional LSTM layers, demonstrating the applicability of deep learning models for a wide range of problems. As in other fields where deep learning made significant impacts, performance of peptide property prediction models critically depends on the availability of large quantity and high-quality data. For example, our HCD sequence ion intensity prediction model was built on a large pool of spectra of excellent quality (18), collected from the whole proteomics community.
EXPERIMENTAL PROCEDURES
Sources of Data
iRT data: Data used for training the iRT model were described in (19). The data were compiled from three DIA technical runs with HeLa and HEK293 cell lysates, with 1-m long column and 4-h acquisition time. According to Bruderer et al. (19), the identification false discovery rate was controlled at 1%. Peptides of duplicate identifications and different charge states were pooled together, and the pooled median iRT values were used. After filtering, such as limiting the peptide length to 40 or less, the total number of peptides for the iRT model is 125,793, in which 10% were randomly selected as the test dataset, and the rest 90% were used to train the iRT model.
Charge state distribution data: For the charge state distribution model, raw data files of data-dependent acquisition runs (19) were downloaded from ProteomExchange. Those data-dependent acquisition experiments included 0.5–4-h single-shot runs and high pH fractionation sample runs. The raw data files were converted to peaklists using previously developed PAVA code (20). MS2 or MS/MS peaklists were searched with the MSGF+ search engine (21). The human proteome (Uniprot Proteome ID: UP000005640) sequences (71,778 entries) were downloaded on February 6, 2018. Trypsin specificity with up to five missed cleavages were allowed (filtered after search). Cysteine carbamidomethylation was the fixed modification. Variable modifications were oxidation of methionine, pyro-glu from peptide n-terminal glutamine, and protein n-terminal acetylation with or without loss of methionine and deamidation of glutamine and asparagine. Precursor mass tolerance was 10 ppm, and fragment ion detection instrument was set to “Q Exactive.” The peptide EValue cutoff of 0.045 was used to obtain the peptide false discovery rate of 1.0%.
For an identified peptide, the ion chromatograms for charge states 1–5 were extracted from survey scans and the XICs (extracted ion chromatograms) were fitted with a polynomial variance Gaussian function (20). The intensities of peptide charge states were normalized by division of their sum. The total number of peptides for the charge state distribution model is 126,876, in which 10% of randomly selected data were used as the test dataset.
HCD sequence ion intensity prediction data: A human HCD spectral library with 2,154,269 peptide ions was compiled by Wang et al. (18). Fragment ions were annotated with in-house code, and data were formatted as the National Institute of Standards and Technology (NIST) MSP format. Monoisotopic and isotopic ions for singly and doubly charged b- and y- ions, internal ions, H2O and NH3 neutral loss ions, and the precursor and their neutral loss ions were annotated. For precursor ions with charge state of 2, b and y ions of charge state one were collected. For precursors with charge state 3 to 6, both charge states of 1 and 2 were considered for b and y ions. The ion intensities of a single spectrum were normalized by division of the highest intensity. Due to the graphic processing unit, GPU's memory limitation, the dataset was separated into two subsets of 1,074,898 spectra each. The membership of the subsets was random. The HCD prediction model was trained sequentially with the two subsets. Out of the 1,074,898 spectra in the second subset, 10% were used as the test dataset.
For the selected test cases, raw files from Bruderer et al. (19) and from ProteomeTools (22) were used.
Feature Structure
Peptide sequence features: A peptide sequence is represented with one-hot encoding array of size, MAX_LENGTH X NO_OF_AAS, in which MAX_LENGTH of 40 is the maximum length of peptides and NO_OF_AAS is the number of amino acid types. In all cases, a deamidated glutamine residue is considered as a glutamic acid and a deamidated asparagine as an aspartic acid. In the iRT model, only modification allowed is the oxidized methionine, encoded separately Therefore, NO_OF_AAS is 21 for the iRT model. For the charge state distribution model, oxidized methionine, pyro-glutamine, and n-terminal acetylation were separated feature types. Its NO_OF_AAS is 23.
For the sequence ion intensity prediction model, peptide n-terminal carbamylation was also considered as a separate type, given its NO_OF_AAS of 24. The charge state of a peptide ion (precursor ion) is one-hot encoded as a six-member long array. Therefore, a peptide ion is encoded as a 40 × 30 matrix (20 natural amino acid types, 4 PTM types, 6 charge state types). As suggested by (12), normalized collision energy was not used as a feature. For peptides of length less than 40, zeros were added to both sides (zero padding on both ends). The one-hot encoding of peptide sequence and modification featurization is illustrated in Supplemental Fig. S2.1.
Learning Models
All the models consist of a masking layer followed by two stacks of bidirectional LSTM layers. For regularization, both dropout and recurrent dropout were applied. The models differ in output (label) layers: for the iRT model the output was a scalar and no final activation was used. Two dense layers with activation function of hyperbolic tangent function (tanh) were used and the final output has a dimension of one. For the charge state distribution model, there were two dense layers with the output as one dimensional array of size 5 (corresponding to charge states 1 to 5). The output activation function was softmax. For the spectral prediction model, a time distributed dense layer with sigmoid activation function was used. The output (label) structure is illustrated in Supplemental Fig. S2.2.
All LSTM models, implemented with Keras framework (23) were trained on a GeForce RTX 2080 Ti Graphics Card with 11 GB video memory. The longest training time was 7.2 h for one subset (1,074,898 spectra) of the HCD sequence ion prediction model.
RESULTS
iRT Prediction
The essence of a deep learning model is to capture the data generation distribution. However, we do not have the independent information about the distribution underlying the iRT dataset. The iRT values were obtained from the DIA experiments (19), and several sources of errors may be incorporated into the dataset. First, the peptide false discovery rate was about 1%, and therefore, at least 1% of the iRT values were incorrect, and those incorrect iRT entries are likely to contribute to outliers. Second, calibration of RT versus iRT can introduce some uncertainties (RT versus iRT calibration is briefly described in Supplemental Material Section S1). Third, when the iRT values were aggregated from different LC runs, errors are introduced due to LC alignment inconsistencies. As can be observed in the Fig. 1, the outliers are mostly likely due to the incorrect peptide identification. The variation becomes higher with increased iRT values. For peptides of later elution or having larger iRT values, intensities may spread out in more charge states and in wider LC peak shapes, and retention time measurement becomes less accurate.
Both training and testing error distributions are close to a Gaussian (Fig. 2), but away from the center there are still significant deviation of the error distributions from the Gaussian-fitted curves. The 95% confidence intervals were computed from distributions numerically, and they are 7.18 and 9.62 iRT units, for training and test datasets, respectively.
A higher capacity model was also investigated with nearly twice as many parameters, and it did not result in any improvement in the confidence intervals, suggesting the current model may have captured the most important characteristics of the iRT data generation distribution. Several different deep learning models have also been surveyed, including conventional CNN (convolutional neural networks), temporal CNN (24), and Capsnet (3). None of the models surveyed performed better than the current LSTM model with the current datasets. It is nearly impossible to obtain an optimal deep model because model hyperparameter optimization is multidimensional and problem dependent. We can, however, make a rough comparison of performance between our model to that of DeepRT(3). Our test confidence interval of 9.62 iRT units (iRT span: −60 to 140) is about 28% narrower than the DeepRT result of 13.4 (iRT span: −40 to 160, Supplemental Fig. S7 of reference (3)) for reverse-phase LC experiments.
Prediction of Charge State Distribution
Survey scan or MS1 charge state distribution provides rich information on relative charge partition among peptide ions. The information may also be helpful in assessing the quality of the associated MSMS spectra and may be used directly for extraction of MS1 ion chromatograms. In two previous studies (4, 5), conventional machine learning models were trained to predict charge states for electron transfer dissociation spectra. A common problem with the conventional machine learning algorithms is the requirement for manual collection of features. In both case, features are extracted from identified ETD (electron transfer dissociation) tandem mass spectra. Liu et al. (25) trained a model with linear combination of amino acid composition with Gaussian probability. In our model, only peptide sequences were used, and the deep learning model is capable of learning the complex contribution of peptide sequence to the charge state distribution. The relative intensities of different charge states detected in a mass spectrum are not only directly proportional to their abundances. In a Fourier transform ion cyclotron resonance mass spectrometer, the detected image current is inversely proportional to the ion's m/z (26). In an Orbitrap, the image current is inversely proportional to the square root of the ion's m/z (27). However, for our purpose, treating the relative intensities as one of LC-MS/MS properties is advantageous: the predicted distribution can facilitate peptide identification and quantification. Therefore, although we refer to the term “charge state distribution,” it is an LC-MS/MS behavior other than the true charge partitions on a peptide. The deep model on the other hand is able to learn the “behavior” or property with great accuracy. However, the property should not be transferred to other mass analyzers without proper intensity calibration.
As shown in Fig. 3, the charge state distribution prediction produces highly accurate results judged by the test Pearson correlation coefficient (PCC) distribution, with median PCCs of 0.998 and 0.997 for the training and testing datasets, respectively. The distribution for the test dataset is similar to that for the training dataset, indicating that overfitting is not a significant issue here.
Two examples of charge state prediction are illustrated in Fig. 4. For peptide VLPGMHHPIQMKPADSEK and its Q-deamidated form, the charge state distributions are quite similar. The missing observed charge state of 2 for the deamidated form may be due to its low abundance for experimental detection.
Prediction of HCD Sequence Ion Intensities
The whole dataset for training the spectrum prediction model contains two subsets of 1,074,898 HCD spectra each. The HCD prediction model was trained sequentially with the two subsets. 10% of spectra in the second subset were used as the testing set. During training, 20% of the training subsets were used to validate the model. In the whole dataset, 3.85% contains n-terminal acetylation, 8.10% methionine oxidation, 2.00% pyro-glutamine, and 5.68% n-terminal carbamylation.
As shown in Fig. 5, the PCC distributions for training and testing are also quite similar, indicating that the model does not overfit much. The median PCCs are 0.955 and 0.953 for training and testing, respectively. The median test PPC for our test dataset is slightly better than that obtained in the pDeep study (12). A breakdown analysis of performance is provided in Supplemental Table S3. For peptide ions of 20 amino acid residues or shorter with no modifications, the median test PCC of 0.965 is a more significant improvement from that (0.950) for the pDeep model. However, our dataset contains peptide sequence length up to 40 amino acid residues compared with 20 in the pDeep study. For human proteome, this adds about 25% more unique peptides (see peptide length distribution of human protein sequences in Supplemental Fig. S3). In addition, the pDeep data do not contain modified peptides, and in our dataset, peptides with the following modifications are allowed: oxidized methionine, peptide n-terminal pyro-glutamic acid, peptide n-terminal carbamylation, and protein n-terminal acetylation. In addition to increase sequence length and inclusion of some common modifications, our model differs from that of pDeep in the number of features: our model uses fewer features. One of the core advantages of deep learning over traditional machine learning is that it uses multiple layers to learn feature characteristics. Therefore, deep models are typically much simpler in terms of feature engineering (feature collection, normalization, reduction, etc.). We designed our models using the simplest possible architecture with only 30 feature dimensions (to represent 20 natural amino acid residues, 2 modified amino acid residues, 2 separately coded modifications, and 6 charge states) versus 88 in the pDeep model (illustrated in Supplemental Fig. S2.1). Our model also aligns n-terminal (b) and c-terminal (y) ions with the peptide sequence (see Supplemental Fig. S2.2 for an example of model label (output) structure), whereas in the pDeep model the length (19) of the series of ions is one less than that (20) of the peptide sequence length.
One of the utilities of HCD sequence ion intensity prediction is to assist isomeric peptide assignments. Isomeric peptides concerned here are peptides with exactly the same chemical composition and nearly identical sequences. The minor difference in sequence makes them difficult to be distinguished using MS2 spectra. The authors of the pDeep model (12) investigated their model's utility to distinguish three classes of the extremely similar peptides: (a) substitution of an isoleucine for a leucine; (b) shared chemical formula of different amino acid combinations, such as GG for N; and (c) local amino acid permutations, such as AF for FA.
In Fig. 6, an experimental HCD spectrum was identified by the MSGF+ search engine as LALDLEIATYR 2+ ion of the keratin, type II cytoskeletal 8 (P05787) or LALDIEIATYR 2+ ion of the keratin, type II cytoskeletal 1 (P04264). Because both sequences were assigned with the same EValue (4.2 × 10−6, 1% false discovery rate cutoff = 0.045), the search engine was not capable of distinguishing them. When the experimental spectrum was matched with the predicted ion intensities of the two sequences, the PCC difference for matching between the experimental and predicted data was small (0.942 for LALDLEIATYR and 0.943 for LALDIEIATYR) if all sequence ions were used. However, if only some “local” ions were considered, such as y6, y7, and y8, there is a clear difference in the relative intensities. The “local” ion PCC values for the three ions were 0.870 and 0.994 for LALDLEIATYR and LALDIEIATYR, respectively. Clearly, using the local product ions near the site of sequence difference can provide more significant discrimination power. The assignment of the isoleucine peptide of LALDIEIATYR was also supported by spectral count information. In the same raw data file, there are 137 spectra including those for the isoleucine peptide (four spectra) belonging to the protein (P04264), whereas only 11 spectra supporting the protein (P05787) contain the leucine peptide. Four out of the 11 spectra were assigned to the leucine peptide. It is unlikely that the experimental spectrum is a mixture of both peptides because they also have different retention times. Both the matching analysis of predicted ion intensities and the spectral counting information support the assignment of the isoleucine peptide of LALDIEIATYR. Another example for leucine/isoleucine peptide discrimination is provided in Supplemental Fig. S4.1. The “local” ions with the most discriminating power can be selected (i) by choosing ions near the site of difference and (ii) by examining local PCC values between the predicted ion intensities of the isomeric sequences.
The isomeric peptide classes (b) and (c) can be easily distinguished if a fragment ion is observed for the peptide bond cleavage between amino acid residue combination or permutation. If this is not the case, the predicted ion intensity difference may be used to distinguish the assignment. For class (b), the pDeep model was able to provide strong support for assignment of an experimental spectrum from the ProteomeTools dataset to the peptide GGFFSFGDLTK (PCC = 0.98) against NFFSFGDLTK (PCC = 0.87). Using our HCD sequence ion intensity prediction model, the separation of the PCC values (0.994 versus 0.838) is even larger, presumably due to the higher accuracy of our prediction model (see Supplemental Fig. S4.2).
To demonstrate the utility of charge state distribution prediction for DIA analysis, a peptide (FVNVVPTFGKK) was selected and its ion chromatograms were extracted from a DIA data file (Fig. 7). Our charge state distribution prediction model predicts that the 2+ precursor charge state (0.32) is X2 less abundant that 3+ (0.68).
Even the 3+ precursor ion has greater abundance, its MS2 product ion abundance levels are low (bottom panel, Fig. 7), and only three of the six product ions display signals and they are very weak. It is therefore more confident to use the MS1 XICs for charge state 3+ and MS2 XICs for charge state 2+ for the peptide identification and quantification.
A web service for LC-MS/MS property prediction is provided (see Supplemental Materials Section S5). The trained models, training and testing data, and associated Python codes can be downloaded from Zenodo (see Supplemental Materials Section S6).
DISCUSSION
Deep learning models have been surveyed for prediction of three key LC-MS/MS properties of peptides: iRT, MS1 charge state distribution, and HCD sequence ion intensities. The three properties characterize peptide behaviors at the three distinct levels of experiment: chromatography, MS1 or precursor detection, and MS/MS or fragmentation of a precursor. The deep learning models share the same peptide sequence encoding strategy (additional charge encoding for spectrum prediction) and structure of the first two layers of recurrent neural network. The output layers and the corresponding activation functions were chosen to reflect the need for the different types of output dimensionality.
Those models provide superior performance. The iRT prediction is ∼28% more accurate compared with that of the DeepRT model(3). The 95% confidence interval for iRT prediction of Prosit (14) is nearly twice as small as that of this work. However, training and testing data of Prosit came from pools of synthetic peptides on the order of 1000 with equal abundance, whereas data used for this study came from a complex sample of cell lysate. It is indeed useful to investigate the impact of sample complexity, gradient length on the iRT prediction accuracy. For the charge state distribution prediction, a median test PCC of 0.997 was achieved. The performance of our HCD sequence ion intensity prediction model outperforms that of the pDeep model (12). Our model is also able to predict peptide sequence length of 40 compared with 20 for pDeep and allows for common peptide modifications. The superior performance of our models may come from two factors. First, our models are simpler with minimal feature engineering. More complicated feature engineering (as the case in the pDeep model) may set constraints on the model to fit data and to generalize well. Second, our training datasets may be more homogeneous.
Prosit's fragment ion intensity model uses bidirectional gated recurrent unit (14), which is less sophisticated and more efficient to train than the LSTM model used in this study. All the prediction models are simpler enough, and the computation effort is not a major practical issue. The other difference is that our spectrum prediction model (and that for pDeep) uses a time distributed layer in combination with a masking layer to discard the useless parameters and output dimensions. The Prosit model sets the useless output labels to −1.
The authors of Prosit (14) claimed to have achieved a median PCC of 0.99 between the experimental and predicted spectra. However, as pointed out by the authors of pDeep, the agreement among their experimental spectra was the PCC of 0.981, which was considered as the upper limit of any prediction model. Tiwary et al. (15) developed two models, DeepMass:Prism and wiNNer. DeepMass:Prism, an LSTM-based model, achieves a median PCC of 0.925 for HCD spectra.
Our models with current trained parameter values are practically useful in the right context: for example the iRT predicted values are only useful in combination of the right iRT to RT calibration. The current prediction model for HCD sequence ion intensities does not account for internal fragmentation and neutral loss from fragment ions. Extension to include simple neutral loss can be accomplished with the expansion of the label dimension (See Supplemental Fig. S2.2).
Our models for prediction of iRT and charge state distribution may have learned major portions of the corresponding data generation distributions because the labels (the measured iRTs and charge state distributions) contain limited information. More training data may not improve their performance. The model for HCD sequence ion intensity prediction on the other hand may be further improved given more data.
We also proposed two encoding schemes to featurize modifications: (A) if a modification occurs on many different types of amino acid residues, a separate feature is created or (B) if a modification occurs in only on one or two types of amino acid residues, the modified residues are used as features. As illustrated in Supplemental Figs. S2.1 and S2.2, featurization Scheme (A) will increase the space occupied by the separately coded modifications in the sequence length dimensions in both sequence code and label. The Scheme (B) entries will increase the feature dimension in sequence code. An optimal design may be a trade-off of the two conflicting effects to minimize the dimensions for both sequence code and label.
Broad applications of those models are subject to future studies. The models can supply DIA data extraction algorithms with predicted retention times, charge state distributions, and spectra for peptides not previously detected. Enhancement of search engines' theoretical fragment pattern generation, expansion of spectral libraries for spectral library search, selected ion monitoring (SRM), MRM, PRM method development can all benefit from the models' capabilities.
Data Availability
Training and testing data, trained models, prediction results, sample codes for training and accessing data are provided in the following link: https://zenodo.org/record/2652602#.XP67VlxKi70.
Supplementary Material
Acknowledgments
We thank Jia Rong Wu for setting up the GPU system for training and testing the deep models of this study.
Footnotes
* S.G. received partial funding for this work from the Pan-Canadian Proteomics Centre funded by Genome Canada and a Genome Canada B/CB grant. M.F.M. was funded in part by Canada Research Chairs, Canada Foundation for Innovation, and National Sciences and Engineering Research Council (NSERC) of Canada. B.M. was funded in part by NSERC and a Genome Canada B/CB grant. The authors declare that they have no conflict of interest.
This aricle contains supplemental material Sections S1 to S6, associated supplemental tables, and supplemental figures.
1 The abbreviations used are:
- LC-MS/MS
- liquid chromatography - tandem mass spectrometry
- DIA
- data independent acquisition
- CNN
- convolutional neural networks
- ETD
- electron transfer dissociation
- GPU
- graphic processing unit
- HCD
- higher-energy collisional dissociation
- iRT
- Indexed retention time
- LSTM
- long-short term memory
- MRM
- multiple reaction monitoring
- MS/MS
- tandem mass spectrometry
- MS1
- the first level of mass analysis
- MS2
- the second level of mass analysis
- NIST
- National Institute of Standards and Technology
- PRM
- parallel reaction monitoring
- PCC
- Pearson correlation coefficient
- RT
- retention time
- SRM
- selected ion monitoring
- XIC
- extracted ion chromatogram.
REFERENCES
- 1. Petritis K., Kangas L. J., Ferguson P. L., Anderson G. A., Pasa-Tolić L., Lipton M. S., Auberry K. J., Strittmatter E. F., Shen Y., Zhao R., and Smith R. D. (2003) Use of artificial neural networks for the accurate prediction of peptide liquid chromatography elution times in proteome analyses. Anal. Chem. 75, 1039–1048 [DOI] [PubMed] [Google Scholar]
- 2. Moruz L., and Käll L. (2017) Peptide retention time prediction. Mass Spectrom. Rev. 36, 615–623 [DOI] [PubMed] [Google Scholar]
- 3. Ma C., Ren Y., Yang J., Ren Z., Yang H., and Liu S. (2018) Improved peptide retention time prediction in liquid chromatography through deep learning. Anal. Chem. 90, 10881–10888 [DOI] [PubMed] [Google Scholar]
- 4. Sharma V., Eng J. K., Feldman S., von Haller P. D., MacCoss M. J., and Noble W. S. (2010) Precursor charge state prediction for electron transfer dissociation tandem mass spectra. J. Proteome Res. 9, 5438–5444 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. Carvalho P. C., Cociorva D., Wong C. C., Carvalho M. da, Barbosa V. C., and Yates J. R. 3rd. (2009) Charge prediction machine: tool for inferring precursor charge states of electron transfer dissociation tandem mass spectra. Anal. Chem. 81, 1996–2003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Zhang Z. (2004) Prediction of low-energy collision-induced dissociation spectra of peptides. Anal. Chem. 76, 3908–3922 [DOI] [PubMed] [Google Scholar]
- 7. Zhang Z. (2005) Prediction of low-energy collision-induced dissociation spectra of peptides with three or more charges. Anal. Chem. 77, 6364–6373 [DOI] [PubMed] [Google Scholar]
- 8. Zhang Z. (2010) Prediction of electron-transfer/capture dissociation spectra of peptides. Anal. Chem. 82, 1990–2005 [DOI] [PubMed] [Google Scholar]
- 9. Zhang Z., and Shah B. (2010) Prediction of collision-induced dissociation spectra of common N-glycopeptides for glycoform identification. Anal. Chem. 82, 10194–10202 [DOI] [PubMed] [Google Scholar]
- 10. Tran N. H., Zhang X., Xin L., Shan B., and Li M. (2017) De novo peptide sequencing by deep learning. Proc. Natl. Acad. Sci. U.S.A. 114, 8247–8252 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Tran N. H., Qiao R., Xin L., Chen X., Liu C., Zhang X., Shan B., Ghodsi A., and Li M. (2019) Deep learning enables de novo peptide sequencing from data-independent-acquisition mass spectrometry. Nat. Methods 16, 63–66 [DOI] [PubMed] [Google Scholar]
- 12. Zhou X.-X., Zeng W.-F., Chi H., Luo C., Liu C., Zhan J., He S.-M., and Zhang Z. (2017) pDeep: Predicting MS/MS spectra of peptides with deep learning. Anal. Chem. 89, 12690–12697 [DOI] [PubMed] [Google Scholar]
- 13. Schoenholz S. S., Hackett S., Deming L., Melamud E., Jaitly N., McAllister F., O'Brien J., Dahl G., Bennett B., Dai A. M., and Koller D. (2018) Peptide-spectra matching from weak supervision. arXiv 1808.06576.v2 [Google Scholar]
- 14. Gessulat S., Schmidt T., Zolg D. P., Samaras P., Schnatbaum K., Zerweck J., Knaute T., Rechenberger J., Delanghe B., Huhmer A., Reimer U., Ehrlich H.-C., Aiche S., Kuster B., and Wilhelm M. (2019) Prosit: Proteome-wide prediction of peptide tandem mass spectra by deep learning. Nat. Methods 16, 509–518 [DOI] [PubMed] [Google Scholar]
- 15. Tiwary S., Levy R., Gutenbrunner P., Salinas Soto F., Palaniappan K. K., Deming L., Berndl M., Brant A., Cimermancic P., and Cox J. (2019) High-quality MS/MS spectrum prediction for data-dependent and data-independent acquisition data analysis. Nat. Methods 16, 519–525 [DOI] [PubMed] [Google Scholar]
- 16. Escher C., Reiter L., MacLean B., Ossola R., Herzog F., Chilton J., MacCoss M. J., and Rinner O. (2012) Using iRT, a normalized retention time for more targeted measurement of peptides. Proteomics 12, 1111–1121 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Goodfellow I., Bengio Y., and Courville A. (2016) Deep learning. MIT Press [Google Scholar]
- 18. Wang M., Wang J., Carver J., Pullman B. S., Cha S. W., and Bandeira N. (2018) Assembling the community-scale discoverable human proteome. Cell Syst. 7, 412–421.e5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Bruderer R., Bernhardt O. M., Gandhi T., Xuan Y., Sondermann J., Schmidt M., Gomez-Varela D., and Reiter L. (2017) Optimization of experimental parameters in data-independent mass spectrometry significantly increases depth and reproducibility of results. Mol. Cell. Proteomics 16, 2296–2309 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Guan S., Price J. C., Prusiner S. B., Ghaemmaghami S., and Burlingame A. L. (2011) A data processing pipeline for mammalian proteome dynamics studies using stable isotope metabolic labeling. Mol. Cell. Proteomics 10, M111.010728 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Kim S., and Pevzner P. A. (2014) MS-GF+ makes progress towards a universal database search tool for proteomics. Nat. Commun. 5, 5277. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Zolg D. P., Wilhelm M., Schnatbaum K., Zerweck J., Knaute T., Delanghe B., Bailey D. J., Gessulat S., Ehrlich H.-C., Weininger M., Yu P., Schlegl J., Kramer K., Schmidt T., Kusebauch U., Deutsch E. W., Aebersold R., Moritz R. L., Wenschuh H., Moehring T., Aiche S., Huhmer A., Reimer U., and Kuster B. (2017) Building ProteomeTools based on a complete synthetic human proteome. Nat. Methods 14, 259–262 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Chollet F. (2017) Deep learning with Python. Shelter Island, NY: Manning Publications. [Google Scholar]
- 24. Bai S., Kolter J. Z., and Koltun V. (2018) An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 1803.01271v2 [Google Scholar]
- 25. Liu H., Zhang J., Sun H., Xu C., Zhu Y., and Xie H. (2011) The prediction of peptide charge states for electrospray ionization in mass spectrometry. Proc. Environ. Sci. 8, 483–491 [Google Scholar]
- 26. Guan S., and Marshall A. G. (1995) Ion traps for Fourier transform ion cyclotron resonance mass spectrometry: Principles and design of geometric and electric configurations. Int. J. Mass Spectrom. Ion Processes 146–147, 261–296 [Google Scholar]
- 27. Makarov A. (2000) Electrostatic axially harmonic orbital trapping: A high-performance technique of mass analysis. Anal. Chem. 72, 1156–1162 [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Training and testing data, trained models, prediction results, sample codes for training and accessing data are provided in the following link: https://zenodo.org/record/2652602#.XP67VlxKi70.