Skip to main content
Bioinformatics logoLink to Bioinformatics
. 2026 Jan 11;42(1):btag003. doi: 10.1093/bioinformatics/btag003

RP3Net: a deep learning model for predicting recombinant protein production in Escherichia coli

Evgeny Tankhilevich 1,, Sergio Martinez Cuesta 2, Ian Barrett 3, Carolina Berg 4, Lovisa Holmberg Schiavone 5, Andrew R Leach 6,1
Editor: Jonathan Wren
PMCID: PMC12857573  PMID: 41520295

Abstract

Motivation

Recombinant protein expression can be a limiting step in the production of protein reagents for drug discovery and other biotechnology applications. We introduce RP3Net (Recombinant Protein Production Prediction Network), an AI model of small-scale heterologous soluble protein expression in Escherichia coli. RP3Net utilizes the most recent protein and genomic foundational models. A curated dataset of internal experimental results from AstraZeneca and publicly available data from the Structural Genomics Consortium was used for training, validation and testing of RP3Net.

Results

RP3Net achieves an increase in area under the receiver operator curve (AUROC) of 0.15, compared to a baseline model. When experimentally validated on an independent, prospective, manually selected set of 97 constructs, RP3Net outperformed currently available models, with an AUROC of 0.83, delivering accurate predictions in 77% of the cases, and correctly identifying successfully expressing constructs in 92% of cases.

Availability and implementation

The model, along with installation and running instructions, is available under an MIT licence at https://github.com/RP3Net/RP3Net, DOI 10.5281/zenodo.17243498.

1 Introduction

1.1 Motivation

The production of protein reagents is an essential part of the research and development process in the pharmaceutical and biotechnology industries. In drug discovery it is often a pre-requisite for screening and hit identification (Zanders 2020, Singh et al. 2023). In living tissues, the target protein may occur in very small amounts alongside numerous other biomolecules. To be used for high-throughput screening of drug candidates, structural determination and the development of functional assays, the target protein needs to be expressed in a cell culture and purified. The ability to express a protein depends on multiple factors. First and foremost is the protein itself, but other factors include the cloning vector, the species and strain of the host cells, the codon optimization algorithm, the use of tags and fusion proteins and other experimental conditions (Burgess-Brown et al. 2008, 2021, Gordon et al. 2008, Gräslund et al. 2008, Hayashi and Kojima 2008, Structural Genomics Consortium et al. 2008; Haacke et al. 2009, Raymond et al. 2009, Francis and Page 2010, Zhong et al. 2015, Cooper and Marsden 2017, Söderberg et al. 2019, Chapple and Dyson 2021, Mahajan et al. 2021a, b, Strain-Damerell et al. 2021, Morão et al. 2022, Simm et al. 2022, Kurashiki et al. 2023, Schütz et al. 2023). The choice of these parameters is often influenced by the details of the downstream experiments (Acton et al. 2011), making protein production time-consuming and error-prone, and often requiring multiple iterations and much trial and error. The purpose of this work is to develop a deep learning model to predict soluble protein expression in Escherichia coli from the construct sequence, thus accelerating the timescales for protein production from months to weeks, cutting costs and reducing environmental impact.

A recombinant protein production experimental pipeline involves several steps, including construct design, cloning, small-scale expression screening, progression of expressing constructs to large-scale purification and quality control (Edfeldt et al. 2024). Small-scale soluble expression screening, shown in Fig. 1, available as supplementary data at Bioinformatics online, is crucial for assessing whether to progress the construct to large-scale production. First, cells are transfected with vectors (e.g. plasmids) carrying the cloned DNA of the protein of interest (Step 1). Recombinant protein production is performed in deep well format (Step 2). Cells are spun down and lysed (Step 3). The lysate contains the total amount of protein produced. After an additional centrifugation step the soluble protein is found in the supernatant and the insoluble material is discarded in the pellet (Step 3). Soluble protein is captured in a one-step purification using the histidine tag and immobilized metal affinity chromatography (IMAC, Step 4) (Sulkowski 1985). The soluble protein yield and correct size are typically assessed by performing denaturing gel electrophoresis (sodium dodecyl sulphate–polyacrylamide gel, SDS–PAGE) where yield and size are compared to a protein standard (Step 5). The yield can be estimated by quantifying the amount of the target protein compared to the protein standard in the stained gel, using densitometric analysis. At this stage, it is important to record both positive (produced) and negative (failed to produce) experimental outcomes (Step 6). Throughout the rest of this publication, unless stated explicitly, terms ‘protein production’ and ‘protein expression’ refer to this step of the experimental pipeline. Constructs that pass this small-scale screening are then typically progressed to large-scale purification and further downstream applications.

Protein and DNA foundation models (FMs) have become ubiquitous tools for predicting structural and functional protein properties from amino acid and/or nucleic acid sequences (Ji et al. 2021, Rives et al. 2021, Brandes et al. 2022, Elnaggar et al. 2022, Lin et al. 2023, Nguyen et al. 2023, Zhou et al. 2023, Outeiral and Deane 2024). These FMs are typically trained on large corpora of sequences, such as UniRef (Suzek et al. 2015, Bateman et al. 2023), GenBank (Benson et al. 2013, Sayers et al. 2024), MGnify (Richardson et al. 2023) or BFD (Jumper et al. 2021) (Table 1, available as supplementary data at Bioinformatics online). During training a portion of the sequence is masked out, i.e. each residue is replaced with a special ‘blank’ character. The training objective then becomes to reconstruct the masked portion of the input or ‘fill in the blanks’. This technique, referred to as language modelling task, originates from natural language processing (Devlin et al. 2018).

ESM (Rives et al. 2021, Lin et al. 2023, Hayes et al. 2025), ProtBert (Elnaggar et al. 2022) and ProteinBert (Brandes et al. 2022) are examples of protein FMs; DNABert (Ji et al. 2021, Zhou et al. 2023) is a popular DNA FM. These models, except for ProteinBert, are based on transformer deep learning architecture (Vaswani et al. 2017) with different number of layers, feature dimensions and other details. ProteinBert uses convolutional layers for sequence embedding and global attention for global protein properties. HyenaDna (Nguyen et al. 2023) is another DNA FM that uses a different architecture, whereas CaLM (Outeiral and Deane 2024) is a transformer-based model that uses codons as inputs.

The intermediate layers of foundational models yield a residue-level sequence representation that can be used to predict the protein property of interest, such as secondary or tertiary structure, binding affinity, fluorescence, thermodynamic stability, solubility, etc. (Rao et al. 2019, Dallago et al. 2021, Jumper et al. 2021, Rives et al. 2021, Thumuluri et al. 2022, Wang and Zhao 2022, Lin et al. 2023, Li et al. 2024, Xiong et al. 2024, Hayes et al. 2025). The experimental datasets that describe these properties typically contain orders of magnitude fewer entries when compared to the sequence corpora. This scarcity of experimental datasets often makes it unfeasible to train large foundational models from scratch for predicting protein properties. Such models are usually pre-trained with the language modelling objective on the large corpora first, and then further trained to predict the property of interest using the smaller dataset (Rao et al. 2019, Dallago et al. 2021, Li et al. 2024). This final training step is referred to as fine-tuning if the FM weights are updated, or downstream training if these weights are frozen and extra layers are trained to predict the property of interest. Recombinant Protein Production Prediction Network (RP3Net) follows the latter architectural blueprint by encoding the biological sequence with a foundational model, feeding this encoding through an aggregation layer to obtain a global representation for the entire construct, and then applying a fully connected classification head to compute the predicted probability of recombinant expression in E. coli as a binary outcome (Fig. 1A).

Figure 1.

Figure 1

(A) Architecture diagram of RP3Net. The input biological sequence is encoded by the foundation model to obtain a sequence representation, where each residue/codon/nucleotide is represented by a vector. The aggregation layer builds a global protein representation vector from the sequence representation. The predicted probability of successful recombinant expression of the protein in E. coli is computed by the fully connected classification head from the protein representation. (B) Training with meta label correction on a mixture of clean and noisy data. The standard training set-up, where the model loss on clean inputs and labels is minimized with gradient descent, is shown in the top row. A special ‘teacher’ model is trained to predict the corrected labels from the noisy input and labels. These corrected labels, along with noisy inputs, serve as inputs for training the ‘student’ model. The latter model has the same architecture and weights as the ‘clean’ model. The bi-level optimization algorithm that makes sure that the corrected labels do not deviate from the (unknown) clean labels relies on using the cross-entropy (CE) loss. Model components with trainable weights are shown as blue boxes. Training data are shown as yellow boxes. Images generated with BioRender.com.

Although in theory, a soluble protein production fine-tuning dataset could be designed and experimentally generated from scratch, in practice it would be too time-consuming and expensive. Moreover, there already exist publicly available datasets of protein expression that contain the results of experiments worth millions of dollars and representing years of lab work (Gabanyi et al. 2011, Berman et al. 2017, Protein production data from the SGC 2022). For training and evaluating RP3Net, internal AstraZeneca (AZ) small-scale expression screen data are combined with datasets from the Structural Genomics Consortium (SGC), specifically their sites in Stockholm (Gräslund et al. 2008, Savitsky et al. 2010) and Toronto (Protein production data from the SGC 2022, Edfeldt et al. 2024). The experimental pipeline for generating data from AZ and SGC Stockholm has been already discussed above, see Fig. 1, available as supplementary data at Bioinformatics online, Step 6. SGC Toronto captures the results of large-scale protein purification.

1.2 Existing work

A number of models that predict soluble expression from construct sequence have been published in recent years. Most of these systems use datasets derived from the Protein Structure Initiative (PSI) compendium, sometimes referred to as TargetTrack (Gabanyi et al. 2011, Berman et al. 2017). PSI was an experimental research effort run across multiple laboratories in 2000–17, with the objective of determining protein structures and depositing them in the Protein Data Bank (PDB) (Berman et al. 2000, Burley et al. 2023). This dataset records the pipeline position, i.e. the experimental stage where the work was terminated, for each target and construct. For example, if a construct was selected and cloned, but could not be expressed, its pipeline position would be recorded as ‘cloned’. For another construct that has been selected, cloned, expressed and purified, but could not be crystallized, the pipeline position would be ‘purified’, etc. One limitation of using TargetTrack data in this work is that in this dataset a genuine inability to express the construct under given experimental conditions can be confused with stopping to pursue the construct for other reasons (for example there being another well-behaving construct for the same target). Different labs that have provided data for TargetTack were using different experimental pipelines: sometimes small-scale expression screening as shown in Fig. 1, available as supplementary data at Bioinformatics online, but also large-scale purification results as SGC Toronto, or the results of running SDS-PAGE on unpurified cell lysate.

NetSolP (Thumuluri et al. 2022) uses PSI data to evaluate multiple transformer-based models available at the time for predicting soluble expression. NetSolP outputs two scores: solubility and ‘usability’, the latter being a combined predictor of solubility and the ability of a protein to be expressed.

It is important to make a distinction between solubility as a general physical property of the protein, which can be measured, for example, as peak concentration in the solution, and the ability to achieve soluble expression of the protein under given experimental conditions, which is a binary outcome that is modelled in this work. A protein that is generally soluble could still fail to express, for example because it is toxic for the host cells, or because the chaperones that are required for forming the correct structure are missing, or due to other reasons. Solubility is thus a necessary but insufficient condition for soluble recombinant protein production.

PLMC (Xiong et al. 2024) and SADeepCry (Wang and Zhao 2022) also use data derived from TargetTrack and a Transformer architecture but output the pipeline position given the construct sequence. PPCPred (Mizianty and Kurgan 2011), PredPPCrys (Wang et al. 2014), Crysalis (Wang et al. 2016), and DCFCrystal (Zhu et al. 2021) are examples of older, simpler models that predict pipeline position, trained on various subsets of TargetTrack. SoluProt (Hon et al. 2021) uses a different PSI-based dataset with a Gradient Boosted Machine model (Friedman 2001) and global features based on relative amino acid frequencies, predicted physicochemical properties, similarity to E. coli proteome and output of various other bioinformatics tools to predict soluble expression.

CamSol (Sormanni et al. 2015, Sormanni et al. 2017) is a well-established relative solubility prediction tool for libraries of similar protein sequences. There are many other solubility predictions that use deep neural networks, such as GPSFun (Yuan et al. 2024) and PLM_Sol (Zhang et al. 2024). A few methods exist for modelling expression and solubility of human antibodies, but their experimental protocols differ substantially from E. coli-based expression analysed in this work (Zhang et al. 2022, Basafa et al. 2024).

2 Materials and methods

2.1 The dataset

Protein production results from AZ, SGC Stockholm and SGC Toronto were used for training and evaluation of the models. AZ and SGC Stockholm report the results of small-scale protein expression testing, after one purification step. In the AZ dataset, the outcome is reported as an expression yield category, manually estimated by the scientist who has expressed the protein. In addition, for a subset of constructs, an estimate of the absolute concentration value in milligrams per litre is provided. The estimate is obtained by comparing the size and intensity of the band on the SDS-PAGE gel for the protein of interest with the band for the reference protein of known concentration. The amino acid sequence of the construct includes affinity and solubility tags; DNA sequences are available for a subset of constructs.

The dataset from SGC Stockholm contains genetic sequences, with tags, annotated with categorical outcomes.

For the bulk of the SGC Toronto data, the outcome is reported as a pipeline position, similarly to PSI/TargetTrack. Importantly, there is no dedicated stage for expression screening: ‘cloned’ is immediately followed by ‘purified’. Although it can generally be assumed that a protein must be expressed before it can be purified, sometimes producing at larger scale (expression volume) can rescue a construct that failed to yield soluble protein at small scale, and vice versa. For a small subset of SGC Toronto data, small-scale expression screening outcome is also provided as a categorical variable, similarly to SGC Stockholm. Genetic sequences are available for a subset of observations, and tags are included in the constructs.

Graphical overview of the datasets is shown in Fig. 2, available as supplementary data at Bioinformatics online. There are a total of 67 055 unique sequences, covering 5077 target proteins. Publicly available datasets are significantly larger than the internal AZ dataset, SGC Toronto being the largest. Datasets vary in terms of number of constructs per target, availability of genetic sequences versus protein sequences and imbalance between positive and negative outcomes.

To normalize the data across multiple sources and to compute the outcome imbalance, the labels were converted to binary form, with ‘True’ indicating successful production and ‘False’ failed production. For AZ, this binary outcome was computed based on the existing category annotation, estimate of the absolute concentration and manual re-annotation. For SGC Stockholm, the binary outcome was derived directly from the existing category annotation. For SGC Toronto, it was derived from the pipeline position. The details of this procedure are given in the Supplementary methods, available as supplementary data at Bioinformatics online.

2.2 The model

RP3Net encodes biological sequences using a foundational model, applies an aggregation layer to generate a global protein representation, and then passes this representation through a classification head to produce the final score (Fig. 1A). Two types of aggregation layer were tested in this work: mean pooling and set transformer pooling (STP). For notation, assume that for a sequence of length N, the output of the foundational model for each residue i is represented with a column vector xiF from a d-dimensional space, xiFRd×1. The matrix representation for the entire protein, XF, is obtained by stacking these residue representations along the sequence dimension: XFRd×N. In this notation, mean pooling, which is just averaging all these residue representations, can be written as

XMG=1Ni=1NxiF. # (1)

The advantage of mean pooling is that it is simple to interpret and fast to compute. The disadvantage is that it does not have any trainable parameters, or weights, so all training must happen upstream, in the foundational model, or downstream, in the classification head. STP is an example of an aggregation layer with trainable weights. Here, the global representation is the result of performing multiheaded attention (MHA) (Vaswani et al. 2017, see Supplementary text, available as supplementary data at Bioinformatics online), with a seed vector, wSRd, as a query, and residue representations XF as keys and values. Thus,

XSTPG=MHA(wS,XF,XF).# (2)

The meta label correction (MLC) (Zheng et al. 2021, Taraday and Baskin 2023) framework is used for training the model on predicted small-scale expression labels, derived from large-scale expression labels when the former are not available in the dataset. This framework utilizes a larger, noisy, poor-quality dataset to augment the training process of the model that would normally use only a smaller, clean, high-quality dataset. A separate, ‘teacher’ model is trained to predict the corrected soft label from the noisy data and labels. These corrected labels, along with the clean inputs and labels, are used to train the original model, which in this set-up is referred to as the ‘student’ model (Fig. 1B). Comprehensive mathematical details of the MLC algorithm are given in the Supplementary text, available as supplementary data at Bioinformatics online.

2.3 Experimental validation

The target set for experimental validation of the model was curated to include viable human drug targets and to exclude proteins that are well known from literature to be successfully expressed. We made sure that neither the protein itself, nor its close homologs, have been deposited in the PDB (Berman 2000, Burley et al. 2023). We have also excluded the target from the validation set if it was referenced from ChEMBL (Zdrazil et al. 2024). OpenTargets (Ochoa et al. 2023) was used to check for viability of a drug target. Twenty thousand human proteins from UniProt (Bateman et al. 2023) were narrowed down to 454 viable targets. These targets were further curated manually, to have distribution across different target classes and avoid too many DNA-binding proteins. In the end, 46 targets were selected for experimental validation.

Two full length constructs were created per target: one with a TEV-cleavable 6His tag and a GS linker at the N-terminal (MHHHHHHENLYFQGS…) and another one with a GS linker and a 6His tag at the C-terminal (…GSHHHHHH). Soluble production of the full-length constructs was predicted with RP3Net. For the targets where both full-length constructs were predicted to fail to be produced, trimmed constructs were generated by iteratively removing residues from N- and C-termini, with a minimum construct length of 50. Trimmed constructs that were predicted to express successfully were included in the experimental validation set. Seventy per cent of the set comprised constructs that were predicted to be produced, with the remaining 30% as negative controls. A total of 97 constructs were available for expression testing after taking into account cost constraints and cloning errors. The details of the experimental procedures are given in the Supplementary text, available as supplementary data at Bioinformatics online.

3 Results and discussion

3.1 RP3Net with fixed FM weights outperforms decision trees with global protein features

The RP3Net architecture with fixed FM weights and mean pooling (Fig. 1A) was used for selecting the best performing FM. This architecture is denoted as Model A (Table 1). The models were trained and evaluated on the SGC Stockholm dataset, with five-fold cross-validation. SGC Stockholm was used because this dataset is of medium size, compared to AZ, which is much smaller, and SGC Toronto, which is much larger. This dataset is also the only one of the three that provides DNA sequences for all the constructs.

Table 1.

RP3Net training and architecture configurations.

RP3Net model Aggregation FM weights Meta label correction
A Mean Frozen No
B STP Frozen No
C STP Fine-tuned, LoRA No
D STP Fine-tuned, LoRA Yes

A gradient boosted decision tree, XGBoost (Friedman 2001), with global protein features as inputs was used as a baseline model. As shown in Fig. 2, Model A with any protein FM outperforms the baseline model. Out of all the tested DNA and codon FMs, only CaLM (Outeiral and Deane 2024) shows better results than the baseline. A plausible explanation is due to the datasets used for pre-training the FMs: CaLM was pre-trained on coding sequences from European Nucleotide Archive, whereas other DNA FMs were pre-trained on a mixture of coding and noncoding sequences.

Figure 2.

Figure 2

Performance of RP3Net with fixed foundation model (FM) weights and mean pooling (Model A) on SGC Stockholm, along with FM parameter count. On the left y-axis, each boxplot shows the area under AUROC of Model A with a particular FM, evaluated on SGC Stockholm test data, with five-fold cross-validation. On the right y-axis, black crosses show the number of trainable parameters of the FM, in log scale. ESM2 (650M) was selected for further analysis based on performance, consistency, parameter count and licencing restrictions. ‘Random embeddings’ means using random residue embeddings instead of an FM.

RP3Net performance also varies depending on the training data subset that was used, sometimes dramatically. For example, for the more consistent FMs, such as ESM2 (650M) and CaLM, the difference between the best and the worst runs is 0.03 and 0.01, respectively, whereas for HyenaDNA Medium the difference is 0.14.

The number of trainable FM parameters is used to indicate the resource requirements for fine-tuning the FM (compute time, memory). The FM for subsequent evaluation was chosen based on the pragmatic trade-off between performance, data availability, training complexity and licencing constraints (Table 1, available as supplementary data at Bioinformatics online). We selected ESM2 with 650 million parameters. The simple Model A training protocol, applied to this FM, achieves an average increase in area under the receiver operator curve (AUROC) of 0.03, compared to the baseline model.

3.2 Performance on different data sources reveals dependency on dataset size

Model A performance on SGC Stockholm dataset can be improved by replacing the mean pooling aggregation layer with a more sophisticated STP (Lee et al. 2019, Buterez et al. 2022). This configuration is denoted as Model B. The main difference between mean pooling and STP is that, whereas the former just takes an average across the sequence, giving each residue the same weight, STP uses context-dependent weights for residue representations, by computing multiheaded attention (MHA) (Vaswani et al. 2017), see Supplementary text, available as supplementary data at Bioinformatics online) between a special parameter, called the seed vector, and the output of the FM. The seed vector is updated during training with gradient descent, along with the rest of the model parameters.

Model B gives an AUROC improvement of 0.02 over Model A when trained and evaluated on SGC Stockholm (Table 2). The performance of Model B on other data sources varies with an AUROC of 0.59 on the AZ dataset and of 0.84 on SGC Toronto. A plausible reason for this variation is dataset size. Training Model B on the combined AZ and SGC Stockholm data improves evaluation of AZ to 0.73, which is almost the same as evaluating the same model on SGC Stockholm. Adding SGC Toronto to the training data does not improve the evaluation results significantly for any data source.

Table 2.

Results of evaluating different RP3Net models trained on different data sources versus the baseline model and third-party predictors.

Model Trained on Tested on AUROC Accuracy Recall Precision
NetSolP solubility AZ 0.64 0.61 0.44 0.64
NetSolP useability AZ 0.64 0.57 0.13 0.87
B AZ AZ 0.59 0.52 0.85 0.50
B SGC Stockholm AZ 0.73 0.65 0.63 0.65
B SGC Stockholm, AZ AZ 0.73 0.66 0.72 0.63
B SGC Stockholm, SGC Toronto, AZ AZ 0.71 0.66 0.50 0.72
C SGC Stockholm, SGC Toronto, AZ AZ 0.72 0.66 0.52 0.71
D SGC Stockholm, AZ, SGC Toronto AZ 0.74 0.70 0.71 0.69
D SGC Stockholm, SGC Toronto AZ 0.69 0.67 0.49 0.75
NetSolP solubility SGC Stockholm 0.48 0.66 0.14 0.42
NetSolP useability SGC Stockholm 0.39 0.68 0.00 0.00
Baseline SGC Stockholm SGC Stockholm 0.62 0.62 0.39 0.40
A SGC Stockholm SGC Stockholm 0.70 0.63 0.74 0.45
B SGC Stockholm SGC Stockholm 0.72 0.63 0.74 0.45
B SGC Stockholm, AZ SGC Stockholm 0.73 0.62 0.78 0.45
B SGC Stockholm, SGC Toronto, AZ SGC Stockholm 0.70 0.60 0.64 0.42
C SGC Stockholm, SGC Toronto, AZ SGC Stockholm 0.68 0.62 0.56 0.43
D SGC Stockholm, SGC Toronto, AZ, SGC Stockholm 0.77 0.73 0.54 0.59
D SGC Stockholm, SGC Toronto SGC Stockholm 0.72 0.75 0.51 0.66
B SGC Stockholm, SGC Toronto, AZ SGC Toronto 0.75 0.87 0.12 0.27
C SGC Stockholm, SGC Toronto, AZ SGC Toronto 0.76 0.88 0.08 0.33

Numbers highlighted in bold indicate the highest value of the metric achieved when testing on a given dataset.

3.3 Meta label correction with purification data yields a 0.04 increase in AUROC on SGC Stockholm

Both Model A and Model B are fine-tuned on soluble protein expression data with frozen parameters of the FM. Unfreezing these parameters (Model C) and training on the full dataset leads to overfitting: perfect performance is quickly achieved on the training dataset (AUROC≈1.0), but on the validation and test sets the AUROC remains below 0.75. Training Model C on individual data sources also leads to overfitting, as expected.

This could be explained by the fact that the datasets contain the results of slightly different experiments. The SGC Toronto dataset reports results of large-scale purification, whereas both AZ and SGC Stockholm report small-scale expression testing captured with one-step purification. Although the exact experimental conditions, materials and methods used for purifications were not available during model development, it is safe to assume that the SGC Toronto conditions are quite different from the small-scale expression testing. A natural question arises: given the construct sequence from SGC Toronto, and its binary purification result, what would be the result of small-scale expression testing this construct under the conditions of SGC Stockholm or AZ?

We address this within the MLC framework (Zheng et al. 2021, Taraday and Baskin 2023), where a large, noisy dataset is used to aid training the model on a small, clean set. Rather than adding the noisy data directly to the training set, a special model is trained to predict the corrected label from the noisy input and noisy label. This is referred to as the ‘teacher model’. The corrected labels are used to train the ‘student model’, along with clean inputs and clean labels (Fig. 1B). In our set-up, SGC Toronto large-scale purification dataset is used to train the teacher model, and a union of SGC Stockholm and AZ small-scale expression data are used to train the student model.

Using MLC with large-scale purification data (Model D) achieves in AUROC of 0.74 on AZ dataset. This is an improvement of 0.01 compared to the second-best result of Model B on AZ data. When evaluating Model D on SGC Stockholm, AUROC reaches 0.77, which is an improvement of 0.04 over the next-best result. We have also observed that the MLC model is more robust across different sequence clusters—training, validation and testing—than other models, which tend to over-fit the training data. The MLC framework thus allows utilizing large-scale purification data to improve modelling of small-scale expression testing, whereas simple transfer learning (Model B or Model C trained on all sources) fails to achieve that outcome.

Adding a relatively small number of constructs from AZ to the training set has a larger impact on model performance than using MLC, when evaluated on SGC Stockholm data. This points to a potential further improvement of the model in the future, by better accounting for the difference in experimental conditions between the labs that have provided the data. Using MLC still results in the highest AUROC when evaluating on all data sources, which justifies using this approach.

3.4 Prospective experimental validation of the model shows AUROC of 0.83

To establish the utility of RP3Net for drug discovery projects, in addition to the normal train–validate–test model development loop, we have conducted prospective model evaluation in a real-life scenario. A set of 46 proteins was curated from the human proteome to include viable drug targets, whilst avoiding proteins with prior published evidence of successful expression. We started by generating two full length constructs per target (with a 6-His affinity tag placed at the N- or C-termini) and running RP3Net on them. If both constructs were predicted not to express, we generated trimmed constructs, ran these through the model, and, if they were predicted to express, included them in the dataset. This resulted in a total of 97 constructs for the experimental validation dataset, eight of which were generated by the trimming process. The constructs were cloned and expressed in E. coli at the AZ protein production facility. Forty-nine per cent of the constructs passed small-scale expression screening, including one-step purification by affinity chromatography (Fig. 3 and Table 2, available as supplementary data at Bioinformatics online).

The performance of RP3Net Models B and D was compared with the baseline model, and two third-party predictors: SoluProt (Hon et al. 2021) and NetSolP (Table 3; Table 2, available as supplementary data at Bioinformatics online). The highest AUROC of 0.83 is achieved by RP3Net Model D. This is 0.08 better than the best third-party predictor (NetSolP useability). RP3Net Model D showed an accuracy of 0.77 when the score cut-off of 0.5 was used, and accuracy of 0.81 with the cut-off set to 0.79.

Table 3.

Results of experimental validation of RP3Net, the baseline model and third-party predictors.

Model Trained on AUROC Accuracy Recall Precision
Soluprot 0.64 0.57 0.83 0.57
NetSolp solubility 0.66 0.60 0.79 0.59
NetSolp useability 0.75 0.57 0.23 0.86
Baseline SGC Stockholm 0.67 0.65 0.58 0.71
RP3Net B AZ 0.69 0.65 0.88 0.62
RP3Net B SGC Stockholm 0.81 0.72 0.96 0.65
RP3Net B AZ, SGC Stockholm, 0.77 0.71 0.94 0.66
RP3Net B AZ, SGC Stockholm, SGC Toronto 0.76 0.69 0.65 0.74
RP3Net D AZ, SGC Stockholm, SGC Toronto MLC 0.83 0.77 0.92 0.73
RP3Net D SGC Stockholm, SGC Toronto MLC 0.83 0.77 0.87 0.75

Numbers highlighted in bold indicate the highest value of the metric achieved when testing on a given dataset.

To assess the impact of having AZ data in the training set, the performance of Model D trained on the amended set that excludes AZ data was compared with the model that was trained on the full set. The small AZ dataset has very little impact on the model performance during experimental validation; both AUROC and accuracy figures are identical, with small differences in precision and recall with score cut-off of 0.5.

For the subset of eight trimmed constructs, RP3Net D shows an accuracy of 0.5 with score cut-off of 0.5, and accuracy of 0.62 with score cut-off of 0.79. This could be an artefact of the small evaluation set, or that RP3Net does not consider if sequences will fold into stable protein domains. Curiously, the trimmed constructs that did result in soluble protein also contained degradation products (Fig. 3, available as supplementary data at Bioinformatics online). Using the score cut-off of 0.5, the model predicts all trimmed constructs to express, whereas in fact only four out of eight were expressed successfully.

Performance on trimmed constructs could thus be considered an area for improvement. However, considering the small number of trimmed constructs, and the model accuracy (0.77) and precision (0.73) on the larger experimental validation set, it could be argued that an experimental scientist would still find the modelling results helpful. An ‘overconfident’, high recall, model that predicts too many positives, which are then partly confirmed in the laboratory, is preferrable to a model that misses out constructs that would have expressed in the lab. Model precision could be increased, at the expense of recall, by increasing the score cut-off threshold.

4 Conclusion

The recombinant production of proteins can require multiple experimental rounds of trial and error. To improve the efficiency of such experiments, we have developed RP3Net, an AI model of heterologous protein expression in E. coli. RP3Net predicts the results of protein expression as a binary outcome. It was built using the latest foundational models and was trained using a combination of internal experimental results from small-scale AZ expression screens, and publicly available data from the SGC. Using an STP aggregation layer and MLC with large-scale purification data enables RP3Net to achieve state-of-the-art performance both on the take-out data from SGC Stockholm and AZ. RP3Net has been experimentally validated on a manually selected set of constructs for viable human drug targets and outperformed third-party predictors on that set as well. Ablation studies show that there is no single method that achieves a large performance increase, but rather many small incremental improvements.

This work also underscores the need for large and well-curated datasets of soluble protein expression and for the scientific community to agree on how the data should be captured following the FAIR principles, and to establish a protein production ontology (Wilkinson et al. 2016, Edfeldt et al. 2024). Unfortunately, in the field of protein production, there is not yet an equivalent of the PDB for structural biology. Significant time in this project was spent on data curation.

The modelling results may also be further improved by making the model more aware of the experimental conditions, such as E. coli host strain, induction methods, time and temperature at which various experimental stages were performed, buffer formulations, etc, This information is largely missing from the available datasets.

RP3Net is already deployed and used by the protein scientists at AZ. This publication and the accompanying code repository at GitHub make the model available to the wider research community, both in industry and in academia (Edwards et al. 2025).

Supplementary Material

btag003_Supplementary_Data

Acknowledgements

We would like to thank Susanne Gräslund and Opher Gileadi from SGC Stockholm, and Matthieu Schapira and Peter Loppnau from SGC Toronto for sharing their respective datasets and helping with the curation. We would like to acknowledge colleagues from the Data Sciences and Quantitative Biology, and Protein Science departments at AstraZeneca for constructive discussions during this project and David Öling from BioPharmaceuticals R&D at AstraZeneca for overseeing the cloning of the experimental constructs. We would like to acknowledge Matthew Hall from the Industry Partnerships team at EMBL-EBI and Birgit Kerber and colleagues from EMBLEM for helping to organize the collaboration; and the EMBL-EBI IT team for maintaining the computational facilities used to train the models.

Contributor Information

Evgeny Tankhilevich, European Bioinformatics Institute (EMBL-EBI), Wellcome Genome Campus, Hinxton, Cambridgeshire, CB10 1SD, United Kingdom.

Sergio Martinez Cuesta, Data Sciences and Quantitative Biology, Discovery Sciences, BioPharmaceuticals R&D, Astra-Zeneca, Cambridge, CB2 0AA, United Kingdom.

Ian Barrett, Data Sciences and Quantitative Biology, Discovery Sciences, BioPharmaceuticals R&D, Astra-Zeneca, Cambridge, CB2 0AA, United Kingdom.

Carolina Berg, Protein Science, Structure and Biophysics, Discovery Sciences, BioPharmaceuticals R&D, Astra-Zeneca, Mölndal, 431 83, Sweden.

Lovisa Holmberg Schiavone, Protein Science, Structure and Biophysics, Discovery Sciences, BioPharmaceuticals R&D, Astra-Zeneca, Mölndal, 431 83, Sweden.

Andrew R Leach, European Bioinformatics Institute (EMBL-EBI), Wellcome Genome Campus, Hinxton, Cambridgeshire, CB10 1SD, United Kingdom.

Author contributions

Evgeny Tankhilevich (Data curation [lead], Methodology [equal], Software [lead], Writing—original draft [lead], Writing—review & editing [lead]), Sergio Martinez Cuesta (Project administration [equal], Supervision [equal], Writing—review & editing [equal]), Ian Barrett (Funding acquisition [equal], Project administration [equal]), Carolina Berg (Methodology [equal], Visualization [equal], Writing—review & editing [equal]), Lovisa Holmberg Schiavone (Conceptualization [lead], Funding acquisition [lead], Investigation [lead], Methodology [lead], Project administration [equal], Supervision [lead], Writing—review & editing [equal]), and Andrew R. Leach (Conceptualization [lead], Funding acquisition [lead], Investigation [equal], Project administration [lead], Supervision [lead], Writing—review & editing [equal])

Supplementary material

Supplementary material is available at Bioinformatics online.

Conflict of interest

S.M.C., I.B., C.B., and L.H.S. are present employees and may be shareholders of AstraZeneca.

Funding

This work has been supported by funding from the Member States of the European Molecular Biology Laboratory.

Data availability

The model code is available under an MIT licence at GitHub, https://github.com/RP3Net/RP3Net, DOI 10.5281/zenodo.17243498. The dataset used for training and evaluating the model, except for AZ data, is available at https://ftp.ebi.ac.uk/pub/software/RP3Net. The AZ data are subject to AZ confidentiality policies.

References

  1. Acton TB, Xiao R, Anderson S  et al.  Preparation of protein samples for NMR structure, function, and small-molecule screening studies. Methods Enzymol  2011;493:21–60. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Basafa M, Hashemi A, Behravan A.  Optimizing recombinant antibody fragment production: a comparison of artificial intelligence and statistical modeling. Biotechnol Appl Biochem  2024;71:1094–104. [DOI] [PubMed] [Google Scholar]
  3. Bateman A, Martin MJ, Orchard S  et al.  UniProt: the universal protein knowledgebase in 2023. Nucleic Acids Res  2023;51:D523–31. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Benson DA, Cavanaugh M, Clark K  et al.  GenBank. Nucleic Acids Res  2013;41:D36–42. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Berman HM, Westbrook J, Feng Z  et al.  The protein data bank. Nucleic Acids Res  2000;28:235–42. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Berman HM, Gabanyi MJ, Kouranov A  et al.  Protein Structure Initiative—TargetTrack 2000-2017—All Data Files.  2017. 10.5281/zenodo.821654. [DOI]
  7. Brandes N, Ofer D, Peleg Y  et al.  ProteinBERT: a universal deep-learning model of protein sequence and function. Bioinformatics  2022;38:2102–10. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Burgess-Brown NA, Sharma S, Sobott F  et al.  Codon optimization can improve expression of human genes in Escherichia coli: a multi-gene study. Protein Expr Purif  2008;59:94–102. [DOI] [PubMed] [Google Scholar]
  9. Burgess-Brown NA, Mahajan P, Strain-Damerell C  et al.  Screening and production of recombinant human proteins: protein production in E. coli. Methods Mol Biol  2021;2199:45–66. [DOI] [PubMed] [Google Scholar]
  10. Burley SK, Bhikadiya C, Bi C  et al.  RCSB protein data bank (RCSB.org): delivery of experimentally-determined PDB structures alongside one million computed structure models of proteins from artificial intelligence/machine learning. Nucleic Acids Res  2023;51:D488–508. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Buterez D, Janet JP, Kiddle SJ  et al.  Graph neural networks with adaptive readouts. In Koyejo S et al. (ed.) Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, New Orleans, LA, USA, November 28–December 9, 2022. Red Hook NY: Curran Associates, Inc., 2022, 19746–58. [Google Scholar]
  12. Chapple SD, Dyson MR.  High-throughput expression screening in mammalian suspension cells. Methods Mol Biol  2021;2199:117–25. [DOI] [PubMed] [Google Scholar]
  13. Cooper CDO, Marsden BD.  N- and C-terminal truncations to enhance protein solubility and crystallization: predicting protein domain boundaries with bioinformatics tools. Methods Mol Biol. 2017;1586:11–31. [DOI] [PubMed] [Google Scholar]
  14. Dallago C, Mou J, Johnston KE  et al. FLIP: Benchmark tasks in fitness landscape inference for proteins. In: Vanschoren J, Yeung S-K (ed.), Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks 1, NeurIPS Datasets and Benchmarks 2021, December 2021, Virtual, 2021.
  15. Devlin J, Chang M-W, Lee K  et al.  BERT: pre-training of deep bidirectional transformers for language understanding. arXiv, 10.48550/arXiv.1810.04805, 2018, preprint: not peer reviewed. [DOI]
  16. Edfeldt K, Edwards AM, Engkvist O  et al.  A data science roadmap for open science organizations engaged in early-stage drug discovery. Nat Commun  2024;15:5640–10. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Edwards AM, Owen DR, Zhang L  et al. ; Structural Genomics Consortium Target 2035 Working Group. Protein–ligand data at scale to support machine learning. Nat Rev Chem  2025;9:634–45. [DOI] [PubMed] [Google Scholar]
  18. Elnaggar A, Heinzinger M, Dallago C  et al.  ProtTrans: toward understanding the language of life through self-supervised learning. IEEE Trans Pattern Anal Mach Intell  2022;44:7112–27. [DOI] [PubMed] [Google Scholar]
  19. Francis DM, Page R.  Strategies to optimize protein expression in E. coli. Curr Protoc Protein Sci  2010;5:5.24.1–29. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Friedman JH.  Greedy function approximation: A gradient boosting machine. Ann Stat  2001;29:1189–232. 10.1214/aos/1013203451 [DOI] [Google Scholar]
  21. Gabanyi MJ, Adams PD, Arnold K  et al.  The structural biology knowledgebase: a portal to protein structures, sequences, functions, and methods. J Struct Funct Genomics  2011;12:45–54. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Gordon E, Horsefield R, Swarts HGP  et al.  Effective high-throughput overproduction of membrane proteins in Escherichia coli. Protein Expr Purif  2008;62:1–8. [DOI] [PubMed] [Google Scholar]
  23. Gräslund S, Sagemark J, Berglund H  et al.  The use of systematic N- and C-terminal deletions to promote production and structural studies of recombinant proteins. Protein Expr Purif  2008;58:210–21. [DOI] [PubMed] [Google Scholar]
  24. Haacke A, Fendrich G, Ramage P  et al.  Chaperone over-expression in Escherichia coli: apparent increased yields of soluble recombinant protein kinases are due mainly to soluble aggregates. Protein Expr Purif  2009;64:185–93. [DOI] [PubMed] [Google Scholar]
  25. Hayashi K, Kojima C.  pCold-GST vector: a novel cold-shock vector containing GST tag for soluble protein production. Protein Expr Purif  2008;62:120–7. [DOI] [PubMed] [Google Scholar]
  26. Hayes T, Rao R, Akin H  et al.  Simulating 500 million years of evolution with a language model. Science  2025;387:850–8. [DOI] [PubMed] [Google Scholar]
  27. Hon J, Marusiak M, Martinek T  et al.  SoluProt: prediction of soluble protein expression in Escherichia coli. Bioinformatics  2021;37:23–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Ji Y, Zhou Z, Liu H  et al.  DNABERT: pre-trained bidirectional encoder representations from transformers model for DNA-language in genome. Bioinformatics  2021;37:2112–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Jumper J, Evans R, Pritzel A  et al.  Highly accurate protein structure prediction with AlphaFold. Nature  2021;596:583–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Kurashiki R, Koyama K, Sakaguchi Y  et al.  Development of a thermophilic host–vector system for the production of recombinant proteins at elevated temperatures. Appl Microbiol Biotechnol  2023;107:7475–88. [DOI] [PubMed] [Google Scholar]
  31. Lee J, Lee Y, Kim J  et al.  Set transformer: a framework for attention-based permutation-invariant neural networks. In: Chaudhuri K, Salakhutdinov R (eds.), Proceedings of the 36th International Conference on Machine Learning in Long Beach. Vol. 97. Cambridge MA: JMLR, 2019, 3744–53. [Google Scholar]
  32. Li F-Z, Amini AP, Yue Y  et al.  Feature reuse and scaling: understanding transfer learning with protein language models. In: Salakhutdinov R, Kolter Z, Heller K  et al. (eds.), Proceedings of the 41st International Conference on Machine Learning in Vienna. Vol. 235. Cambridge MA: JMLR, 2024, 27351–75. [Google Scholar]
  33. Lin Z, Akin H, Rao R  et al.  Evolutionary-scale prediction of atomic-level protein structure with a language model. Science (1979)  2023;379:1123–30. [DOI] [PubMed] [Google Scholar]
  34. Mahajan P, Ellis K, Mukhopadhyay S  et al.  Expression screening of human integral membrane proteins using BacMam. Methods Mol Biol  2021a;2199:95–115. [DOI] [PubMed] [Google Scholar]
  35. Mahajan P, Strain-Damerell C, Mukhopadhyay S  et al.  Screening and production of recombinant human proteins: protein production in insect cells. Methods Mol Biol  2021b;2199:67–94. [DOI] [PubMed] [Google Scholar]
  36. Mizianty MJ, Kurgan L.  Sequence-based prediction of protein crystallization, purification and production propensity. Bioinformatics  2011;27:i24–33. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Morão LG, Manzine LR, Clementino LOD  et al.  A scalable screening of E. coli strains for recombinant protein expression. PLoS One  2022;17:e0271403. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Nguyen E, Poli M, Faizi M  et al. HyenaDNA: long-range genomic sequence modeling at single nucleotide resolution. In: Oh A et al. (ed.) Proceedings of the 37th International Conference on Neural Information Processing Systems. Red Hook NY: Curran Associates, Inc., 2023, 43177–201.
  39. Ochoa D, Hercules A, Carmona M  et al.  The next-generation open targets platform: reimagined, redesigned, rebuilt. Nucleic Acids Res  2023;51:D1353–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Outeiral C, Deane CM.  Codon language embeddings provide strong signals for use in protein engineering. Nat Mach Intell  2024;6:170–9. [Google Scholar]
  41. Rao R, Bhattacharya N, Thomas N  et al.  Evaluating protein transfer learning with TAPE. Adv Neural Inf Process Syst  2019;32:9689–701. [PMC free article] [PubMed] [Google Scholar]
  42. Raymond A, Lovell S, Lorimer D  et al.  Combined protein construct and synthetic gene engineering for heterologous protein expression and crystallization using gene composer. BMC Biotechnol  2009;9:37–15. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Richardson L, Allen B, Baldi G  et al.  MGnify: the microbiome sequence data analysis resource in 2023. Nucleic Acids Res  2023;51:D753–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Rives A, Meier J, Sercu T  et al.  Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences. Proc Natl Acad Sci USA  2021;118:e2016239118. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Savitsky P, Bray J, Cooper CDO  et al.  High-throughput production of human proteins for crystallization: the SGC experience. J Struct Biol  2010;172:3–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Sayers EW, Cavanaugh M, Clark K  et al.  GenBank 2024 update. Nucleic Acids Res  2024;52:D134–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Schütz A, Bernhard F, Berrow N  et al.  A concise guide to choosing suitable gene expression systems for recombinant protein production. STAR Protoc  2023;4:102572. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Simm D, Popova B, Braus GH  et al.  Design of typical genes for heterologous gene expression. Sci Rep  2022;12:9625–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Singh N, Vayer P, Tanwar S  et al.  Drug discovery and development: introduction to the general public and patient groups. Front Drug Discov  2023;3:1201419. [Google Scholar]
  50. Söderberg JJ, Grgic M, Hjerde E  et al.  Aliivibrio wodanis as a production host: development of genetic tools for expression of cold-active enzymes. Microb Cell Fact  2019;18:1–16. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Sormanni P, Aprile FA, Vendruscolo M.  The CamSol method of rational design of protein mutants with enhanced solubility. J Mol Biol  2015; 427:478–90. [DOI] [PubMed] [Google Scholar]
  52. Sormanni P, Amery L, Ekizoglou S  et al.  Rapid and accurate in silico solubility screening of a monoclonal antibody library. Sci Rep  2017;7:8200–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Strain-Damerell C, Mahajan P, Fernandez-Cid A  et al.  Screening and production of recombinant human proteins: ligation-independent cloning. Methods Mol Biol  2021;2199:23–43. [DOI] [PubMed] [Google Scholar]
  54. Structural Genomics Consortium, China Structural Genomics Consortium, Northeast Structural Genomics Consortium  et al.  Protein production and purification. Nat Methods  2008; 5:135–46. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Sulkowski E.  Purification of proteins by IMAC. Trends Biotechnol  1985;3:1–7. [Google Scholar]
  56. Suzek BE, Wang Y, Huang H  et al. ; UniProt Consortium. UniRef clusters: a comprehensive and scalable alternative for improving sequence similarity searches. Bioinformatics  2015;31:926–32. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Taraday MK, Baskin C. Enhanced meta label correction for coping with label corruption. In: 2023 IEEE/CVF International Conference on Computer Vision (ICCV) in Paris. NY: IEEE, 2023. 10.1109/ICCV51070.2023.01493 [DOI]
  58. The SGC Consortium. Protein production data from the SGC. BioStudies, 2022. https://www.ebi.ac.uk/biostudies/SGC/studies/SBSST681 (1 October 2024, date last accessed).
  59. Thumuluri V, Martiny H-M, Almagro Armenteros JJ  et al.  NetSolP: predicting protein solubility in Escherichia coli using language models. Bioinformatics  2022;38:941–6. [DOI] [PubMed] [Google Scholar]
  60. Vaswani A, Shazeer N, Parmar N  et al.  Attention is all you need. Adv Neural Inf Process Syst  2017;2017:5999–6009. [Google Scholar]
  61. Wang H, Feng L, Zhang Z  et al.  Crysalis: an integrated server for computational analysis and design of protein crystallization. Sci Rep  2016;6:21383. [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Wang H, Wang M, Tan H  et al.  PredPPCrys: accurate prediction of sequence cloning, protein production, purification and crystallization propensity from protein sequences using multi-step heterogeneous feature fusion and selection. PLoS One  2014;9:e105902. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Wang S, Zhao H.  SADeepcry: a deep learning framework for protein crystallization propensity prediction using self-attention and auto-encoder networks. Brief Bioinform  2022;23. 10.1093/bib/bbac352 [DOI] [PubMed] [Google Scholar]
  64. Wilkinson MD, Dumontier M, Aalbersberg IJ  et al.  The FAIR guiding principles for scientific data management and stewardship. Sci Data  2016;3:160018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Xiong D, Kaicheng U, Sun J  et al.  PLMC: language model of protein sequences enhances protein crystallization prediction. Interdiscip Sci  2024;16:802–13. [DOI] [PubMed] [Google Scholar]
  66. Yuan Q, Tian C, Song Y  et al.  GPSFun: geometry-aware protein sequence function predictions with language models. Nucleic Acids Res  2024;52:W248–55. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Zanders ED.  The Science and Business of Drug Discovery: Demystifying the Jargon. 2nd edn. London: Springer International Publishing, 2020. [Google Scholar]
  68. Zdrazil B, Felix E, Hunter F  et al.  The ChEMBL database in 2023: a drug discovery platform spanning multiple bioactivity data types and time periods. Nucleic Acids Res  2024;52:D1180–92. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Zhang JH, Shan LL, Liang F  et al.  Strategies and considerations for improving recombinant antibody production and Squality in Chinese hamster ovary cells. Front Bioeng Biotechnol  2022;10:856049. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Zhang X, Hu X, Zhang T  et al.  PLM_Sol: predicting protein solubility by benchmarking multiple protein language models with the updated escherichia coli protein solubility dataset. Brief Bioinform  2024;25:1–10. 10.1093/bib/bbae404 [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Zheng G, Awadallah AH, Dumais S.  Meta label correction for noisy label learning. In: Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), held online. Vol. 3. Red Hook NY: Curran Associates, Inc., 2021, 11053–61.
  72. Zhong N, Loppnau P, Seitova A  et al.  Optimizing production of antigens and fabs in the context of generating recombinant antibodies to human proteins. PLoS One  2015;10:e0139695. [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Zhou Z, Ji Y, Li W  et al. DNABERT-2: efficient foundation model and benchmark for multi-species genome. arXiv, 2023.
  74. Zhu Y-H, Hu J, Ge F  et al.  Accurate multistage prediction of protein crystallization propensity using deep-Cascade Forest with sequence-based features. Brief Bioinform  2021;22:1–14. 10.1093/bib/bbaa076 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Citations

  1. Berman HM, Gabanyi MJ, Kouranov A  et al.  Protein Structure Initiative—TargetTrack 2000-2017—All Data Files.  2017. 10.5281/zenodo.821654. [DOI]

Supplementary Materials

btag003_Supplementary_Data

Data Availability Statement

The model code is available under an MIT licence at GitHub, https://github.com/RP3Net/RP3Net, DOI 10.5281/zenodo.17243498. The dataset used for training and evaluating the model, except for AZ data, is available at https://ftp.ebi.ac.uk/pub/software/RP3Net. The AZ data are subject to AZ confidentiality policies.


Articles from Bioinformatics are provided here courtesy of Oxford University Press

RESOURCES