Skip to main content
International Journal of Molecular Sciences logoLink to International Journal of Molecular Sciences
. 2023 Jan 30;24(3):2595. doi: 10.3390/ijms24032595

MetaSEM: Gene Regulatory Network Inference from Single-Cell RNA Data by Meta-Learning

Yongqing Zhang 1, Maocheng Wang 1, Zixuan Wang 1, Yuhang Liu 1, Shuwen Xiong 1, Quan Zou 2,*
Editor: Sarath Chandra Janga
PMCID: PMC9916710  PMID: 36768917

Abstract

Regulators in gene regulatory networks (GRNs) are crucial for identifying cell states. However, GRN inference based on scRNA-seq data has several problems, including high dimensionality and sparsity, and requires more label data. Therefore, we propose a meta-learning GRN inference framework to identify regulatory factors. Specifically, meta-learning solves the parameter optimization problem caused by high-dimensional sparse data features. In addition, a few-shot solution was used to solve the problem of lack of label data. A structural equation model (SEM) was embedded in the model to identify important regulators. We integrated the parameter optimization strategy into the bi-level optimization to extract the feature consistent with GRN reasoning. This unique design makes our model robust to small-scale data. By studying the GRN inference task, we confirmed that the selected regulators were closely related to gene expression specificity. We further analyzed the GRN inferred to find the important regulators in cell type identification. Extensive experimental results showed that our model effectively captured the regulator in single-cell GRN inference. Finally, the visualization results verified the importance of the selected regulators for cell type recognition.

Keywords: meta-learning, gene regulator network inference, structural equation model, bi-level optimization

1. Introduction

The inference of gene regulatory networks (GRNs) allows for a better understanding of transcriptional regulation and how it works in cell-type identification. The models of GRN inference based on scRNA-seq achieved success in cancer treatment [1], the recognition of cellular homeostasis [2], and single-cell multi-omic studies [3]. However, single-cell RNA sequencing technology has many limitations, such as technical noise [4], high gene variability [5], and batch effect [6]. Those regulators whose activities are highly variable among different cell types and predict a small set of essential regulators for significant cell types still need more attention. Therefore, inferring gene regulatory networks to study cell-specific phenomena on computational methods is a challenging problem in bioinformatics.

Recently, deep learning has brought new solutions [7,8] to single-cell GRN inference based on coexpression [9]. The supervised approach has two ways to infer GRNs. One directly sets a determined ground-truth label as the model convergence target, such as DGRN [10] and DeepDRIM [11]. The other is to embed a ground-truth network into the model, such as the GRGNN [12] and scSGL [13]. These models could predict the potential gene regulatory relationships in high-dimensional scRNA-seq data. However, the supervised method can only be applied to general tasks with label data. The unsupervised approach facilitates GRN inference without a ground-truth label, such as VEGA [14] and DeepSEM [15]. SCODE [16] is a machine learning algorithm based on linear ordinal differential equations. GENIE3 [17] and GRNBoost2 [18] can also complete GRN reasoning without the label. However, the problem of highly sparse data features still needs to be resolved better.

Meta-learning is adept at solving parameter initialization and insufficient data labels when using small-scale samples [19]. The meta-learning model is composed of a basic learner and a meta-learner. The basic learner extracts feature information through representation study. The meta-learner guides the basic learner to complete training tasks by learning the parameters of the basic learner. The meta-learner synthesizes the training experience of all modules and provides the initial parameters for new task training. By introducing meta-learning, the generalization ability of deep learning models can be improved to solve the highly sparse data features. For example, Kun Fu et al. employed transfer learning to alleviate the training problem of meta-learning on few-shot tasks [20]. Arkabandhu Chowdhury et al. proposed a meta-learning method to complete the data classification on small-scale samples [21]. Zitian Chen et al. solved the low data samples of one-shot learning by meta-learning in the image deformation problem [22]. However, the existing meta-learning methods ought to sufficiently adapt to the single-cell GRN inference task.

To address the problems above, we proposed a meta-learning framework called MetaSEM (Figure 1) to infer GRN from scRNA-seq. Specifically, we adopted meta-learning to optimize the parameters of each module for learning high-dimensional data features. Next, we employed a meta-decoder to provide pseudo-data labels for the encoder. Then, we put the feature vectors extracted from the encoder into the training process. After that, considering the gene regulatory relationship is an endogenous variable and the gene expression information is an exogenous variable, we initialized a structural equation model (SEM) [23] adjacency matrix as the GRN layer. We considered the matrix as the regulatory weight matrix and embedded it into the meta-decoder. Finally, all the parameters were optimized using a Bi-Level optimization.

Figure 1.

Figure 1

(A) The overview of MetaSEM: the meta-decoder extracts the regulatory relationship to output a pseudo-data label. The encoder transforms the data feature into feature vectors. The GRN Layer is a specially designed layer for embedding the SEM matrix. The red arrows indicate the outer loop, and the yellow arrows indicate the inner loop. With hyperparameter optimization, MetaSEM integrates the outer and inner loop based on gradient. The θF represents the hyperparameters of the encoder, and θA represents the hyperparameters of the meta-decoder. (B) By analyzing the SEM matrix, MetaSEM performs three major functions: identification of the regulators, GRN visualization, and cell-type identification.

In this work, we verified the reliability of MetaSEM from the following aspects. We first compared the model performance of MetaSEM with several state-of-the-art methods. The experimental results showed that MetaSEM was significantly better than current methods in terms of EPR, AUPR, and AUROC. Next, to analyze the robustness of MetaSEMwe, we explored essential data features in single-cell data at different data scales. We then generated cell-type-specific GRNs on the bulk RNA-seq dataset for further research. Pearson correlation analysis of GRNs and gene expression data analysis indicated that the GRNs were cell-type-specific. Finally, the visualization of cell-type-specific GRNs within the HNSCC dataset demonstrated the regulators’ importance in identifying cell types.

2. Results and Discussion

2.1. Comparison with Existing Methods

To verify the performance of MetaSEM, we compared the model with four methods: DeepSEM [15], DGRN [10], GENIE3 [17], and PIDC [2] on the BEELINE dataset. As shown on the 1000 gene datasets in Table 1, MetaSEM outperformed the existing methods on the three evaluation metrics. The EPR of MetaSEM in mHSC-L, mHSC-G, and mHSC-E datasets were 1.36, 1.41, and 1.21, respectively. The EPR of MetaSEM in the mHSCs dataset was 0.15 higher than that of DeepSEM on average. MetaSEM was 0.41 higher than DGRN in AUPR and AUROC. In addition, GENIE3 and PIDC are far inferior to deep learning methods because they are unsupervised machine learning methods. The results show that MetaSEM can effectively memorize gene regulatory relationships and take such relationships to guide the model to extract essential information.

Table 1.

Performance comparison of four competing methods on 1000 gene datasets and 500 gene datasets.

EPR
Methods 1000 Gene Datasets 500 Gene Datasets
mHSC-L mHSC-GM mHSC-E hESC mESC hHep mDC mHSC-L mHSC-GM mHSC-E
DeepSEM 1.09 1.14 1.24 1.43 1.06 1.14 2.55 1.07 1.06 1.30
DGRN - - - - - - - - - -
GENIE3 <1 1.03 1.01 1.00 1.06 1.12 1.01 1.00 1.06 1.03
PIDC <1 <1 <1 <1 1.01 1.03 <1 <1 <1 <1
MetaSEM 1.36 1.41 1.24 2.29 1.24 1.53 3.36 1.13 1.20 1.69
AUPR
Methods 1000 Gene Datasets 500 Gene Datasets
mHSC-L mHSC-GM mHSC-E hESC mESC hHep mDC mHSC-L mHSC-GM mHSC-E
DeepSEM 0.63 0.56 0.42 0.30 0.38 0.48 0.43 0.66 0.63 0.58
DGRN 0.15 0.16 0.25 - - - - 0.15 0.27 0.25
GENIE3 0.09 0.12 0.09 - - - - 0.14 0.15 0.14
PIDC 0.07 0.12 0.10 - - - - 0.16 0.12 0.19
MetaSEM 0.70 0.73 0.66 0.48 0.43 0.70 0.33 0.75 0.77 0.84
AUROC
Methods 1000 Gene Datasets 500 Gene Datasets
mHSC-L mHSC-GM mHSC-E hESC mESC hHep mDC mHSC-L mHSC-GM mHSC-E
DeepSEM 0.51 0.57 0.63 0.52 0.51 0.54 0.77 0.52 0.52 0.67
DGRN 0.63 0.67 0.75 - - - - 0.63 0.71 0.77
GENIE3 0.63 0.63 0.59 - - - - 0.52 0.52 0.55
PIDC 0.57 0.61 0.60 - - - - 0.47 0.47 0.59
MetaSEM 0.75 0.77 0.76 0.81 0.61 0.77 0.71 0.67 0.72 0.87

Part of the results are from the following two papers [10,15]; “-” indicates that the experimental result is missing, and “<1” indicates that the result here is random prediction. The bold number refers to the best performer.

To further evaluate the performance of MetaSEM, we trained MetaSEM on the 500 gene datasets. It should be noted that the label data in 500 gene datasets was just as 30% of that in 1000 gene datasets. The results on the 500 gene datasets in Table 1 show that MetaSEM still surpassed the other methods. The average performance of the model on the three metrics was 1.34, 0.78, and 0.75. As a result, we can observe that MetaSEM can perform well in the shortage of label data.

2.2. MetaSEM Can Adapt to High Dimensions and Sparse Characteristics

In this section, we present the results of the robustness of MetaSEM in different data scales. We first set the sampling points according to the standard deviation of gene expression and sample size. Then, we built sub-datasets using mHSC-L, mHSC-E, and mHSC-GM. In addition, we also created another dataset of the same points through random selection as a reference. We used the same hyperparameters and five-fold cross-validation during model training. As shown in Figure 2, the red area is larger than the blue area. The fluctuation range of the model on the three metrics was 11.5%, 15%, and 30.7%, respectively. This is because genes with a large number of 0 expression values also had a lower standard deviation. This means that removing these genes would reduce the sparsity of scRNA-seq data. By comparing the two gene selections, we found that the degree of data dispersion is an essential feature.

Figure 2.

Figure 2

The robustness of our model on different data scales. Each column corresponds to a cell’s sub-dataset (left: mHSC-L, middle: mHSC-GM, and right: mHSC-E), and each row corresponds to an evaluation index (top: EPR, middle: AUPR, and bottom: AUROC). The red region of the figure is the result of standard deviation selection, and the blue region of the figure is the result of random selection. Pretraining and fine-tuning were not conducted for each test.

2.3. MetaSEM Reveals That GRN Specificity Is Related to Gene Expression

To verify whether MetaSEM captures specific information, we analyzed the inferred GRN and gene expression data. We generated the cell-type-specific GRN based on eight HNSCC sub-datasets. Figure 3 shows the heatmap of the Pearson correlation coefficient between the different GRNs. In this matrix, two observations are made. First, the average correlation coefficient is less than 0.1, which indicates that the inferred GRNs have a very low correlation. Second, as shown in the cancer row, the highest correlation coefficient is Fibroblast. There are two reasons for this: the cancer dataset contains some cancer-associated fibroblasts (CAFs) [24], and the correlation between Fibroblast and Cancer subsets is the highest, which corresponds to the findings of [25].

Figure 3.

Figure 3

The Pearson correlation of different cell-type GRNs. Each element in the matrix represents the Pearson correlation of the GRN corresponding to two different cells. We do not show the results with a p-value greater than 0.05.

Given the phenomenon reflected in Figure 3, we further analyzed the differences in the gene expression in the datasets. As shown in Figure 4, the gene expression data of the Fibroblast and Endothelial subsets has little difference with the cancer subset. However, the results of the B cell and Mast subsets show significant differences. The p-values obtain an independent t-test of Fibroblast and Endothelial subsets are 0.17 and 0.37, respectively. In contrast, the Mast cell and B cell datasets are 0.05 and 0.03, respectively. These results indicate that MetaSEM can capture the specificity information, which is essential for general GRN inference tasks.

Figure 4.

Figure 4

The divergence of gene expression on different cell types. The red dot represents the genes with a positive correlation, the blue dot represents the genes with a negative correlation, and the black dot represents the gene with no difference in expression level. The grey dot represents the gene below the threshold.

2.4. The Selected Regulators from the SEM Model Have a Higher Expression Level

Next, we analyzed how MetaSEM extracts regulators. When sorting out the output of the GRN Layer, we found that the regulatory weight of some genes was very high. Therefore, we collected the regulatory weights of these genes. The boxplots in Figure 5 show the results. These genes had higher regulatory weight in Fibroclast, T cell, Cancer, and Endothelial. As shown in the t-SNE plots of Figure 5, the selected genes had significant weight distribution on different samples. The log2 (transcripts per kilobase per million (TPM) + 1) of ATF4, JUN, RPL7A, and RPS4X on HNSCC cells were 8.1, 6.5, 6.8, and 9.8, respectively. Finally, we selected ATF4, JUN, RPL7A, RPS4X, and other genes as regulators through cross-comparison.

Figure 5.

Figure 5

The regulatory weight of the different genes in the eight cells. Four regulators are presented: ATF4, JUN, RPL7A, and RPS4X. The boxplots show the weight distribution of the regulators on different SEM matrices. The t-SNE plots represent the weight distribution of regulators on the datasets.

2.5. The Selected Regulators Are the Main Factors of Cell-Type Identification

In this section, we present the relationship between the selected regulators and cell-type identification. Figure 6 visualizes the GRNs of Cancer and Fibroblast datasets together. The blue edge represents the typical regulatory relationship. Green and red edges are the regulatory relationships only in Cancer GRNs or Fibroblast GRNs, respectively. We marked several regulators in GRNs, such as STAT1, JUN and JUNB. The protein encoded by STAT1 is a member of the STAT protein family. STAT1 mediates the expression of various genes, which is essential for cell viability in response to different cell stimuli and pathogens [26]. JUN and JUNB belong to the same gene family and are related to human malignancies. JUN often occurs in chromosomal regions of translocation and deletion in human malignancies [27].

Figure 6.

Figure 6

Visualization of the GRN inference by MetaSEM on Cancer and Fibroblast datasets. The size of nodes indicates the regulatory weight. The blue edges are the main part of GRN, indicating the common regulatory relationship between the two cells. Green and red regulatory relationships only exist in Cancer GRNs or Fibroblast GRNs.

In order to verify the importance of the regulatory weight of cell-type recognition. We collected the cell types’ SEM matrix for each gene’s regulatory weight. Then, all genes were divided into three equal parts according to the regulatory weight for cell clustering. Then, we used the Louvain and Leiden methods to cluster cell types. As shown in Figure 7, the effect of clustering positively correlated with regulatory weight. The performance of the clustering methods on normalized mutual information (NMI), v-score, and adjusted rand index (ARI) proved this point.

Figure 7.

Figure 7

Visualization of selected regulators with different regulatory weights. Each row represents a clustering method (top row: Louvain, bottom row: Leiden). Each column represents the regulatory weight of the selected data by ascending ranking. The dimension reduction method of the graph is TSNE.

3. Materials and Methods

3.1. Data Preparation

The BEELINE [28] dataset is used to evaluate the model’s performance. The single-cell dataset contains seven cell types, including five mouse cells and two human cells. For the BEELINE dataset, we excluded the cells annotated as low quality and genes expressed in less than 10% of cells. Then, we logarithmically normalized the remaining data. Each cell only retained the top 1000 standard deviation genes. The sub-datasets were divided according to different cell types. We further built 500 gene datasets in the same way to evaluate the model performance in insufficient label data. The ground-truth GRNs were preprocessed and normalized according to the descriptions in [28].

The head and neck squamous cell carcinoma (HNSCC) dataset [25] is used to study GRN cell specificity, a heterogeneous epithelial tumor closely related to the long-term exposure of cells to the environment of alcohol and tobacco. We divided the dataset into ten subsets based on known cell-type annotations (Fibroblast, B cell, T cell, Endothelial, Dendritic, Mast, Cancer,-Fibroblast, Myocyte, and Macrophage). We discarded the subsets without annotations, and the number of samples was less than 50. We also pitched the genes expressed in less than 30 samples. Finally, the -Fibroblast and Myocyte are removed. The genes with the top 1000 standard deviations were utilized as the training dataset. The ground-truth label corresponding to the HNSCC dataset was obtained from the TCGA database [29].

The datasets were stored in matrices with the vertical axis gene expression and the sample as the horizontal axis. One cell type corresponds to one matrix, and the values in the matrix represent the expression values of genes on the samples. Similar to DeepSEM’s training process [15], we took 64 samples of the dataset as a batch. Each batch was considered a small sample learning task, and the learning goal was unique through the loss function. There was no need to divide the test and verification sets on GRN inference.

3.2. Model Description

The proposed MetaSEM consisted of three parts: an encoder, a meta-decoder, and a GRN layer. (i) Encoder: This part encodes the gene expression data as the feature vector using a three-layer MLP. (ii) Meta-decoder: This part models the regulatory relationship via a two-layer MLP and uses the GRN layer to find the optimal pseudo-data labels. (iii) GRN Layer: With the SEM model, the GRN layer infers the gene regulatory relationships and transforms these relationships into pseudo-data labels.

3.2.1. GRN Layer

We generalized the SEM as a GRN layer to model the conditional dependencies among random variables. We employed the meta-decoder to iterate the GRN Layer for extracting regulatory information. The final output of this module was an adjacency matrix representing GRN, and the elements within the matrix described directed edge weights. The GRN Layer iteration formula is as follows:

A*=θA×A+α×A (1)

where A represents the adjacency matrix resulting from modeling based on the SEM, θA represents the model parameters of the meta-decoder, and α is utilized to control the learning rate of the matrix in the new iteration.

3.2.2. Encoder

We built a decoder to capture the data feature from gene expression. The encoder reads the natural gene expression data X in batches. The data Xi of batch i first passes through a full connection layer to learn the gene expression feature [30]. Then, the feature vector Xip is obtained using a double-layer full connection layer.

Xip=JF(θF)(Xi) (2)

where JF(θF)(Xi) represents the working process of the encoder to calculate feature vector Xip. Yip represents the pseudo-data label from the meta-decoder. The process obtained with the parameter of the encoder θF from each epoch can be expressed by Formula (3):

θF*=argminθFEviVLLJF(θF)Xi,Yip (3)

where L represents the cross-entropy loss of the network, Yp represents the pseudo-data label provided by the meta-decoder, and vi is any gene belonging to the dataset VL.

3.2.3. Meta-Decoder

We embedded the GRN Layer into a two-layer MLP to construct the meta-decoder. Unlike the attention mechanism [31,32], we employed the meta-decoder to guide the feature extraction. Specifically, the objective of the meta-decoder is to find a prediction matrix Yip that agrees with the label matrix Y. The meta-decoder learns the gene regulatory relationships from the scRNA-seq. Then, the regulatory relationships are stored in the GRN Layer. Finally, the meta-decoder outputs the pseudo-data labels to represent the potential regulatory relationships. This procedure can improve the efficiency of the encoder by pseudo-data labels.

Yip=JA(θA)(Xip) (4)

where JA(θA)(Xip) indicates the process of calculating pseudo-data labels with the meta-decoder. In each round of training, the process of calculating and updating from the feature vector set Vp is shown in Formula (5):

θA*=argminθAEviVpLJA(θA)Xip,Yip (5)

3.2.4. Hyperparameter Optimization

We employed a bi-Level optimization for hyperparameter optimization. Specifically, the bi-Level optimization calculates the model’s parameters twice and compares the gradients generated. If the gradients drop, the parameters are updated. If not, the second parameter calculation is executed again [33]. The bi-Level optimization operation can help the model approximate the optimal parameter solution, reducing the impact of the erroneous feature vector on the model. We applied a one-step gradient-descent updating strategy with a significant learning rate to obtain the approximate optimal solution. After the approximate optimal parameters are received, the model parameters need to be updated. μθF and μθA represent the learning rate of the encoder and meta-decoder, respectively. The parameter update operation of the encoder and meta-decoder is expressed by Formulas (6) and (7):

θF*=θFμθFθFJF(Xi) (6)
θA*=θAμθAθAJA(Xip) (7)

Finally, the encoder and meta-decoder outputs are extracted for model loss calculation. The overall loss calculation of the model is expressed in Equation (8):

Loss=argminθEviVLLXip,Yip,:+β×(W1)22 (8)

where (W1) represents the weights of the first fully connected layer of the encoder. β is the weight of the encoder. Considering the feature vector and the pseudo-data labels as the main factors in the loss calculation guarantees that the contents learned by the encoder and meta-decoder remain consistent.

Xp*=θF(X)×0.5+Xp (9)

Formula (9) represents the method of the encoder to the output feature vector.

3.3. Implementation

MetaSEM is mainly implemented based on the PyTorch framework. The determined model configurations are as follows: the encoder is built from a three-layer MLP model, and the meta-decoder is built from a two-layer MLP with the GRN Layer. Hyperparameter optimization is a one-step gradient-updating strategy based on bi-Level optimization. We employed the grid search method to determine the model’s architecture and hyperparameters. The number of hidden-layer neurons set 128, and the batch size set 64. The encoder and meta-decoder optimization employed the Adam optimizer method. The learning rate is 1×104 for the encoder, 5×104 for the meta-decoder, and 1×102 for the one-step gradient-updating strategy. In addition, we set the learning decay rate and the other parameters to default. This article presents the codes of the model and can be downloaded from GitHub.

3.4. Evaluation Metrics

Following the BEELINE framework [34], we applied the top K edges to evaluate the performance of MetaSEM on three indices: early precision ratio (EPR), Area Under the precision-recall curve (AUPR), and area under the receiver-operating characteristic curve (AUROC), where the value of K is equal to the number of edges of the ground-truth GRN. The above metrics were developed as indices to evaluate model performance in previous studies [10,15].

4. Conclusions

This paper proposed a new GRN inference algorithm based on meta-learning to analyze the importance of the selected regulators in cell-type identification. The MetaSEM learns the potential regulatory relationships from gene expression data. Moreover, meta-learning was also used to optimize the process of feature extraction. Extensive experiments in different single-cell datasets showed that MetaSEM performed better than several advanced computational methods in GRN inference tasks. Finally, by visualizing the inferred GRN, we systematically analyzed the importance of the data dispersion. We proved the importance of the selected regulator in cell-type identification. In the future, we intend to construct GRN by fusing scATAC-seq data and scRNA-seq data to explore the effect of GRNs on single cells.

Author Contributions

Conceptualization, Y.Z. and Q.Z.; methodology, Y.Z., M.W. and Z.W.; software, M.W. and Z.W.; validation, M.W. and Y.Z.; formal analysis, M.W. and Y.L.; resources, M.W.; data preparation, S.X.; writing—original draft preparation, Y.Z. and M.W.; writing—review and editing, Y.Z. and Q.Z.; supervision, Q.Z. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset for model training and testing is available at https://github.com/ZhangLab312/MetaSEM (accessed on 1 January 2023).

Conflicts of Interest

The authors declare no conflict of interest.

Funding Statement

This research was funded by the National Natural Science Foundation of China Grant Nos. 62272067, 62131004, 62250028; 2011 Collaborative Innovation Center for Image and Geospatial Information of Sichuan Province.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Wouters J., Atak Z.K., Aerts S. Decoding transcriptional states in cancer. Curr. Opin. Genet. Dev. 2017;43:82–92. doi: 10.1016/j.gde.2017.01.003. [DOI] [PubMed] [Google Scholar]
  • 2.Chan T.E., Stumpf M.P., Babtie A.C. Gene regulatory network inference from single-cell data using multivariate information measures. Cell Syst. 2017;5:251–267. doi: 10.1016/j.cels.2017.08.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Fiers M.W., Minnoye L., Aibar S., Bravo González-Blas C., Kalender Atak Z., Aerts S. Mapping gene regulatory networks from single-cell omics data. Briefings Funct. Genom. 2018;17:246–254. doi: 10.1093/bfgp/elx046. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Wu Y., Zhang K. Tools for the analysis of high-dimensional single-cell RNA sequencing data. Nat. Rev. Nephrol. 2020;16:408–421. doi: 10.1038/s41581-020-0262-0. [DOI] [PubMed] [Google Scholar]
  • 5.Chen G., Ning B., Shi T. Single-cell RNA-seq technologies and related computational data analysis. Front. Genet. 2019:317. doi: 10.3389/fgene.2019.00317. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Chai L.E., Loh S.K., Low S.T., Mohamad M.S., Deris S., Zakaria Z. A review on the computational approaches for gene regulatory network construction. Comput. Biol. Med. 2014;48:55–65. doi: 10.1016/j.compbiomed.2014.02.011. [DOI] [PubMed] [Google Scholar]
  • 7.Kang T., Ding W., Zhang L., Ziemek D., Zarringhalam K. A biological network-based regularized artificial neural network model for robust phenotype prediction from gene expression data. BMC Bioinform. 2017;18:565. doi: 10.1186/s12859-017-1984-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Biswas S., Acharyya S. Neural model of gene regulatory network: A survey on supportive meta-heuristics. Theory Biosci. 2016;135:1–19. doi: 10.1007/s12064-016-0224-z. [DOI] [PubMed] [Google Scholar]
  • 9.Van de Sande B., Flerin C., Davie K., De Waegeneer M., Hulselmans G., Aibar S., Seurinck R., Saelens W., Cannoodt R., Rouchon Q., et al. A scalable SCENIC workflow for single-cell gene regulatory network analysis. Nat. Protoc. 2020;15:2247–2276. doi: 10.1038/s41596-020-0336-2. [DOI] [PubMed] [Google Scholar]
  • 10.Zhao M., He W., Tang J., Zou Q., Guo F. A hybrid deep learning framework for gene regulatory network inference from single-cell transcriptomic data. Briefings Bioinform. 2022;23:bbab568. doi: 10.1093/bib/bbab568. [DOI] [PubMed] [Google Scholar]
  • 11.Chen J., Cheong C., Lan L., Zhou X., Liu J., Lyu A., Cheung W.K., Zhang L. DeepDRIM: A deep neural network to reconstruct cell-type-specific gene regulatory network using single-cell RNA-seq data. Briefings Bioinform. 2021;22:bbab325. doi: 10.1093/bib/bbab325. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Wang J., Ma A., Ma Q., Xu D., Joshi T.J.C., Journal S.B. Inductive inference of gene regulatory network using supervised and semi-supervised graph neural networks. Comput. Struct. Biotechnol. J. 2020;18:3335–3343. doi: 10.1016/j.csbj.2020.10.022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Karaaslanli A., Saha S., Aviyente S., Maiti T. scSGL: Kernelized signed graph learning for single-cell gene regulatory network inference. Bioinformatics. 2022;38:3011–3019. doi: 10.1093/bioinformatics/btac288. [DOI] [PubMed] [Google Scholar]
  • 14.Seninge L., Anastopoulos I., Ding H., Stuart J. VEGA is an interpretable generative model for inferring biological network activity in single-cell transcriptomics. Nat. Commun. 2021;12:5684. doi: 10.1038/s41467-021-26017-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Shu H., Zhou J., Lian Q., Li H., Zhao D., Zeng J., Ma J. Modeling gene regulatory networks using neural network architectures. Nat. Comput. Sci. 2021;1:491–501. doi: 10.1038/s43588-021-00099-8. [DOI] [PubMed] [Google Scholar]
  • 16.Matsumoto H., Kiryu H., Furusawa C., Ko M.S., Ko S.B., Gouda N., Hayashi T., Nikaido I.J.B. SCODE: An efficient regulatory network inference algorithm from single-cell RNA-Seq during differentiation. Bioinformatics. 2017;33:2314–2321. doi: 10.1093/bioinformatics/btx194. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Huynh-Thu V.A., Geurts P. dynGENIE3: Dynamical GENIE3 for the inference of gene networks from time-series expression data. Sci. Rep. 2018;8:3384. doi: 10.1038/s41598-018-21715-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Moerman T., Aibar Santos S., Bravo González-Blas C., Simm J., Moreau Y., Aerts J., Aerts S. GRNBoost2 and Arboreto: Efficient and scalable inference of gene regulatory networks. Bioinformatics. 2019;35:2159–2161. doi: 10.1093/bioinformatics/bty916. [DOI] [PubMed] [Google Scholar]
  • 19.Vilalta R., Drissi Y.J. A perspective view and survey of meta-learning. Artif. Intell. Rev. 2002;18:77–95. doi: 10.1023/A:1019956318069. [DOI] [Google Scholar]
  • 20.Fu K., Zhang T., Zhang Y., Wang Z., Sun X. Few-shot SAR target classification via meta-learning. IEEE Trans. Geosci. Remote Sens. 2021;60:1–14. [Google Scholar]
  • 21.Chowdhury A., Chaudhari D., Chaudhuri S., Jermaine C. Meta-Meta Classification for One-Shot Learning; Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision; Waikoloa, HI, USA. 3–8 January 2022; pp. 177–186. [Google Scholar]
  • 22.Chen Z., Fu Y., Wang Y.X., Ma L., Liu W., Hebert M. Image deformation meta-networks for one-shot learning; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; Long Beach, CA, USA. 15–20 June 2019; pp. 8680–8689. [Google Scholar]
  • 23.Yu Y., Chen J., Gao T., Yu M. DAG-GNN: DAG structure learning with graph neural networks; Proceedings of the International Conference on Machine Learning, PMLR; Long Beach, CA, USA. 9–15 June 2019; pp. 7154–7163. [Google Scholar]
  • 24.Zheng S., Zou Y., Tang Y., Yang A., Liang J.Y., Wu L., Tian W., Xiao W., Xie X., Yang L., et al. Landscape of cancer-associated fibroblasts identifies the secreted biglycan as a protumor and immunosuppressive factor in triple-negative breast cancer. Oncoimmunology. 2022;11:2020984. doi: 10.1080/2162402X.2021.2020984. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Puram S.V., Tirosh I., Parikh A.S., Patel A.P., Yizhak K., Gillespie S., Rodman C., Luo C.L., Mroz E.A., Emerick K.S. Single-cell transcriptomic analysis of primary and metastatic tumor ecosystems in head and neck cancer. Cell. 2017;171:1611–1624.e24. doi: 10.1016/j.cell.2017.10.044. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Kaviany S., Bartkowiak T., Dulek D.E., Khan Y.W., Hayes M.J., Schaefer S.G., Ye X., Dahunsi D.O., Connelly J.A., Irish J.M.J.I. Systems Immunology Analyses of STAT1 Gain-of-Function Immune Phenotypes Reveal Heterogeneous Response to IL-6 and Broad Immunometabolic Roles for STAT1. ImmunoHorizons. 2022;6:447–464. doi: 10.4049/immunohorizons.2200041. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Tsunoda M., Fukasawa M., Nishihara A., Takada L., Asano M. JunB can enhance the transcription of IL-8 in oral squamous cell carcinoma. J. Cell. Physiol. 2021;236:309–317. doi: 10.1002/jcp.29843. [DOI] [PubMed] [Google Scholar]
  • 28.Pratapa A., Jalihal A.P., Law J.N., Bharadwaj A., Murali T.J. Benchmarking algorithms for gene regulatory network inference from single-cell transcriptomic data. Nat. Methods. 2020;17:147–154. doi: 10.1038/s41592-019-0690-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Tomczak K., Czerwińska P., Wiznerowicz M. Review The Cancer Genome Atlas (TCGA): An immeasurable source of knowledge. Contemp. Oncol. Współczesna Onkol. 2015;2015:68–77. doi: 10.5114/wo.2014.47136. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Zhang Y., Wang Z., Zeng Y., Zhou J., Zou Q. High-resolution transcription factor binding sites prediction improved performance and interpretability by the deep learning method. Briefings Bioinform. 2021;22:bbab273. doi: 10.1093/bib/bbab273. [DOI] [PubMed] [Google Scholar]
  • 31.Zhang Y., Wang Z., Zeng Y., Liu Y., Xiong S., Wang M., Zhou J., Zou Q. A novel convolution attention model for predicting transcription factor binding sites by combining sequence and shape. Briefings Bioinform. 2022;23:bbab525. doi: 10.1093/bib/bbab525. [DOI] [PubMed] [Google Scholar]
  • 32.Zhang Y., Qiao S., Zeng Y., Gao D., Han N., Zhou J. CAE-CNN: Predicting transcription factor binding site with convolutional autoencoder and convolutional neural network. Expert Syst. Appl. 2021;183:115404. doi: 10.1016/j.eswa.2021.115404. [DOI] [Google Scholar]
  • 33.Liu R., Gao J., Zhang J., Meng D., Lin Z. Investigating Bi-Level Optimization for Learning and Vision from a Unified Perspective: A Survey and Beyond. IEEE Trans. Pattern Anal. Mach. Intell. 2021 doi: 10.1109/TPAMI.2021.3132674. [DOI] [PubMed] [Google Scholar]
  • 34.Li X., Wang K., Lyu Y., Pan H., Zhang J., Stambolian D., Susztak K., Reilly M.P., Hu G., Li M. Deep learning enables accurate clustering with batch effect removal in single-cell RNA-seq analysis. Nat. Commun. 2020;11:2338. doi: 10.1038/s41467-020-15851-3. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The dataset for model training and testing is available at https://github.com/ZhangLab312/MetaSEM (accessed on 1 January 2023).


Articles from International Journal of Molecular Sciences are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES