Abstract
The rapid advancement of high-throughput sequencing and other assay technologies has resulted in the generation of large and complex multi-omics datasets, offering unprecedented opportunities for advancing precision medicine. However, multi-omics data integration remains challenging due to the high-dimensionality, heterogeneity, and frequency of missing values across data types. Computational methods leveraging statistical and machine learning approaches have been developed to address these issues and uncover complex biological patterns, improving our understanding of disease mechanisms. Here, we comprehensively review state-of-the-art multi-omics integration methods with a focus on deep generative models, particularly variational autoencoders (VAEs) that have been widely used for data imputation, augmentation, and batch effect correction. We explore the technical aspects of VAE loss functions and regularisation techniques, including adversarial training, disentanglement, and contrastive learning. Moreover, we highlight recent advancements in foundation models and multimodal data integration, outlining future directions in precision medicine research.
Keywords: deep generative models, machine learning, multi-omics integration, precision medicine
Introduction
Recent advances in high-throughput technologies have enabled the comprehensive characterisation of cellular models across multiple molecular layers—omics (Fig. 1). Multi-omics studies have become commonplace in precision medicine research, providing a holistic perspective of biological systems [1, 2], uncovering disease mechanisms, identifying molecular subtypes [3–5], and discovering new drug targets [6] and biomarkers for clinical applications [7–14].
Figure 1.
Multi-omics data integration. Left: Illustration of diverse omics layers (rows) for three different groups of samples (columns), highlighting four integration strategies. Vertical integration combines different omics modalities within the same group of samples; horizontal integration aligns datasets from the same omics layer across different sample groups (e.g. batches, cellular models), typically addressing batch effect correction; diagonal integration combines distinct omics modalities from different sample groups to explore inter-modality relationships across groups, and mosaic integration leverages overlapping modalities across samples to infer relationships and impute missing modalities. Dashed boxes indicate missing data. Right: Overview of common tasks in multi-omics analysis. Dimensionality reduction infers low-dimensional embeddings that facilitate downstream tasks like clustering and classification. Batch effect correction ensures that samples cluster based on biological attributes, such as tissue or cell type, rather than technical artifacts. Imputation addresses missing data, both for randomly missing features and for entire missing modalities. Symbol? Inside dashed boxes denote missing values or modalities.
Several consortia, including TCGA/ICGC [15] and ProCan [16], have generated invaluable multi-omics datasets and resources, particularly for cancer studies. Multi-omics data repositories and portals were reviewed in [17–19]. Despite the potential, integrating these datasets remains challenging due to their high-dimensionality, heterogeneity, and sparsity [20, 21]. Multi-omics datasets often comprise thousands of features and are generated through diverse laboratory techniques, leading to inconsistent data distributions [20, 22]. Moreover, due to experimental limitations, data quality issues, or incomplete sampling, these datasets are often unbalanced and incomplete [23].
To address these issues, statistical and machine learning models focusing on dimensionality reduction, batch effect correction, and data imputation have been developed (Fig. 1). Dimensionality reduction techniques infer low-dimensional spaces that capture variability across omics, facilitating downstream tasks like clustering and classification [24, 25]. Batch effect correction attenuates technical biases while preserving critical biological signals [26, 27] and imputation techniques enhance data quality through data denoising and augmentation [28–30].
Many authors have reviewed multi-omics integration methods, offering diverse perspectives on approaches, challenges, and applications. Several of these reviews have focused on state-of-the-art statistical and machine learning approaches, categorising multi-omics integration methods based on different criteria. For instance, some classify these methods by the intrinsic nature of multi-omics experiments [31, 32], while others focus on fusion strategies [21, 33]. Other authors emphasized the applications of multi-omics integration methods and the supported omics data types [17, 19], with a particular focus on oncology [11, 18, 34, 35]. In Vahabi and Michailidis [36], unsupervised learning methods were reviewed based on their underlying approach. Recently, deep learning-based approaches have gained prominence, and several authors reviewed both traditional architectures, emerging trends, and their applications [37–46].
This review provides a comprehensive technical overview of the methods developed for multi-omics data integration, categorising them into correlation-based, matrix factorisation, probabilistic, network, kernel-based, or deep learning approaches (Fig. 2, Table 1). Instead of reviewing a specific model category, multi-omics data types, or biological applications, we emphasize the architectural and computational innovations driving this broad range of methods and categories. Recent advances in the field have shifted the focus from more classical statistical to deep learning approaches, particularly generative methods. Therefore, we place particular emphasis on variational autoencoders (VAEs), which have gained prominence since 2020 for tasks such as imputation, denoising, and creating joint embeddings of multi-omics data [47–51]. Beyond describing VAE applications, we explore training strategies and regularisation techniques proposed for adversarial training, cycle-consistency, contrastive, and disentangled representation learning.
Figure 2.
Schematic representation of the multi-omics integration approaches reviewed. (a) Canonical correlation analysis; (b) matrix factorisation of omics matrices into a shared and omics-specific matrices; (c) Bayesian approaches for probabilistic modelling; (d) multiple kernel learning; (e) similarity network fusion; (f) artificial neuron, the fundamental unit of neural networks: For each sample vector each input feature
is associated with a weight
. To compute a weighted sum. The resulting value is passed through an activation function to produce the neuron’s output.
Table 1.
Overview of multi-omics integration methods categories, outlining their general strengths, limitations, and typical applications.
| Model approach | Strengths | Limitations | Typical applications in multi-omics analysis |
|---|---|---|---|
| Correlation / Covariance-based | Captures relationships across omics based on correlation or covariance, interpretable, flexible sparse and regularised extensions | Limited to linear associations, typically requires matched samples across omics | Disease subtyping, detection of co-regulated modules |
| Matrix Factorisation | Efficient dimensionality reduction, identifies shared and omic-specific factors, interpretable, scalable | Assumes linearity, does not explicitly model uncertainty or noise | Disease subtyping, identification of shared molecular patterns, biomarker discovery |
| Probabilistic-based | Efficient dimensionality reduction, captures uncertainty in latent factors, probabilistic inference | Computationally intensive, may require careful tuning and strong model assumptions | Disease subtyping, latent factors discovery, biomarker discovery |
| Multiple Kernel Learning | Can capture nonlinear relationships, well-suited for heterogeneous data types | Sensitive to kernel choice and parameters, limited interpretability | Disease subtyping, patient similarity analysis |
| Network-based | Represents samples or omics relationships as networks, typically robust to missing data | Sensitive to similarity metrics choice, may require extensive tuning | Disease subtyping, patient similarity analysis, identification of regulatory mechanisms |
| Deep generative learning | Learns complex nonlinear patterns, flexible architecture designs, can support missing data and denoising | High computational demands, limited interpretability, requires large data to train | High-dimensional omics integration, data augmentation and imputation, disease subtyping, biomarker discovery |
Here, we standardize terminology and provide clarity in a field with a wide range of different methods and definitions. Finally, we highlight promising future directions, including foundation models and multimodal data integration, which have the potential to advance the field and further enhance its impact on precision medicine.
Classical statistical and machine-learning approaches
In this section, we introduce multi-omics integration methods, ranging from correlation and covariance-based to matrix decomposition methods, and probabilistic or Bayesian approaches. Lastly, network and kernel-based methods are highlighted. Generally, this considers
different omics matrices
,
each with
samples and
features.
Correlation/covariance-based methods
Canonical Correlation Analysis (CCA) [52] is a classical statistical method designed to explore the relationships between two sets of variables
and
(
), with the same set of samples
CCA aims to find column vectors
and
that maximise the correlation between the linear combinations
and
:
![]() |
(1) |
where
and
are the first canonical vectors, and
and
are the corresponding canonical variables (Fig. 2a) used for further multi-omics analysis.
CCA has proven particularly useful as a joint dimensionality reduction and information extraction method in genomic studies, where multiple types of data, such as DNA copy number or mutation, are often collected from the same set of samples [53–57]. However, the high-dimensional nature of multi-omics data presents a major challenge, often resulting in ill-defined problems that limit the direct applicability of classical CCA. Therefore, several CCA extensions were proposed to induce sparsity in the solution of the optimisation problem [54–56, 58, 59]. To extend the application of CCA-based methods to more than two datasets, sparse and regularized Generalised CCA (sGCCA/rGCCA) were proposed [60, 61], being currently one of the most widely used generalisations of CCA to multi-omics data.
DIABLO [62] extends sGCCA to a supervised framework. It simultaneously maximizes common or correlated information between multiple omics datasets and minimizes the prediction error of a response variable. This approach is particularly effective for selecting co-varying modules that explain the outcome, usually phenotypic traits.
Recently, deep learning-based extensions of traditional CCA have been proposed to handle nonlinearity and scalability in multi-omics data integration, including SDGCCA (Supervised Deep GCCA) [63], and VIPCCA [64] or VIMCCA [65] for unpaired and paired single-cell data, respectively.
Partial Least Squares (PLS) is an alternative approach for data integration that aims to maximize the covariance between components. Various methods were also proposed to find sparse solutions [66] and to extend the application to more than two datasets [67]. Several implementations of PLS, which optimize different objective functions with different constraints, have been described and reviewed in [68].
Matrix factorisation methods
Matrix decomposition is a powerful method for joint dimensionality reduction, condensing datasets into fewer factors to reveal important patterns that can be used, e.g. to identify disease-associated biomarkers or cancer subtypes (Table 1).
JIVE [69] is considered an extension of Principal Component Analysis (PCA) that decomposes each omics matrix into joint and individual low-rank approximation and a residual noise by minimising the overall sum of squared residuals. JIVE quantifies the variation across and within datasets, reduces the dimensionality of the data, and avoids overfitting.
Non-Negative Matrix Factorisation (NMF) is a popular technique for decomposing datasets into two non-negative matrices. Several extensions of NMF have been developed to address the specific challenges of multi-omics datasets.
jNMF [70] decomposes multiple omics datasets into a shared basis matrix
and specific omics coefficient matrices
(Fig. 2b):
![]() |
(2) |
The objective function is formulated as
, where
denotes the Frobenius norm, and the constraint
ensures that all entries in matrices
and
are non-negative.
intNMF [71] is an extension of NMF for clustering analysis of multi-omics data. Once the matrix
has been computed, each sample is associated with one of the
clusters, determined by the highest entry in the matrix.
For single-cell data, LIGER [72] applies iNMF [73] to decompose each omics dataset into dataset-specific weights (
), shared weights (
), and sample specific factors (
). The objective function is defined as:
![]() |
(3) |
An additional regularisation term is added to handle omics-specific noise and heterogeneity. After performing iNMF, each cell is assigned to the factor with the highest loading, and a shared-factor neighborhood graph is constructed by connecting cells with similar factor loading profiles, improving the robustness of joint clustering. UINMF [74] extends iNMF by adding an unshared weights matrix term to the objective function. This method incorporates features that belong to only one or a subset of the omics datasets, performing mosaic integration.
Probabilistic-based methods
Matrix factorisation is a robust approach for dimensionality reduction but has several limitations, particularly in handling missing data. Probabilistic matrix factorisation offers substantial advantages by incorporating uncertainty estimates and allowing for flexible regularisation (Table 1).
iCluster [75] is a joint latent variable model designed to identify latent cancer subtypes based on multi-omics data. This method decomposes each omics into a shared latent factor matrix
and omics-specific weight matrices
:
![]() |
(4) |
assuming both the errors
and the latent factor matrix
follow a normal distribution. Latent variables are estimated using the expectation-maximisation method [76], and clusters are obtained by applying K-means to matrix
. iClusterPlus [77] extends iCluster by modelling different statistical distributions to handle discrete omics datasets but it was criticized for its computational intensity. iClusterBayes [78] further extends iClusterPlus by introducing a fully Bayesian inference approach, improving both stability and computational efficiency.
LRAcluster [79] is another example of a clustering probabilistic method for continuous and categorical data integration, but it is based on a low-rank probabilistic approach to improve latent variables estimation efficiency. moCluster [80] is also a joint latent model that uses modified consensus PCA [81] for latent variable estimation, providing a more stable and efficient alternative to the expectation-maximisation algorithm used in iCluster.
MOFA [82] is a probabilistic Bayesian framework for multi-omics integration, designed to handle diverse data distributions and missing values automatically (Fig. 2b-c). It decomposes omics datasets as in Equation 4, placing prior distributions on all unobserved variables. To enhance interpretability and disentangle variation across omics, MOFA employs a two-step regularisation process to enforce sparsity in the weight matrix and capture both omics-specific and shared factors. MOFA is optimized by maximising the Evidence Lower Bound (ELBO) [83], enhancing its generative capabilities. MOFA+ [84] further extends this framework to improve scalability for both bulk and single-cell datasets.
Kernel-based methods
The previously described models rely predominantly on linear combinations to integrate multi-omics data. In contrast, kernel- and network-based approaches enable the modelling of nonlinear relationships across omics layers in a structured way.
Kernel learning approaches [85, 86] use kernel functions to map omics data into higher-dimensional feature spaces. This mapping is defined by a kernel matrix that represents pairwise similarities, computed as their inner product in the feature space
, where
maps the original data to the feature space. The kernel function
is the only required definition for the kernel method. In multi-omics integration, a kernel matrix is computed for each omics layer and multiple kernel learning (MKL) combines them into a final kernel matrix by minimising an objective function. The final kernel matrix is used for downstream analysis (Fig. 2d). In MKL, the integration problem shifts from heterogeneous feature spaces to a unified sample space.
Several MKL-based methods have been proposed for multi-omics integration. For instance, rMKL-LPP [87] uses a linear combination of kernels constructed using an objective function based on the Locality Preserving criterion, while web-rMKL [88] provides a web-based implementation. pairwiseMKL [89] is a time and memory-efficient version of MKL with applications to drug response prediction.
Network-based methods
Network-based methods leverage graphs and other network structures to represent omics data and their relationship, capturing topological structures and interactions to identify key biomarkers and reveal biological pathways. Similarity Network Fusion (SNF) [90] is a popular network-based method for multi-omics integration. SNF constructs patient similarity networks for each omics layer, where nodes represent samples and edges their similarity. These networks are iteratively fused into a unified similarity network using a nonlinear combination method based on message passing theory (Fig. 2e). The unified similarity network captures shared patterns across omics layers, enhancing subtype identification for several diseases [91–94]. As MKL, SNF is computationally efficient as it depends primarily on the number of samples rather than features. However, it does not distinguish between data types and relies on Euclidean distance to calculate sample similarity, which may not fully capture the complex relationships in omics data. Several extensions and deep learning-based adaptations have been proposed to overcome these limitations [95–99].
NEMO [100] is a popular similarity network-based method that handles unmatched samples without requiring imputation. It constructs a patient similarity matrix for each omics using a radial basis function kernel, adjusts for local neighborhoods, and computes an average relative similarity matrix shared across all omics. Spectral clustering is then applied to identify disease subtypes, demonstrating superior clustering performance compared to other state-of-the-art methods.
Overall, similarity kernel and network-based methods focusing on sample similarities and relationships offer a flexible strategy to integrate heterogeneous datasets (Table 1).
Deep learning approaches
Deep learning approaches have emerged as powerful and flexible tools for multi-omics integration, capturing nonlinear and complex patterns from data (Table 1). The following sections explore various neural network architectures for multi-omics integration, from non-generative to generative approaches, placing particular focus on VAEs.
Non-generative models
Non-generative approaches focus on learning direct mappings or relationships between input features and outputs, often prioritising tasks like classification, regression, or dimensionality reduction.
Feed forward neural networks
The feed-forward neural network (FFNN) is the most common neural network architecture, consisting of fully interconnected layers of neurons. The individual neurons compute weighted sums of their inputs, apply an activation function and propagate it forward to the next layer (Fig. 2f). The activation function typically introduces nonlinearity, increasing the expressive power of the network. These models are trained to minimize a loss function using optimisation techniques such as backpropagation. Despite being computationally intensive and having a high number of parameters, neural networks have demonstrated significant potential in biological data analysis [101, 102].
MOLI [103] is a supervised FFNN designed for drug response prediction. This method employs separate subnetworks for each omics layer, extracts features, and concatenates them into a unified representation for the final classification network. MOLI incorporates a binary cross-entropy and a triplet loss function for training. FFNNs have also been applied to other tasks, such as synergistic drug combination prediction [104, 105], survival analysis [106] or trajectory inference [107].
Graph neural networks
Graph neural networks (GNNs) are a powerful framework for processing data structured as graphs, making them particularly valuable in biological research where entities are intrinsically linked, such as in protein–protein interactions (PPIs) or gene regulatory networks. Graph Convolutional Networks (GCNs) are the most dominant GNNs introducing convolution operations to the graph structure. MOGONET [108] is designed for supervised multi-omics integration and classification by constructing a sample similarity network for each omics and leveraging GCNs to predict labels based on individual modalities. scMoGNN [109] is an example of a framework that leverages GNNs for single-cell multi-omics data integration to tackle modality prediction, matching, and joint embedding tasks.
While methods like MOGONET focus on sample similarity networks, they do not incorporate biological interaction data, such as PPIs, which could provide additional meaningful context. For instance, Zhuang et al. [110] proposed a GCN method for disease classification integrating transcriptomics and proteomics data with PPI networks.
Autoencoders
Autoencoders are an unsupervised deep learning model widely used for dimensionality reduction and feature extraction tasks—embeddings. This model architecture leverages neural networks to compress the input data into a lower dimensional latent space via an encoder and attempts to reconstruct it back to the original space through a decoder. Several autoencoder-based models have been developed for multi-omics integration for cross-modality translation tasks [111, 112] or for joint dimensionality reduction, where the latent space is used for downstream tasks like disease prognosis and subtyping [113–118], clustering [119, 120], synergistic drug combination prediction [121], or batch correction [122, 123].
Among several extensions of autoencoders, VAEs [124, 125] are the most prominent for multi-omics data analysis due to their probabilistic framework and generative capability.
Variational autoencoders
In contrast with non-generative approaches, deep generative models (DGMs) learn the underlying data distribution, enabling the generation of realistic synthetic data. DGMs have significantly impacted molecular biology, including the multi-omics data integration field [126]. The majority of DGMs published and reviewed here are based on VAEs [124, 125]. However, other frameworks including generative adversarial networks (GANs) [127] and, more recently, generative pre-trained transformer (GPT) approaches [128] have been proposed.
In a common Bayesian approach, each sample vector
, with
features, is assumed to be generated by a latent vector
, with
. The latent vector is drawn from a prior distribution
, and the sample vector is generated from the conditional distribution
, where θ are the parameters of the decoder network. Since the estimation of the marginal data likelihood
is computationally intractable, approximation inference methods are employed to efficiently estimate model parameters.
VAEs provide a principled way for performing variational inference by approximating the true posterior distribution
using a variational posterior
, where
are the parameters of the encoder. This encoder network processes the input data and outputs two layers representing the mean μ and standard deviation σ of the variational posterior. The reparameterisation trick is used to allow backpropagation by sampling
as
,
. Finally, the latent vector is the input of the decoder network that will reconstruct the input data (Fig. 3a).
Figure 3.
VAE architectures. (a) VAE with a supervised module for task-specific supervision. Each rectangle represents a fully connected block. Data from two omics are concatenated in the second hidden layer. The parameters
and
represent the mean and standard deviation of the learned posterior distribution, and
. The reparameterisation trick is depicted in the dashed box. (b) Multimodal VAE architecture, highlighting three strategies to build the joint latent space: concatenation, MoE, and PoE. (c) Adversarial training strategies in VAEs to align the latent spaces of different omics modalities. (d) Cross-modal cycle consistency. (e) Contrastive learning to self-supervise VAEs by gathering positive pairs and separating negative pairs.
The parameters
and
of the encoder and decoder networks, respectively, are jointly optimised by maximising the ELBO:
![]() |
(5) |
where
is the Kullback–Leibler (KL) divergence, weighted by the hyperparameter
. The overall VAE loss function is the negative ELBO. The first term aims to minimise the decoder reconstruction error, e.g. using the mean squared error. The KL divergence term regularizes the model by minimising the divergence between the variational posterior and prior distribution. The variational posterior typically follows a Gaussian distribution. However, in single-cell data applications other probability distributions, such as the negative binomial, can better handle the sparsity and count nature of the data [129, 130].
VAEs are widely used for multi-omics analysis, particularly for single-cell experiments, due to their flexibility in handling high-dimensional and incomplete datasets while balancing dimensionality reduction and generative capabilities. The following sections detail various methodological improvements that have been proposed to enhance VAE integration performance during recent years (Table 2).
Table 2.
Overview of deep generative models for multi-omics data integration. Omics modalities listed correspond to those demonstrated in the original publications, although some models may be extended to other omics layers. Similarly, the integration type refers to the categorisation provided by the authors or to the integration setting inferred from the original work; however, several methods may be adapted for alternative integration scenarios. CyTOF: cytometry by time of flight; CNV: copy number variation; scATAC-seq: single-cell sequencing assay for transposase-accessible chromatin; scRNA-seq: single-cell RNA-sequencing; snRNA-seq: single-nucleus RNA-sequencing; snmC-seq: single-nucleus methylation sequencing.
| Name | Method | Omics modalities demonstrated | Integration type demonstrated | Key tasks and applications |
|---|---|---|---|---|
| General multi-omics methods | ||||
| OmiVAE / XOmiVAE [136,137] | VAE with a supervised module / extension incorporating Deep SHAP | gene expression, DNA methylation | joint embedding, classification, clustering | |
| MMD-VAE [48] | VAE with a supervised module and a MMD instead of KL regularizer | DNA methylation, CNV, mRNA, or RNAseq | joint embedding, classification, survival analysis, clustering | |
| OmiEmbed [139] | VAE with supervised modules | miRNA, gene expression, DNA methylation | joint embedding, classification, regression, survival prediction, clustering | |
| omicsGAN [148] | GAN to integrate two omics modalities and their interaction networks | mRNA, miRNA expression | classification, survival prediction | |
| MOSA [23] | Conditional MVAE with concatenation of latent spaces and a contrastive loss; SHAP values | CNV, proteomics, gene expression, methylomics, metabolomics, drug response, CRISPR-Cas9 | joint embedding, data augmentation, clustering, biomarker identification related to drug and gene dependencies | |
| Single-cell multi-omics methods | ||||
| scMVAE [142] | MVAE with three strategies of joint-learning (direct concatenation, neural network, PoE) | scRNA-seq, scATAC-seq | Vertical | joint embedding, data denoising and imputation, regulatory mechanisms analysis, clustering |
| scMM [129] | MVAE with MoE and a pseudo cell generation strategy for model interpretability | scRNA-seq, scATAC-seq, or surface protein | joint embedding, cross-modal generation, regulatory inference, clustering | |
| DCCA [147] | Separate VAEs mutually supervised by cross-omics cycle attention | scRNA-seq, scATAC-seq | separate embeddings, transfer learning, data denoising, regulatory inference, missing cells and omics generation, clustering | |
| scMVP [154] | MVAE with GMM prior, attention-based channels, and intra-modal consistency modules | scRNA-seq, scATAC-seq | joint embedding, clustering, trajectory inference, cis-regulatory element prediction, data denoising and imputation | |
| scDisInFact [135] | VAE with an additional MMD term for disentanglement learning | scRNA-seq | Horizontal | shared-bio embedding, clustering, batch effect correction, condition-associated key gene detection, perturbation prediction |
| SCIM [47] | VAE with adversarial training to distinguish between omics based on latent space | scRNA-seq, CyTOF | joint embedding, cell matching across different omics technologies, batch correction, clustering | |
| Cobolt [146] | MVAE with PoE for the integration of data from multi and single-modality platforms | scRNA-seq, scATAC-seq | joint embedding, clustering | |
| Multigrate [49] | MVAE with a shared decoder, PoE strategy, and an additional MMD regularizer | scRNA-seq, scATAC-seq, or surface protein | joint embedding, multimodal reference atlas building, transfer learning for multi-omic query datasets mapping, imputation of missing omics, clustering | |
| GLUE [149] | Omics-specific VAEs and a graph VAE; adversarial training to distinguish between omics based on cell embeddings | scRNA-seq, scATAC-seq, snmC-seq | joint embedding, regulatory inference, multi-omics atlas building, clustering | |
| Portal [152] | Encoder and GAN: modality-specific encoders, cross-modal generators, and discriminators; additional regularizers for cross-modal embeddings and samples consistency, and intra-modality reconstructions | scRNA-seq, scATAC-seq, or snRNA-seq | joint embedding, batch correction, label transfer, clustering | |
| sciCAN [155] | Encoder and GAN: one discriminator to distinguish between omics latent spaces, one discriminator to distinguish between real and cross-modal generated data, cross-modal embeddings cycle-consistency | scRNA-seq, scATAC-seq | Diagonal or Mosaic | joint embedding, co-trajectory inference, label transfer, clustering |
| scVAEIT [164] | VAE with conditional variational inference using missing masks | scRNA-seq, scATAC-seq, surface proteins | joint embedding, data denoising, imputation of missing features and omics modalities, transfer learning to new datasets, clustering | |
| JAMIE [50] | VAEs with cross-modal correspondence and correlation-based latent aggregation; SHAP values | scRNA-seq, scATAC-seq, electrophysiology data | separate aggregate embeddings, cross-modal imputation, phenotype prediction, clustering | |
| MultiVI [134] | MVAE using distributional average and penalisation to mix latent embeddings; symmetric Jeffrey’s divergence term; adversarial training to distinguish between batches or modalities based on the shared latent space | scRNA-seq, scATAC-seq, surface protein | joint embedding, differential expression analysis, imputation of missing features and omics modalities, batch correction, clustering | |
| MIDAS [51] | MVAE with PoE that employs self-supervised modality alignment, information-theoretic latent disentanglement, and masking techniques to handle missing modalities | scRNA-seq, scATAC-seq, surface protein | joint embedding, batch correction, transfer learning, imputation of missing features and omics modalities, clustering, trajectory inference | |
| scCross [150] | Omics-specific VAEs with an FFNN aligner; one discriminator to distinguish between omics based on the shared latent space and omics specific discriminators to distinguish between original and reconstructed data | scRNA-seq, scATAC-seq, snmC-seq | joint embedding, cross-modal generation, data simulation, in silico cellular perturbations, clustering | |
Maximum mean discrepancy regularizer
VAEs can face issues with the usual ELBO-based loss function, leading to failures in learning a variational posterior distribution that approximates the true posterior and meaningful or informative latent features [48, 131]. Inference failures can occur due to an imbalance in ELBO optimisation or modelling bias. When the input data is high-dimensional compared to the latent space, the model can prioritize data reconstruction over learning a distribution that approximates the true posterior, potentially leading to poor generalisation and overfitting. Additionally, VAEs can reconstruct input data without relying on the latent variables, making them uninformative and failing to capture meaningful information about the input data.
To address these issues, several models [48, 132] have replaced the traditional KL divergence term with the Maximum Mean Discrepancy (MMD) [133] in their loss function. The MMD-based regularizer estimates the divergence by how different the moments of two distributions
and
are and can be defined as:
![]() |
(6) |
where
is any positive-definite kernel, with the Gaussian kernel a popular choice.
Several models reviewed in the following sections incorporate MMD terms into their loss functions. For instance, MMD-VAE [48] replaces the KL divergence with an MMD regularizer. Multigrate [49] and MultiVI [134] add MMD terms into the VAE loss to ensure alignment and consistency across omics modalities. Additionally, scDisInFact [135] introduces a MMD term to ensure disentanglement of latent factors.
Supervised learning tasks with VAEs
In a standard VAE, the bottleneck layer extracts essential features for accurate input data reconstruction. However, these features are often too general and may not be relevant to particular downstream analyses. To address this, several VAE-based models incorporate supervised modules. These models typically connect the mean vector
or the latent space
to a neural network to perform a specific task (Fig. 3a). This introduces a task-specific regularizer, by adding the loss of the downstream task into the overall VAE loss function. The total loss ensures the latent factors extracted by the VAE are informative for both accurate data reconstruction and supervised tasks.
One example is OmiVAE [136], which integrates gene expression and DNA methylation data to classify pan-cancer tumour samples. XOmiVAE [137] extends this by incorporating explainability through deep SHAP [138], providing insights into individual features and latent dimensions contributions for cancer classification to identify potential biomarkers.
OmiEmbed [139] extends previous models by incorporating additional downstream tasks such as demographic and clinical features reconstruction and survival prediction using a multi-task training strategy. The latent space is shared across tasks and the downstream module loss is the weighted sum of all downstream losses, improving the overall model performance through joint learning.
MMD-VAE [48] integrates tri-omics data for ovarian cancer analysis, focusing on molecular subtypes clustering, classification, and survival prediction. By replacing the KL divergence with an MMD regularizer, MMD-VAE improves performance on most omics datasets.
Overall, VAEs with supervised modules improve omics analysis by integrating generative modelling with supervised learning. This approach enhances the extraction of biologically relevant features and optimizes classification or regression performance, addressing limitations of traditional unsupervised methods.
Inferring joint latent representations with multimodal VAEs
Multimodal VAEs (MVAEs) are a common approach for multi-omics integration where each omics modality is assigned its own encoder-decoder and a shared latent space is inferred. There are several strategies to combine the latent variables from individual modalities into a unified latent representation, including concatenation or probabilistic methods such as the mixture of experts (MoE), product of experts (PoE) [140], and mixture-of-product-of-experts [141] (Fig. 3b).
In MoE approaches, the joint variational posteriors for
omics modalities is defined as
, usually with
, where
represents a sample vector for omics
. The resulting ELBO is the weighted average of each modality’s ELBO. The PoE is an alternative approach that infers the joint posterior as the product of the individual variational posteriors
.
scMVAE [142] integrates multi-omics data with a multimodal encoder that infers the joint latent space with three different strategies: direct concatenation, neural network-based feature concatenation, and PoE. scMVAE includes single-modal encoders/decoders for data normalisation, denoising, and imputation and a Gaussian Mixture Model (GMM) as the prior to learn more disentangled and interpretable latent representations.
scMM [129] extends scMVAE using modality-specific encoders and decoders and a MoE strategy to infer the joint latent space. scMM enhances interpretability by generating pseudo cells with controlled variations in one latent dimension while keeping others fixed (latent traversals) and assessing correlations between traversed dimensions and features in each omics, identifying features strongly associated with each latent dimension. Additionally, scMM enables bidirectional cross-modal generation, allowing missing data in one modality to be inferred from another.
Multigrate [49] also applies a PoE approach for mosaic integration but is trained conditionally on a set of study labels. One key innovation is an additional MMD loss to minimize the distance between joint latent representations for pairs of datasets. The model architecture features modality-specific encoders and decoders, along with a shared decoder that captures both common biological and omics-specific patterns. Multigrate allows multimodal reference building, query data mapping into a reference atlas using transfer learning [143], and missing modalities imputation.
MultiVI [134] from the scvi-tools library [144] is built on earlier VAE-based models [130, 145] and is conceptually similar to Cobolt [146], a MVAE that employs a PoE approach. However, MultiVI is trained conditionally on a set of covariates, uses tailored noise models for each omics and infers the latent representation using a distributional average and penalisation strategy. For two omics modalities sample vectors
and
, the shared latent space is defined as:
![]() |
(7) |
where
and
are the weights for each modality, sample-specific or the same for all samples, also optimised during training. If only one modality is available for a specific sample, the latent space is inferred directly from that modality. MultiVI also minimises the distance between two latent representations using symmetric Jeffrey’s divergence:
![]() |
(8) |
An alternative MMD penalty was also explored by the authors. Additionally, MultiVI incorporates an adversarial penalty in the loss function, as detailed in the following sections, to minimize batch effects.
The projection of multi-omics data into a common latent space using MVAEs has become the dominant strategy for integration. The unified latent representation captures shared biological patterns across omics and facilitates downstream analyses such as clustering and visualisation. However, joint embeddings inevitably attenuate omic-specific patterns, potentially hiding relevant molecular insights. Therefore, several models have incorporated cross-learning approaches to retain omics-specific patterns while leveraging shared information.
Cross-learning approaches
The authors of scMVAE have extended this model to address limitations such as the need for all modalities to be present during training and omics-specific patterns attenuation. Therefore, DCCA [147] was proposed to combine multiple modalities into separate but coordinated latent space. DCCA processes each omics modality with a separate VAE that learns from each other with mutual supervision through cross-omics attention transfer. In DCCA, a well-trained teacher network on one modality guides the training of a student network of another modality. The model employs an additional term to each VAE loss to minimize the differences between the latent variables of each VAE, ensuring embeddings alignment:
![]() |
(9) |
where
and
are
-dimensional latent vectors for two omics modalities, and
the weight of the added term.
JAMIE [50] is a VAE framework for di-omics integration and imputation. It uses a cross-modal correspondence matrix
to handle partially aligned samples. JAMIE processes each modality with separate encoders, and their latent spaces are combined using correlation-based latent aggregation based on matrix
. Combination and alignment terms are added to the common VAE loss. The combination loss enforces similarity between separate and aggregate latent spaces, while the alignment term shapes the aggregated latent spaces to ensure similarity between cross-modal cell representations. JAMIE supports imputation and the prioritisation of input features for cross-modal imputation using SHAP. Its ability to adaptively learn correspondences and generate aggregate latent spaces makes it a versatile tool for multi-omics analysis.
Adversarial training strategies
With the advance of DGMs, GANs have inspired multi-omics integration methods by leveraging adversarial training to generate realistic synthetic data while distinguishing it from real data. GAN [127] architecture consists of two competing neural networks jointly optimized: a generator that learns to transform input noise into real data, and a discriminator that distinguishes between real and synthetic generated data.
omicsGAN [148] leverages Wasserstein GANs to integrate two omics modalities and their interaction networks to learn inter-modality relationships. The generator synthesizes data for one modality using the other modality and the interaction network adjacency matrix and the discriminator is adversarially trained to distinguish real from synthetic data.
Despite their flexibility and ability to learn complex data distributions, GANs face challenges such as the complexity of training dual networks, scalability issues with larger numbers of modalities, and the need for large sample sizes. To address these, adversarial training is often integrated into VAE frameworks as a regularisation strategy.
In this adversarial approach, a discriminator commonly distinguishes between omics modalities or batches based on latent space samples (Fig. 3c), while the encoder-decoders are trained to fool the discriminator by producing indistinguishable samples. The adversarial penalty in the standard VAE loss function encourages latent space alignment across modalities, attenuating batch effects.
SCIM [47] applies adversarial training for diagonal integration, aligning single-cell data from different omics technologies. Separate encoders generate a technology-invariant latent space using a discriminator to ensure that latent representations from different omics sources are comparable. Additionally, SCIM pairs cells across different technologies via their latent representations and a fast bipartite matching algorithm.
GLUE [149] also focuses on multi-omics diagonal integration using graph-guided embeddings and adversarial alignment. It employs omics-specific VAEs to learn sample latent spaces and a graph VAE incorporating prior biological knowledge to learn feature latent spaces for each omic. These latent spaces are combined via inner product to reconstruct omics data. A discriminator is used in the sample space to ensure cell embeddings alignment across omics and batch effects attenuation.
scCross [150] is a recent model designed for cross-modality translation and to perform in silico perturbations. scCross trains modality-specific VAEs to extract low-dimensional cell embeddings which are converted to a shared latent space using a FFNN aligner. A global discriminator operates on this shared latent space to identify the omics origin of the cells, ensuring proper alignment. For cross-modality translation, an omics-specific encoder maps input data into the shared latent space, which is then decoded into another omics modality. Modality-specific discriminators distinguish original from cross-modal reconstructed data. Several models have been developed for cross-modality translation [129, 151, 152], with scCross being one of the most recent.
Cycle-consistency training
Building on adversarial training in the latent space, cycle-consistency terms can also be incorporated into the VAE loss function to enforce intra-modal and cross-modal consistency, inspired by cycleGAN [153]. For intra-modal consistency, decoded representations are re-encoded with the omics-specific encoder and aligned with the original latent embeddings. For cross-modal consistency, embeddings from one modality are decoded and re-encoded using another modality’s encoder-decoder pair (Fig. 3d), ensuring that cross-modal translations remain consistent with the original latent space.
scMVP [154] is a VAE for di-omics vertical integration. It employs modality-specific encoders and decoders and a GMM prior to derive the shared latent space. This model uses multi-head self-attention transformer modules for scATAC-seq data to highlight the most informative features. Simpler attention blocks are used for scRNA-seq data to dynamically weight features, emphasising their importance during training. scMVP integrates single-modal encoders to ensure clustering consistency by minimising the KL-divergence between the joint embeddings and the modality-specific re-embeddings from the decoder output:
![]() |
(10) |
where
are two different omics sample vectors,
and
the corresponding reconstructed vectors, and
the common embedding. This intra-modal consistency loss ensures robust integration, the alignment of omics reconstructions and is used to impute missing data.
sciCAN [155] integrates two omics modalities by combining a shared encoder to derive latent representations with a cycle-GAN for modality alignment. The encoder uses noise contrastive estimation to preserve data structure and two discriminators are employed: one to distinguish between latent spaces of the two modalities and another to distinguish between real and generated data from a cross-modal generator. A cycle-consistency loss further ensures alignment between the embeddings produced by the generated data and the original ones.
Portal [152] performs di-omics integration using dual encoders, generators and discriminators. Encoders learn latent embeddings for each omics modality and the cross-modal generators generate synthetic omics data. Discriminators distinguish between original and generated data. To enhance consistency, Portal employs an autoencoder loss for intra-modality reconstruction, an alignment loss for consistency between the cross-modal embedding and the original one, and a cosine similarity loss to preserve correspondence between original and cross-modal reconstructed samples.
Adversarial and cycle consistency training strategies have proven effective in aligning and integrating multi-omics data by attenuating batch effects, encouraging consistency across modalities, improving reconstruction accuracy, and enabling the translation between different omics modalities.
Contrastive learning
Contrastive learning offers an alternative and complementary approach to extract meaningful representations in unsupervised learning methods. The main idea is to ensure that similar samples (positive pairs) are represented closer together in the latent space, while dissimilar samples (negative pairs) are pushed further apart [156, 157]. By enforcing these relationships during training, the model effectively learns to align similar sample types, improving accuracy and robustness in downstream tasks (Fig. 3e). To distinguish between positive and negative pairs, several metrics and functions can be used [157–159].
MOSA [23] is a conditional MVAE that adopts an embedding concatenation strategy to build the joint latent space. This model added a self-supervised contrastive loss to the standard VAE loss, defined as:
![]() |
(11) |
where
and
represent the cosine similarity between positive pairs and negative pairs defined by whether two samples have the same tissue type, and
and
are positive and negative margins tuned during model training. Given that MOSA is one of the first models trying to systematically integrate over seven different omics, it adopts a whole-omic dropout layer strategy that masks entire omic during training, improving model generalization and reducing complexity. The model further enhances biological relevance by concatenating key features such as cancer driver mutations as conditionals to the model, and by incorporating SHAP for model interpretation.
Contrastive learning refines the representation of shared structures across omics modalities. However, to further unravel independent and interpretable factors driving biological variation, disentanglement representation learning (DRL) emerges as a complementary approach.
Disentanglement learning
DRL [160] is a machine learning strategy designed to extract latent representations that separate independent and informative factors of variation. VAEs are especially well-suited for DRL due to their flexibility, allowing the incorporation of regularizers to promote disentanglement [160, 161]. In multi-omics integration, DRL is particularly effective at disentangling complex molecular processes, improving interpretability and model generalisability.
scDisInFact [135] is a VAE framework designed to learn latent factors that disentangle conditions from batch effects in scRNA-seq data, enabling it to simultaneously remove batch effects and identify condition-associated key genes. The goal is to disentangle shared-bio factors (
) and unshared-bio factors (
) using two additional MMD loss terms. One term ensures that
is independent of condition and batch labels:
![]() |
(12) |
where
is the set of batches under condition label
, and
the total number of conditions.
is the latent representation of a reference batch and condition and
the remaining ones. Similarly, the MMD term applied to
enforces independence from batch effects while preserving condition-specific variability:
![]() |
(13) |
where
is the latent representation of a reference batch under condition label
. The disentanglement is further enhanced with group lasso for feature selection and cross-entropy loss for condition prediction using
.
MIDAS [51] is a MVAE employing the PoE approach for mosaic integration of single-cell data. It incorporates self-supervised learning and information-theoretic latent disentanglement [162] to perform dimensionality reduction, imputation, and batch correction. A disentanglement and an alignment term are added to the VAE loss. The disentanglement term is based on the Information Bottleneck loss and aims to disentangle the joint embedding into biological states and technical noise. MIDAS also highlights the need for approaches that can effectively handle missing modalities in mosaic multi-omics data.
Missing modalities in mosaic multi-omics data
Mosaic integration methods address the challenges of combining multi-omics datasets with incomplete and overlapping modalities across samples (Fig. 1), addressing scalability, cost, and modality coverage limitations. However, these methods face difficulties such as batch technical variation or robust imputation. Recent approaches, including matrix factorisation [74, 163], VAEs [49–51, 164], or StabMap [165], offer promising solutions.
For example, Multigrate [49] previously described takes the PoE approach and deals with the missing data by setting the posterior of the missing modality to 1. This allows the joint latent distribution to be determined from the available modalities, enabling joint embedding generation and reconstruction of all modalities, even if some are missing.
scVAEIT [164] uses conditional variational inference to handle missing features in multi-omics data. It generates random masks during training to encourage the model to predict the missing features based on the remaining observed data. This process allows the imputation of the unobserved values and denoising of the observed features.
MIDAS [51] addresses missing features by padding them with zeros for each cell, ensuring consistent feature vector size. The learned joint disentangled latent variables are passed to modality decoders, and a masking function removes the padded values, returning imputed values for missing features.
The flexibility of VAEs underscores their key role in advancing deep learning-based multi-omics integration, addressing challenges like handling missing modalities and enabling interpretable latent spaces. While these models provide robust frameworks for extracting meaningful biological insights, further refinement of their architectures is essential to enhance scalability, broaden applicability, and maximize their potential impact on precision medicine and synthetic biology.
Discussion and promising perspectives
Multi-omics data integration is a rapidly advancing area in computational biology, driven by the need to interpret complex biological systems across molecular layers. As outlined in Table 1, integration methods range from classical statistical to deep generative learning approaches, each offering unique trade-offs in terms of interpretability, scalability, computational complexity, and generative capacity. The choice of an appropriate integration strategy depends on multiple factors, including the characteristics of the datasets, and the specific research and biological questions being addressed.
Classical approaches, such as correlation-based and matrix factorisation techniques, offer interpretable and computationally efficient solutions, making them well-suited for joint dimensionality reduction exploratory analyses or as foundational steps in more complex frameworks. Probabilistic models offer the benefit of explicitly modelling uncertainty, while kernel- and network-based methods enable integration in a unified sample similarity space, either through kernel functions or graph-based representations. Deep generative models, particularly VAEs, have shown considerable success in addressing the challenges of high-dimensional, sparse, and noisy data. These models are especially powerful for learning joint embeddings, imputation, and denoising. However, current implementations are largely focused on scRNA-seq and scATAC-seq data, leveraging their availability and complementarity, and the maturity of single-cell analysis workflows. While effective in these domains, this focus limits their generalisability to broader experimental settings. Furthermore, VAEs typically require greater computational resources and are sensitive to training data size. As model complexity increases through deeper networks, larger inner layers, or more sophisticated loss functions, GPU support and large memory are often needed. In the multi-omics context, where large datasets across omics layers and conditions are often scarce, model performance may degrade. Data scarcity can lead to unstable training, poor convergence, and reduced robustness, ultimately constraining the applicability of deep generative models in real-world biomedical scenarios.
While several benchmarking studies have examined the runtime and computational resource requirements of specific multi-omics integration tools [18, 41, 166–169], providing a precise comparative evaluation across all categories and methods remains challenging. Nevertheless, all approaches reviewed here are feasible to run on standard research computing infrastructures. As the field advances, continued efforts to enhance model generalisability, computational efficiency, and flexibility across diverse omics types will be essential. In this context, emerging paradigms such as foundation models and the integration of increasingly diverse data modalities hold great promise for enhancing the scope and impact of multi-omics research in precision medicine.
Beyond omics: multimodal integration
The rapid advancement of biomedical technologies and the increasing diversity of data modalities present unprecedented opportunities for precision medicine and synthetic biology. Beyond molecular omics data, modern approaches increasingly integrate phenotypic datasets, imaging modalities, electronic health records, or bio-signals from wearable devices (Fig. 4). These multimodal datasets hold the promise of providing an even more holistic view of biological processes and diseases [170–172]. However, their integration remains challenging due to significant differences in data scale, format, and structure.
Figure 4.
Integration of multiple data modalities—molecular, phenotypic, imaging and electronic health records datasets (left)—to uncover molecular and regulatory mechanisms, enable disease subtyping and classification, enhance clinical outcome predictions for diagnosis and prognosis, and identify biomarkers of therapeutic response (right).
Integrating structured, tabular data, such as gene expression or clinical measurements, is comparatively straightforward, as these data modalities typically share compatible data distributions and can be aligned across samples. In contrast, unstructured data — such as spatially encoded medical images or clinical text, like electronic health records that require language models [173]—demands more advanced modelling strategies to extract meaningful biological insights.
Machine learning models have proven effective in this context. For example, Cheerla and Gevaert [174] developed a deep learning framework that integrates clinical data, molecular omics, and histopathology images using modality-specific FFNNs and a convolutional neural network (CNN) to learn a unified embedding used to predict survival through a Cox loss. Similarly, Carrillo-Perez et al. [175] proposed a late fusion strategy for non-small cell lung cancer (NSCLC) classification, where histology image features extracted by a CNN and molecular omics features modelled by support vector machines (SVMs) are later combined to produce the final output. Sammut et al. [176] introduced an ensemble machine learning model that integrates clinical, genomic, transcriptomic, digital pathology, and treatment data from breast cancer biopsies to predict therapeutic response. By combining modality-specific feature selection with an unweighted ensemble of logistic regression, SVMs, and random forests, the model achieved high predictive performance, validating the benefit of comprehensive multimodal data fusion.
In a more dynamic setting, DyAM [177], an attention-based multimodal framework designed to integrate clinical, genomic, radiomic features from CT scans, and pathology data for predicting immunotherapy response in NSCLC was proposed. DyAM processes each modality through logistic regression models and integrates them via a dynamic attention mechanism that adjusts weights based on modality relevance per patient. A masking function handles missing modalities by assigning them zero attention, enabling robust predictions without imputation.
Together, these studies highlight the increasing sophistication and potential of multimodal machine learning frameworks in precision medicine. Realizing this potential will require continued refinement of machine learning and advanced artificial intelligence (AI) methods to ensure their effective translation into clinical practice, ultimately enabling robust multimodal biomarker discovery and advancing data-driven, personalized patient care.
Transformers and foundation models
Foundation models, initially designed for natural language processing [178], are now being expanded to biological datasets [179–183], demonstrating versatility in multi-batch and multi-omics integration, perturbation response prediction, and gene regulatory network inference. These models, typically based on the self-attention transformer architecture, are pre-trained on large, diverse datasets, enabling them to generalize across various domains. By leveraging transfer learning, the pre-trained model can be fine-tuned on a specific domain or task with much less data than what would be needed to train from scratch. Several transformers-based models for single-cell omics and foundation models for bioinformatics applications are reviewed in [184, 185].
In single-cell omics, scFoundation [182], pre-trained on tens of millions of scRNA-seq profiles, demonstrated strong performance across a variety of downstream tasks including cell type annotation, perturbation response prediction, and gene module inference. Similarly, scGPT [180], a generative pre-trained transformer trained on over 33 million single-cell transcriptomic profiles, supports batch correction, perturbation modelling, gene network inference, and is designed to extend toward multi-omics integration.
Recent advances increasingly emphasize explicit integration of multiple omics layers. For instance, mosGraphGPT [179] represents multi-omics profiles as molecular graphs and uses a generative transformer to predict disease phenotypes and uncover key signalling pathways. scmFormer [186] uses a multi-task transformer to integrate transcriptomics and proteomics, supporting label transfer, imputation, and spatial analysis. GET [183] is an interpretable foundation model for transcriptional regulation analysis that integrates chromatin accessibility and genomic sequence data to predict gene expression. It enables applications such as regulatory activity prediction and inference of transcription factor networks.
Beyond molecular profiles, foundation models have been extended to integrate biological networks [187, 188], pathway-level information [189], and unstructured clinical information such as electronic health records [173] or digital pathology data [190], broadening their impact in translational medicine.
A notable example is PathChat [191], a multimodal large language model (MLLM) designed for computational pathology. It integrates a vision encoder pre-trained on over 100 million histopathology image patches, with a 13-billion-parameter language model. To align image and pathology text representations, the encoder was further pre-trained on over 1.18 million pathology image–caption pairs. The combined MLLM was fine-tuned on over 450 000 pathology-specific instructions. PathChat is a vision-language AI assistant, capable of jointly interpreting pathology images and text, and achieves state-of-the-art performance on diagnostic tasks across diverse tissue types and disease contexts.
Despite their promise, concerns remain about the applicability and performance of foundation models compared to state-of-the-art machine learning approaches [192, 193], highlighting the need for further refinement of these models and careful identification of their most relevant applications. Nonetheless, foundation models represent a promising direction for integrating complex multimodal biological data and advancing multi-omics-driven discovery in precision medicine.
Conclusions
This review explores diverse approaches to integrate multi-omics data, highlighting both classical methods and recent advancements in deep learning. We categorize integration techniques by their underlying approaches, offering a comprehensive technical overview of the models developed and highlighting their strengths, limitations, and applications. Classical methods remain valuable for their interpretability, serving as the foundational basis for the development of more complex models. Deep generative models, particularly VAEs, have significantly advanced the integration of high-dimensional and incomplete data, offering flexible designs and generative capabilities. Specifically, we review various regularisation strategies incorporated into VAE frameworks to address key challenges in multi-omics analysis and allow the extraction of biologically meaningful representations. Additionally, we highlight the emerging potential of foundation models and multimodal data to advance precision medicine research. Overall, this review provides a detailed overview of the current state of multi-omics integration methods and outlines promising directions for future advancements in understanding complex biological systems.
Key Points
This review systematically covers multi-omics integration, from classical statistics to deep learning, emphasising on technical descriptions and standardisation of notations.
Key focus in the generative potential of multi-omics models in precision medicine and synthetic biology, alongside recent technical advancements improving their performance.
Given the central role of VAEs, we provide a detailed technical analysis of their architectures, loss functions, and regularisation, unifying terminology. This is the first systematic review of VAEs in multi-omics integration.
We highlight emerging data modalities and foundation models, outlining future directions in multi-omics research.
Contributor Information
Ana R Baião, INESC-ID, Rua Alves Redol 9, 1000-029 Lisboa, Portugal; Instituto Superior Técnico (IST), Universidade de Lisboa, Av. Rovisco Pais, 1049-001 Lisboa, Portugal.
Zhaoxiang Cai, ProCan®, Children’s Medical Research Institute, Faculty of Medicine and Health, The University of Sydney, 214 Hawkesbury Road, Westmead, NSW 2145, Australia.
Rebecca C Poulos, ProCan®, Children’s Medical Research Institute, Faculty of Medicine and Health, The University of Sydney, 214 Hawkesbury Road, Westmead, NSW 2145, Australia.
Phillip J Robinson, ProCan®, Children’s Medical Research Institute, Faculty of Medicine and Health, The University of Sydney, 214 Hawkesbury Road, Westmead, NSW 2145, Australia.
Roger R Reddel, ProCan®, Children’s Medical Research Institute, Faculty of Medicine and Health, The University of Sydney, 214 Hawkesbury Road, Westmead, NSW 2145, Australia.
Qing Zhong, ProCan®, Children’s Medical Research Institute, Faculty of Medicine and Health, The University of Sydney, 214 Hawkesbury Road, Westmead, NSW 2145, Australia.
Susana Vinga, INESC-ID, Rua Alves Redol 9, 1000-029 Lisboa, Portugal; Instituto Superior Técnico (IST), Universidade de Lisboa, Av. Rovisco Pais, 1049-001 Lisboa, Portugal; IDMEC, Instituto Superior Técnico, Universidade de Lisboa, Av. Rovisco Pais, 1049-001 Lisboa, Portugal.
Emanuel Gonçalves, INESC-ID, Rua Alves Redol 9, 1000-029 Lisboa, Portugal; Instituto Superior Técnico (IST), Universidade de Lisboa, Av. Rovisco Pais, 1049-001 Lisboa, Portugal.
Conflict of interest
None declared.
Funding
ARB, SV, and EG work is partially supported by national funds through FCT, Fundação para a Ciência e a Tecnologia, under projects UIDB/50021/2020 (DOI:10.54499/UIDB/50021/2020), LAETA (DOI:10.54499/UIDB/50022/2020), 2024.07252.IACDC (10.54499/2024.07252.IACDC, through RE-C05-i08.M04), and grant number 15030 (10.54499/LISBOA2030-FEDER-00868200). ARB is also funded by the Portuguese national agency FCT, through the research grant UI/BD/154599/2022. Rebecca Poulos is supported by a Sydney Cancer Partners Translational Partners Fellowship with funding from a Cancer Institute NSW Capacity Building Grant (Grant ID. 2021/CBG0002) and the NHMRC of Australia (GNT2000855). ProCan is supported by the Australian Cancer Research Foundation, Cancer Institute New South Wales (NSW) (2017/TPG001, REG171150, 2021/CBG0002), NSW Ministry of Health (CMP-01), the University of Sydney, Cancer Council NSW (IG 18-01), Ian Potter Foundation, the Medical Research Future Fund (MRFF-PD), National Health and Medical Research Council (NHMRC) of Australia European Union grant (GNT1170739, a companion grant to support the ‘iPC-individualized Paediatric Cure’ [ref. 826121]), and National Breast Cancer Foundation (IIRS-18-164). Work at ProCan is done under the auspices of a Memorandum of Understanding between Children’s Medical Research Institute and the U.S. National Cancer Institute’s International Cancer Proteogenome Consortium (ICPC) that encourages cooperation among institutions and nations in proteogenomic cancer research in which datasets are made available to the public.
Data availability
No new data were generated or analyzed in this review.
References
- 1. Hasin Y, Seldin M, Lusis A. Multi-omics approaches to disease. Genome Biol 2017;18:83. 10.1186/s13059-017-1215-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2. Jiang P, Sinha S, Aldape K, et al. Big data in basic and translational cancer research. Nat Rev Cancer 2022;22:625–39. 10.1038/s41568-022-00502-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Zhang Z, Huang K, Gu C, et al. Molecular subtyping of serous ovarian cancer based on multi-omics data. Sci Rep 2016;6:26001. 10.1038/srep26001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Lindskrog SV, Prip F, Lamy P, et al. An integrated multi-omics analysis identifies prognostic molecular subtypes of non-muscle-invasive bladder cancer. Nat Commun 2021;12:2301. 10.1038/s41467-021-22465-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. Lehmann BD, Colaprico A, Silva TC, et al. Multi-omics analysis identifies therapeutic vulnerabilities in triple-negative breast cancer subtypes. Nat Commun 2021;12:6276. 10.1038/s41467-021-26502-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Dugger SA, Platt A, Goldstein DB. Drug development in the era of precision medicine. Nat Rev Drug Discov 2018;17:183–96. 10.1038/nrd.2017.226. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Iorio F, Knijnenburg TA, Vis DJ, et al. A landscape of pharmacogenomic interactions in cancer. Cell 2016;166:740–54. 10.1016/j.cell.2016.06.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Cohen JD, Javed AA, Thoburn C, et al. Combined circulating tumor DNA and protein biomarker-based liquid biopsy for the earlier detection of pancreatic cancers. Proc Natl Acad Sci U S A 2017;114:10202–7. 10.1073/pnas.1704961114. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Mishra NK, Southekal S, Guda C. Survival analysis of multi-omics data identifies potential prognostic markers of pancreatic ductal adenocarcinoma. Front Genet 2019;10:624. 10.3389/fgene.2019.00624. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Montaner J, Ramiro L, Simats A, et al. Multilevel omics for the discovery of biomarkers and therapeutic targets for stroke. Nat Rev Neurol 2020;16:247–64. 10.1038/s41582-020-0350-6. [DOI] [PubMed] [Google Scholar]
- 11. Menyhárt O, Győrffy B. Multi-omics approaches in cancer research with applications in tumor subtyping, prognosis, and diagnosis. Comput Struct Biotechnol J 2021;19:949–60. 10.1016/j.csbj.2021.01.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Chai H, Zhou X, Zhang Z, et al. Integrating multi-omics data through deep learning for accurate cancer prognosis prediction. Comput Biol Med 2021;134:104481. 10.1016/j.compbiomed.2021.104481. [DOI] [PubMed] [Google Scholar]
- 13. Gonçalves E, Poulos RC, Cai Z, et al. Pan-cancer proteomic map of 949 human cell lines. Cancer Cell 2022;40:835–849.e8. 10.1016/j.ccell.2022.06.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Rosellini M, Marchetti A, Mollica V, et al. Prognostic and predictive biomarkers for immunotherapy in advanced renal cell carcinoma. Nat Rev Urol 2023;20:133–57. 10.1038/s41585-022-00676-0. [DOI] [PubMed] [Google Scholar]
- 15. ICGC/TCGA Pan-Cancer Analysis of Whole Genomes Consortium . Pan-cancer analysis of whole genomes. Nature 2020;578:82–93. 10.1038/s41586-020-1969-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Tully B, Balleine RL, Hains PG, et al. Addressing the challenges of high-throughput cancer tissue proteomics for clinical application: ProCan. Proteomics 2019;19:e1900109. 10.1002/pmic.201900109. [DOI] [PubMed] [Google Scholar]
- 17. Subramanian I, Verma S, Kumar S, et al. Multi-omics data integration, interpretation, and its application. Bioinform Biol Insights 2020;14:1177932219899051. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Cai Z, Poulos RC, Liu J, et al. Machine learning for multi-omics data integration in cancer. iScience 2022;25:103798. 10.1016/j.isci.2022.103798. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Athieniti E, Spyrou GM. A guide to multi-omics data collection and integration for translational medicine. Comput Struct Biotechnol J 2023;21:134–49. 10.1016/j.csbj.2022.11.050. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Mirza B, Wang W, Wang J, et al. Machine learning and integrative analysis of biomedical big data. Genes 2019;10:10. 10.3390/genes10020087. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Reel PS, Reel S, Pearson E, et al. Using machine learning approaches for multi-omics data analysis: a review. Biotechnol Adv 2021;49:107739. 10.1016/j.biotechadv.2021.107739. [DOI] [PubMed] [Google Scholar]
- 22. Zitnik M, Nguyen F, Wang B, et al. Machine learning for integrating data in biology and medicine: principles, practice, and opportunities. Inf Fusion 2019;50:71–91. 10.1016/j.inffus.2018.09.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Cai Z, Apolinário S, Baião AR, et al. Synthetic augmentation of cancer cell line multi-omic datasets using unsupervised deep learning. Nat Commun 2024;15:10390. 10.1038/s41467-024-54771-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Meng C, Zeleznik OA, Thallinger GG, et al. Dimension reduction techniques for the integrative analysis of multi-omics data. Brief Bioinform 2016;17:628–41. 10.1093/bib/bbv108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Cantini L, Zakeri P, Hernandez C, et al. Benchmarking joint multi-omics dimensionality reduction approaches for the study of cancer. Nat Commun 2021;12:124. 10.1038/s41467-020-20430-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Poulos RC, Hains PG, Shah R, et al. Strategies to enable large-scale proteomics for reproducible research. Nat Commun 2020;11:3793. 10.1038/s41467-020-17641-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Yu Y, Mai Y, Zheng Y, et al. Assessing and mitigating batch effects in large-scale omics studies. Genome Biol 2024;25:254. 10.1186/s13059-024-03401-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Song M, Greenbaum J, Luttrell J, 4th, et al. A review of integrative imputation for multi-omics datasets. Front Genet 2020;11:570255. 10.3389/fgene.2020.570255. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Flores JE, Claborne DM, Weller ZD, et al. Missing data in multi-omics integration: recent advances through artificial intelligence. Front Artif Intell 2023;6:1098308. 10.3389/frai.2023.1098308. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. Huang L, Song M, Shen H, et al. Deep learning methods for omics data imputation. Biology 2023;12:12. 10.3390/biology12101313. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31. Argelaguet R, Cuomo ASE, Stegle O, et al. Computational principles and challenges in single-cell data integration. Nat Biotechnol 2021;39:1202–15. 10.1038/s41587-021-00895-7. [DOI] [PubMed] [Google Scholar]
- 32. Fouché A, Zinovyev A. Omics data integration in computational biology viewed through the prism of machine learning paradigms. Front Bioinform 2023;3:1191961. 10.3389/fbinf.2023.1191961. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33. Picard M, Scott-Boyer M-P, Bodein A, et al. Integration strategies of multi-omics data for machine learning analysis. Comput Struct Biotechnol J 2021;19:3735–46. 10.1016/j.csbj.2021.06.030. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Nicora G, Vitali F, Dagliati A, et al. Integrated multi-omics analyses in oncology: a review of machine learning methods and tools. Front Oncol 2020;10:1030. 10.3389/fonc.2020.01030. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Correa-Aguila R, Alonso-Pupo N, Hernández-Rodríguez EW. Multi-omics data integration approaches for precision oncology. Mol Omics 2022;18:469–79. 10.1039/D1MO00411E. [DOI] [PubMed] [Google Scholar]
- 36. Vahabi N, Michailidis G. Unsupervised multi-omics data integration methods: a comprehensive review. Front Genet 2022;13:854752. 10.3389/fgene.2022.854752. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37. Zhang Z, Zhao Y, Liao X, et al. Deep learning in omics: a survey and guideline. Brief Funct Genomics 2019;18:41–57. 10.1093/bfgp/ely030. [DOI] [PubMed] [Google Scholar]
- 38. Tran KA, Kondrashova O, Bradley A, et al. Deep learning in cancer diagnosis, prognosis and treatment selection. Genome Med 2021;13:152. 10.1186/s13073-021-00968-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39. Stahlschmidt SR, Ulfenborg B, Synnergren J. Multimodal deep learning for biomedical data fusion: a review. Brief Bioinform 2022;23:bbab569. 10.1093/bib/bbab569. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Kang M, Ko E, Mersha TB. A roadmap for multi-omics data integration using deep learning. Brief Bioinform 2022;23:bbab454. 10.1093/bib/bbab454. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Brombacher E, Hackenberg M, Kreutz C, et al. The performance of deep generative models for learning joint embeddings of single-cell multi-omics data. Front Mol Biosci 2022;9:962644. 10.3389/fmolb.2022.962644. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42. Athaya T, Ripan RC, Li X, et al. Multimodal deep learning approaches for single-cell multi-omics data integration. Brief Bioinform 2023;24:bbad313. 10.1093/bib/bbad313. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. Wen Y, Zheng L, Leng D, et al. Deep learning-based multiomics data integration methods for biomedical application. Adv Intell Syst 2023;5:5. 10.1002/aisy.202200247. [DOI] [Google Scholar]
- 44. Erfanian N, Heydari AA, Feriz AM, et al. Deep learning applications in single-cell genomics and transcriptomics data analysis. Biomed Pharmacother 2023;165:115077. 10.1016/j.biopha.2023.115077. [DOI] [PubMed] [Google Scholar]
- 45. Wekesa JS, Kimwele M. A review of multi-omics data integration through deep learning approaches for disease diagnosis, prognosis, and treatment. Front Genet 2023;14:1199087. 10.3389/fgene.2023.1199087. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46. Ballard JL, Wang Z, Li W, et al. Deep learning-based approaches for multi-omics data integration and analysis. BioData Min 2024;17:38. 10.1186/s13040-024-00391-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47. Stark SG, Ficek J, Locatello F, et al. SCIM: universal single-cell matching with unpaired feature sets. Bioinformatics 2020; 36:i919–27. 10.1093/bioinformatics/btaa843. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48. Hira MT, Razzaque MA, Angione C, et al. Integrated multi-omics analysis of ovarian cancer using variational autoencoders. Sci Rep 2021;11:6265. 10.1038/s41598-021-85285-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49. Lotfollahi M, Litinetskaya A, Theis FJ. Multigrate: Single-cell multi-omic data integration. bioRxiv 2022. 10.1101/2022.03.16.484643. [DOI] [Google Scholar]
- 50. Cohen Kalafut N, Huang X, Wang D. Joint variational autoencoders for multimodal imputation and embedding. Nature Machine Intelligence 2023;5:631–42. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51. He Z, Hu S, Chen Y, et al. Mosaic integration and knowledge transfer of single-cell multimodal data with MIDAS. Nat Biotechnol 2024;42:1594–605. 10.1038/s41587-023-02040-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52. Hotelling H. Relations between two sets of variates. Biometrika 1936;28:321. [Google Scholar]
- 53. Waaijenborg S, de Witt V, Hamer PC, et al. Quantifying the association between gene expressions and DNA-markers by penalized canonical correlation analysis. Stat Appl Genet Mol Biol 2008;7: Article3. [DOI] [PubMed] [Google Scholar]
- 54. Parkhomenko E, Tritchler D, Beyene J. Sparse canonical correlation analysis with application to genomic data integration. Stat Appl Genet Mol Biol 2009;8: Article 1. [DOI] [PubMed] [Google Scholar]
- 55. Witten DM, Tibshirani RJ. Extensions of sparse canonical correlation analysis with applications to genomic data. Stat Appl Genet Mol Biol 2009;8: Article28. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56. Lin D, Zhang J, Li J, et al. Group sparse canonical correlation analysis for genomic data integration. BMC Bioinformatics 2013;14:245. 10.1186/1471-2105-14-245. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57. Meng C, Kuster B, Culhane AC, et al. A multivariate approach to the integration of multi-omics datasets. BMC Bioinformatics 2014;15:162. 10.1186/1471-2105-15-162. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58. Witten DM, Tibshirani R, Hastie T. A penalized matrix decomposition, with applications to sparse principal components and canonical correlation analysis. Biostatistics 2009;10:515–34. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59. Lê Cao K-A, Martin PGP, Robert-Granié C, et al. Sparse canonical methods for biological data integration: application to a cross-platform study. BMC Bioinformatics 2009;10:34. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60. Tenenhaus A, Tenenhaus M. Regularized generalized canonical correlation analysis. Psychometrika 2011;76:257–84. [DOI] [PubMed] [Google Scholar]
- 61. Tenenhaus A, Tenenhaus M. Regularized generalized canonical correlation analysis for multiblock or multigroup data analysis. Eur J Oper Res 2014;238:391–403. [Google Scholar]
- 62. Singh A, Shannon CP, Gautier B, et al. DIABLO: an integrative approach for identifying key molecular drivers from multi-omics assays. Bioinformatics 2019;35:3055–62. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63. Moon S, Hwang J, Lee H. SDGCCA: supervised deep generalized canonical correlation analysis for multi-omics integration. J Comput Biol 2022;29:892–907. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64. Hu J, Chen M, Zhou X. Effective and scalable single-cell data alignment with non-linear canonical correlation analysis. Nucleic Acids Res 2022;50:e21. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65. Wang Y, Lian B, Zhang H, et al. A multi-view latent variable model reveals cellular heterogeneity in complex tissues for paired multimodal single-cell data. Bioinformatics 2023;39:btad005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66. Lê Cao K-A, Boitard S, Besse P. Sparse PLS discriminant analysis: biologically relevant feature selection and graphical displays for multiclass problems. BMC Bioinformatics 2011;12:253. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67. Li W, Zhang S, Liu C-C, et al. Identifying multi-layer gene regulatory modules from multi-dimensional genomic data. Bioinformatics 2012;28:2458–66. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68. Boulesteix A-L, Strimmer K. Partial least squares: a versatile tool for the analysis of high-dimensional genomic data. Brief Bioinform 2007;8:32–44. [DOI] [PubMed] [Google Scholar]
- 69. Lock EF, Hoadley KA, Marron JS, et al. Joint and individual variation explained (jive) for integrated analysis of multiple data types. Ann Appl Stat 2013;7:523–42. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70. Zhang S, Liu C-C, Li W, et al. Discovery of multi-dimensional modules by integrative analysis of cancer genomic data. Nucleic Acids Res 2012;40:9379–91. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71. Chalise P, Fridley BL. Integrative clustering of multi-level ‘omic data based on non-negative matrix factorization algorithm. PloS One 2017;12:e0176278. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72. Welch JD, Kozareva V, Ferreira A, et al. Single-cell multi-omic integration compares and contrasts features of brain cell identity. Cell 2019;177:1873–1887.e17. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73. Yang Z, Michailidis G. A non-negative matrix factorization method for detecting modules in heterogeneous omics multi-modal data. Bioinformatics 2016;32:1–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74. Kriebel AR, Welch JD. UINMF performs mosaic integration of single-cell multi-omic datasets using nonnegative matrix factorization. Nat Commun 2022;13:780. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75. Shen R, Olshen AB, Ladanyi M. Integrative clustering of multiple genomic data types using a joint latent variable model with application to breast and lung cancer subtype analysis. Bioinformatics 2009;25:2906–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76. Moon TK. The expectation-maximization algorithm. IEEE Signal Process Mag 1996;13:47–60. [Google Scholar]
- 77. Mo Q, Wang S, Seshan VE, et al. Pattern discovery and cancer gene identification in integrated cancer genomic data. Proc Natl Acad Sci U S A 2013;110:4245–50. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78. Mo Q, Shen R, Guo C, et al. A fully Bayesian latent variable model for integrative clustering analysis of multi-type omics data. Biostatistics 2018;19:71–86. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79. Wu D, Wang D, Zhang MQ, et al. Fast dimension reduction and integrative clustering of multi-omics data using low-rank approximation: application to cancer molecular classification. BMC Genomics 2015;16:1022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80. Meng C, Helm D, Frejno M, et al. moCluster: identifying joint patterns across multiple omics data sets. J Proteome Res 2016;15:755–65. [DOI] [PubMed] [Google Scholar]
- 81. Westerhuis JA, Kourti T, MacGregor JF. Analysis of multiblock and hierarchical PCA and PLS models. J Chemometr 1998;12:301–21. [Google Scholar]
- 82. Argelaguet R, Velten B, Arnol D, et al. Multi-omics factor analysis-a framework for unsupervised integration of multi-omics data sets. Mol Syst Biol 2018;14:e8124. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83. Blei DM, Kucukelbir A, McAuliffe JD. Variational inference: a review for statisticians. J Am Stat Assoc 2017;112:859–77. [Google Scholar]
- 84. Argelaguet R, Arnol D, Bredikhin D, et al. MOFA+: a statistical framework for comprehensive integration of multi-modal single-cell data. Genome Biol 2020;21:111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85. Schölkopf B, Tsuda K, Vert J-P. Kernel Methods in Computational Biology. Cambridge: MIT press, 2004.
- 86. Gönen M, Alpaydin E. Multiple kernel learning algorithms. J Mach Learn Res 2011;12:2211–68. [Google Scholar]
- 87. Speicher NK, Pfeifer N. Integrating different data types by regularized unsupervised multiple kernel learning with application to cancer subtype discovery. Bioinformatics 2015;31:i268–75. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88. Röder B, Kersten N, Herr M, et al. Web-rMKL: a web server for dimensionality reduction and sample clustering of multi-view data based on unsupervised multiple kernel learning. Nucleic Acids Res 2019;47:W605–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 89. Cichonska A, Pahikkala T, Szedmak S, et al. Learning with multiple pairwise kernels for drug bioactivity prediction. Bioinformatics 2018;34:i509–18. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 90. Wang B, Mezlini AM, Demir F, et al. Similarity network fusion for aggregating data types on a genomic scale. Nat Methods 2014;11:333–7. [DOI] [PubMed] [Google Scholar]
- 91. Li C-X, Wheelock CE, Sköld CM, et al. Integration of multi-omics datasets enables molecular classification of COPD. Eur Respir J 2018;51:1701930. [DOI] [PubMed] [Google Scholar]
- 92. Chiu AM, Mitra M, Boymoushakian L, et al. Integrative analysis of the inter-tumoral heterogeneity of triple-negative breast cancer. Sci Rep 2018;8:11807. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 93. Wang C, Lue W, Kaalia R, et al. Network-based integration of multi-omics data for clinical outcome prediction in neuroblastoma. Sci Rep 2022;12:15425. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 94. Yang M, Matan-Lithwick S, Wang Y, et al. Multi-omic integration via similarity network fusion to detect molecular subtypes of ageing. Brain Commun 2023;5:fcad110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 95. Chierici M, Bussola N, Marcolini A, et al. Integrative network fusion: a multi-omics approach in molecular profiling. Front Oncol 2020;10:1065. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96. Wang T-H, Lee C-Y, Lee T-Y, et al. Biomarker identification through multiomics data analysis of prostate cancer prognostication using a deep learning model and similarity network fusion. Cancers (Basel) 2021;13:2528. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97. Li X, Ma J, Leng L, et al. MoGCN: a multi-omics integration method based on graph convolutional network for cancer subtype analysis. Front Genet 2022;13:806842. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 98. Li L, Wei Y, Shi G, et al. Multi-omics data integration for subtype identification of Chinese lower-grade gliomas: a joint similarity network fusion approach. Comput Struct Biotechnol J 2022;20:3482–92. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99. Liu X-Y, Mei X-Y. Prediction of drug sensitivity based on multi-omics data using deep learning and similarity network fusion approaches. Front Bioeng Biotechnol 2023;11:1156372. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 100. Rappoport N, Shamir R. NEMO: Cancer subtyping by integration of partial multi-omic data. Bioinformatics 2019;35:3348–56. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101. Angermueller C, Pärnamaa T, Parts L, et al. Deep learning for computational biology. Mol Syst Biol 2016;12:878. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 102. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature 2015;521:436–44. [DOI] [PubMed] [Google Scholar]
- 103. Sharifi-Noghabi H, Zolotareva O, Collins CC, et al. MOLI: multi-omics late integration with deep neural networks for drug response prediction. Bioinformatics 2019;35:i501–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 104. Preuer K, Lewis RPI, Hochreiter S, et al. DeepSynergy: predicting anti-cancer drug synergy with deep learning. Bioinformatics 2018;34:1538–46. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 105. Kuru HI, Tastan O, Cicek AE. MatchMaker: a deep learning framework for drug synergy prediction. IEEE/ACM Trans Comput Biol Bioinform 2022;19:2334–44. [DOI] [PubMed] [Google Scholar]
- 106. Huang Z, Zhan X, Xiang S, et al. SALMON: survival analysis learning with multi-omics neural networks on breast cancer. Front Genet 2019;10:166. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 107. Zhang Z, Yang C, Zhang X. scDART: integrating unmatched scRNA-seq and scATAC-seq data and learning cross-modality relationship simultaneously. Genome Biol 2022;23:139. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 108. Wang T, Shao W, Huang Z, et al. MOGONET integrates multi-omics data using graph convolutional networks allowing patient classification and biomarker identification. Nat Commun 2021;12:3445. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 109. Wen H, Ding J, Jin W. et al. Graph neural networks for multimodal single-cell data integration. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD '22). Association for Computing Machinery, New York, NY, USA, 4153–4163. 10.1145/3534678.3539213. [DOI]
- 110. Zhuang Y, Xing F, Ghosh D, et al. Deep learning on graphs for multi-omics classification of COPD. PloS One 2023;18:e0284563. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 111. Wu KE, Yost KE, Chang HY, et al. BABEL enables cross-modality translation between multiomic profiles at single-cell resolution. Proc Natl Acad Sci U S A 2021;118:e2023070118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 112. Yang KD, Belyaeva A, Venkatachalapathy S, et al. Multi-domain translation between single-cell imaging and sequencing data using autoencoders. Nat Commun 2021;12:31. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 113. Chaudhary K, Poirion OB, Lu L, et al. Deep learning-based multi-omics integration robustly predicts survival in liver cancer. Clin Cancer Res 2018;24:1248–59. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 114. Zhang L, Lv C, Jin Y, et al. Deep learning-based multi-omics data integration reveals two prognostic subtypes in high-risk neuroblastoma. Front Genet 2018;9:477. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 115. Poirion OB, Jing Z, Chaudhary K, et al. DeepProg: an ensemble of deep-learning and machine-learning models for prognosis prediction using multi-omics data. Genome Med 2021;13:112. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 116. Song H, Ruan C, Xu Y, et al. Survival stratification for colorectal cancer via multi-omics integration using an autoencoder-based model. Exp Biol Med (Maywood) 2022;247:898–909. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 117. Zhang G, Peng Z, Yan C, et al. MultiGATAE: a novel cancer subtype identification method based on multi-omics and attention mechanism. Front Genet 2022;13:855629. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 118. Munquad S, Das AB. DeepAutoGlioma: a deep learning autoencoder-based multi-omics data integration and classification tools for glioma subtyping. BioData Min 2023;16:32. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 119. Lin X, Tian T, Wei Z, et al. Clustering of single-cell multi-omics data with a multimodal deep learning method. Nat Commun 2022;13:7705. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 120. Chen Y, Wen Y, Xie C, et al. MOCSS: multi-omics data clustering and cancer subtyping via shared and specific representation learning. iScience 2023;26:107378. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 121. Zhang T, Zhang L, Payne PRO, et al. Synergistic drug combination prediction by integrating multiomics data in deep learning models. Methods Mol Biol 2021;2194:223–38. [DOI] [PubMed] [Google Scholar]
- 122. Xu Y, Das P, McCord RP. SMILE: mutual information learning for integration of single-cell omics data. Bioinformatics 2022;38:476–86. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 123. Yu T. AIME: autoencoder-based integrative multi-omics data embedding that allows for confounder adjustments. PLoS Comput Biol 2022;18:e1009826. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 124. Kingma DP, Welling M. Auto-encoding variational bayes. In Proceedings of the 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14–16, 2014. [Google Scholar]
- 125. Kingma DP, Welling M. An introduction to variational autoencoders. Found Trends® Mach Learn 2019;12:307–92. [Google Scholar]
- 126. Lopez R, Gayoso A, Yosef N. Enhancing scientific discoveries in molecular biology with deep generative models. Mol Syst Biol 2020;16:e9198. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 127. Goodfellow I, Pouget-Abadie J, Mirza M, et al. Generative adversarial networks. Commun ACM 2020;63:139–44. [Google Scholar]
- 128. Vaswani A, Shazeer N, Parmar N, et al. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems 2017. Red Hook, NY, USA: Curran Associates Inc., 6000–6010.
- 129. Minoura K, Abe K, Nam H, et al. A mixture-of-experts deep generative model for integrated analysis of single-cell multiomics data. Cell Rep Methods 2021;1:100071. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 130. Gayoso A, Steier Z, Lopez R, et al. Joint probabilistic modeling of single-cell multi-omic data with totalVI. Nat Methods 2021;18:272–82. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 131. Zhao S, Song J, Ermon S. InfoVAE: balancing learning and inference in variational autoencoders. Proc Conf AAAI Artif Intell 2019;33:5885–92. [Google Scholar]
- 132. Benkirane H, Pradat Y, Michiels S, et al. CustOmics: a versatile deep-learning based strategy for multi-omics integration. PLoS Comput Biol 2023;19:e1010921. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 133. Gretton A, Borgwardt KM, Rasch M, et al. A kernel method for the two-sample-problem. Neural Inf Process Syst 2006;19:513–20. [Google Scholar]
- 134. Ashuach T, Gabitto MI, Koodli RV, et al. MultiVI: deep generative model for the integration of multimodal data. Nat Methods 2023;20:1222–31. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 135. Zhang Z, Zhao X, Bindra M, et al. scDisInFact: disentangled learning for integration and prediction of multi-batch multi-condition single-cell RNA-sequencing data. Nat Commun 2024;15:912. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 136. Zhang X, Zhang J, Sun K, et al. Integrated multi-omics analysis using variational autoencoders: application to pan-cancer classification. IEEE International Conference on Bioinformatics and Biomedicine (BIBM) 2019;2019:765–9. [Google Scholar]
- 137. Withnell E, Zhang X, Sun K, et al. XOmiVAE: an interpretable deep learning model for cancer classification using high-dimensional omics data. Brief Bioinform 2021;22:bbab315. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 138. Lundberg SM, Lee S-I. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems 2017. Red Hook, NY, USA: Curran Associates Inc., 4768–77.
- 139. Zhang X, Xing Y, Sun K, et al. OmiEmbed: a unified multi-task deep learning framework for multi-omics data. Cancers 2021;13:3047. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 140. Shi Y, Narayanaswamy S, Paige B, et al. Variational mixture-of-experts autoencoders for multi-modal deep generative models. In Proceedings of the 33rd International Conference on Neural Information Processing Systems 2019. Red Hook, NY, USA: Curran Associates Inc., Article 1408, 15718–15729. [Google Scholar]
- 141. Sutter TM, Daunhawer I, Vogt JE. Generalized multimodal ELBO. In Proceedings of the 9th International Conference on Learning Representations, ICLR 2021. [Google Scholar]
- 142. Zuo C, Chen L. Deep-joint-learning analysis model of single cell transcriptome and open chromatin accessibility data. Brief Bioinform 2021;22:bbaa287. 10.1093/bib/bbaa287. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 143. Lotfollahi M, Naghipourfar M, Luecken MD, et al. Mapping single-cell data to reference atlases by transfer learning. Nat Biotechnol 2022;40:121–30. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 144. Gayoso A, Lopez R, Xing G, et al. A python library for probabilistic analysis of single-cell omics data. Nat Biotechnol 2022;40:163–6. [DOI] [PubMed] [Google Scholar]
- 145. Lopez R, Regier J, Cole MB, et al. Deep generative modeling for single-cell transcriptomics. Nat Methods 2018;15:1053–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 146. Gong B, Zhou Y, Purdom E. Cobolt: integrative analysis of multimodal single-cell sequencing data. Genome Biol 2021;22:351. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 147. Zuo C, Dai H, Chen L. Deep cross-omics cycle attention model for joint analysis of single-cell multi-omics data. Bioinformatics 2021;37:4091–9. [DOI] [PubMed] [Google Scholar]
- 148. Ahmed KT, Sun J, Cheng S, et al. Multi-omics data integration by generative adversarial network. Bioinformatics 2021;38:179–86. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 149. Cao Z-J, Gao G. Multi-omics single-cell data integration and regulatory inference with graph-linked embedding. Nat Biotechnol 2022;40:1458–66. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 150. Yang X, Mann KK, Wu H, et al. scCross: a deep generative model for unifying single-cell multi-omics with seamless integration, cross-modal generation, and in silico exploration. Genome Biol 2024;25:198. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 151. Zhang R, Meng-Papaxanthos L, Vert J-P, et al. Semi-supervised single-cell cross-modality translation using polarbear. In Proceedings of the 26th Annual International Conference Research in Computational Molecular Biology, RECOMB 2022, San Diego, CA, USA, May 22–25, 2022. Berlin, Heidelberg: Springer-Verlag, 20–35. 10.1007/978-3-031-04749-7_2. [DOI] [Google Scholar]
- 152. Zhao J, Wang G, Ming J, et al. Adversarial domain translation networks for integrating large-scale atlas-level single-cell datasets. Nat Comput Sci 2022;2:317–30. [DOI] [PubMed] [Google Scholar]
- 153. Zhu J-Y, Park T, Isola P, et al. Unpaired image-to-image translation using cycle-consistent adversarial networks. IEEE International Conference on Computer Vision (ICCV) 2017;2017:2242–51. [Google Scholar]
- 154. Li G, Fu S, Wang S, et al. A deep generative model for multi-view profiling of single-cell RNA-seq and ATAC-seq data. Genome Biol 2022;23:20. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 155. Xu Y, Begoli E, McCord RP. sciCAN: single-cell chromatin accessibility and gene expression data integration via cycle-consistent adversarial network. NPJ Syst Biol Appl 2022;8:33. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 156. Tian Y, Krishnan D, Isola P. Contrastive Multiview coding. Computer Vision – ECCV 2020;2020:776–94. [Google Scholar]
- 157. Han W, Cheng Y, Chen J, et al. Self-supervised contrastive learning for integrative single cell RNA-seq data analysis. Brief Bioinform 2022;23:bbac377. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 158. Wang X, Hu Z, Yu T, et al. Con-AAE: contrastive cycle adversarial autoencoders for single-cell multi-omics alignment and integration. Bioinformatics 2023;39:btad162. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 159. Zhao C, Liu A, Zhang X, et al. CLCLSA: cross-omics linked embedding with contrastive learning and self attention for integration with incomplete multi-omics data. Comput Biol Med 2024;170:108058. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 160. Wang X, Chen H, Tang S, et al. Disentangled representation learning. IEEE Trans Pattern Anal Mach Intell 2024;46:9677–96. [DOI] [PubMed] [Google Scholar]
- 161. Kumar A, Sattigeri P, Balakrishnan A. Variational inference of disentangled latent concepts from Unlabeled observations. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018.
- 162. Yu S, Sanchez Giraldo L, Principe J. Information-theoretic methods in deep neural networks: Recent advances and emerging opportunities. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 2021, Montreal, Canada, August 19-27, 2021. 10.24963/ijcai.2021/633. [DOI]
- 163. Zhang Z, Sun H, Mariappan R, et al. scMoMaT jointly performs single cell mosaic integration and multi-modal bio-marker detection. Nat Commun 2023;14:384. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 164. Du J-H, Cai Z, Roeder K. Robust probabilistic modeling for single-cell multimodal mosaic integration and imputation via scVAEIT. Proc Natl Acad Sci U S A 2022;119:e2214414119. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 165. Ghazanfar S, Guibentif C, Marioni JC. Stabilized mosaic single-cell data integration using unshared features. Nat Biotechnol 2024;42:284–92. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 166. Rappoport N, Shamir R. Multi-omic and multi-view clustering algorithms: review and cancer benchmark. Nucleic Acids Res 2018;46:10546–62. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 167. Lee MYY, Kaestner KH, Li M. Benchmarking algorithms for joint integration of unpaired and paired single-cell RNA-seq and ATAC-seq data. Genome Biol 2023;24:244. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 168. Xiao C, Chen Y, Meng Q, et al. Benchmarking multi-omics integration algorithms across single-cell RNA and ATAC data. Brief Bioinform 2024;25:bbae095. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 169. Hu Y, Wan S, Luo Y, et al. Benchmarking algorithms for single-cell multi-omics prediction and integration. Nat Methods 2024;21:2182–94. [DOI] [PubMed] [Google Scholar]
- 170. Boehm KM, Khosravi P, Vanguri R, et al. Harnessing multimodal data integration to advance precision oncology. Nat Rev Cancer 2022;22:114–26. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 171. Tong L, Shi W, Isgut M, et al. Integrating multi-omics data with EHR for precision medicine using advanced artificial intelligence. IEEE Rev Biomed Eng 2024;17:80–97. [DOI] [PubMed] [Google Scholar]
- 172. Nam Y, Kim J, Jung S-H, et al. Harnessing artificial intelligence in multimodal omics data integration: paving the path for the next frontier in precision medicine. Annu Rev Biomed Data Sci 2024;7:225–50. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 173. Yang X, Chen A, PourNejatian N, et al. A large language model for electronic health records. NPJ Digit Med 2022;5:194. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 174. Cheerla A, Gevaert O. Deep learning with multimodal representation for pancancer prognosis prediction. Bioinformatics 2019;35:i446–54. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 175. Carrillo-Perez F, Morales JC, Castillo-Secilla D, et al. Machine-learning-based late fusion on multi-omics and multi-scale data for non-small-cell lung cancer diagnosis. J Pers Med 2022;12:601. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 176. Sammut S-J, Crispin-Ortuzar M, Chin S-F, et al. Multi-omic machine learning predictor of breast cancer therapy response. Nature 2022;601:623–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 177. Vanguri RS, Luo J, Aukerman AT, et al. Multimodal integration of radiology, pathology and genomics for prediction of response to PD-(L)1 blockade in patients with non-small cell lung cancer. Nat Cancer 2022;3:1151–64. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 178. Bommasani R, Hudson DA, Adeli E, et al. On the opportunities and risks of foundation models. arXiv [csLG] 2021. 10.48550/arXiv.2108.07258. [DOI] [Google Scholar]
- 179. Zhang H, Huang D, Chen E, et al. mosGraphGPT: a foundation model for multi-omic signaling graphs using generative AI. bioRxiv 2024. 10.1101/2024.08.01.606222. [DOI] [Google Scholar]
- 180. Cui H, Wang C, Maan H, et al. scGPT: toward building a foundation model for single-cell multi-omics using generative AI. Nat Methods 2024;21:1470–80. [DOI] [PubMed] [Google Scholar]
- 181. Méndez-Lucio O, Nicolaou CA, Earnshaw B. MolE: a foundation model for molecular graphs using disentangled attention. Nat Commun 2024;15:9431. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 182. Hao M, Gong J, Zeng X, et al. Large-scale foundation model on single-cell transcriptomics. Nat Methods 2024;21:1481–91. [DOI] [PubMed] [Google Scholar]
- 183. Fu X, Mo S, Buendia A, et al. A foundation model of transcription across human cell types. Nature 2025;637:965–73.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 184. Szałata A, Hrovatin K, Becker S, et al. Transformers in single-cell omics: a review and new perspectives. Nat Methods 2024;21:1430–43. [DOI] [PubMed] [Google Scholar]
- 185. Guo F, Guan R, Li Y, et al. Foundation models in bioinformatics. Natl Sci Rev 2025;12:nwaf028. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 186. Xu J, Huang D-S, Zhang X. ScmFormer integrates large-scale single-cell proteomics and transcriptomics data by multi-task transformer. Adv Sci (Weinh) 2024;11:e2307835. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 187. Hetzel L, Fischer DS, Günnemann S, et al. Graph representation learning for single-cell biology. Curr Opin Syst Biol 2021;28:100347. [Google Scholar]
- 188. Liu J, Yang C, Lu Z, et al. Towards graph foundation models: a survey and beyond. arXiv [csLG] 2023. 10.48550/arXiv.2310.11829. [DOI] [Google Scholar]
- 189. Cai Z, Poulos RC, Aref A, et al. DeePathNet: A Transformer-Based Deep Learning Model Integrating Multi-Omic Data with Cancer Pathways. Commun: Cancer Res, 2024;4:3151-3164. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 190. Li D, Wan G, Wu X, et al. Multi-modal foundation models for computational pathology: a survey. arXiv [csCV] 2025. 10.48550/arXiv.2503.09091. [DOI] [Google Scholar]
- 191. Lu MY, Chen B, Williamson DFK, et al. A multimodal generative AI copilot for human pathology. Nature 2024;634:466–73. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 192. Wenteler A, Occhetta M, Branson N, et al. PertEval-scFM: benchmarking single-cell foundation models for perturbation effect prediction. bioRxiv 2024. 10.1101/2024.10.02.616248. [DOI] [Google Scholar]
- 193. Ahlmann-Eltze C, Huber W, Anders S. Deep learning-based predictions of gene perturbation effects do not yet outperform simple linear methods. bioRxiv 2024. 10.1101/2024.09.16.613342. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
No new data were generated or analyzed in this review.

















