Skip to main content
Science Advances logoLink to Science Advances
. 2025 Aug 1;11(31):eadv0741. doi: 10.1126/sciadv.adv0741

Virtual staining of label-free tissue in imaging mass spectrometry

Yijie Zhang 1,2,3,, Luzhe Huang 1,2,3,, Nir Pillar 1,2,3, Yuzhu Li 1,2,3, Yuhang Li 1,2,3, Lukasz G Migas 4,5, Raf Van de Plas 4,5,6, Jeffrey M Spraggins 4,6,7,8,9, Aydogan Ozcan 1,2,3,10,*
PMCID: PMC12315973  PMID: 40749063

Abstract

Imaging mass spectrometry (IMS) enables untargeted, highly multiplexed mapping of molecular species in biological tissue with unparalleled chemical specificity and sensitivity. However, most IMS platforms lack microscopy-level spatial resolution and cellular morphological contrast, necessitating subsequent histochemical staining, microscopic imaging, and advanced image registration to correlate/link molecular distributions with specific tissue features and cell types. We present a diffusion model–based virtual histological staining approach that enhances spatial resolution and digitally introduces cellular morphological contrast into mass spectrometry images of label-free human tissue. Blind testing on human kidney tissue demonstrated that the virtually stained images of label-free samples closely match their histochemically stained counterparts (with periodic acid–Schiff staining), showing high concordance in identifying key renal pathology structures despite using IMS data with 10-fold larger pixel size. Additionally, our approach uses optimized noise sampling during the diffusion model’s inference to achieve reliable and repeatable virtual staining. We believe this virtual staining method will open avenues for IMS-based biomedical research.


Diffusion model–based staining enhances resolution and introduces cellular contrast into mass spectrometry images of tissue.

INTRODUCTION

Imaging mass spectrometry (IMS) (1) is a powerful tool for spatial biology, enabling the discovery of intricate relationships between molecular distributions and key tissue features and cell types (24). IMS offers in situ pixel-wise mapping of hundreds to thousands of molecular species with high chemical specificity and sensitivity. By combining mass spectrometric analysis with spatial mapping, IMS has great potential as a discovery platform for biomedical research. Although multiple surface sample approaches have been used for imaging, matrix-assisted laser desorption/ionization (MALDI) is one of the most common IMS technologies due to its applicability to a wide range of biomolecular classes and high–spatial resolution capabilities. Commercially available MALDI IMS platforms can routinely achieve 5- to 30-μm pixel sizes (57), and custom platforms have been developed to enable imaging at cellular resolution with pixel sizes approaching 1 μm (810). Briefly, MALDI IMS is performed by mounting a thin tissue section to a glass slide and coating it with a ultraviolet-absorbing chemical matrix, which assists with the desorption and ionization of endogenous molecules during laser irradiation. The sample stage is scanned from one location to another, recording a mass spectrum at every pixel. An intensity heatmap is then plotted across the measurement region, resulting in an ion image for each molecule detected. MALDI IMS spatial resolution is determined by several factors, including the laser diameter at the sample surface, the precision of the stage movement between the pixels (i.e., pitch), and the extent of molecular delocalization during the sample preparation (5, 11, 12).

While MALDI IMS provides broad molecular coverage and high chemical specificity, it lacks inherent histological context, making it challenging to directly link molecular profiles to precise cellular features without additional information. The primary reasons for this are (i) the relatively low spatial resolution of IMS and (ii) the absence of color and cellular morphological contrast that experts are familiar with. Both limitations can be addressed by multimodal methods integrating IMS data with optical microscopy, often requiring advanced experimental and computational methods. For example, IMS–microscopy integrative methods have been developed to enhance human interpretability of IMS data through applications such as spatial sharpening of IMS data (13), out-of-sample prediction of molecular distributions (14), mining of IMS-derived molecular profiles based on microscopy features (15), and others (1618). Each of these examples requires the collection of complimentary optical microscopy data such as histochemically stained (HS) bright-field or immunofluorescence images. However, in cases where chemical staining is necessary, it can introduce additional experimental complexity, posing a major limitation, as it precludes the preservation of the unlabeled tissue, thereby hindering any further molecular analysis on the same sample. Although serial tissue sections can be used, this requires complicated coregistration processes to account for tissue heterogeneity and cellular differences between adjacent sections. Therefore, a technique capable of predicting high-resolution cellular histopathology cues on the basis of low-resolution IMS data is urgently needed to streamline workflows and improve molecular histology.

Over the past decade, generative artificial intelligence (AI) models have made significant strides, finding wide-ranging applications for vision analysis. In the biomedical domain, one of the most notable uses of these models has been the virtual histological staining of label-free tissues, where generative models are trained to transform microscopic images of label-free tissues into their HS (1930) or fluorescence stained counterpart (3135). Most of these approaches rely on label-free imaging modalities that exhibit a spatial resolution comparable to the resolution of the images of the stained tissue captured by a digital histology scanner, which serves as the ground truth (GT) from the perspective of the clinical workflow. Recent efforts have also generated virtual histopathology images using segmentation masks (36), domain-specific knowledge, and tissue genomics (37).

In this work, we introduce a diffusion model–based virtual histological staining technique that transforms IMS-measured ion images reporting molecular species’ distributions in label-free tissue samples into super-resolved bright-field microscopy images, closely matching their HS counterparts [as illustrated in Fig. 1 (A and B)]. Our approach uses an image-conditional diffusion model, underpinned by the Brownian bridge process (Fig. 1C) (38, 39), which integrates the low-resolution conditional input with noise estimation through an attention-based U-Net (Fig. 1, F to H) (40) that incorporates the time-step information to reconstruct the high-resolution histological image of the label-free sample. Following a one-time training effort, the diffusion-based virtual staining (VS) model was able to generate bright-field microscopy equivalent periodic acid–Schiff (PAS)–stained images from IMS-measured ion images of label-free human kidney tissue samples (never seen before) despite the fact that the input IMS data had a 10-fold larger pixel size. Quantitative evaluations confirmed that our VS approach effectively overcomes key limitations of IMS for histological interpretation by digitally creating high-resolution histological stain images using low-resolution IMS data without the need for chemical staining and microscopic imaging of stained tissue, also eliminating image registration steps because the virtually stained images are automatically registered with the IMS data used as the label-free input of the VS model. A board-certified pathologist was able to identify key pathological structures directly from the virtually stained PAS images, demonstrating a high degree of concordance with the features observed in the corresponding HS images and emphasizing the human interpretability of this approach. Furthermore, to mitigate the inherent high variance associated with diffusion models, we used an optimized noise sampling strategy, which eliminates additive random noise in the final stages of the reverse diffusion process. This approach not only quantitatively reduces output variance but also ensures that the pathological features in the diffusion model outputs from different test runs remain consistent and histologically equivalent. In summary, our diffusion-based VS approach overcomes some of the limitations of traditional IMS, including its relatively low spatial resolution and the absence of cellular morphological contrast, also eliminating the need for time-intensive histological staining process and complex image coregistration after the IMS data acquisition. We believe that this technique offers significant benefits for mass spectrometry–based molecular histology and will help accelerate IMS-enabled histological analyses in life sciences.

Fig. 1. Diffusion model–based VS of label-free IMS-measured ion images.

Fig. 1.

(A) The workflow for histochemical staining of label-free tissue sections and the bright-field optical microscope scanning for digitization. (B) The diffusion model–based VS pipeline. The Brownian bridge process is used for both the forward and reverse processes. (C) Schematic diagram of the forward process of our Brownian bridge diffusion model (BBDM). (D and E) Schematic diagrams of the diffusion model reverse process that uses the vanilla and mean sampling strategies. (F) Detailed workflow for training the diffusion-based VS model. (G and H) Workflow for a single vanilla and mean sampling step, used individually or in combination during the reverse sampling process in (D) and (E).

RESULTS

Virtual histological staining of IMS-measured ion images of label-free tissue using a diffusion model

The dataset used in this study comprises IMS-measured ion images of label-free human kidney tissues and high-resolution bright-field images of the HS versions of the same tissue samples. As shown in Fig. 1B, the IMS data of each tissue sample were acquired using pixel-wise raster scanning with 10-μm lateral spacing. Subsequent data processing selected the most representative channels (41), resulting in individual ion images containing 1453 mass/charge ratio (m/z) channels, each with a pixel size of 10 μm. After the IMS data acquisition, the label-free tissue slides were subjected to histopathological PAS staining and digitally imaged using a benchtop bright-field optical microscope. The bright-field images of the HS tissue samples were then registered to their corresponding ion images, which formed our label-free input (IMS) and GT (bright-field and labeled) image pairs. It is worth noting that we also introduced autofluorescence (AF) microscopy imaging acquired both before and following the IMS data acquisition to facilitate the registration between the IMS data and the bright-field microscopy images of HS tissue. Details of the dataset collection and preprocessing are provided in Methods.

The training and sampling process of the Brownian bridge diffusion model (BBDM) (38) reflect the two-directional propagation of a Brownian bridge process, as depicted in Fig. 1 (C to E). The forward process starts from x0, representing the target image domain of the HS bright-field images (GT), and progresses toward xT, the input image domain, corresponding to lower-resolution IMS images (with 1453 m/z channels), processed through a shallow convolutional neural network for image dimension alignment/match, as shown in Fig. 1B. The mean of the forward process is linearly scheduled from x0 to xT, while the variance evolves quadratically over time steps (t). In contrast, the reverse process aims to denoise the input IMS images of label-free tissue step by step, gradually refining the data without directly using the GT image x0 . A U-Net–based denoising neural network (38, 40) is trained to estimate the posterior mean of xt based on xT. As shown in Fig. 1 (F to H), the denoising network is used to consistently estimate the difference between the current state xt and the target image x0 at arbitrary time steps between 0 and T (see Methods). This Brownian bridge framework departs fundamentally from the classic denoising diffusion probabilistic model (DDPM) (42). In unconditional DDPM (4244), the forward process degrades an initial sample x0 by adding Gaussian noise at each timestep according to a fixed variance schedule, ultimately producing a pure-noise state xT . A denoising neural network is then trained to invert this process: beginning from xT𝒩(0,I) , it iteratively predicts the noise component and partially removes it, gradually reconstructing x0 . When adapted for conditional image translation, the conditioning signal, whether a text embedding or an image, is concatenated with the noisy image at every forward and reverse step, ensuring that the model retains direct access to the context throughout the diffusion trajectory (4346). In contrast, the BBDM reframes the forward process so that the terminal state is the conditional image xT itself (the label-free IMS data). Noise is incrementally injected into x0 (HS, GT image) until it converges to xT, yielding intermediate states xt that stochastically interpolate between x0 and xT along a Brownian bridge. Consequently, in the learning stage, the denoising neural network learns to denoise these blended states. Crucially, for the Brownian bridge reverse process, the initial input is solely the conditional image (label-free IMS data). These forward and reverse processes of the BBDM framework result in more stable image translation outputs, as detailed in prior work (38).

To improve the consistency/repeatability of the VS process and minimize stochasticity in the diffusion process-generated images, we implemented a deterministic noise sampling strategy alongside the standard noise sampling method. Specifically, we used a mean sampling strategy (Fig. 1, E and H) that eliminates, after a certain time point is reached, the posterior noise introduced during the vanilla sampling process (Fig. 1, D and G). Detailed sampling algorithms for this mean sampling strategy are provided in Methods. The performance advantages of this deterministic sampling are further evaluated in the “Repeatable diffusion model inference via noise sampling engineering” section.

Following the training phase, the BBDM-based VS model was tested on human kidney tissue excluded from both the training and validation datasets. Figure 2 showcases the VS results generated by the BBDM-based model using IMS data from different patients. The virtually stained outputs, as presented in Fig. 2B, exhibit a good resemblance to the GT bright-field images, despite being generated from 10-fold larger pixel size IMS images, as illustrated in Fig. 2B. Furthermore, a board-certified pathologist annotated key renal structures—glomeruli (denoted as G), as well as proximal (P) and distal (D) convoluted tubules—on both the VS and HS images. These structures are of clinical importance because they are involved in most renal pathological conditions. As depicted in Fig. 2 (B and C), there was very good concordance in the identification of these structures between the VS and HS GT PAS images. This alignment was consistently demonstrated across multiple fields of view (FOVs) of tissue, underscoring the robustness and generalizability of our framework for the VS of low-resolution IMS data.

Fig. 2. Visual comparisons between the virtually stained PAS images generated from label-free IMS data of different patients and their HS counterparts.

Fig. 2.

(A) Imaging mass spectrometry data of label-free tissue, consisting of 1453 ion (m/z) channels with a pixel size of 10 μm. (B) Virtually stained images digitally generated from the IMS data using our diffusion-based VS model. (C) HS GT images. Both the virtually stained and the HS images have a pixel size of 1 μm. The concordant localization of glomeruli (G), proximal convoluted tubules (P), and distal convoluted tubules (D), as annotated by a board-certified pathologist, can be visualized on both the VS and HS images.

To quantitatively evaluate the fidelity of the virtually generated high-resolution PAS images in replicating their HS counterparts, we conducted a comparative analysis on a test dataset comprising 201 distinct tissue FOVs, each with 640 × 640 pixels, from 10 patients. This evaluation focused on three critical aspects: (i) image color distribution; (ii) image contrast (47); and (iii) spatial frequency spectrum (detailed in Methods). These quantitative metrics reported in Fig. 3 were selected to evaluate whether the VS images generated from ion images can meaningfully contribute to the interpretation of IMS-measured tissue samples without physically performing the chemical staining and microscopic imaging of stained tissue and, thus, effectively safeguarding the tissue for other assay types and subsequent analysis. While our approach requires multimodal image registration between IMS data and microscopy images during the training phase, which is a one-time effort, this eliminates cumbersome image registration steps during the inference phase. During image inference, the VS images are inherently aligned with their input IMS data, removing the need for additional registration. As depicted in Fig. 3A, the color distribution histograms in the YCbCr color space across all the test FOVs demonstrate strong color agreement between the VS images and the HS bright-field images. This agreement is further supported by the small Hellinger distances (48, 49) calculated between the color histograms of each VS and HS image pair. Moreover, as illustrated in Fig. 3C, the generated VS images achieve image contrast comparable to that of the bright-field HS images.

Fig. 3. Quantitative comparisons between virtually stained PAS images generated from label-free IMS data and their HS counterparts.

Fig. 3.

(A) The color histogram comparisons of virtually stained PAS images and their HS counterparts in Y, Cb, and Cr channels, calculated separately. (B) The Hellinger distances between the color histograms of VS and HS image pairs in Y, Cb, and Cr channels. (C) The box plots of image contrast of virtually stained PAS images and their HS counterparts across a test dataset comprising 201 distinct FOVs. A statistically significant improvement of the image contrast values of the VS images over the HS images was determined by a two-tailed paired t test (P = 2.60 × 10−15). (D) The low-resolution single MS channel image, the corresponding high-resolution grayscale virtually and HS images (top left), and their respective spatial frequency spectra (amplitude, displayed in bottom left). The radially averaged normalized power spectrum [log scale, arbitrary unit (a.u.)] cross section for each case is also presented (right).

Furthermore, to showcase the super-resolution capabilities of our VS framework, we conducted a spatial frequency spectrum analysis on the raw single-channel MS images (picked from 1453 m/z channels based on the contrast of glomeruli), network-generated VS images, and their corresponding HS GT images; see Methods for details. This analysis, illustrated in Fig. 3D, includes cross sections of the radially averaged power spectra (50, 51), which demonstrate an excellent match between the spatial frequency spectra of the VS and HS image pairs, as desired. The relative error between the radially averaged power spectra (log scale, normalized) of the VS and the corresponding HS GT images is 6.70 ± 4.84% (means ± SD) across all the frequencies (blue and red curves in Fig. 3D). These results further confirm the effectiveness of the VS output images, which successfully align with the spatial frequency spectra of the high-resolution HS images. Additionally, the relative increase of the radially averaged power spectra (log scale, normalized) of the VS images compared to their lower resolution MS input is 144.23 ± 81.19% (means ± SD) over all the frequencies (blue and green curves in Fig. 3D). These results demonstrate a marked improvement over the spatial frequency spectra of low-resolution MS images, underscoring the diffusion-based VS framework’s capacity to enhance spatial resolution significantly.

To further illustrate the effectiveness of the spatial frequency analysis in quantifying the agreement between the virtually stained images generated by our framework and their corresponding paired HS images, we conducted an additional comparison involving two unpaired HS image FOVs with the paired VS/HS images presented in Fig. 3. As demonstrated in fig. S1, there is a clear discrepancy in the radially averaged power spectra between the unpaired HS images and the paired VS/HS images. This result further underscores the effectiveness and specificity of our VS framework in accurately enhancing spatial frequency features to closely match the corresponding HS GT image. Similarly, to validate the significance of the color distribution agreement quantified using the Hellinger distance, we performed an additional comparative analysis involving unpaired HS images and their virtually stained counterparts. Figure S2 presents the calculated Hellinger distances between the color histograms of the paired VS/HS images and those of unpaired VS/HS image pairs across the Y, Cb, and Cr channels. As illustrated in this figure, there is a statistically significant increase in the Hellinger distances for the unpaired VS/HS images compared to the paired VS/HS images, reinforcing the specificity of our VS method in reproducing the color characteristics inherent in histochemical staining.

Reduction analysis on mass spectrometry image channels

The high-resolution image translation capability of VS for IMS data relies on the rich molecular information captured in ion images. To further shed light on this, we trained a series of diffusion-based VS models using different numbers of mass spectrometry (m/z) channels and evaluated their performance on the same test dataset. For the channel selection, we evaluated three distinct strategies. In our primary approach, the IMS channel indices were selected from the top-ranking channels of the sorted list of 1453 channels, prioritized on the basis of signal-to-noise ratio (SNR) values (in descending order). The SNR definition used in this study entails that, for each IMS channel, the mean is divided by the SD across all pixels within that channel. In addition to this SNR-based ranking and channel selection, we explored two alternative channel selection strategies: frequency-based selection and uniform selection. For the frequency-based selection method, we applied a two-dimensional (2D) Fourier transform to the ion image of each channel, ranking the channels by the ratio of the average power in the angular spectrum of the high-frequency components to that of the low-frequency components. We set the half-maximum frequency as the threshold to differentiate between high and low frequencies. For the uniform selection strategy, we sampled the channels at fixed intervals (every 4th, 16th, or 64th channel for reductions of 4-fold, 16-fold, and 64-fold, respectively), starting from the first channel. Using these three distinct selection approaches, we selected channels from the full set of 1453 IMS channels, progressively reducing the total number of channels used per input image from 363 down to 23, corresponding to 4-fold, 16-fold, and 64-fold reductions. These selected IMS channels remained consistent throughout the training and testing of each VS model for a given channel selection strategy.

As shown in Fig. 4A, reducing the number of MS channels from 1453 to 23 resulted in a gradual degradation of the VS performance for all three channel selection strategies, with noticeable losses in critical features such as nuclear morphology. The relationship between the number of IMS channels and VS quality was further quantified using the peak SNR (PSNR) and learned perceptual image patch similarity (LPIPS) (52) and Pearson correlation coefficient (PCC). Figure 4B presents the PSNR, LPIPS, and PCCs metrics for the 10 VS models, each trained with a combination of a different number of IMS channels and channel selection approaches, evaluated across a test dataset comprising 201 distinct tissue FOVs from 10 patients, each with 640 × 640 pixels. These results confirmed that as the number of used MS channels increased, the diffusion-based VS models achieved a statistically significantly higher VS fidelity, underscoring the rich molecular information present in MS data and its strong utility for label-free histological staining. Among the three distinct channel selection strategies, the SNR-based selection consistently delivered the best performance compared to the other two channel sampling approaches, further validating its effectiveness in preserving the critical molecular information for high-quality VS with a reduced number of IMS channels.

Fig. 4. Comparison of VS performance for diffusion models trained with different numbers of IMS channels and channel selection strategies.

Fig. 4.

(A) Visual comparisons of virtually stained PAS images generated from diffusion-based VS models trained using all the IMS channels and three reduced channel sets obtained through three different selection strategies (resulting in nine configurations in total). (B) Bar plots displaying the quantitative PSNR, LPIPS, and PCC metrics averaged across testing virtually stained images for all the configurations. Higher PSNR, PCC, and lower LPIPS values are desired.

Repeatable diffusion model inference via noise sampling engineering

Tissue heterogeneity and inherent histochemical staining variability often lead to minor differences between adjacent tissue sections. These spatial changes usually do not influence the overall slide-level diagnosis and are well tolerated by human pathologists in their clinical workflow. However, computational models applied for tissue analysis are frequently misled by these slide-to-slide variations, resulting in lower performance. For our VS model, the primary source of variation in the output VS results comes from the stochastic nature of the noise sampling process during the backward diffusion. To address this, we applied noise sampling process engineering techniques to improve VS consistency and reduce output variance without the need for fine-tuning or transfer learning on the trained model. In addition to the mean sampling strategy illustrated in Fig. 1, we also introduced an alternative “skip sampling” strategy for comparison, which directly estimates x0 from the denoising network’s output.

The rationale behind the mean and skip sampling strategies arises from the significant increase in the additional variance δ~t during the final stages of the reverse diffusion process (see fig. S3). Both of these sampling strategies, mean and skip, avoid this increasing noise variance in the reverse path after an engineered exit point te (see Methods); this effectively reduces stochastic variations (observed from run to run) in the output of the diffusion model for the same label-free tissue FOV, which is highly desired for VS applications. Detailed descriptions of these engineered noise sampling techniques are provided in Methods and figs. S3 to S5.

To quantitatively assess the effectiveness of these noise sampling strategies, we tested the trained BBDM using the vanilla, mean, and skip sampling strategies, repeating each method five times and calculating the pixel-wise coefficient of variance (CV) across these repetitions of the diffusion-based VS process. Figure 5A shows the CV maps for the three distinct strategies in the YCbCr color channels, demonstrating that both the mean and skip sampling strategies effectively reduce the variance in the sampled VS images compared to the vanilla method. Additionally, we computed the average CV values across all the pixels in the test image FOVs and plotted them for each YCbCr channel, as shown in Fig. 5B. Our results further corroborate that the mean and skip sampling strategies are effective in achieving lower output variances, indicating the repeatability of the diffusion-based VS process using these engineered noise sampling techniques.

Fig. 5. Comparison of the CV for different noise sampling engineering approaches.

Fig. 5.

(A) Visualization of the coefficient of variation (CV) maps for the YCbCr channels of virtually stained images obtained using three different approaches: vanilla, mean, and skip sampling approaches. The CV was calculated using five repetitions of VS inference on the same FOV. (B) Plots of the mean CV calculated across all pixels of all test image FOVs for the three sampling approaches.

This comparative analysis in Fig. 5 further reveals that the skip sampling strategy yields a lower average CV value compared to the mean sampling method. However, the mean sampling strategy produces results that exhibit better perceptual similarity to the GT HS PAS images. Additional visual and quantitative comparisons between different noise sampling strategies in diffusion-based VS models are presented in fig. S4. These results confirm that the mean sampling strategy is superior to both the vanilla and skip sampling strategies, achieving a lower average LPIPS as desired. Additionally, we evaluated the performance of an averaging strategy, which involves averaging independent test runs of different inferences for the same tissue FOV. While fig. S4C highlights the advantage of the averaging strategy in reducing pixel-level CV, the resulting images exhibit relatively lower contrast with a significantly worse LPIPS score, falling short of the performance achieved by the mean sampling strategy alone. Therefore, this averaging strategy would only be preferable in scenarios where high consistency is prioritized over image contrast.

The noise sampling strategies demonstrated here can be further optimized. For instance, we evaluated the performance of the mean and skip sampling strategies across eight different exit points ( te ), ranging from 0 (equivalent to the vanilla sampling strategy) to 100, as illustrated in fig. S5. Our findings indicate that both the mean and skip sampling strategies consistently outperformed the vanilla strategy across a range of exit points. Specifically, both strategies achieved optimal LPIPS performance at an exit time point te ~ 10. Consequently, selecting a well-suited evaluation metric and determining the appropriate exit point are critical for optimizing the performance of diffusion-based VS models, particularly for clinical applications.

DISCUSSION

The effectiveness of the presented label-free VS of IMS-measured ion images, despite the 10-fold larger pixel size of the input IMS data, can be attributed to the capabilities of diffusion models in effectively capturing and modeling complex data distributions (42, 43, 53). Historically, generative adversarial network (GAN)–based approaches (5456) have been a predominant choice in image restoration for biomedical applications, particularly in super-resolution image reconstruction tasks (51, 57, 58). However, recent advancements (59, 60) in the field have revealed that GANs might struggle with highly challenging image reconstruction problems, especially at extreme super-resolution factors. In contrast, diffusion models have emerged as a superior alternative, consistently producing more realistic and accurate spatial features even at these high super-resolution factors. Moreover, the complexity of multiplexed inference tasks that require simultaneous resolution enhancement and cross-domain image translation has further underscored the limitations of GAN-based techniques. It has been demonstrated that diffusion models outperform GANs in such multiplexed tasks (30, 61), owing to their robustness in generating high-fidelity images. Additionally, the issue of mode collapse (62), which is a well-documented limitation of GANs when confronted with sparse or low-quality data, is well mitigated in diffusion models. Diffusion models, in general, exhibit more stable training dynamics, even when faced with significant discrepancies between the input data and GT images, making them more resilient in handling challenging datasets (63). For the same super-resolution VS task presented in this work, we also evaluated the cGAN framework (29); however, the model collapsed at an early stage and failed to converge effectively, resulting in images containing unacceptable staining artifacts. Taking these factors into account, we can conclude that the success of our VS models in generating high-resolution virtual stains from low-resolution IMS data of label-free tissue samples is a testament to the diffusion models’ inherent strengths.

The repeatability and consistency of the IMS-generated virtually stained images are crucial for digital pathology interpretation and were achieved in this work using the mean and skip sampling strategies used during the reverse process of the diffusion model. Another notable achievement of this study is its ability to enable medical experts to directly identify diagnostically relevant renal pathology structures—such as glomeruli, proximal, and distal convoluted tubules—using virtually stained images generated from lower-resolution IMS images. Historically, recognizing these structures directly from IMS images was not feasible. Conventional histology-directed IMS analysis requires both IMS data and bright-field microscopy images of the HS tissue, followed by a complex and time-consuming registration process to link the IMS data with pathology-annotated regions in optical microscopy images of stained tissue. Furthermore, histochemical staining of IMS slides prevents their utilization for genomics/epigenetics analyses and hinders IMS-molecular comparison studies. Our diffusion-based VS framework, however, enables us to bypass these limitations by offering a means of histological interpretation of IMS data for regions of interest within the kidney tissue, immediately after an IMS scan, without the need for histochemical staining. Moreover, once the diffusion model is trained, this technique can be seamlessly integrated into the post-IMS data processing pipeline without requiring any modifications to the existing hardware/setup.

It is also important to note that the quality of IMS-based VS can be further enhanced. In this study, we demonstrated the virtual histological staining of ion images with a pixel size of ~10 μm, as this is a common spatial resolution for MALDI IMS systems. However, with recent advances in IMS technology, achieving higher spatial resolution is now possible (810). For instance, transition mode ion sources allow IMS to achieve 1-μm pixel sizes. Unfortunately, these specialized workflows often inflict significant tissue damage, as the tissue is ablated during the sampling process, making the post-IMS multimodal analysis impossible. By applying our VS framework to these high-resolution IMS images, we not only overcome the post-IMS multimodal analysis challenge of tissue damage but also enhance the fidelity of the virtually stained images, further elevating the precision and quality of the diffusion-model results. Furthermore, although we demonstrated the efficacy of our technique using PAS staining on human kidney samples, this label-free approach can be extended to other types of histochemical stains and various organs. This adaptability is supported by the prior success of various VS techniques (19), making it versatile across different staining protocols and tissue types.

We believe that this advancement has the potential to drive further biomedical research, particularly in the study of glomerular and tubular diseases through IMS. Furthermore, this method can augment existing IMS datasets by generating virtually stained images, offering advantages to the broader biomedical research community. In addition, our reduction analyses on IMS channels further confirm the rich information encoded in IMS data and can potentially reveal the relationship of IMS channels with histochemical staining, which might prospectively facilitate a better understanding of both the virtual and histochemical staining processes. We demonstrated that our VS method remains effective even when using a reduced number of IMS channels, as selected via a SNR-based reduction strategy. This provides several practical advantages. First, computational load and hardware requirements for both model training and inference scale with the number of input channels, thus reducing the channel count enhances the method’s accessibility and efficiency. Second, our results suggest that this SNR-based feature reduction makes the method more compatible with high–spatial resolution IMS experiments, such as those performed using MALDI IMS. One of the key limitations to achieving high spatial resolution in MALDI IMS is the trade-off between spatial resolution and molecular sensitivity. As the laser beam diameter is reduced (e.g., via decreased fluence, improved focus, or oversampling), the sampled tissue area—and hence the amount of ablated material—also decreases. This results in lower signal intensities and the potential loss of low-SNR molecules, as they may fall below the detection threshold. In such scenarios, only high-intensity, high-SNR molecules remain detectable. Our findings suggest that, even under these conditions, the VS method could remain viable and effective, further underscoring its applicability.

Study limitations

In our study, we identified a few mismatched areas where our diffusion-based model did not accurately reconstruct uncommon structures, such as an occasional large collecting duct, as illustrated in fig. S7 (A and B). This limitation arises because our training and testing datasets were composed of kidney samples obtained from uninvolved cortical areas during total nephrectomy procedures, which typically lack hypocellular regions. Consequently, the model was not effectively trained to reconstruct such sparse areas, leading to hallucinations in such regions. Expanding the training dataset to include diverse tissue morphologies with pathological changes in future studies could alleviate this issue. Additionally, fig. S7 (C and D) highlights another representative FOV, where both virtually stained and HS images exhibit staining artifacts, particularly within the glomerular regions. These artifacts can be attributed to tissue damage that occurred during tissue processing and the IMS imaging pipeline.

As shown in Fig. 2, the virtually stained images sometimes exhibit minor variations with respect to the HS GT images at the cellular level. These differences primarily arise from the inherently lower spatial resolution of the input IMS data, which have an ~10-fold larger pixel size relative to bright-field microscopy, GT images used in this work. However, such discrepancies generally do not affect diagnostic outcomes at the structure level, as pathologists typically examine multiple serial sections, each of which naturally exhibits minor variations in cellular or tissue morphology due to tissue heterogeneity. Consequently, even with slight misalignments in the virtually stained images compared to the histochemical stains, pathologists and researchers can potentially recognize the tissue features and make accurate diagnoses at the structure level.

However, we also acknowledge that our current VS approach may face limitations in clinical scenarios where precise morphological details, such as the exact shape and geometry of glomeruli, play a critical role in diagnosis (64). Additionally, certain single-cell features, particularly rare but diagnostically significant, such as immune cells within the glomeruli, could potentially be obscured or misrepresented by the VS model, due to the necessity of inferring missing spatial information from lower-resolution molecular inputs. To effectively address these limitations in diseases requiring meticulous cellular and geometric detail, using higher-resolution IMS systems with pixel sizes as small as ~1 μm would likely yield substantially improved VS results, as previously discussed in our manuscript.

It is also worth noting that our approach is not intended for primary diagnosis in clinical settings. First, IMS is not a US Food and Drug Administration (FDA)–approved imaging modality for digital pathology. In current digital pathology workflow, the standard FDA-approved whole-slide scanners typically offer much faster imaging speeds and higher spatial resolution compared to IMS systems. Consequently, our IMS-based VS approach is not practical for widespread clinical diagnostic applications in digital pathology workflow but remains a valuable tool for IMS-focused biomedical research. We believe that our IMS-based label-free VS framework will become an essential tool for IMS-driven molecular pathology research, effectively bridging the gap between fundamental IMS investigations and potential clinical diagnostic applications.

METHODS

Sample preparation

Human kidney tissue was surgically removed during a full nephrectomy, and remnant tissue was processed for research purposes by the Cooperative Human Tissue Network at Vanderbilt University Medical Center. Remnant biospecimens were collected in compliance with the Cooperative Human Tissue Network standard protocols and the National Cancer Institute’s Best Practices for the procurement of remnant surgical research material. Participants were consented for remnant tissue collection in accordance with Institutional Review Board (IRB) policies. The study received ethical approval from Vanderbilt University’s IRB (no. 210190). The study involved kidney specimens from five individual patients. Kidney tissue was flash frozen over an isopentane-dry ice slurry, embedded in carboxymethylcellulose, and stored at −80°C. The tissue was cryosectioned into 10-μm-thick sections using a CM3050 S cryostat (Leica Biosystems, Wetzlar, Germany). The sections were then thaw mounted onto indium tin oxide–coated glass slides (Delta Technologies, Loveland, CO) for IMS analysis or regular glass slides for histological staining. Slides were stored at −80°C and returned to ~20°C within a vacuum desiccator before further processing. To remove endogenous salt for IMS, the section was washed three times with chilled (4°C) 150 mM ammonium formate three times for 45 s each. It was then dried with nitrogen gas to remove excess moisture.

AF microscopy and histochemical staining

To enable coregistration of IMS data with HS images, we introduced AF images as an intermediate modality, to which both IMS and HS images can be registered, as shown in fig. S6A (also see details in the “Multimodal image registration” section). AF microscopy images were acquired on each tissue before IMS analysis using 4′,6-diamidino-2-phenylindole, enhanced green fluorescent protein, and DsRed filters on a Zeiss AxioScan.Z1 slide scanner (Carl Zeiss Microscopy GmbH, Oberkochen, Germany). Additionally, AF was also collected after IMS data acquisition before matrix removal, enabling visualization of ablation marks created by the MALDI laser (16). The resulting images have a pixel size of 0.65 μm.

It is worth noting that acquiring AF microscopy images after the IMS data acquisition but before the chemical matrix is removed offers a clear visualization of the ablation craters for each IMS pixel in the microscopy coordinate space, thereby facilitating precise IMS-microscopy registration. During the MALDI process, the tissue surface, including its chemical matrix layer, is irradiated with a laser that causes the ablation and desorption of material and the ionization of endogenous molecules for mass spectrometry detection. Because the laser is focused on a diameter slightly smaller than that of an IMS pixel to avoid oversampling, this results in an array of distinct ablation marks on the matrix-coated tissue surface. In most experiments, only the chemical matrix layer is ablated during MALDI, while the underlying tissue remains intact.

After the acquisition of the post-IMS AF images, the same unlabeled tissue was stained using a standard PAS staining protocol (65). The PAS staining process takes 30 to 60 min for each batch of processed slides. The stained tissue slides were scanned and digitized using a bright-field slide scanner (Leica Biosystems Aperio AT2). The resulting PAS-stained bright-field microscopy images have a pixel size of 0.22 μm.

Imaging mass spectrometry

Samples for IMS analysis were coated with solution (20 mg/ml) of DAN dissolved in tetrahydrofuran using a TM Sprayer M3 (HTX Technologies LLC, Chapel Hill, NC, USA), yielding a coating of 1.67 mg/cm2 (0.05 ml/hour, four passes, 40°C spray nozzle). MALDI IMS was performed on a prototype timsTOF fleX mass spectrometer (Bruker Daltonik, Bremen, Germany). The ion images were collected in negative ionization mode at 10-μm pixel size with a beam scan set to 6 μm, using 150 laser shots per pixel and 18.6% laser power (30% global attenuator and 62% local laser power) at 10 Hz. Data were acquired in negative ionization quadrupole Time-of-Flight (qTOF) mode, covering an m/z range from 150 to 2000. The acquisition speed of IMS is ~5 hours per slide.

Multimodal image registration

The multimodal registration pipeline is illustrated in fig. S6B. Due to the 10-fold difference in the pixel size and inherent differences in the imaging modalities, direct registration of the HS images with IMS data is not feasible. To address this challenge, we introduced prior-IMS and post-IMS AF images to facilitate the registration process. The pipeline consists of three sequential image registration steps: (i) the first step aligns the HS images with the prior-IMS AF images; (ii) next, the prior-IMS AF images are registered to the post-IMS AF images; and (iii) lastly, the IMS data are registered to the post-IMS AF images. Notably, the third step leverages the laser ablation craters left on the chemical matrix after the IMS acquisition to guide image registration. However, this ablation process causes localized feature loss in the post-IMS AF images, making them unsuitable for direct alignment with the HS images. Therefore, the prior-IMS AF images serve as an intermediary, as they share the same modality as the post-IMS AF images and retain structural features needed for robust image registration. The microscopy modalities were coregistered using the elastix (66) framework integrated into wsireg (67) software. The post-IMS AF was selected as the target modality to enable integration with IMS, with the PAS-stained bright-field microscopy images becoming the source modalities. The rigid and affine transformations were used because the microscopy modalities were collected on the same tissue section. All registered whole-slide images (WSIs) were stored in the vendor-neutral pyramidal Open Microscopy Environment Tagged Image File Format (OME-TIFF) at a common pixel size (i.e., the PAS image was resampled to the resolution of the prior-IMS AF and post-IMS AF). Furthermore, the MALDI IMS datasets were manually registered to the post-IMS AF images using the laser ablation marks and IMS pixels. The manual registration was performed using IMS Microlink software (68), where 8 to 12 fiducial markers were selected in both modalities to estimate the affine transformation.

Data division and preparation

The collected IMS data were exported from the Bruker timsTOF file format (.d) to a custom binary format for ease of access and improved performance. Each pixel/frame contains between 104 and 105 centroid peaks covering the entire acquisition range, which can be reconstructed into a pseudoprofile mass spectrum using Bruker’s SDK (v2.21). The dataset was m/z-aligned using six internally identified peaks (appearing in at least 50% of the pixels) through the msalign library (v0.2.0). This step corrects spectral misalignment (drift along the m/z axis), resulting in increased overlap between spectral features (peaks) across the experiment. Subsequently, the mass axis of the dataset was calibrated using the theoretical masses of the six peaks, achieving a precision of ~±1 ppm. Following the preprocessing steps, normalization correction factors were computed, and a total ion current approach was used for mass spectral and ion image normalization. Subsequently, an average mass spectrum based on all pixels was calculated for each dataset. Because samples from multiple donors were used in this study, an average mass spectrum of all samples was generated and peak-picked, identifying 1453 that were used for further analysis. The resulting average spectrum had a resolving power of ~40,000 at m/z 885.55.

To better match the dimensions of the IMS data and facilitate the transformation from IMS to HS images, we downsampled the HS images to achieve a pixel size of 1 μm, which is 10 times smaller than the pixel size of IMS. The WSIs of IMS and HS images, acquired from four patients, were then segmented into smaller FOV pairs of ~1400 μm by 1400 μm (each IMS image with 140 × 140 pixels and each HS image with 1400 × 1400 pixels), with 10% overlap between neighboring regions. To augment the data for robust model training, the paired WSIs were spatially transformed using a combination of rotation and flipping. The transformed WSIs were segmented as described above. This process generated a training dataset containing 712 paired IMS-PAS microscopic image patches obtained from four deidentified patients. Additionally, 201 paired IMS-PAS microscopic image patches without data augmentation (each FOV with 64 × 64 pixels for IMS and 640 × 640 pixels for the HS image) were reserved for blind testing, obtained from 10 deidentified patient not included in the training set. During each training epoch, the paired image FOVs were further subdivided.

Brownian bridge diffusion process

The processed dataset provided input-output image pairs, where x0H×W×3 corresponds to HS image, and y0H10×W10×1453 represents the label-free IMS-measured ion images of the same tissue FOV. Before training the diffusion model, the IMS dataset y0 was preprocessed by a shallow neural network gs , as shown in Fig. 1B. This shallow network consisted of two convolutional layers followed by a pixel-shuffle (69) layer, which upscaled the image dimensions to match the target HS image, i.e., the corresponding GT image. Specifically, the processed output yH×W×3 has the same dimension as the HS target image. This shallow network can be denoted as

y=gs(y0) (1)

Subsequently, the processed label-free image y and the corresponding HS image x0 were used as the starting and ending points of the Brownian bridge, respectively, thereby forming an image-conditional diffusion process. In this Brownian bridge diffusion process, y acts as the image condition input, while x0 represents the target image. The forward diffusion process of the Brownian bridge with an initial state x0 and terminal state y is defined as

q(xtx0,xT=y)=𝒩(1at)x0+aty,δtI (2)

where at and the variance δt can be calculated as

at=tTδt=2st(Tt)T2 (3)

Here, t refers to the intermediate time step and T is the total sampling steps. T was set to 1000 during both the forward and reverse processes. Factor s here is used to control the sampling diversity and is set to 1 in the forward process. During the forward sampling process, the intermediate state xt can be derived from Eq. 2 as

xt=x0+at(yx0)+δtϵt,ϵt𝒩(0,I) (4)

As shown in Eq. 4, the target image x0 can be recovered from any intermediate state xt by removing the add-on term at(yx0)+δtϵt . Therefore, a denoising network ϵθ was also trained as part of the BBDM to estimate this add-on term to recover the target image x0 . Because the term at(yx0)+δtϵt is a function of time step t , the denoising network takes both the noisy image xt and the time step t as input. This network is built upon a U-Net (40) architecture, as shown in Fig. 6B. The U-Net structure features two connected parts: a downsampling path and an upsampling path, with a central layer and skip connections linking them. The downsampling path comprises four levels, each consisting of a residual block followed by an attention block, and adjacent levels are connected via 2 × 2 average pooling. In parallel, the upsampling path mirrors this structure but with 2 × 2 nearest-neighbor interpolation connecting the levels. At the core of the U-Net, the middle block incorporates two residual blocks and one attention block, using multihead attention (70) for enhanced feature processing. Additionally, the time step t is embedded through a linear layer with a SiLU (71, 72) activation function and integrated into the residual blocks as an added feature.

Fig. 6. The network architecture for the diffusion-based VS model.

Fig. 6.

(A) The pipeline of the forward and reverse sampling processes. The detailed architecture of the shallow convolutional neural network used for dimension reduction is also presented. (B) Detailed architecture of the denoising network used at each step of both the forward and reverse sampling processes.

The loss function of the denoising network ϵθ is defined as

L=ΣtγtE(x0,y),ϵtat(yx0)+δtϵtϵθ(xt,t)22 (5)

where γt is the weight for each t . During the training, γt was empirically set as 1 for all the sampling steps, t.

During the blind testing, the reverse sampling process of BBDM starts from the conditional label-free input, xT=y=gs(y0) . Based on the main idea of the denoising diffusion method (42), the reverse process aims to estimate xt1 from xt . Hence, the vanilla reverse process can be defined as

p(xt1xt,y)=𝒩xt1;μt(xt,y),δtI (6)
μt(xt,y)=cxtxt+cytycϵtϵθ(xt,t) (7)
δ~t=δtt1δt1δt (8)

where μt(xt,y) is the estimated mean relying on the trained denoising network, while the δ~t is the additional variance added during the sampling process. The expressions of cxt,cyt,cϵt , and δtt1 can be formulated as

cxt=δt1δt(1at)(1at1)+δtt1δt(1at1) (9)
cyt=at1at(1at)(1at1)δt1δt (10)
cϵt=(1at1)δtt1δt (11)
δtt1=δtδt1(1at)2(1at1)2 (12)

During the blind testing phase, the reverse process of the diffusion model predicts the target image x0 starting from xT=y. The vanilla sampling strategy operates as a Markov chain with learned Gaussian transitions, initiating from p(xT)=𝒩(xT;y,0) . This process can be expressed as

pθv=p(xT)t=1Tp(xt1xt,y) (13)

Each transition probability p(xt1xt,y) estimates the intermediate state xt1 from the previous state xt and can be formulated as

xt1=cxtxt+cytycϵtϵθ(xt,t)+δtz,z𝒩(0,I) (14)

By iterating Eq. 14, each intermediate state is sequentially estimated, starting from xT and continuing until the final target image x0 is obtained, revealing the virtually stained PAS image of the label-free test tissue. This iterative process is referred to as the vanilla sampling strategy and is depicted in Fig. 1 (D and G).

For the mean sampling strategy, shown in Fig. 1 (E and H), the procedure initially mirrors the vanilla sampling approach up until the time step te<T , where te is defined as the exit point in the reverse sampling process. Once the process reaches tte , this strategy estimates xt1 from xt using a mean sampling step as follows

pm(xt1xt,y)=𝒩xt1;μt(xt,y),0 (15)

where the estimation of xt1 relies solely on μt(xt,y) , without incorporating additional variance δ~t , i.e., xt1=cxtxt+cytycϵtϵθ(xt,t) for tte . Thus, the mean sampling strategy can be formulated as

pθm=p(xT)t=1tepm(xt1xt,y)t=teTp(xt1xt,y) (16)

The skip sampling strategy follows the same vanilla sampling steps up to an exit point t=te , i.e.

pθs=p(xT)t=teTp(xt1xt,y) (17)

However, instead of continuing the iterative inference process, this skip sampling method estimates the final virtual stained image directly from the state at xte

x0s=xteϵθ(xte,te) (18)

During the training phase, both the shallow convolutional network and the denoising network were trained jointly. The parameters of these networks were optimized using the AdamW optimizer (73) with an initial learning rate of 1×104 . The training process used a batch size of 16 and took ~72 hours to converge. For blind testing, the exit point, te was empirically set to 10 , where both the mean and skip sampling strategies achieved better performance, as illustrated in fig. S5.

Quantitative performance evaluation metrics

To quantitatively evaluate the performance of PAS VS results for the study reported in Figs. 3 and 4, we used 201 FOVs of virtually stained PAS images together with their corresponding HS images for paired image comparisons. In the quantitative evaluation illustrated in Fig. 3, we conducted a comprehensive comparative analysis using several features: image contrast, histogram distributions in the YCbCr color space, Hellinger distance calculated between the color histograms of the VS and HS image pairs, and spatial frequency spectrum analysis. The image contrast is defined as

Contrast=A90%A10%A90%+A10% (19)

where A90% and A10% represents the 90th percentile and 10th percentile of the intensity values in image A , respectively. In our case, image A refers to the grayscale version of the VS or HS image. For the color analysis, the paired VS and HS images were converted from RGB to YCbCr color space to facilitate a detailed comparison of the distributions/histograms in the Y, Cb, and Cr channels, performed separately. The Hellinger distances were calculated between the YCbCr color histograms for each VS and GT image pair as follows

Hellinger distance[p(x),q(x)]=1xXp(x)q(x) (20)

where p(x) and q(x) are the probability distributions of the YCbCr color (approximated using the extracted color distribution points). The Hellinger distance ranges between 0 (identical) and 1 (completely different). As for the spatial frequency spectrum analysis, the single channel raw MS image was bilinearly upsampled by a factor of 10, from 64 × 64 pixels to 640 × 640 pixels, matching the dimensions of the grayscale VS and HS images. The frequency spectrum of each image was obtained by performing a 2D Fourier Transform on the 10× bilinearly upsampled single-channel IMS image (selected on the basis of the contrast of glomeruli), the VS output and the corresponding HS GT image. For this analysis, both the VS and HS images were processed in gray scale. The radially averaged power spectrum was calculated according to Wang et al. (57) Moreover, the relative error between the radially averaged power spectra (log scale, normalized) of the VS image and their HS GT (represented by blue and red curves in Fig. 3D) is defined as

Relativeerr=(PVSPHS)PHS·100% (21)

where PVS and PHS denote the radially averaged power spectra (log scale, normalized) of the VS image and its HS counterpart. Similarly, the relative gain of the radially averaged power spectra (log scale, normalized) of the VS images compared to their lower resolution MS input (blue and green curves in Fig. 3D) is defined as

Relative gain=(PVSPMS)PMS·100% (22)

where PVS and PMS represent the radially averaged power spectra (log scale, normalized) of the VS image and the corresponding single-channel MS input.

For the evaluations reported in Fig. 4, we used the metrics of PSNR, PCC, and LPIPS. The PSNR is defined on the basis of mean squared error (MSE)

MSE=1MNmMnN[AmnBmn]2 (23)

where A and B present the histochemically and virtually stained bright-field PAS images, respectively. m and n are the pixel indices, and MN denotes the total number of pixels in each image. PSNR can be denoted as

PSNR=10log10(max(A)2MSE) (24)

where max(A) is the maximum pixel value of the GT HS PAS image.

The PCC was used to quantify the linear correlation between the HS PAS images and the corresponding VS images. PCC is calculated as

PCC=mMnN(AmnA¯)(BmnB¯)mMnN(AmnA¯)2mMnN(BmnB¯)2 (25)

where A¯ and B¯ represent the mean pixel value of A and B.

The calculation of the LPIPS metrics used a pretrained VGG network (74) to evaluate the learned perceptual similarity between the generated VS images n and their corresponding HS images n0 . These compared image pairs were fed into the pretrained VGG network and their feature stack from L layers can be extracted as m^l,m^0lHl×Wl×Cl for layer l . The LPIPS score can be denoted as

d(n,n0)=l1HlWlh,w(m^hwlm^0hwl)22 (26)

The Frechet inception distance [FID; (75)] values, reported in fig. S5, are calculated as follows

FID=μμω2+tr(Σ+Σω2(ΣΣω)12) (27)

where N(μ,Σ) represents the multivariate normal distribution estimated from the Inception v3 features (75, 76) of the GT PAS-stained images and N(μω,Σω) represents the distribution estimated from the Inception v3 features of the generated virtually stained images. We used the 36 VS-HS image pairs (each with 640 × 640 pixels) as input to the Inception v3 network to extract 2048-dimensional feature vectors. From these extracted features, the mean vectors (μ,μω) and covariance matrices (Σ,Σω) were estimated.

The naturalness image quality evaluator [NIQE; (77)] values reported in fig. S5, are calculated as

D(ν1,ν2,Σ1,Σ2)=(ν1ν2)T(Σ1+Σ22)1(ν1ν2) (28)

where ν1 and Σ1 represent the mean vector and covariance matrix of our customized reference multivariate Gaussian (MVG) model, which was fitted using 800 GT PAS image patches (each with 640 × 640 pixels) sampled from both the training and testing datasets. Meanwhile, ν2 and Σ2 are the mean vector and covariance matrix for the MVG model of a generated virtually stained image. The MATLAB functions fitniqe and niqe were used to fit the reference MVG model and compute the final NIQE values for the virtually stained images (78).

Statistical analysis

In Fig. 3, a two-tailed paired t test was conducted to assess whether the image contrast between the virtually stained output and its HS PAS counterpart was statistically equivalent, with a statistical significance level of 0.05. This analysis was performed on 201 paired VS and HS images. A P value greater than 0.05 indicates no statistically significant difference in the image contrast between the VS images and their HS counterparts.

Other implementation details

All network training and testing tasks were conducted on a desktop computer equipped with an Intel Core i9-13900K CPU, 64 GB of memory, and an NVIDIA GeForce RTX 4090 GPU. The code for training the diffusion models was developed in Python 3.9.19 using PyTorch 2.2.1.

Acknowledgments

The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. We report the use of Grammarly (through MS Word), and we take full responsibility for the content of this paper.

Funding: The Ozcan Research Group at UCLA acknowledges the support of NIH P41, the National Center for Interventional Biophotonic Technologies (NCIBT). We also acknowledge the NIH Common Fund, the National Institute of Diabetes and Digestive and Kidney Diseases (NIDDK), and the Office of the Director (OD) under Award Numbers U54DK134302 and U01DK133766 (J.M.S. and R.V.d.P.) and, in part, by grant numbers 2021-240339 and 2022-309518 (L.G.M. and R.V.d.P.) from the Chan Zuckerberg Initiative DAF, an advised fund of Silicon Valley Community Foundation.

Author contributions: A.O. conceived the idea. Y.Z. developed the image processing pipeline and prepared the training/validation and test datasets. Y.Z., L.H., and Yuzhu Li, trained/tested the neural networks. Y.Z., L.H., N.P., Yuzhu Li, and Yuhang Li performed the result analysis and statistical studies. L.G.M., R.V.d.P., and J.M.S. contributed optical microscopy and IMS-related data and image registration methods. Y.Z., L.H., and A.O. prepared the manuscript. A.O. supervised the research.

Competing interests: Y.Z., L.H., and A.O. have a pending patent application filed by UCLA on VS technology. Patent status: pending; coinventors: Y.Z., L.H., and A.O.; application no. 63/712,348; filing date: 25 October 2024. The authors declare that they have no other competing interests.

Data and materials availability: The AF images (7983), IMS data (8488), and PAS-stained images (8993) can be accessed through the Human Biomolecular Atlas Program (HuBMAP): https://hubmapconsortium.org. Deep learning models reported in this work used standard libraries and scripts that are publicly available in PyTorch. The codes for our VS framework can be found at https://doi.org/10.5281/zenodo.15043291. All other data needed to evaluate the conclusions in the paper are present in the paper and/or the Supplementary Materials.

Supplementary Materials

This PDF file includes:

Figs. S1 to S7

sciadv.adv0741_sm.pdf (1.4MB, pdf)

REFERENCES AND NOTES

  • 1.Caprioli R. M., Farmer T. B., Gile J., Molecular imaging of biological samples: Localization of peptides and proteins using MALDI-TOF MS. Anal. Chem. 69, 4751–4760 (1997). [DOI] [PubMed] [Google Scholar]
  • 2.Colley M. E., Esselman A. B., Scott C. F., Spraggins J. M., High-specificity imaging mass spectrometry. Annu. Rev. Anal. Chem. 17, 1–24 (2024). [DOI] [PubMed] [Google Scholar]
  • 3.Djambazova K. V., van Ardenne J. M., Spraggins J. M., Advances in imaging mass spectrometry for biomedical and clinical research. TrAC Trends Anal. Chem. 169, 117344 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Krestensen K. K., Heeren R. M. A., Balluff B., State-of-the-art mass spectrometry imaging applications in biomedical research. Analyst 148, 6161–6187 (2023). [DOI] [PubMed] [Google Scholar]
  • 5.Spraggins J. M., Djambazova K. V., Rivera E. S., Migas L. G., Neumann E. K., Fuetterer A., Suetering J., Goedecke N., Ly A., Van De Plas R., Caprioli R. M., High-performance molecular imaging with MALDI trapped ion-mobility time-of-flight (timsTOF) mass spectrometry. Anal. Chem. 91, 14552–14560 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Neumann E. K., Migas L. G., Allen J. L., Caprioli R. M., Van De Plas R., Spraggins J. M., Spatial metabolomics of the human kidney using MALDI trapped ion mobility imaging mass spectrometry. Anal. Chem. 92, 13084–13091 (2020). [DOI] [PubMed] [Google Scholar]
  • 7.Taylor M. J., Lukowski J. K., Anderton C. R., Spatially resolved mass spectrometry at the single cell: Recent innovations in proteomics and metabolomics. J. Am. Soc. Mass Spectrom. 32, 872–894 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Spivey E. C., McMillen J. C., Ryan D. J., Spraggins J. M., Caprioli R. M., Combining MALDI-2 and transmission geometry laser optics to achieve high sensitivity for ultra-high spatial resolution surface analysis. J. Mass Spectrom. 54, 366–370 (2019). [DOI] [PubMed] [Google Scholar]
  • 9.Niehaus M., Soltwisch J., Belov M. E., Dreisewerd K., Transmission-mode MALDI-2 mass spectrometry imaging of cells and tissues at subcellular resolution. Nat. Methods 16, 925–931 (2019). [DOI] [PubMed] [Google Scholar]
  • 10.Bien T., Bessler S., Dreisewerd K., Soltwisch J., Transmission-mode MALDI mass spectrometry imaging of single cells: Optimizing sample preparation protocols. Anal. Chem. 93, 4513–4520 (2021). [DOI] [PubMed] [Google Scholar]
  • 11.Puolitaival S. M., Burnum K. E., Cornett D. S., Caprioli R. M., Solvent-free matrix dry-coating for MALDI imaging of phospholipids. J. Am. Soc. Mass Spectrom. 19, 882–886 (2008). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Anderson D. M. G., Floyd K. A., Barnes S., Clark J. M., Clark J. I., Mchaourab H., Schey K. L., A method to prevent protein delocalization in imaging mass spectrometry of non-adherent tissues: Application to small vertebrate lens imaging. Anal. Bioanal. Chem. 407, 2311–2320 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Van de Plas R., Yang J., Spraggins J., Caprioli R. M., Image fusion of mass spectrometry and microscopy: A multimodality paradigm for molecular tissue mapping. Nat. Methods 12, 366–372 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Tideman L. E., Migas L. G., Djambazova K. V., Patterson N. H., Caprioli R. M., Spraggins J. M., Van de Plas R., Automated biomarker candidate discovery in imaging mass spectrometry data through spatially localized Shapley additive explanations. Anal. Chim. Acta 1177, 338522 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Verbeeck N., Spraggins J. M., Murphy M. J., Wang H., Deutch A. Y., Caprioli R. M., Van de Plas R., Connecting imaging mass spectrometry and magnetic resonance imaging-based anatomical atlases for automated anatomical interpretation and differential analysis. Biochim. Biophys. Acta Proteins Proteom. 1865, 967–977 (2017). [DOI] [PubMed] [Google Scholar]
  • 16.Patterson N. H., Tuck M., Lewis A., Kaushansky A., Norris J. L., Van De Plas R., Caprioli R. M., Next generation histology-directed imaging mass spectrometry driven by autofluorescence microscopy. Anal. Chem. 90, 12404–12413 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Abdelmoula W. M., Regan M. S., Lopez B. G. C., Randall E. C., Lawler S., Mladek A. C., Nowicki M. O., Marin B. M., Agar J. N., Swanson K. R., Kapur T., Sarkaria J. N., Wells W., Agar N. Y. R., Automatic 3D nonlinear registration of mass spectrometry imaging and magnetic resonance imaging data. Anal. Chem. 91, 6206–6216 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Jones M. A., Cho S. H., Patterson N. H., Van De Plas R., Spraggins J. M., Boothby M. R., Caprioli R. M., Discovering new lipidomic features using cell type specific fluorophore expression to provide spatial and biological specificity in a multimodal workflow with MALDI imaging mass spectrometry. Anal. Chem. 92, 7079–7086 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Bai B., Yang X., Li Y., Zhang Y., Pillar N., Ozcan A., Deep learning-enabled virtual histological staining of biological samples. Light Sci. Appl. 12, 57 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Rivenson Y., Wang H., Wei Z., de Haan K., Zhang Y., Wu Y., Günaydın H., Zuckerman J. E., Chong T., Sisk A. E., Westbrook L. M., Wallace W. D., Ozcan A., Virtual histological staining of unlabelled tissue-autofluorescence images via deep learning. Nat. Biomed. Eng. 3, 466–477 (2019). [DOI] [PubMed] [Google Scholar]
  • 21.Borhani N., Bower A. J., Boppart S. A., Psaltis D., Digital staining through the application of deep neural networks to multi-modal multi-photon microscopy. Biomed. Opt. Express 10, 1339–1350 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Zhang Y., de Haan K., Rivenson Y., Li J., Delis A., Ozcan A., Digital synthesis of histological stains using micro-structured and multiplexed virtual staining of label-free tissue. Light Sci. Appl. 9, 78 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Li J., Garfinkel J., Zhang X., Wu D., Zhang Y., De Haan K., Wang H., Liu T., Bai B., Rivenson Y., Biopsy-free in vivo virtual histology of skin using deep learning. Light Sci. Appl. 10, 233 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Pradhan P., Meyer T., Vieth M., Stallmach A., Waldner M., Schmitt M., Popp J., Bocklitz T., Computational tissue staining of non-linear multimodal imaging using supervised and unsupervised deep learning. Biomed. Opt. Express 12, 2280–2298 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Bai B., Wang H., Li Y., De Haan K., Colonnese F., Wan Y., Zuo J., Doan N. B., Zhang X., Zhang Y., Li J., Yang X., Dong W., Darrow M. A., Kamangar E., Lee H. S., Rivenson Y., Ozcan A., Label-free virtual HER2 immunohistochemical staining of breast tissue using deep learning. BME Front. 2022, 9786242 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Abraham T. M., Costa P. C., Filan C., Guang Z., Zhang Z., Neill S., Olson J. J., Levenson R., Robles F. E., Label-and slide-free tissue histology using 3D epi-mode quantitative phase imaging and virtual hematoxylin and eosin staining. Optica 10, 1605–1618 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Cao R., Nelson S. D., Davis S., Liang Y., Luo Y., Zhang Y., Crawford B., Wang L. V., Label-free intraoperative histology of bone tissue via deep-learning-assisted ultraviolet photoacoustic microscopy. Nat. Biomed. Eng. 7, 124–134 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Li Y., Pillar N., Li J., Liu T., Wu D., Sun S., Ma G., de Haan K., Huang L., Zhang Y., Hamidi S., Urisman A., Keidar Haran T., Wallace W. D., Zuckerman J. E., Ozcan A., Virtual histological staining of unlabeled autopsy tissue. Nat. Commun. 15, 1684 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Yang X., Bai B., Zhang Y., Aydin M., Li Y., Selcuk S. Y., Casteleiro Costa P., Guo Z., Fishbein G. A., Atlan K., Wallace W. D., Pillar N., Ozcan A., Virtual birefringence imaging and histological staining of amyloid deposits in label-free tissue using autofluorescence microscopy and deep learning. Nat. Commun. 15, 7978 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Y. Zhang, L. Huang, N. Pillar, Y. Li, H. Chen, A. Ozcan, Super-resolved virtual staining of label-free tissue using diffusion models. arXiv:2410.20073v1 (2024). [DOI] [PMC free article] [PubMed]
  • 31.Christiansen E. M., Yang S. J., Ando D. M., Javaherian A., Skibinski G., Lipnick S., Mount E., O’Neil A., Shah K., Lee A. K., Goyal P., Fedus W., Poplin R., Esteva A., Berndl M., Rubin L. L., Nelson P., Finkbeiner S., In silico labeling: Predicting fluorescent labels in unlabeled images. Cell 173, 792–803.e19 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Ounkomol C., Seshamani S., Maleckar M. M., Collman F., Johnson G. R., Label-free prediction of three-dimensional fluorescence images from transmitted-light microscopy. Nat. Methods 15, 917–920 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Kandel M. E., He Y. R., Lee Y. J., Chen T. H.-Y., Sullivan K. M., Aydin O., Saif M. T. A., Kong H., Sobh N., Popescu G., Phase imaging with computational specificity (PICS) for measuring dry mass changes in sub-cellular compartments. Nat. Commun. 11, 6256 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Somani A., Ahmed Sekh A., Opstad I. S., Birna Birgisdottir Å., Myrmel T., Singh Ahluwalia B., Horsch A., Agarwal K., Prasad D. K., Virtual labeling of mitochondria in living cells using correlative imaging and physics-guided deep learning. Biomed. Opt. Express 13, 5495–5516 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Kreiss L., Jiang S., Li X., Xu S., Zhou K. C., Lee K. C., Mühlberg A., Kim K., Chaware A., Ando M., Barisoni L., Lee S. A., Zheng G., Lafata K. J., Friedrich O., Horstmeyer R., Digital staining in optical microscopy using deep learning - a review. PhotoniX 4, 34 (2023). [Google Scholar]
  • 36.D. Winter, N. Triltsch, M. Rosati, A. Shumilov, Z. Kokaragac, Y. Popov, T. Padel, L. S. Monasor, R. Hill, M. Schick, N. Brieu, Mask-guided cross-image attention for zero-shot in-silico histopathologic image generation with a diffusion model. arXiv:2407.11664 (2024).
  • 37.Z. Li, T. Su, B. Zhang, W. Han, S. Zhang, G. Sun, Y. Cong, X. Chen, J. Qi, Y. Wang, His-MMDM: Multi-domain and multi-omics translation of histopathology images with diffusion models. medRxiv 24310294 [Preprint] (2024). 10.1101/2024.07.11.24310294. [DOI]
  • 38.B. Li, K. Xue, B. Liu, Y.-K. Lai, “Bbdm: Image-to-image translation with brownian bridge diffusion models,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (IEEE, 2023), pp. 1952–1961; http://openaccess.thecvf.com/content/CVPR2023/html/Li_BBDM_Image-to-Image_Translation_With_Brownian_Bridge_Diffusion_Models_CVPR_2023_paper.html. [Google Scholar]
  • 39.D. Revuz, M. Yor, Continuous Martingales and Brownian Motion (Springer Science & Business Media, 2013), vol. 293; https://books.google.com/books?hl=en&lr=&id=IWjsCAAAQBAJ&oi=fnd&pg=PA1&dq=Revuz,+Daniel%3B+Yor,+Marc+(1999).+Continuous+Martingales+and+Brownian+Motion+(2nd+ed.).+New+York:+Springer-Verlag.&ots=hzriIbNqul&sig=DqvBYBTle9EEUFzeRascv_U70bI. [Google Scholar]
  • 40.O. Ronneberger, P. Fischer, T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention (Springer, 2015), pp. 234–241. [Google Scholar]
  • 41.Esselman A. B., Patterson N. H., Migas L. G., Dufresne M., Djambazova K. V., Colley M. E., Van de Plas R., Spraggins J. M., Microscopy-directed imaging mass spectrometry for rapid high spatial resolution molecular imaging of glomeruli. J. Am. Soc. Mass Spectrom. 34, 1305–1314 (2023). [DOI] [PubMed] [Google Scholar]
  • 42.Ho J., Jain A., Abbeel P., Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 33, 6840–6851 (2020). [Google Scholar]
  • 43.R. Rombach, A. Blattmann, D. Lorenz, P. Esser, B. Ommer, “High-resolution image synthesis with latent diffusion models,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (IEEE, 2022), pp. 10684–10695; http://openaccess.thecvf.com/content/CVPR2022/html/Rombach_High-Resolution_Image_Synthesis_With_Latent_Diffusion_Models_CVPR_2022_paper.html. [Google Scholar]
  • 44.D. Podell, Z. English, K. Lacey, A. Blattmann, T. Dockhorn, J. Müller, J. Penna, R. Rombach, SDXL: Improving latent diffusion models for high-resolution image synthesis. arXiv:2307.01952 (2023).
  • 45.L. Zhang, A. Rao, M. Agrawala, “Adding conditional control to text-to-image diffusion models,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (IEEE, 2023), pp. 3836–3847; http://openaccess.thecvf.com/content/ICCV2023/html/Zhang_Adding_Conditional_Control_to_Text-to-Image_Diffusion_Models_ICCV_2023_paper.html. [Google Scholar]
  • 46.T. Brooks, A. Holynski, A. A. Efros, “Instructpix2pix: Learning to follow image editing instructions,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (IEEE, 2023), pp. 18392–18402; http://openaccess.thecvf.com/content/CVPR2023/html/Brooks_InstructPix2Pix_Learning_To_Follow_Image_Editing_Instructions_CVPR_2023_paper.html. [Google Scholar]
  • 47.Janowczyk A., Zuo R., Gilmore H., Feldman M., Madabhushi A., HistoQC: An open-source quality control tool for digital pathology slides. JCO Clin. Cancer Inform. 3, 1–7 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Hellinger E., Neue Begründung der Theorie quadratischer Formen von unendlichvielen Veränderlichen. J. Für Reine Angew. Math. 1909, 210–271 (1909). [Google Scholar]
  • 49.Jeffreys H., An invariant form for the prior probability in estimation problems. Proc. R. Soc. Lond. A Math. Phys. Sci. 186, 453–461 (1946). [DOI] [PubMed] [Google Scholar]
  • 50.Rivenson Y., Göröcs Z., Günaydin H., Zhang Y., Wang H., Ozcan A., Deep learning microscopy. Optica 4, 1437–1443 (2017). [Google Scholar]
  • 51.de Haan K., Ballard Z. S., Rivenson Y., Wu Y., Ozcan A., Resolution enhancement in scanning electron microscopy using deep learning. Sci. Rep. 9, 12050 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.R. Zhang, P. Isola, A. A. Efros, E. Shechtman, O. Wang, “The unreasonable effectiveness of deep features as a perceptual metric,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (IEEE, 2018), pp. 586–595; http://openaccess.thecvf.com/content_cvpr_2018/html/Zhang_The_Unreasonable_Effectiveness_CVPR_2018_paper.html. [Google Scholar]
  • 53.A. Q. Nichol, P. Dhariwal, “Improved denoising diffusion probabilistic models,” in International Conference on Machine Learning (PMLR, 2021), pp. 8162–8171; https://proceedings.mlr.press/v139/nichol21a.html. [Google Scholar]
  • 54.I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, Y. Bengio, “Generative adversarial nets,” in Advances in Neural Information Processing Systems, Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, K. Q. Weinberger, Eds. (Curran Associates Inc., 2014), vol. 27; 10.1145/3422622. [DOI] [Google Scholar]
  • 55.M. Mirza, Conditional generative adversarial nets. arXiv:1411.1784 (2014).
  • 56.P. Isola, J.-Y. Zhu, T. Zhou, A. A. Efros, “Image-to-image translation with conditional adversarial networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (IEEE, 2017), pp. 1125–1134. [Google Scholar]
  • 57.Wang H., Rivenson Y., Jin Y., Wei Z., Gao R., Günaydın H., Bentolila L. A., Kural C., Ozcan A., Deep learning enables cross-modality super-resolution in fluorescence microscopy. Nat. Methods 16, 103–110 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.de Haan K., Rivenson Y., Wu Y., Ozcan A., Deep-learning-based image reconstruction and enhancement in optical microscopy. Proc. IEEE 108, 30–50 (2020). [Google Scholar]
  • 59.Saharia C., Ho J., Chan W., Salimans T., Fleet D. J., Norouzi M., Image super-resolution via iterative refinement. IEEE Trans. Pattern Anal. Mach. Intell. 45, 4713–4726 (2022). [DOI] [PubMed] [Google Scholar]
  • 60.S. Gao, X. Liu, B. Zeng, S. Xu, Y. Li, X. Luo, J. Liu, X. Zhen, B. Zhang, “Implicit diffusion models for continuous super-resolution,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (IEEE, 2023), pp. 10021–10030; http://openaccess.thecvf.com/content/CVPR2023/html/Gao_Implicit_Diffusion_Models_for_Continuous_Super-Resolution_CVPR_2023_paper.html. [Google Scholar]
  • 61.L. Huang, X. Xiao, S. Li, J. Sun, Y. Huang, A. Ozcan, P. Bogdan, Multi-scale conditional generative modeling for microscopic image restoration. arXiv:2407.05259 (2024).
  • 62.L. Metz, B. Poole, D. Pfau, J. Sohl-Dickstein, Unrolled generative adversarial networks. arXiv:1611.02163 (2017).
  • 63.Z. Wang, H. Zheng, P. He, W. Chen, M. Zhou, Diffusion-GAN: Training GANs with diffusion. arXiv:2206.02262 (2022).
  • 64.Asghar M. S., Denic A., Rule A. D., Morphometric analysis of chronicity on kidney biopsy: A useful prognostic exercise. Clin. Kidney J. 17, sfad226 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.E. Neumann, J. Allen, A. Kruse, J. Harvey, M. Brewer, C. Romer, M. De Caestecker, J. Spraggins, PAS staining of fresh frozen or paraffin embedded human kidney tissue, protocols.io (2023); 10.17504/protocols.io.x54v98nkzl3e/v3. [DOI]
  • 66.Klein S., Staring M., Murphy K., Viergever M. A., Pluim J. P., Elastix: A toolbox for intensity-based medical image registration. IEEE Trans. Med. Imaging 29, 196–205 (2009). [DOI] [PubMed] [Google Scholar]
  • 67.H. Patterson, NHPatterson/wsireg, GitHub (2022); https://github.com/NHPatterson/wsireg.
  • 68.H. Patterson, NHPatterson/napari-imsmicrolink, GitHub (2022); https://github.com/NHPatterson/napari-imsmicrolink.
  • 69.W. Shi, J. Caballero, F. Huszár, J. Totz, A. P. Aitken, R. Bishop, D. Rueckert, Z. Wang, “Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (IEEE, 2016), pp. 1874–1883; www.cv-foundation.org/openaccess/content_cvpr_2016/html/Shi_Real-Time_Single_Image_CVPR_2016_paper.html. [Google Scholar]
  • 70.Vaswani A., Attention is all you need. Adv. Neural Inf. Process. Syst. 30, 6000–6010 (2017). [Google Scholar]
  • 71.D. Hendrycks, K. Gimpel, Gaussian error linear units (gelus). arXiv:1606.08415 (2016).
  • 72.Elfwing S., Uchibe E., Doya K., Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 107, 3–11 (2018). [DOI] [PubMed] [Google Scholar]
  • 73.I. Loshchilov, F. Hutter, Decoupled weight decay regularization. arXiv:1711.05101 (2017).
  • 74.K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556 (2015).
  • 75.C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, Z. Wojna, “Rethinking the inception architecture for computer vision,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (IEEE, 2016), pp. 2818–2826; www.cv-foundation.org/openaccess/content_cvpr_2016/html/Szegedy_Rethinking_the_Inception_CVPR_2016_paper.html. [Google Scholar]
  • 76.Heusel M., Ramsauer H., Unterthiner T., Nessler B., Hochreiter S., GANs trained by a two time-scale update rule converge to a local nash equilibrium. Adv. Neural Inf. Process. Syst. 30, 6629–6640 (2017). [Google Scholar]
  • 77.Mittal A., Soundararajan R., Bovik A. C., Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 20, 209–212 (2012). [Google Scholar]
  • 78.niqe, Naturalness Image Quality Evaluator (NIQE) no-reference image quality score, MATLAB; www.mathworks.com/help/images/ref/niqe.html.
  • 79.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, Auto-fluorescence microscopy data right kidney, HuBMAP Consortium (2024); 10.35079/HBM888.PHHS.299. [DOI]
  • 80.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, Auto-fluorescence microscopy data left kidney, HuBMAP Consortium (2024); 10.35079/HBM578.WQFJ.765. [DOI]
  • 81.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, Auto-fluorescence microscopy data right kidney, HuBMAP Consortium (2024); 10.35079/HBM576.CBTX.826. [DOI]
  • 82.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, Auto-fluorescence microscopy data right kidney, HuBMAP Consortium (2024); 10.35079/HBM353.XGRQ.833. [DOI]
  • 83.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, Auto-fluorescence microscopy data right kidney, HuBMAP Consortium (2024); 10.35079/HBM836.DHDQ.639. [DOI]
  • 84.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, MALDI Imaging mass spectrometry data right kidney, HuBMAP Consortium (2023); 10.35079/HBM437.BGCD.226. [DOI]
  • 85.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, MALDI Imaging mass spectrometry data right kidney, HuBMAP Consortium (2023); 10.35079/HBM644.HLRW.739. [DOI]
  • 86.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, MALDI Imaging mass spectrometry data right kidney, HuBMAP Consortium (2023); 10.35079/HBM286.ZZNV.536. [DOI]
  • 87.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, MALDI Imaging mass spectrometry data right kidney, HuBMAP Consortium (2023); 10.35079/HBM946.ZHFC.683. [DOI]
  • 88.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, MALDI Imaging mass spectrometry data Left kidney, HuBMAP Consortium (2023); 10.35079/HBM794.JRXQ.455. [DOI]
  • 89.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, Periodic acid-Schiff stained microscopy data right kidney, HuBMAP Consortium (2024); 10.35079/HBM396.XMSB.623. [DOI]
  • 90.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, Periodic acid-Schiff stained microscopy data right kidney, HuBMAP Consortium (2024); 10.35079/HBM964.CZLM.657. [DOI]
  • 91.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, Periodic acid-Schiff stained microscopy data right kidney, HuBMAP Consortium (2024); 10.35079/HBM738.MXDD.328. [DOI]
  • 92.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, Periodic acid-Schiff stained microscopy data right kidney, HuBMAP Consortium (2024); 10.35079/HBM864.MFDZ.983. [DOI]
  • 93.J. M. Spraggins, A. B. Fogo, D. B. Gutierrez, D. M. G. Anderson, E. K. Neumann, H. Yang, J. L. Allen, L. Tideman, L. Migas, M. E. Colley, M. deCaestecker, M. Brewer, M. A. Farrow, N. H. Patterson, R. Van de Plas, R. C. Harris, R. M. Caprioli, Periodic acid-Schiff stained microscopy data left kidney, HuBMAP Consortium (2024); 10.35079/HBM593.DVGN.469. [DOI]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Figs. S1 to S7

sciadv.adv0741_sm.pdf (1.4MB, pdf)

Articles from Science Advances are provided here courtesy of American Association for the Advancement of Science

RESOURCES