Skip to main content
Nature Portfolio logoLink to Nature Portfolio
. 2026 Mar 20;23(4):728–731. doi: 10.1038/s41592-026-03044-7

LazySlide: accessible and interoperable whole-slide image analysis

Yimin Zheng 1, Ernesto Abila 1,2, Eva Chrenková 3, Iva Buljan 1,2, Juliane Winkler 3, André F Rendeiro 1,2,
PMCID: PMC13076205  PMID: 41862659

Abstract

Histopathological data are foundational in both biological research and clinical diagnostics but remain siloed from modern multimodal and single-cell frameworks. Here we introduce LazySlide, an open-source Python package built on the scverse ecosystem for efficient whole-slide image analysis and multimodal integration. By leveraging vision–language foundation models and adhering to scverse data standards, LazySlide bridges histopathology with omics workflows. It supports tissue and cell segmentation, feature extraction, cross-modal querying and zero-shot classification, with minimal setup.

Subject terms: Software, Computational platforms and environments


LazySlide combines scverse with foundation models to enable efficient whole-slide image analysis.

Main

Histopathology is a cornerstone of biomedical research and clinical practice, offering high-resolution insights into the spatial organization of tissues, cellular morphology and pathological alterations. Advances in digital pathology have enabled the collection of large repositories of whole-slide images (WSIs), creating unprecedented opportunities to study the tissue basis of human health and disease. However, computational analysis of WSIs remains challenging. Existing toolkits for WSI analysis, such as QuPath1, CLAM2, PathML3, Slideflow4 and TIAToolbox5, offer partial solutions but are often limited by fragmented data structures, platform-specific constraints and high technical barriers to entry. These limitations hinder integration with the multimodal and single-cell workflows that are increasingly standard in modern biology.

To address these limitations, we developed LazySlide, a general-purpose Python framework for WSI analysis, fully interoperable with the scverse ecosystem6 widely used in biomedical research. LazySlide streamlines WSI preprocessing, segmentation, feature extraction and multimodal integration using state-of-the-art deep learning models—effectively bridging histopathology with omics workflows, data structures and operations.

At its core, LazySlide introduces a custom data structure, WSIData (Fig. 1), built upon scverse’s SpatialData7 but optimized for the diverse formats and scale of WSIs. While SpatialData supports only OME-TIFF pyramids, most WSIs are stored in proprietary formats requiring specialized readers. Existing extensions such as SOPA8 adapt SpatialData to WSIs but incur substantial disk overhead due to serialization (5–10× increase). By contrast, WSIData enables efficient, direct access to standard WSI formats without duplication, while maintaining compatibility with scverse tools and deep learning pipelines. The framework follows familiar application programming interface (API) conventions from AnnData9, Scanpy10 and Squidpy11, reducing the barrier for computational pathologists and genomics researchers alike.

Fig. 1. LazySlide framework overview.

Fig. 1

WSIData structure enables WSI preprocessing, quantification of cellular, morphological and microanatomical tissue features, and multimodal integration with deep learning frameworks. scverse logo reproduced from ref. 7 under a Creative Commons license CC BY 4.0.

LazySlide includes a full suite of tools for WSI analysis, from preprocessing, quantification, visualization and multi-omics integration to advanced capabilities that support insightful interpretation and interactive exploration, leveraging zero-shot approaches and genomic data integration (Fig. 1). Specifically, slides are segmented into tissue regions, tiled into memory-efficient patches and embedded via vision foundation models into high-dimensional feature spaces. These embeddings support downstream analyses such as unsupervised clustering, which can reveal spatial tissue architectures—for example, the mucosa, submucosa, muscularis and lymphoid tissue in the human small intestine (Fig. 1). Built-in cell segmentation and classification tools enable the quantification of cellular composition and morphology.

Beyond image-based analysis, LazySlide leverages vision–language models to support natural language queries, enabling users to localize image regions most similar to prompts, such as ‘lymphocyte’. These features also enable multimodal integration with transcriptomics and other data types, supporting disease scoring and mechanistic interpretation. LazySlide further provides zero-shot capabilities for slide captioning, relevance scoring and text-guided segmentation, as well as efficient, declarative visualization of WSIs and annotations, including segmented regions, cells and extracted features.

LazySlide supports a diverse range of applications and facilitates rigorous benchmarking against existing methodologies, thereby showcasing its robust capabilities and broad utility. To showcase its capabilities in combining WSI data with other modalities, we demonstrate three representative applications: (1) zero-shot vision–language querying, (2) multimodal integration of histology and transcriptomics and (3) zero-shot organ classification—each illustrating cutting-edge use cases for computational pathology and omics data integration.

First, we assembled a dataset of human artery slides from the GTEx project12, including healthy (n = 24) and calcified (n = 21) tissues with matched RNA sequencing (RNA-seq) profiles (Fig. 2a). With only five lines of code, users can compute text-to-image similarity maps across each slide. Terms related to ‘calcification’ show higher enrichment in calcified samples, whereas anatomical terms predominate in healthy tissues (Fig. 2b,c). A differential analysis of image–text features highlights terms such as gap junction, vascular niche and apoptosis as significantly enriched in calcified arteries, consistent with observed morphological changes. Building on this, a slide-level ‘calcification score’ is computed using top-k pooling over similarity maps with the term calcification (Fig. 2d), yielding significantly elevated scores in calcified tissues.

Fig. 2. LazySlide applications and benchmarks.

Fig. 2

a, WSIs of healthy and calcified arteries. b, Text term scores across WSIs showing calcification-associated patterns. c, Differential text features between conditions (two-sided Mann–Whitney U test, Padj is calculated with the Bonferroni method). d, Calcification scores per WSI with representative tiles. ****P < 0.0001 (P = 0.00001, two-sided Mann–Whitney U test). e, UMAP of RNA-seq (top) and WSI features (bottom) from paired samples. f, Variance explained by MOFA of RNA and WSI modalities. g, Enriched pathways from WSI + RNA versus RNA-only analysis. h, Count of highly enriched pathways by analysis type. i, Representative tissues for zero-shot classification. j, Confusion matrix of zero-shot organ classification. k, Zero-shot calcification prediction probabilities. l, Code complexity comparison across frameworks. m, Classification accuracy versus QuPath. NS, non-significant; *P < 0.05; **P < 0.01 (unpaired two-sided Student’s t-test; QuPath versus ResNet50 P = 0.22, versus Titan P = 0.044, versus h0-mini P = 0.0059, versus UNI2 P = 0.0028). n, Tissue segmentation time versus QuPath. **P < 0.01; ***P < 0.001 (two-sided Mann–Whitney U test; LazySlide versus QuPath-auto P = 0.0024, versus QuPath-manual P = 0.0007). Box plots: medians, quartiles, 1.5× interquartile range. Error bars: mean ± 95% confidence interval.

To explore joint imaging-transcriptomic patterns, we use the calcification score to anchor a multimodal integration of WSI-derived and RNA-seq embeddings (Fig. 2e). This analysis, which is executable in a single line of code, shows that image features separate healthy and calcified groups more distinctly in Uniform Manifold Approximation and Projection (UMAP) space than RNA-seq alone. This pattern is also captured via multi-omics factor analysis (MOFA)13 integration using Muon14, the scverse multi-omics framework (Fig. 2f). We then compared gene-level relevance with calcification by ranking genes from RNA-only differential expression or joint WSI + RNA integration. Pathway enrichment on the top 300 up- and downregulated genes from each method revealed that only the WSI and RNA integration identifies key calcification-related pathways, including IL-18 signaling (Fig. 2g), and yields a greater number of significant pathway associations overall (Fig. 2h).

Finally, we demonstrate the zero-shot classification capabilities of LazySlide. Using vision–language models, WSIs from nine distinct human organs were queried against their organ names (Fig. 2i). With a single line of code, LazySlide correctly identified the majority of organ sources (Fig. 2j). Applying the same approach to the artery calcification dataset also yields high classification performance (Fig. 2k), which illustrates the ability to extract valuable insights without tissue- or context-specific training.

To assess usability and performance, we benchmarked LazySlide against established tools across a standard preprocessing pipeline: tissue segmentation, tiling, tile dataset preparation for PyTorch15 and feature extraction. LazySlide completes this workflow with fewer lines of code, lower token count and a simpler API, facilitating rapid development and code maintenance (Fig. 2l). We also compared LazySlide with QuPath in a classification setting of semantically labeled microanatomical domains of murine breast cancer lung metastasis using extracted features by different vision models. Excluding ResNet50, which is not trained on tissue images, LazySlide consistently outperforms QuPath-derived features (Fig. 2m). Moreover, tissue segmentation is markedly faster in LazySlide compared with QuPath using either automated or manual workflows (Fig. 2n). Finally, we performed a qualitative comparison of LazySlide’s features with existing software (Extended Data Table 1), highlighting its remarkable flexibility and the number of unique tasks it supports, including text–image queries, RNA data integration, zero-shot learning, unsupervised spatial domain detection, declarative visualization and virtual staining.

Extended Data Table 1.

Software feature comparison

graphic file with name 41592_2026_3044_Tab1_ESM.jpg

Comparison of features between LazySlide and existing software.

Despite its advantages, LazySlide’s strict use of the Zarr format maximizes interoperability but its distributed file structure may limit flexibility in compute environments with file count restrictions. Future developments will focus on co-registration capabilities across imaging modalities and for three-dimensional tissue reconstruction.

Altogether, LazySlide represents a substantial advancement in artificial-intelligence-enabled histopathology and tissue biology, bridging computational pathology with multimodal omics through a modular, user-friendly and open-source framework. By supporting a wide range of foundational models and ensuring seamless interoperability with the scverse ecosystem, LazySlide enables integrative, scalable and interpretable analysis of WSIs. It empowers both computational pathologists and genomics researchers to uncover new insights into tissue biology and disease, accelerating the development of data-driven, clinically meaningful models.

Methods

Data structure

The WSIData object was designed as an inheritance of SpatialData, allowing it to behave exactly like SpatialData7. This design enables users familiar with scverse and SpatialData to readily use WSIData across scverse packages such as Lazyslide. The SpatialData component of WSIData is a data container, storing all analysis metadata and results generated within LazySlide. An additional layer and corresponding APIs were implemented to specifically interact with WSIs. WSIData was built to support multiple WSI backends, including OpenSlide16, tiffslide17, bioformats18 and cuCIM19, ensuring compatibility with a broad range of image formats. An accessor interface, similar to those found in pandas20, xarray21 and SpatialData, was also implemented to provide a variety of features. For instance, users can easily retrieve useful information, such as the number of tissue pieces found in each slide, the number of tiles in a tesselated image, and an AnnData object with morphological features through a ‘fetch’ accessor. The ‘iter’ accessor allows users to iterate through each tissue piece, tile or their features. The ‘dataset’ accessor provides an interface to create various PyTorch datasets that could be directly used for training or running deep learning models.

Tissue segmentation

Tissue segmentation was performed using a multistep image processing pipeline, which was implemented with OpenCV. The algorithm initiates processing of WSIs at an automatically determined optimal resolution level. For artifact filtering, specialized thresholding techniques were applied to identify and exclude nontissue regions on the basis of color properties. Alternatively, images can be either converted to grayscale or transformed to HSV (Hue, Saturation, Value) color space, with the saturation channel extracted to enhance tissue visibility. A median blur filter was first applied to reduce noise, followed by binary thresholding, using either Otsu’s method or a fixed threshold value. The resulting binary mask undergoes morphological closing operations to fill small gaps within tissue regions while preserving overall tissue structure. This process effectively separated tissue regions from background and artifacts, enabling subsequent region identification and analysis. Tissue masks can be further refined using higher-resolution image data when specified, and regions below a configurable minimum area threshold are excluded. As an alternative, the GrandQC22 pretrained tissue segmentation model was also integrated into LazySlide to perform deep-learning-based tissue segmentation.

Tissue tiling

Tissue regions were tiled using an automated approach that generated a grid of tiles within identified tissue masks or contours. The algorithm allows users to request arbitrary tile sizes no lower than the raw image resolution and accepted user-specified tile dimensions along with customizable stride or overlap between adjacent tiles. For each tissue region, tiles were initially generated within its bounding box, then filtered to retain only those intersecting with the tissue contour. To eliminate tiles containing excessive background, a configurable background fraction threshold is applied, discarding tiles where the background exceeded this threshold. The tiling process supports multiresolution analysis by allowing the user to specify a resolution in micrometers per pixel, ensuring a consistent physical scale across slides scanned at different magnifications.

Quality control

Following tiling, quality control was implemented through a configurable scoring system that evaluated tiles on the basis of metrics including focus quality, contrast, brightness and tissue characteristics. This scoring facilitated subsequent filtering of low-quality tiles containing artifacts or out-of-focus regions, thereby ensuring the reliability of downstream analysis. All operations supported parallel processing for computational efficiency when handling large datasets. In addition, GrandQC22 was incorporated in LazySlide to detect common artifacts such as bubbles, tissue folds and out-of-focus regions.

Feature extraction and aggregation

The feature extraction module supports a variety of pretrained vision models through Torch Image Models (timm, https://github.com/huggingface/pytorch-image-models). The tiled images are processed through selected models, including classic vision architectures pretrained on ImageNet (for example ResNet, DenseNet and ViT) and pathology-specific foundation models (for example UNI/UNI223, Virchow/Virchow224,25, H-Optimus-0/1 and GigaPath26), to extract high-dimensional feature representations. Feature extraction supports automatic device selection (central processing unit/graphics processing unit), mixed-precision inference and configurable batch processing for computational efficiency. For each tissue tile, the deep learning model extracts feature vectors that capture complex visual patterns not easily quantifiable through traditional image analysis. Following extraction, LazySlide implements multiple feature aggregation strategies to generate slide-level or tissue-level representations from tile-level features. These include statistical aggregation methods (mean and median) and specialized neural encoders that incorporate spatial information (for example, PRISM27 and TITAN28).

Natural language query of tissue images

Natural language query capabilities in LazySlide leverage multimodal foundation models to bridge vision and language domains, enabling content retrieval through text-based queries. The framework uses state-of-the-art pathology-specific models, such as PLIP29 and CONCH30, to facilitate semantic search within WSIs. The query process follows a two-stage approach: first, text queries are transformed into dense vector representations using the text encoder component of the chosen model. Second, these text embeddings are compared with pre-extracted image feature vectors using cosine similarity to identify regions most relevant to the query. Similarity scores are computed via the dot product between normalized text and image embeddings, with higher scores indicating stronger semantic correspondence. This powerful functionality enables users to search for specific histological patterns, cell types or tissue structures using natural language descriptions, offering a remarkable advantage over traditional visual inspection methods.

Spatial domain detection

LazySlide implements unsupervised spatial domain identification through a multistage, graph-based clustering approach. Initially, a neighborhood graph is constructed from the extracted tiles, where each tile represents a node and edges connect spatially adjacent tiles. Optionally, morphological features can be smoothed by incorporating neighboring information, thereby enhancing feature consistency. Following graph construction, LazySlide uses a dimensionality reduction pipeline, as introduced in Scanpy, which first scales the feature data and then applies principal component analysis to capture dominant variation patterns. The resulting low-dimensional representation is then used to construct a weighted k-nearest neighbor graph. Finally, the Leiden community detection algorithm31 (implemented via igraph) identifies coherent tissue domains by partitioning this graph into communities, with the clustering resolution controlled by a user-adjustable parameter. This comprehensive approach effectively segments histologically distinct regions within tissue slides without requiring manual annotation, thereby enabling automated identification of intricate tissue architectural features.

Whole-slide cell segmentation and classification

LazySlide supports cell segmentation of WSIs through Instanseg32 and Cellpose33. Joint cell segmentation with classification through Nulite34 and HistoPLUS35. The segmentation stages operate within a dedicated SegmentationRunner class, which efficiently handles batch processing across tiles, utilizing graphics processing unit acceleration when available, and performs the crucial merging of results from individual tiles. For each tiled image, cells are identified and their contours are transformed into a polygon representation. To merge results across tiles of the WSI, results are aggregated by addressing potential boundary artifacts, via a sophisticated polygon merging algorithm that leverages a spatial indexing tree (STR tree). All initially overlapping polygons are identified and then merged by preserving their class labels while calculating probability-weighted averages for cells spanning multiple tiles, thereby ensuring spatial continuity across the entire tissue section. Newly merged polygons are then pruned from the tree, and the process iterates until the tree contains no remaining branches, ensuring all overlapping polygons have been successfully merged into cohesive cell instances.

Integration with RNA-seq data

LazySlide enables multimodal data integration through the RNALinker class, which links histopathological features with paired transcriptomic profiles. Morphological features extracted from WSIs are first aggregated into slide-level representations that are matched to corresponding genomic samples. The integration proceeds in two steps: (1) samples are grouped and scored on the basis of morphological features, either through differential analysis between defined sample groups or by using morphologically derived annotations, such as text-based scores; (2) these morphology-based scores are statistically associated with gene expression data using correlation metrics (for example, Pearson, Spearman and Kendall) and regression models (for example, linear regression and lasso). This approach identifies genes whose expression is strongly associated with specific histological patterns. By linking morphology to transcriptomic signatures, this integration supports hypothesis generation about the molecular mechanisms driving tissue phenotypes and facilitates the discovery of biomarkers that bridge computer vision-derived features with biological interpretation.

Zero-shot learning

LazySlide supports zero-shot learning through multimodal vision–language foundation models such as PRISM27 and TITAN28, supporting various biologically relevant questions without task-specific training. For zero-shot classification, slide-level feature embeddings are compared with arbitrary, user-defined text prompts, yielding probability scores that quantify the semantic alignment between tissue morphology and natural language descriptions. Alternatively, for descriptive analysis, LazySlide generates natural language summaries of histological content from slide embeddings and text prompts, enabling automated preliminary reporting and assisting pathologists with concise, context-aware interpretations. For zero-shot segmentation of tissue, LazySlide uses text–image similarity metrics to produce a binary mask based on a user-defined threshold. Detected object bounding boxes from this mask are then used to guide segmentation via the Segment Anything Model 2 (SAM2)36, enabling flexible and label-free identification of tissue regions of interest.

Benchmarking

To assess the usability and interoperability of LazySlide within the scverse ecosystem, we conducted a benchmark comparison against several widely used WSI processing libraries: CLAM2, TRIDENT37, PathML3, Tiatoolbox5, Histolab38 and Slideflow4. Each library was evaluated using a standardized digital pathology workflow comprising tissue segmentation and tiling, feature extraction with ResNet50, construction of a PyTorch dataset for tile-level access (to assess ease of integration with deep learning), and generation of an AnnData object containing tile features and spatial metadata (to assess compatibility with scverse workflows). Benchmark scripts were implemented for each library, formatted using ruff, and executed in isolated Docker environments to ensure consistency and reproducibility. To quantify implementation complexity, we measured token counts, lines of code and API entropy, providing a comparative view of conceptual and practical overhead.

For direct benchmarking against QuPath, we used four mouse lung tissue slides from patient-derived xenograft models of breast cancer with metastasis. These slides were previously semantically annotated for three tissue classes: tumor, airways and blood vessels. Feature extraction was performed using both QuPath and LazySlide (with ResNet50, h0-mini39, UNI223 and TITAN28 backbones), and classification accuracy was used for performance comparison. Tissue segmentation was benchmarked by comparing QuPath (manual and automated methods) with LazySlide. Following a standardized protocol, four experienced volunteers used QuPath’s manual labeling and thresholding tools, and the time needed to perform tissue segmentation was recorded for two slides. The time performance of LazySlide was measured using the find_tissues function.

Reporting summary

Further information on research design is available in the Nature Portfolio Reporting Summary linked to this article.

Online content

Any methods, additional references, Nature Portfolio reporting summaries, source data, extended data, supplementary information, acknowledgements, peer review information; details of author contributions and competing interests; and statements of data and code availability are available at 10.1038/s41592-026-03044-7.

Supplementary information

Reporting Summary (1.6MB, pdf)
Peer Review File (1.6MB, pdf)

Acknowledgements

This research was funded by a grant from the Vienna Science and Technology Foundation (WWTF-LS23-067). The Rendeiro group was supported by funding from the European Research Council (ERC) under the European Union’s Horizon Europe research and innovation program (grant agreement no. 101220825), and by Angelini Ventures S.p.A. Rome, Italy. The Genotype-Tissue Expression (GTEx) Project was supported by the Common Fund of the Office of the Director of the National Institutes of Health, and by NCI, NHGRI, NHLBI, NIDA, NIMH and NINDS. We thank all Rendeiro Lab members for their precious input and for testing LazySlide. We thank E. Weigert, E. Gurnhofer, P. Wagner and G. Timelthaler for their technical support and pathologists Z. Bagó-Horváth and U. Heber for their feedback on tissue annotations.

Extended data

Author contributions

Y.Z. and A.F.R. conceptualized the study; Y.Z. developed LazySlide with contributions from E.A.; E.C. and I.B. provided annotations for WSIs; J.W. and A.F.R. supervised the research. All authors reviewed and approved the paper.

Peer review

Peer review information

Nature Methods thanks Muhammad Shaban and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. Primary Handling Editor: Madhura Mukhopadhyay, in collaboration with the Nature Methods team. Peer reviewer reports are available.

Data availability

GTEx WSIs and RNA expression data are available from the GTEx portal (https://gtexportal.org). WSIs of patient-derived xenograft models of breast cancer and their semantic segmentation are available via Zenodo at 10.5281/zenodo.15497223 (ref. 40).

Code availability

Lazyslide and WSIData are publicly available via GitHub at https://github.com/RendeiroLab/LazySlide and https://github.com/RendeiroLab/WSIData, published on the Python package index (PyPI) and conda-forge, available via Zenodo at 10.5281/zenodo.14939805 (ref. 41) and 10.5281/zenodo.15083425 (ref. 42), and documented at https://lazyslide.readthedocs.io and https://wsidata.readthedocs.io. A companion repository on benchmarking is also publicly available via GitHub at https://github.com/RendeiroLab/LazySlide-benchmark.

Competing interests

The authors declare no competing interests.

Footnotes

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Extended data

is available for this paper at 10.1038/s41592-026-03044-7.

Supplementary information

The online version contains supplementary material available at 10.1038/s41592-026-03044-7.

References

  • 1.Bankhead, P. et al. QuPath: open source software for digital pathology image analysis. Sci. Rep.7, 16878 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Lu, M. Y. et al. Data-efficient and weakly supervised computational pathology on whole-slide images. Nat. Biomed. Eng.5, 555–570 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Omar, M. et al. Antibody-based multiplex image analysis: standard analytical workflows and artificial intelligence tools for pathologists. Lab. Invest.105, 104220 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Dolezal, J. M. et al. Slideflow: deep learning for digital histopathology with real-time whole-slide visualization. BMC Bioinformatics25, 134 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Pocock, J. et al. TIAToolbox as an end-to-end library for advanced tissue image analytics. Commun. Med.2, 120 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Virshup, I. et al. The scverse project provides a computational ecosystem for single-cell omics data analysis. Nat. Biotechnol.41, 604–606 (2023). [DOI] [PubMed] [Google Scholar]
  • 7.Marconato, L. et al. SpatialData: an open and universal data framework for spatial omics. Nat. Methods22, 58–62 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Blampey, Q. et al. Sopa: a technology-invariant pipeline for analyses of image-based spatial omics. Nat. Commun.15, 4981 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Virshup, I., Rybakov, S., Theis, F. J., Angerer, P. & Wolf, F. A. anndata: access and store annotated data matrices. J. Open Source Softw.9, 4371 (2024). [Google Scholar]
  • 10.Wolf, F. A., Angerer, P. & Theis, F. J. SCANPY: large-scale single-cell gene expression data analysis. Genome Biol.19, 15 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Palla, G. et al. Squidpy: a scalable framework for spatial omics analysis. Nat. Methods19, 171–178 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Lonsdale, J. et al. The Genotype-Tissue Expression (GTEx) project. Nat. Genet.45, 580–585 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Argelaguet, R. et al. Multi-Omics Factor Analysis—a framework for unsupervised integration of multi-omics data sets. Mol. Syst. Biol.14, e8124 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Bredikhin, D., Kats, I. & Stegle, O. MUON: multimodal omics analysis framework. Genome Biol.23, 42 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Paszke, A. et al. PyTorch: an imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst.32, (2019).
  • 16.Goode, A., Gilbert, B., Harkes, J., Jukic, D. & Satyanarayanan, M. OpenSlide: a vendor-neutral software foundation for digital pathology. J. Pathol. Inform.4, 27 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Poehlmann, A. & Villalba, S. TiffSlide - Cloud Native Openslide-Python Replacement. Zenodo10.5281/ZENODO.6327079 (2022).
  • 18.Linkert, M. et al. Metadata matters: access to image data in the real world. J. Cell Biol.189, 777–782 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Cucim: CuCIM - RAPIDS GPU-Accelerated Image Processing Library. GitHub. https://github.com/rapidsai/cucim (2021).
  • 20.Mc Kinney, W. Pandas: a foundational Python library for data analysis and statistics. Python for high performance and scientific computing14, 9 (2011). [Google Scholar]
  • 21.Hoyer, S. & Hamman, J. xarray: N-D labeled arrays and datasets in Python. J. Open Res. Softw.5, 10 (2017). [Google Scholar]
  • 22.Weng, Z. et al. GrandQC: a comprehensive solution to quality control problem in digital pathology. Nat. Commun.15, 10685 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Chen, R. J. et al. Towards a general-purpose foundation model for computational pathology. Nat. Med.30, 850–862 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Vorontsov, E. et al. A foundation model for clinical-grade computational pathology and rare cancers detection. Nat. Med.30, 2924–2935 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Zimmermann, E. et al. Virchow2: scaling self-supervised mixed magnification models in pathology. Preprint at https://arxiv.org/abs/2408.00738 (2024).
  • 26.Xu, H. et al. A whole-slide foundation model for digital pathology from real-world data. Nature630, 181–188 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Shaikovski, G. et al. PRISM: a multi-modal generative foundation model for slide-level histopathology. Preprint at https://arxiv.org/abs/2405.10254 (2024).
  • 28.Ding, T. et al. A multimodal whole-slide foundation model for pathology. Nat. Med. 31, 3749–3761 (2025). [DOI] [PMC free article] [PubMed]
  • 29.Huang, Z., Bianchi, F., Yuksekgonul, M., Montine, T. J. & Zou, J. A visual–language foundation model for pathology image analysis using medical Twitter. Nat. Med.29, 2307–2316 (2023). [DOI] [PubMed] [Google Scholar]
  • 30.Lu, M. Y. et al. A visual–language foundation model for computational pathology. Nat. Med.30, 863–874 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Traag, V. A., Waltman, L. & van Eck, N. J. From Louvain to Leiden: guaranteeing well-connected communities. Sci. Rep.9, 5233 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Goldsborough, T. et al. InstanSeg: an embedding-based instance segmentation algorithm optimized for accurate, efficient and portable cell segmentation. Preprint at https://arxiv.org/abs/2408.15954 (2024).
  • 33.Stringer, C., Wang, T., Michaelos, M. & Pachitariu, M. Cellpose: a generalist algorithm for cellular segmentation. Nat. Methods18, 100–106 (2021). [DOI] [PubMed] [Google Scholar]
  • 34.Tommasino, C., Russo, C. & Rinaldi, A. M. NuLite—lightweight and fast model for nuclei instance segmentation and classification. Biomed. Signal Process. Control.114, 109333 (2026). [Google Scholar]
  • 35.Adjadj, B. et al. Towards Comprehensive cellular characterisation of H&E slides. Preprint at https://arxiv.org/abs/2508.09926 (2025).
  • 36.Ravi, N. et al. SAM 2: Segment Anything in Images and Videos. Preprint at https://arxiv.org/abs/2408.00714 (2024).
  • 37.Zhang, A., Jaume, G., Vaidya, A., Ding, T. & Mahmood, F. Accelerating data processing and benchmarking of AI models for pathology. Preprint at https://arxiv.org/abs/2502.06750 (2025).
  • 38.Marcolini, A. et al. histolab: a Python library for reproducible Digital Pathology preprocessing with automated testing. SoftwareX20, 101237 (2022). [Google Scholar]
  • 39.Filiot, A. et al. Distilling foundation models for robust and efficient models in digital pathology. Preprint at https://arxiv.org/abs/2501.16239 (2025).
  • 40.Zheng, Y. WSIs for breast cancer PDX models. Zenodo10.5281/zenodo.15497224 (2025).
  • 41.Zheng, Y. et al. rendeirolab/LazySlide: v0.10.0. Zenodo10.5281/zenodo.14939805 (2025).
  • 42.Zheng, Y. et al. rendeirolab/wsidata: v0.8.1. Zenodo10.5281/zenodo.17737183 (2025).

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Reporting Summary (1.6MB, pdf)
Peer Review File (1.6MB, pdf)

Data Availability Statement

GTEx WSIs and RNA expression data are available from the GTEx portal (https://gtexportal.org). WSIs of patient-derived xenograft models of breast cancer and their semantic segmentation are available via Zenodo at 10.5281/zenodo.15497223 (ref. 40).

Lazyslide and WSIData are publicly available via GitHub at https://github.com/RendeiroLab/LazySlide and https://github.com/RendeiroLab/WSIData, published on the Python package index (PyPI) and conda-forge, available via Zenodo at 10.5281/zenodo.14939805 (ref. 41) and 10.5281/zenodo.15083425 (ref. 42), and documented at https://lazyslide.readthedocs.io and https://wsidata.readthedocs.io. A companion repository on benchmarking is also publicly available via GitHub at https://github.com/RendeiroLab/LazySlide-benchmark.


Articles from Nature Methods are provided here courtesy of Nature Publishing Group

RESOURCES