Abstract
Multiplex imaging is a powerful tool to analyze the structural and functional states of cells in their morphological and pathological contexts. However, hypothesis testing with multiplex imaging data is a challenging task due to the extent and complexity of the information obtained. Various computational pipelines have been developed and validated to extract knowledge from specific imaging platforms. A common problem with customized pipelines is their reduced applicability across different imaging platforms: Every multiplex imaging technique exhibits platform-specific characteristics in terms of signal-to-noise ratio and acquisition artifacts that need to be accounted for to yield reliable and reproducible results. We propose a pixel classifier-based image preprocessing step that aims to minimize platform-dependency for all multiplex image analysis pipelines. Signal detection and noise reduction as well as artifact removal can be posed as a pixel classification problem in which all pixels in multiplex images can be assigned to two general classes of either I) signal of interest or II) artifacts and noise. The resulting feature representation maps contain pixel-scale representations of the input data, but exhibit significantly increased signal-to-noise ratios with normalized pixel values as output data. We demonstrate the validity of our proposed image preprocessing approach by comparing the results of two well-accepted and widely-used image analysis pipelines.
Author summary
Multiplex tissue imaging techniques are powerful tools increasingly used to characterize the structural and functional states of cells in situ. Each multiplex imaging platform exhibits unique characteristics such as signal-to-noise ratio and artifacts that need to be accounted for before to accurately analyze the data. Here, we present an image preprocessing framework that removes noise, artifacts, and platform-specific characteristics of the images and generates normalized, high-quality, and reproducible data for subsequent stages of analyses. Our platform-independent solution performs multiple tasks of denoising, artifact correction, and normalization using a single pixel classification step, eliminates the need for any further normalization process, and works across all tested multiplex imaging technologies.
Introduction
Multiplex tissue imaging technologies such as Multiplexed Ion Beam Imaging (MIBI) [1, 2], Imaging Mass Cytometry (IMC) [3], CO-Detection by indEXing (CODEX) [4, 5], Multiplexed Immunofluorescence (MxIF) [6, 7], and cyclic Immunofluorescence (cycIF) [8] provide researchers with a wealth of information on the single cell level that illustrates the complexity and heterogeneity of tissue samples. Comprehensive surveys providing an overview of these technologies are readily available [9, 10]. By preserving the spatial context of >40 markers measured simultaneously on a tissue, these advanced technologies have opened new avenues for biological discoveries in healthy and diseased microenvironments [11–19]. A substantial challenge remains, however, in developing accurate, robust, and automated computational pipelines for the analyses and interpretation of these complex high-dimensional imaging data.
A high-dimensional tissue image consists of a set of antibody-based visualizations of multiple parameters (markers) measured with fluorescence or mass spectrometry readouts. The signal intensity in each image channel is directly proportional to the expression level of the corresponding marker bound to its target. Since expression of specific marker combinations is the key determinant of cellular phenotypes, a primary step in the identification of cellular phenotypes is to accurately quantify signal intensities across channels. Accurate quantification of these signal intensities can be hindered by instrumental noise, acquisition artifacts, or experimental variability that are often incorporated in the imaging data during sample preparation and data acquisition. Therefore, most imaging platforms require denoising and artifact removal techniques prior to accurate quantification of signal intensities [2, 20–23]. Remaining artifacts may have far-reaching consequences in downstream analyses, potentially leading to inaccurate cellular identification and false conclusions in statistical comparisons or spatial analyses. Particularly for high-parameter imaging techniques that rely on unsupervised clustering algorithms for cell-type identification, high-quality images are required for robust and accurate quantitative outputs. The data must also be appropriately normalized to limit tissue and batch variability. Without a proper normalization, sample-to-sample intensity variations can cause cells to cluster by individual samples rather than by cell types [24, 25]. However, choosing an appropriate normalization method is a critical task, as the accuracy of cell-type identification has been shown to depend more on the choice of normalization approach than on the clustering algorithm [26]. Machine learning methods that can extract features less sensitive to intensity are therefore adopted [27]. Ideally, high-quality imaging data should satisfy the following criteria: 1) the data should exhibit high signal-to-noise ratios (SNR) and be as free as possible from artifacts; and 2) be appropriately normalized to remove non-biological signal variability within and across acquisition batches and tissues.
Various image processing and filtering methods can be used for denoising and artifact removal [2, 20–22, 28, 29]. For instance, variants of Gaussian filters are used for noise reduction [30–32]. Binary masks can be generated to specifically remove artifacts that differ from real signal by size and pixel distribution. Such techniques have been used for segmentation of microarray imaging data where binary masks are generated following pixel classification that separates the signal (spots in a microarray image) from background (noise and artifact) [33, 34]. It has been shown that such techniques can limit data loss in preprocessing of the mass cytometry data [34]. Recently, a five-step computational pipeline was proposed to prepare MIBI data for downstream analysis [2]. Steps in the pipeline include background subtraction, necrosis removal, batch normalization, denoising, and aggregate removal. In the first step, non-specific background is removed by subtracting counts from pixels where background noise is present. To specify those pixels with background noise, a binary mask (which is 1 where background exists and 0 elsewhere) is generated by filtering and thresholding a blank channel that is specifically included in the data acquisition process for this purpose. Both the thresholding parameter and the number of counts subtracted are determined manually based on visual inspection by the user. In the second step, artifacts such as areas of necrosis are removed by estimating another binary mask for the necrotic region using morphological opening and closing [2]. The number of counts subtracted and the thresholding parameters are decided by the user. Batch effects are then removed in the third step using quantile normalization. In the fourth step for denoising, a k-nearest-neighbor approach is used to generate a binary mask that separates the noisy pixels based on density. The parameter k and the thresholding parameter that separates the high-density pixels (signal) from the low-density pixels (noise) as well as the subtracting pixel counts are determined by the user. In the last step, the images are first smoothed using a Gaussian kernel and then binarized using Otsu’s method [35]. This five-step pipeline was recently wrapped in a graphical user interface called MAUI (Mass based imaging Analysis User Interface) [28].
While MAUI facilitates MIBI image preprocessing, it relies on accurate estimation of binary masks with user-defined thresholding parameters that are prone to bias. Thresholding methods separate two distinct features using a single value as the thresholding parameter. However, given the variabilities in complex multiplex tissue imaging data, a single threshold is unlikely to obtain a parameter configuration that perform well across all the images in a dataset. Additionally, estimating binary masks for each artifact in several sequential steps is labor-intensive and not scalable. Finally, new artifacts may appear as data are acquired in new tissue types or with new technologies, requiring additional steps to be added to the aforementioned five-step pipeline.
Here, we aim to create a simple automated framework that provides a single-step unified solution that works across imaging platforms. Our approach combines denoising and removal of various artifacts into a single pixel classification step, outputs a feature representation map (FR map) that eliminates the need for any further normalization process, does not rely on estimation of any thresholding parameters and therefore is more robust, and provides a unified solution across multiplex imaging platforms (both mass spectrometry and fluorescence-based imaging). We validate our proposed approach by comparing our results with two well-accepted and widely-used baselines: a) inForm software, a commercial software package for fluorescence-based multiplex imaging analyses and b) MAUI [28], a publicly available computational pipeline used to analyze MIBI data. All code and data have been made available as a resource for the research community [35].
Results
A single-step platform-independent framework for preprocessing of multiplex tissue imaging data
Our proposed framework for cellular phenotyping of multiplex tissue imaging data formulates denoising and artifact removal as a pixel classification problem in which the pixels in an image are classified into two classes (Fig 1):
class I: pixels with signal of interest (marker signal/positive signal).
class II: pixels with noise, artifacts, and other platform-specific properties that are not useful for downstream analyses.
The input of the classifier is a training set (labeled data) comprised of user-provided examples for both class I (desired signal) and class II (noise and artifacts) pixels. Features can be extracted by convolving various kernels with the image such as noise reduction filters, edge detectors, and texture kernels. Classifying algorithms, such as Random Forest [36], can then be used to classify each pixel given the extracted features. Several well-established and user-friendly tools have been developed for pixel classification such as Trainable WEKA Segmentation (TWS) [37], QuPath [38], or ilastik [39] and can be used to interactively label training data for signal and noise during image preprocessing. The output of the classifier consists of two FR maps, one for each class. The pixel values in each FR map represent the probability of that pixel belonging to the corresponding class [37]. Pixels with noise and artifacts have either zero or very low values in the FR map of class I, meaning FR maps of class I are free from noise and artifacts. In addition, the values in the FR maps range between 0 to 1. Therefore, the FR maps are normalized and this eliminates the need for further normalization processes.
Noise and artifact removal from multiplex imaging data
Image noise is generally defined as unwanted random variation that obscures the desired information in an image. Multiplex images are visualizations of bound antibodies, detected either by heavy metal ions conjugated to antibodies in mass spectrometry-based imaging or by fluorescent particles directly or indirectly associated with antibodies in light microscopy-based imaging. Thus, noise in multiplex imaging can be defined as any signal that differs from the biological structures to which antibodies are specifically bound [28]; and it can be classified into several categories: 1) channel crosstalk, 2) nonspecific antibody staining, and 3) aggregates.
Channel crosstalk or cross-channel contamination is the variable presence of signal from a contaminating channel in a target channel. For fluorescence microscopy imaging techniques, channel crosstalk is caused by the wide and often overlapping emissions spectra of fluorochromes. For mass-spectrometry imaging techniques, impurities in the heavy metal ion source used for antibody labeling, modification of the heavy metal ions (by hydrogenation, oxygenation, hydroxylation, etc), or for MIBI specifically, gold ions from the slide surface can introduce channel crosstalk to the image. For example, a low intensity contaminating signal that mirrors the structure of the real higher intensity signal in the hepatocyte antigen-145 channel (Fig 2A-top) is visible in CD20–161 channel (Fig 2A-middle). To correct this cross-talk, a few examples of the contaminating hepatocyte and real CD20 signals were used to train the pixel classifier and generate an FR map for the positive signal (Fig 2-bottom), in which the contaminating hepatocyte signal is removed. As demonstrated in the histograms (Fig 2A), the SNR is significantly enhanced in the FR map (green box) compared to the raw image (red box). That is, the values of pixels with positive signal in the FR map group to the far right of the histogram, whereas the rest of the pixels, including those with cross-talk contamination, group to the far left of the histogram and thus are cleanly separated. Another example of cross-talk contamination is gold background (Fig 2B) where signal from the exposed part of the gold slide (contaminating channel) is observed in the target channel, here Ki-67 (Fig 2B-left). Our approach allows easy removal of this artifact by training the pixel classifier with a few examples of contaminating pixels (pixels with low intensity that correlate with the bare gold slide) as well as the real Ki-67 signal. The contaminating artifact is completely removed in the resulting FR map (Fig 2B-right) and the Ki-67 signal is clearly separated from the noise pixels, as shown in the histograms.
Non-specific antibody binding is caused by cross-reactivity of antibodies, tissue features that non-specifically bind to antibodies, or instrumental noise. This artifact can appear as either random signal or as dim patterns superimposed on top of biological structures and may have correlation with the histological structure of the tissue [28]. Regardless of the source, the intensity and spatial distribution of this artifact is different from the true positive signal. Therefore, regions with signal from non-specific antibody binding can be easily labeled as undesired signal (class II) using our approach. An example of non-specific antibody staining is necrosis where a necrotic tissue region that exhibits non-specific signal in many channels appears in the pan-cytokeratin channel of a breast cancer tissue (Fig 2C-left). After pixel classification, the artifact is removed in the resulting FR map (Fig 2C-right) and the signal to noise ratio is enhanced, as shown in the histograms.
Aggregates are small high-intensity specks of signal caused by aggregation of antibodies, secondary reagents, or fluorochromes used in the staining process. Due to the concentration of high-intensity pixels, aggregates can be falsely interpreted as positive signal. Fig 2D shows an example of CD163 staining surrounded by noise and aggregates (left). To compare the impact of signal caused by aggregates, pseudo-cells of the same area were drawn by hand (610 pixels) and the mean expression of CD163 was calculated. While area A (m = 1.18) is clearly a CD163+ cell, the mean signal for area B (m = 0.43), which is likely a CD163+ cell, is difficult to distinguish from area C (m = 0.38) that contains positive signal due to an aggregate in the raw image (left). However, after removing noise and aggregates (Fig 2D-right), the mean signal for area C is greatly reduced (m = 0.02) while the mean signal for area B remains high (m = 0.34). Increased SNR after noise and aggregate removal is also demonstrated in the overlaid histograms, where the pixel intensities for each pseudo-cell are plotted for the raw image (top) and the FR map (bottom). Therefore, our approach can prevent identification of false positive cells from aggregate signal.
Mass based technologies generate pixelated imaging data with low SNR compared to fluorescence-based technologies. Another important advantage of our framework is that the sparse and pixelated signal of Mass-based imaging data (Fig 2D-left) is converted to a continuous signal with high SNR (Fig 2D-right). This property makes the processing of the such data a challenging task because the pixel intensity alone does not carry sufficient information to separate positive signal from noise. The pixel density should be considered along with the intensity [2]. That is, while area A (Fig 2D-left) can be effortlessly called a CD163+ cell given its consistent and high pixel intensity, the signal intensity in area B is much lower and it is the density of the pixels that contributes to forming a cell-like structure. The low intensity signal values of the individual pixels in area B are difficult to distinguish from noise in the raw image (see the corresponding histograms in Fig 2D). However, the spatial information of the pixels contributes to the extracted features in the FR map and thus the classification output. As a result, low-density and low-intensity pixels of noise have very low pixel values in the FR map, while high-density and low-intensity pixels of area B have high pixel values (see the Fig 2D-right and the corresponding histograms). For further illustration, two additional examples of converting mass-based imaging raw data from IMC and MIBI platforms to FR maps are shown Fig 2E and 2F.
Accurate phenotyping of tissue imaging data across imaging platforms
To further evaluate the performance of our image preprocessing framework, we compared the results of unsupervised clustering after denoising and artifacts correction using two previously published datasets collected by MIBI and Vectra Polaris. The MIBI data are a set of publicly available images from a triple-negative breast cancer cohort (TNBC) with 41 patients [2]. This dataset was analyzed by a computational pipeline [2] that was later wrapped into a user-friendly graphical interface, MAUI [28]. The Vectra dataset includes a subset of 6 patients with ovarian cancer collected in-house and analyzed using inForm [40, 41], a widely-used commercial software package for analyses of multiplex fluorescence imaging data.
We compared the output of our analysis framework with the MAUI pipeline using t-SNE plots comprised of about 200,000 cells from the MIBI breast cancer dataset (Fig 3A and Fig A in S1 Text). The overlaid heatmaps indicate the signal intensities for a subset of non-immune (Keratin6, CD31, SMA, Vimentin) and immune (CD45, CD3, CD4, CD8, CD20, CD68) markers measured from the raw images (top row) or FR maps (middle row). FR maps generated signal intensities ranging between 0 and 1 with improved SNR, as indicated by the more differentiated positive regions in the heat map. The scatter plots (bottom row) demonstrate the non-linear mapping of the average expression values per cell measured from the raw image (x-axis) and the FR maps (y-axis). The Spearman’s rank correlation coefficient confirms the presence of a monotonically increasing correlation between the level of expression measured on the raw image compared to the FR map. Additionally, it is noteworthy that as these values increase in the raw image, the values in the FR map level off. Fig B in S1 Text further illustrates this correlation specifically for positive cells of a given marker. We then compared the outcomes of unsupervised clustering (Fig 3B) using normalized marker expression scaled from zero to one (left) or measured from FR maps (right). To directly compare the two heatmaps and avoid discrepancies introduced by segmentation or clustering, we used the same single-cell segmentation maps and the same cellular phenotypes identified in the TNBC study [2]. Qualitative comparison of the two heatmaps confirms that similar cell populations can be identified using FR maps. For quantitative comparison, we applied the same unsupervised clustering algorithm, FlowSOM [42], to the single-cell information extracted using FR maps. The applicability of our analysis pipline is confirmed by the significant correlation (quantified by Pearson correlation coefficient) in the counts of different cell types identified on a patient-level in the TNBC study and using our framework (Fig 3C). Furthermore, there is a reasonable agreement between the predicted cell types in a comparison of individual cells (Fig A in S1 Text). The discrepancies between the predictions of the two pipelines can be further reduced by manual quality control of the clustering results.
Next, we analyzed ovarian cancer fluorescence imaging data from 6 patients (about 30,000 cells) using inForm software as a benchmark. To avoid any discrepancies caused by differences in the single-cell segmentation, we used the same cell segmentation maps produced by inForm. Fig 4A shows the signal intensities measured from the raw images (top row) or FR maps (middle row) overlaid on tSNE plots. Scatter plots (bottom row) visualize the non-linear mapping of the average expression values per cell measured from the raw image (x-axis) to the FR maps (y-axis) and demonstrate the correlation of signal intensities between the methods. We extracted single-cell information from the FR maps as input to FlowSOM [42] and identified tumor cells and four groups of immune cells. Fig 4B displays a cell-cell comparison of the cell types identified by inForm and our framework. The table entries represent the percentages of cells in the dataset, with the columns listing the identified cell types by the baseline (inForm), and the rows listing the identified cell types by our framework. The pipelines exhibit an agreement of approximately 85% (sum of diagonal entries) in identifying cellular phenotypes. Around 12% of the cells remain unidentified by inForm but are assigned to a cell type by our pipeline, which is indicated in red. Many of the cells in inForm’s unidentified cluster (Fig 4B-right) have measurable signal intensities that are more clearly defined in the measured counts from the FR map (Fig C in S1 Text). Therefore, our pipeline was better able to discern the phenotypes of these unidentified cells with enhanced SNR. To further investigate these discrepancies, we generated a color overlay of markers from the raw image and the corresponding FR map (Fig 4C-top). Comparing the pseudo-coloring of cell populations predicted by inForm and our pipeline (Fig 4C-bottom) with the protein expression on top, we can easily spot many CK+ tumor cells that are unassigned by inForm.
Discussion
A goal of image quantification pipelines is to remove noise and artifacts and normalize signal intensities to eliminate batch variations so that imaging data can be combined. The amount of manual curation, dependency on user-defined thresholding parameters, and lack of a unified solution that works across different imaging platforms slows analyses and hinders robust and reproducible results. We developed a framework that overcomes these challenges by replacing the multi-step low-level image processing with a single-step pixel classification. We classify each image such that all categories of undesired signal are placed in a different class from the desired marker signal and continue the downstream analysis using the generated feature representation map of the marker signal. This pipeline opens a new window for integrating data across multiplex imaging platforms and constructing and training generalizable cell-type annotators that can be used in a clinical study with patient and experimental variation.
Using FR maps instead of the raw images has several advantages beyond noise and artifact removal. First, the positive signal intensity values are mapped to normalized values between 0 and 1 in the FR map. Therefore, the artifact removal and normalization are combined into a single step and the normalization is completed at the image level in a very early stage of the analysis.
In addition to noise and artifacts, the platform-specific properties of images such as differences in signal-to-noise ratio, autofluorescence, and background staining are removed from the data. These platform-specific features do not carry useful information and only play confounding roles in downstream analysis. For instance, platform-specific features prevent single-cell segmentation models trained using MIBI data from generalizing on fluorescence-based imaging data [43]. Using our method, discretized and low-SNR signal of MIBI data is converted to continuous and high-SNR signal using FR maps, making the images more similar to fluorescence-based imaging data and making it easier to identify phenotypes. The general idea of pixel classification works across all imaging platforms and thus provides a unified solution for all the preprocessing required for robust cell clustering. Our platform-independent solution eliminates the need for alternative approaches that would otherwise be chosen according to which works best for each imaging technology. For instance, while fluorescence-based imaging data can be denoised using intensity-based methods, MIBI data require density-based methods for denoising.
Previous computational methods for denoising and artifact removal require substantial manual curation and parameter tuning and there is little consensus about which denoising pipeline produces high quality and reproducible results [44]. Many parameters of the algorithms used in the existing pipelines were tuned by hand and evaluated visually, leveraging the expert knowledge of pathologists and biologists. In addition, based on the level and composition of noise, custom-built approaches were optimized by the investigators for different imaging data [44]. As a result, those methods are prone to human bias and may lack reproducibility. While our proposed method requires expert knowledge for labeling the training data, it offers a unified solution for denoising MIBI data. This is achieved by redefining the problem as a pixel classification problem, where the final value of a pixel is determined by an algorithm that incorporates various imaging features extracted from the raw data. Unlike thresholding parameters that are inferred by the user, our approach eliminates the need for manual thresholding and instead relies on the algorithm’s decision-making process to determine the pixel values. Evaluation of functional markers can still be performed with manual thresholding of expression levels, however this process is easier and more robust using FR maps compared to thresholding the raw signal data.
Although advantageous in many ways, supervised pixel level classification is computationally expensive. For instance, generating FR maps for a 41-page image of 800 × 800 μm (2048 × 2048 pixels) of a highly-expressed marker (e.g. tumor cell markers) in the TNBC data took 140 minutes using an Intel Core i9 CPU-3.30GHz with 128 GB RAM. This time was measured to be less than 90 minutes for low-count markers (e.g. NK cell marker). The time required for labeling the training data should also be considered as 10–15 minutes for adding 20–100 annotations for positive and negative classes depending on the variability in different channels of a marker in a data set. For instance for labeling a marker like Foxp3, 10 annotations per class is sufficient for the classifier to remove the noise from the data. For more ubiquitous signals such as Beta-tubulin or Vimentin more examples are needed. While the run time of our proposed framework is relatively high compared to conventional image preprocessing pipelines, the significantly enhanced SNR reduces manual efforts required downstream for quality control.
Methods
Ovarian tumor sample preparation
Images from 12 previously described ovarian tumors were analyzed [41, 45]. Tumors were formalin-fixed, paraffin embedded (FFPE) and assembled in a tissue microarray (TMA) consisting of 2 mm cores Akoya Polaris instrument. TMA construction was approved by Colorado Multiple Institutional Review Board (#17–7788).
Computational framework
Data preparation: Multiplex imaging platforms commonly generate multi-page TIFF or OME TIFF as their primary output. Every page represents a region of interest recorded with a distinct optical or mass channel. For the purpose of pixel classification with our pipeline, a representative training set consisting of multi-page TIFF should be prepared. Every training set TIFF contains the information from a single optical or mass channel, but with representative regions of interest taken from the dataset. To generate a representative training set we use quantitative scores obtained from the CU-IScore scoring system. CU-IScore [46] is a scoring system developed in-house and publicly available, based on the IHC-Profiler method for quantitative evaluation and automated scoring of immunohistochemistry (IHC) images [47]. The CU-IScore algorithm assigns a score from 0–300 based on the pixel intensity and abundance, with higher scores indicating a higher SNR. By utilizing these quantitative scores, we ensure that our training stacks include images with a range of scores, creating a representative subset of the entire dataset for a given marker.
Pixel classifier training: Pixel classification has been extensively researched with versatile, interactive, and user-friendly tools available for the analysis of microscopic imaging data [37–39, 48, 49]. Therefore, we did not develop any tool for pixel classification and instead used TWS [37], a series of library methods that combines Fiji [49] and WEKA [50] and provides an interactive tool for extracting non-linear features and statistical properties of imaging data from user-provided examples. To use the TWS, two sets of inputs are needed from the user: a) multipage TIFF training sets that includes both examples or labels from class I (signal) and class II (noise and artifacts) for each channel of interest, and b) the manual selection of nonlinear features that should be used for classification: To generate label data for the pixel classification training, we provide annotations on the images of the training stack using the TWS interface. Notably, these annotations can be conveniently generated in a single step, encompassing positive signals categorized in class I, as well as all types of noise and artifacts classified within class II. A list of nonlinear features, their parameters, and information gain score that quantifies the importance of each feature in our classification as well as a visualization of the 5 top features are given, respectively, in Tables A and B and Fig D in S1 Text.
Pixel classifier evaluation: TWS provides library functions to evaluate the effectiveness of the extracted features in classification as well as the accuracy of used classification algorithms. When using the Random Forest (RF) algorithm for pixel classifcation, we maintain the out-of-bag (OOB) error as a performance measure below a certain threshold. Each decision tree within the RF algorithm is then trained using only a subset of the labeled data, with some samples held out from each tree’s training process. The OOB error provides an estimate of the model’s performance on unseen data. Additionally, we evaluate the model performance using multiple performance measures including areas under the receiver operating characteristic and the precision-recall curves to explore the sensitivity and specificity of the trained classifier. These built-in performance measures in WEKA package can guide the user in building an appropriate classifier.
Pixel classification of our datasets: Since the size of the image sets in both datasets we benchmarked is fairly limited, we did not generate a training subset and instead built a stack from all available regions of interest of a given channel. We note that this approach is only feasible when the dataset size is not very large. Furthermore, we maintained the out-of-bag error below a certain value (around 1%) during training with a Random Forest algorithm with 1000 decision trees. An additional visual inspection of the resulting FR maps was used to correct the pixel labels until optimal results were obtained.
We have developed a series of code that generates the required data structure for the classifier and integrates the output of the classifier with the single-cell segmentation results. Using the the FR maps and the cell segmentation maps, we generate a cell table with cells in row and summary statistics of expression as well as spatial information and morphological properties of cells in columns.
Data analysis
Ovarian tumor data collected by Vectra: using inForm software (Akoya Biosciences), the images were spectrally unmixed, individual cells were identified using DAPI+ nuclei, and the phenotyping algorithms were trained by marking over 100 cells as positive or negative for each of the phenotypic markers in the panel (CD3, CD8, CD68, CD19, and CK). The algorithms were applied to the entire dataset and the data were merged and consolidated in Phenoptr Reports, an open access software package by Akoya Biosciences. Using our proposed pipeline, we generated five stacks from the phenotypic markers where each stack contained 12 images. We then used the TWS tool [37] to train a pixel classifier for each stack and produced a set of FR maps for each marker. We selected Gaussian blur, Sobel filter, mean, median, and entropy kernels for noise reduction and texture filtering. Finally, we extracted single-cell expression values by measuring the signal counts on the FR maps of the positive class. To avoid any discrepancies caused by differences in the single-cell segmentation, we used the same cell segmentation maps produced by inForm. We used the extracted single-cell information as input to FlowSOM [42] and identified tumor cells and four groups of immune cells. The unidentified cluster includes cells with no positive signal for the present markers or cells with co-expression of mutually exclusive markers.
Breast cancer tissue collected by MIBI: We trained a Random Forest classifier using the following features: Gaussian blur, for denoising, mean, median, and entropy kernels for texture filtering, and Sobel filter and difference of Gaussians as edge detectors. We used the single-cell segmentation maps that were generated in TNBC study [2] to extract the single-cell information from our FR maps. Then we followed the exact hierarchical clustering scheme that was originally used to identify the cell types. In doing so, initially FlowSOM was used to cluster the cells into “Immune” and “Non-immune” using 16 markers (CD45, FoxP3, CD4, CD8, CD3, CD20, CD16, CD68, MPO, HLA-DR, Pan-Keratin, Keratin17, Keratin6, p53, Beta catenin, EGFR). Then, using 8 markers (Vimentin, SMA, CD31, Beta-catenin, EGFR, Keratin 17, Keratin 6, Pan-keratin) non-immune cells were clustered to into Epithelial, Mesenchymal, Endothelial and Unidentified. Immune cells were clustered into 12 groups using 13 markers (CD4, CD16, CD56, CD209, CD11c, CD68, CD8, CD3, CD20, HLA-DR, CD11b, MPO and FoxP3). We then manually labeled merged and labeled the clusters with different cell types.
Supporting information
Acknowledgments
Technical expertise was provided by the Human Immune Monitoring Shared Resource (RRID:SCR021985) within the University of Colorado Human Immunology and Immunotherapy Initiative and the University of Colorado Cancer Center.
Data Availability
The authors confirm that all data underlying the findings are fully available without restriction. This work describes two distinct datasets: breast cancer and ovarian cancer data. The breast cancer dataset, previously studied in Keren et al. [https://doi.org/10.1016/j.cell.2018.08.039], is publicly accessible at https://mibi-share.ionpath.com. Comprehensive information, including channel images, segmentation masks, and cell identities for the ovarian cancer data, can be obtained from https://github.com/himsr-lab/CU-PhenoNorm. The analysis code can also be found in the same GitHub repository.
Funding Statement
KRJ was supported by NIH (P30CA046934). BGB supported by DOD, Award OC170228 and the Kay L. Dunton Endowed Memorial Professorship In Ovarian Cancer Research. MA supported by NIH/NCATS Colorado CTSA Grant Number UL1 TR002535. EWYH was supported by the Lupus Research Alliance, Lupus Innovation Award. The funding agencies had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1. Angelo M, Bendall SC, Finck R, Hale MB, Hitzman C, Borowsky AD, et al. Multiplexed ion beam imaging of human breast tumors. Nature medicine. 2014;20(4):436–42. doi: 10.1038/nm.3488 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2. Keren L, Bosse M, Marquez D, Angoshtari R, Jain S, Varma S, et al. A structured tumor-immune microenvironment in triple negative breast cancer revealed by multiplexed ion beam imaging. Cell. 2018;174(6):1373–87. doi: 10.1016/j.cell.2018.08.039 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Giesen C, Wang HA, Schapiro D, Zivanovic N, Jacobs A, Hattendorf B, et al. Highly multiplexed imaging of tumor tissues with subcellular resolution by mass cytometry. Nature methods. 2014;11(4):417–22. doi: 10.1038/nmeth.2869 [DOI] [PubMed] [Google Scholar]
- 4. Goltsev Y, Samusik N, Kennedy-Darling J, Bhate S, Hale M, Vazquez G, et al. Deep profiling of mouse splenic architecture with CODEX multiplexed imaging. Cell. 2018;174(4):968–81. doi: 10.1016/j.cell.2018.07.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. Schürch CM, Bhate SS, Barlow GL, Phillips DJ, Noti L, Zlobec I, et al. Coordinated cellular neighborhoods orchestrate antitumoral immunity at the colorectal cancer invasive front. Cell. 2020;182(5):1341–59. doi: 10.1016/j.cell.2020.07.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Stack EC, Wang C, Roman KA, Hoyt CC. Multiplexed immunohistochemistry, imaging, and quantitation: a review, with an assessment of Tyramide signal amplification, multispectral imaging and multiplex analysis. Methods. 2014;70(1):46–58. doi: 10.1016/j.ymeth.2014.08.016 [DOI] [PubMed] [Google Scholar]
- 7. Lin JR, Fallahi-Sichani M, Sorger PK. Highly multiplexed imaging of single cells using a high-throughput cyclic immunofluorescence method. Nature communications. 2015;6(1):1–7. doi: 10.1038/ncomms9390 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Lin JR, Izar B, Wang S, Yapp C, Mei S, Shah PM, et al. Highly multiplexed immunofluorescence imaging of human tissues and tumors using t-CyCIF and conventional optical microscopes. Elife. 2018;7. doi: 10.7554/eLife.31657 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Parra ER, Francisco-Cruz A, Wistuba II. State-of-the-art of profiling immune contexture in the era of multiplexed staining and digital analysis to study paraffin tumor tissues. Cancers. 2019;11(2):247. doi: 10.3390/cancers11020247 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. van Dam S, Baars MJ, Vercoulen Y. Multiplex Tissue Imaging: Spatial Revelations in the Tumor Microenvironment. Cancers. 2022;14(13):3170. doi: 10.3390/cancers14133170 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Binnewies M, Roberts EW, Kersten K, Chan V, Fearon DF, Merad M, et al. Understanding the tumor immune microenvironment (TIME) for effective therapy. Nature medicine. 2018;24(5):541–50. doi: 10.1038/s41591-018-0014-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Tsujikawa T, Kumar S, Borkar RN, Azimi V, Thibault G, Chang YH, et al. Quantitative multiplex immunohistochemistry reveals myeloid-inflamed tumor-immune complexity associated with poor prognosis. Cell reports. 2017;19(1):203–17. doi: 10.1016/j.celrep.2017.03.037 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Gerdtsson E, Pore M, Thiele JA, Gerdtsson AS, Malihi PD, Nevarez R, et al. Multiplex protein detection on circulating tumor cells from liquid biopsies using imaging mass cytometry. Convergent science physical oncology. 2018;4(1):015002. doi: 10.1088/2057-1739/aaa013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Rovira-Clavé X, Jiang S, Bai Y, Zhu B, Barlow G, Bhate S, et al. Subcellular localization of biomolecules and drug distribution by high-definition ion beam imaging. Nature Communications. 2021;12(1):4628. doi: 10.1038/s41467-021-24822-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Chang Q, Ornatsky OI, Siddiqui I, Straus R, Baranov VI, Hedley DW. Biodistribution of cisplatin revealed by imaging mass cytometry identifies extensive collagen binding in tumor and normal tissues. Scientific reports. 2016;6(1):1–11. doi: 10.1038/srep36641 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Zhao Y, Uduman M, Siu JH, Tull TJ, Sanderson JD, Wu YCB, et al. Spatiotemporal segregation of human marginal zone and memory B cell populations in lymphoid tissue. Nature communications. 2018;9(1):1–15. doi: 10.1038/s41467-018-06089-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Li N, van Unen V, Abdelaal T, Guo N, Kasatskaya SA, Ladell K, et al. Memory CD4+ T cells are generated in the human fetal intestine. Nature immunology. 2019;20(3):301–12. doi: 10.1038/s41590-018-0294-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Snijder B, Pelkmans L. Origins of regulated cell-to-cell variability. Nature reviews Molecular cell biology. 2011;12(2):119–25. doi: 10.1038/nrm3044 [DOI] [PubMed] [Google Scholar]
- 19. Rani B, Cao Y, Malfettone A, Tomuleasa C, Fabregat I, Giannelli G. Role of the tissue microenvironment as a therapeutic target in hepatocellular carcinoma. World Journal of Gastroenterology: WJG. 2014;20(15):4128. doi: 10.3748/wjg.v20.i15.4128 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Chevrier S, Crowell HL, Zanotelli VR, Engler S, Robinson MD, Bodenmiller B. Compensation of signal spillover in suspension and imaging mass cytometry. Cell systems. 2018;6(5):612–20. doi: 10.1016/j.cels.2018.02.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Damond N, Engler S, Zanotelli VR, Schapiro D, Wasserfall CH, Kusmartseva I, et al. A map of human type 1 diabetes progression by imaging mass cytometry. Cell metabolism. 2019;29(3):755–68. doi: 10.1016/j.cmet.2018.11.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Wang YJ, Traum D, Schug J, Gao L, Liu C, Atkinson MA, et al. Multiplexed in situ imaging mass cytometry analysis of the human endocrine pancreas and immune system in type 1 diabetes. Cell metabolism. 2019;29(3):769–83. doi: 10.1016/j.cmet.2019.01.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Lu P, Oetjen KA, Bender DE, Ruzinova MB, Fisher DA, Shim KG, et al. IMC-Denoise: a content aware denoising pipeline to enhance Imaging Mass Cytometry. Nature communications. 2023;14(1):1601. doi: 10.1038/s41467-023-37123-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Harris CR, McKinley ET, Roland JT, Liu Q, Shrubsole MJ, Lau KS, et al. Quantifying and correcting slide-to-slide variation in multiplexed immunofluorescence images. Bioinformatics. 2022;38(6):1700–7. doi: 10.1093/bioinformatics/btab877 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Ahmed Raza SE, Langenkämper D, Sirinukunwattana K, Epstein D, Nattkemper TW, Rajpoot NM. Robust normalization protocols for multiplexed fluorescence bioimage analysis. BioData mining. 2016;9(1):1–13. doi: 10.1186/s13040-016-0088-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Hickey JW, Tan Y, Nolan GP, Goltsev Y. Strategies for accurate cell type identification in CODEX multiplexed imaging data. Frontiers in Immunology. 2021:3317. doi: 10.3389/fimmu.2021.727626 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Solorzano L, Wik L, Olsson Bontell T, Wang Y, Klemm AH, Öfverstedt J, et al. Machine learning for cell classification and neighborhood analysis in glioma tissue. Cytometry Part A. 2021;99(12):1176–86. doi: 10.1002/cyto.a.24467 [DOI] [PubMed] [Google Scholar]
- 28. Baranski A, Milo I, Greenbaum S, Oliveria JP, Mrdjen D, Angelo M, et al. MAUI (MBI Analysis User Interface)—An image processing pipeline for Multiplexed Mass Based Imaging. PLoS computational biology. 2021;17(4):e1008887. doi: 10.1371/journal.pcbi.1008887 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Jorge XP, Li X, Risom T, Zubkov A, Tabatsky E, Prilipko A, et al. SPEX: A modular end-to-end analytics tool for spatially resolved omics of tissues. bioRxiv. 2022. [Google Scholar]
- 30. Hamilton N. Quantification and its applications in fluorescent microscopy imaging. Traffic. 2009;10(8):951–61. doi: 10.1111/j.1600-0854.2009.00938.x [DOI] [PubMed] [Google Scholar]
- 31. Byun J, Verardo MR, Sumengen B, Lewis GP, Manjunath B, Fisher SK. Automated tool for the detection of cell nuclei in digital microscopic images: application to retinal images. Mol Vis. 2006;12(105-07):949–60. [PubMed] [Google Scholar]
- 32. Kong H, Akakin HC, Sarma SE. A generalized Laplacian of Gaussian filter for blob detection and its applications. IEEE transactions on cybernetics. 2013;43(6):1719–33. doi: 10.1109/TSMCB.2012.2228639 [DOI] [PubMed] [Google Scholar]
- 33. Giannakeas N, Karvelis PS, Exarchos TP, Kalatzis FG, Fotiadis DI. Segmentation of microarray images using pixel classification—Comparison with clustering-based methods. Computers in biology and medicine. 2013;43(6):705–16. doi: 10.1016/j.compbiomed.2013.03.003 [DOI] [PubMed] [Google Scholar]
- 34. Ijsselsteijn ME, Somarakis A, Lelieveldt BP, Höllt T, de Miranda NF. Semi-automated background removal limits data loss and normalizes imaging mass cytometry data. Cytometry Part A. 2021;99(12):1187–97. doi: 10.1002/cyto.a.24480 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Sezgin M, Sankur B. Survey over image thresholding techniques and quantitative performance evaluation. Journal of Electronic imaging. 2004;13(1):146–65. doi: 10.1117/1.1631315 [DOI] [Google Scholar]
- 36. Breiman L. Random forests. Machine learning. 2001;45(1):5–32. doi: 10.1023/A:1010933404324 [DOI] [Google Scholar]
- 37. Arganda-Carreras I, Kaynig V, Rueden C, Eliceiri KW, Schindelin J, Cardona A, et al. Trainable Weka Segmentation: a machine learning tool for microscopy pixel classification. Bioinformatics. 2017;33(15):2424–6. doi: 10.1093/bioinformatics/btx180 [DOI] [PubMed] [Google Scholar]
- 38. Bankhead P, Loughrey MB, Fernández JA, Dombrowski Y, McArt DG, Dunne PD, et al. QuPath: Open source software for digital pathology image analysis. Scientific reports. 2017;7(1):1–7. doi: 10.1038/s41598-017-17204-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Sommer C, Straehle C, Koethe U, Hamprecht FA. Ilastik: Interactive learning and segmentation toolkit. In: 2011 IEEE international symposium on biomedical imaging: From nano to macro. IEEE; 2011. p.230-3.
- 40. Steinhart B, Jordan KR, Bapat J, Post MD, Brubaker LW, Bitler BG, et al. The Spatial Context of Tumor-Infiltrating Immune Cells Associates with Improved Ovarian Cancer SurvivalSpatial Interactions in the Tumor Immune Microenvironment. Molecular Cancer Research. 2021;19(12):1973–9. doi: 10.1158/1541-7786.MCR-21-0411 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Jordan KR, Sikora MJ, Slansky JE, Minic A, Richer JK, Moroney MR, et al. The Capacity of the Ovarian Cancer Tumor Microenvironment to Integrate Inflammation Signaling Conveys a Shorter Disease-free Interval. Clinical Cancer Research. 2020;26(23):6362–73. doi: 10.1158/1078-0432.CCR-20-1762 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42. Van Gassen S, Callebaut B, Van Helden MJ, Lambrecht BN, Demeester P, Dhaene T, et al. FlowSOM: Using self-organizing maps for visualization and interpretation of cytometry data. Cytometry Part A. 2015;87(7):636–45. doi: 10.1002/cyto.a.22625 [DOI] [PubMed] [Google Scholar]
- 43. Greenwald NF, Miller G, Moen E, Kong A, Kagel A, Dougherty T, et al. Whole-cell segmentation of tissue images with human-level performance using large-scale data annotation and deep learning. Nature biotechnology. 2022;40(4):555–65. doi: 10.1038/s41587-021-01094-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44. Baharlou H, Canete NP, Cunningham AL, Harman AN, Patrick E. Mass cytometry imaging for the study of human diseases—applications and data analysis strategies. Frontiers in immunology. 2019;10:2657. doi: 10.3389/fimmu.2019.02657 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45. Watson ZL, Yamamoto TM, McMellen A, Kim H, Hughes CJ, Wheeler LJ, et al. Histone methyltransferases EHMT1 and EHMT2 (GLP/G9A) maintain PARP inhibitor resistance in high-grade serous ovarian carcinoma. Clinical epigenetics. 2019;11(1):1–16. doi: 10.1186/s13148-019-0758-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Rickert C, Jordan KR. GitHub; 2021. Available from: 10.5281/zenodo.4599591. [DOI]
- 47. Varghese F, Bukhari AB, Malhotra R, De A. IHC Profiler: an open source plugin for the quantitative evaluation and automated scoring of immunohistochemistry images of human tissue samples. PloS one. 2014;9(5):e96801. doi: 10.1371/journal.pone.0096801 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48. Berg S, Kutra D, Kroeger T, Straehle CN, Kausler BX, Haubold C, et al. Ilastik: interactive machine learning for (bio) image analysis. Nature Methods. 2019;16(12):1226–32. doi: 10.1038/s41592-019-0582-9 [DOI] [PubMed] [Google Scholar]
- 49. Schneider CA, Rasband WS, Eliceiri KW. NIH Image to ImageJ: 25 years of image analysis. Nature methods. 2012;9(7):671–5. doi: 10.1038/nmeth.2089 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50. Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH. The WEKA data mining software: an update. ACM SIGKDD explorations newsletter. 2009;11(1):10–8. doi: 10.1145/1656274.1656278 [DOI] [Google Scholar]