Skip to main content
PLOS One logoLink to PLOS One
. 2025 Dec 3;20(12):e0317865. doi: 10.1371/journal.pone.0317865

PixlMap: A generalisable pixel classifier for cellular phenotyping in multiplex immunofluorescence images

Rachel L Baird 1, David Mason 2, Kai Rakovic 1,3,4, Fiona Ballantyne 1, Ian R Powley 1, Anastasia Georgakopoulou 1,3, Lucy Hillary 3, Thomas G Bird 1,3,5, Leah Officer-Jones 1,*, John Le Quesne 1,3,4,*
Editor: Jordan Robin Yaron6
PMCID: PMC12674528  PMID: 41335594

Abstract

Multiplexed methods for the detection of protein expression generate extremely data-rich images of intact tissue sections. These images are invaluable for the quantification and analysis of complex biology and biomarker development. However, their interpretation presents a considerable analytical challenge. Cell segmentation from images is a key bottleneck and a major focus of research activity in artificial intelligence. Most current methods depend initially on the use of a nuclear counterstain to identify nuclear boundaries, which is a relatively straightforward task. The cellular boundary is then assigned either by expansion of the nuclear outline, or by the use of membrane or cytoplasm-specific stains to delineate cell boundaries, or by some combination of the two. The task is critical, as inaccurate segmentation leads to information loss and data contamination from neighbouring cells. Increasingly sophisticated methods are being developed to address these issues, but each has its own shortcomings. We present an alternative method which is inspired by the fact that the assignation of a cellular phenotype ‘by eye’ does not depend upon the accurate identification of cell boundaries. We present an easy-to-use deep learning-based cellular phenotyping method which leverages this human capacity to assign phenotypes without segmenting the entire cell, and which can accurately phenotype cells based on nuclear segmentation alone. Using human ground truth annotations of entire cellular regions, we developed a classifier leveraging the U-Net architecture within a commercially available deep learning image analysis platform, but the principle is transferrable to any deep-learning framework. Crucially, training requires only a single example of each compartmental stain (nuclear/cytoplasmic/membranous). The resulting algorithm assigns class identities to cells with nuclear labelling alone, without the need for whole cell expansion. The method is highly novel, broadly generalisable, and comparable in accuracy to intensity-based phenotyping methods, bridging the gap between inaccurate cellular segmentation and accurate phenotype generation.

Introduction

Chromogenic immunohistochemistry (IHC) remains a gold standard clinical pathology test, revealing the expression of single proteins in situ to diagnose disease and stratify patients. However, newer methods using multiple colour channels to simultaneously quantify many target molecules are becoming increasingly common. This is largely driven by the need to understand the cellular composition and spatial interactions within complex tissue environments, such as the tumour microenvironment (TME) in cancer, as well as cellular niches in other disease types. This has driven the development and commercialisation of multiplex methods including multiplex immunofluorescence (mIF), various RNA detection methods, and hybrid multimodal approaches. These allow researchers to interrogate these disease types at the cellular level uncovering spatial interactions [1] and previously unknown phenotypes [2] in intact tissue samples. Such methods generate extremely data-rich images, but most analytical pipelines rely heavily on accurate cell segmentation to generate meaningful results. Accurate segmentation of multiplex images presents many challenges, and when done poorly errors can lead to misclassification and reduced accuracy, for example by combining information in overlapping or adjacent cells, therefore creating false phenotypes. Automated image analysis approaches are becoming essential, as manual methods, while potentially more accurate, become more laborious, subjective and time-consuming as image complexity grows. Automation removes subjectivity and the opportunity for human error, improving consistency whilst increasing throughput.

The current de facto standard approach used by almost all cell segmentation strategies relies upon the detection of a nuclear outline, typically achieved using a DNA stain such as DAPI, followed by an expansion step to include as much of the surrounding cytoplasmic area as possible. This permits the calculation of whole cell mean intensity per image channel. These values would then typically be used for cellular phenotyping, sorting cells into lineage categories by thresholding, clustering, or alternative computational techniques. Regardless of whether an intensity threshold or clustering technique is used to assign phenotypes [3] the foundational input remains the mean intensity per biomarker per cell. There have been various attempts to improve the accuracy of cell detection, including spillover correction [4], combinations of markers for cell segmentation [5,6] and the introduction of deep learning methods, which have led to the publication of numerous algorithms [79]. Cell segmentation using antibody/probe cocktails, although offering improved segmentation in several cell types, does not yet offer a universal solution. Currently the most effective methods combine several approaches, though optimal solutions typically need to be tailored to each experiment.

We have developed a novel deep learning approach, PixlMap, to assign accurate cellular phenotypes in a way which requires only nuclear segmentation without the need for whole cell segmentation at all. Whilst our bespoke analysis protocol package (APP) uses the deep learning tools in the Visiopharm digital image analysis platform, the same principle can be applied within any sufficiently powerful framework. Our method has been trained to recognise patterns of positivity for each of the major cellular compartments, i.e., cytoplasm, surface membrane, and nucleus, and assigns a resultant likelihood of belonging to a particular cell class to each individual pixel, based upon the broader local context. This permits accurate cellular phenotyping from assessment of just a small number of pixels within each cell, eliminating the need for accurate cell boundary determination and minimising even the requirement for accurate nucleus boundary segmentation.

Materials and methods

Multiplex immunofluorescence staining and image collection

Human samples.

The LATTICe cohort was originally constructed under REC 14/EM/1159 (East Midlands REC), and the ongoing management of the resource by the Greater Glasgow and Clyde Biorepository was approved under an amendment granted by the Leicester South REC. Ongoing use of the collection is now managed under REC 16/WS/0207, biorepository approval number 644.

Multiplex images of lung adenocarcinoma (LUAD) tissue microarray (TMA) samples from the well-characterised LATTICe-A cohort [10] were used to train the deep learning classifier (Panels 1–4 in S1 Table). All multiplex assays were optimised on relevant control tissues using the Roche Ventana Discovery ULTRA (Roche Tissue Diagnostics). 4 µm thick sections were stained with multiplex immunofluorescent (mIF) panels as previously described [1115]. Detailed information on the mIF assays and antibodies is provided in Supplementary S1 Table Individual core images in each mIF panel were collected at 20x magnification (except for panel 7 in S1 Table where core images were collected at 40x magnification) using the Akoya PhenoImager HT multispectral slide scanner (Akoya Biosciences) and spectrally unmixed using Inform software (Akoya Biosciences version 2.6.0).

For the deep learning training substrate, single fluorescent channels from panels 1–4 in S1 Table were extracted from mIF component images using QuPath (version 0.4.x) and individual channel images were subsequently imported into Visiopharm.

Murine samples.

All animal experiments were conducted in accordance with UK Home Office licences (70/8891, PP0604995, 70/8646, 70/8468, and PP8854860) and in accordance with UK Animal (Scientific Procedures) Act 1986 and EU direction 2010. They were subject to review by the animal welfare and ethical review board of the University of Glasgow and the University of Newcastle upon Tyne. ARRIVE guidelines were followed for reporting of animal experiments, as previously described [16]. Single use of needles and non-adverse handling techniques were employed to minimize distress, pain and suffering. Mice were housed in a pathogen-free environment under standard conditions (19–22°C, 45–65% humidity, and 12 hours light-dark cycle) with access to food and water ad libitum. Environmental enrichment, including shredded paper, cardboard houses, wooden chews, and tunnels, was provided in all cages to encourage burrowing, natural hiding, and nesting behaviours. The following transgenic mouse strains were used: Ctnnb1tm1Mmt (Ctnnb1ex3) [17], Gt(ROSA)26Sortm1(MYC)Djmy (R26LSL-MYC) [18]. Following ear-notch for identification purposes at weaning (3 weeks of age), ear tissue was sent to Transnetyx, Inc (Cordova, TN 38016) for genotyping using real-time PCR assay.

Male Ctnnb1ex3/wt; R26LSL-MYC/LSL-MYC mice of mixed background were induced between 8–12 weeks of age, with AAV8.TBG.PI.Cre.rBG (AAV8-TBG-Cre) (Addgene, #107787-AAV8) as described previously [19]. Virus was diluted in sterile PBS to the desired working titre (6.4*10^8 GC/ml) and 100μl was administered to each mouse vial tail vein injection. Mice were sampled as day 120 time-point. All mice were humanely euthanised using CO2 inhalation followed by cervical dislocation. Liver tissue was collected and placed in 10% neutral buffered formalin (NBF). After 24 hours, formalin was replaced with 70% ethanol prior to tissue processing and embedding in paraffin blocks.

4 µm thick murine sample sections were prepared and stained (Panel 5, S1 Table) as previously described [1114]. Individual liver lobe regions were collected at 20x magnification using the Akoya PhenoImager HT multispectral slide scanner (Akoya Biosciences) and spectrally unmixed using Inform software (Akoya Biosciences version 2.6.0).

Image analysis

All image analysis was carried out in Visiopharm (version 2022.12.0.12865) using a physical Supermicro server running Windows 2016 Datacentre, with dual AMD EPYC 7713 processors, 256 GB RAM, Sabrent Rocket Q 8TB fast disk and four NVIDIA Quadro RTX 5000 GPU cards.

Deep learning training.

PixlMap was trained on single-channel information extracted from mIF panels using the workflow shown in Fig 1. Training was performed on a virtualised server running Windows server 2016 Datacentre with a grid NVIDIA Tesla V100-PCIE-32Gb card Virtualised processors providing 32 cores and 64GB RAM. The magnification for training was set at 20x magnification at 0.5 µm per pixel. Ground truth regions were manually annotated using freehand image segmentation on single-channel images giving labelled examples of positivity for each cellular compartment, with negative regions labelled as background. The total training set included 26 training regions from 26 different LUAD patient cores, from 4 separate mIF panels. Markers used to train each cellular compartment were as follows: pan-cytokeratin (AE1/AE3), RPS6 (5G10), and p-eIF2a (D9G8) for cytoplasmic labelling; PD-L1 (E1L2N), CAIX (poly), CD4 (SP35), and CD8 (C8/144B) for membranous labelling; and FOXP3 (236A/E7) and Ki-67 (30−9) for nuclear labelling. The chosen markers were selected based on their robust and compartment-specific staining properties, ensuring high confidence in annotation. Additionally, LUAD was selected as a training substrate due to its well-characterised tumour microenvironment and availability of multiplexed data. All annotated training images were verified by a pathologist (KR, JLQ) prior to training the classifier.

Fig 1. Deep learning training pipeline.

Fig 1

A schematic map showing the deep learning training pipeline. The training set comprised regions from 26 different lung adenocarcinoma patient cores, from 4 separate experimental mIF panels. Markers used to train each cellular compartment were as follows: pan-cytokeratin, RPS6, and p-eIF2a for cytoplasmic labelling; PD-L1, CAIX, CD4, and CD8 for membranous labelling; and FOXP3, and Ki-67 for nuclear labelling. All training labels were verified by a pathologist prior to training.

Using our fully annotated training set, we developed a new Visiopharm APP using the deep learning module with the U-Net architecture. Due to all images in the training set containing just one feature, the architecture of the classifier was set to use 1 input channel. The size of the receptive field sent to the training network was set to 512 x 512 pixels (1 pixel = 0.5 µm2). The training parameters included a learning rate of 1.0e-06 and a mini batch size of 2 to increase stepwise training accuracy without sacrificing speed. The APP underwent 539,000 iterations for training with the freeze depth being adjusted every 100,000 iterations for enhanced refinement. The final model achieved a loss function value of 0.035. The trained classifier outputs two features: Positive and Background and the resulting phenotypic heatmaps represent the probability that each pixel belongs to a specific cellular phenotype, with white areas indicating a high probability of belonging to a given phenotype. These probability values can be dynamically thresholded for downstream phenotyping analysis, enabling robust and adaptable cellular classification.

Nuclear segmentation.

PixlMap was applied to previously classified images that were segmented into labelled nuclei with associated X/Y coordinates. Nucleus labelling was generated by re-training the pre-trained APP developed by Visiopharm (Nuclei Detection, AI (Fluorescence), version 2021.02.0.9284). To ensure accuracy, we manually annotated nuclei and nuclear boundaries on 1 mm TMA core images (n = 12 patient cores), providing additional ground truth data for training. The APP was further refined with 255,000 additional training iterations. This re-trained APP was subsequently applied to all images used in this study, ensuring consistent nucleus detection across the dataset.

Assigning Phenotypes Using PixlMap Probability Heatmaps.

Nuclei were assigned to phenotypes based on probability heatmaps generated from PixlMap. We created a version of PixlMap for each phenotype that we wished to assign. Each of these APPs assigns nuclei phenotype labels (positive or negative) based on their heatmap intensity for marker of interest, using post-processing steps within Visiopharm. We defined an object as positive if 80–100% of the object’s pixels had a positive probability of between 0.5 and 1.0, corresponding to a confidence level of 50–100%. The decision to set a threshold of 80% coverage or higher was made by observing the automated classifier against the ground truth. To accommodate mIF data, a separate APP was developed for each channel. While this modular approach creates separate APPs for each channel, its scalability allows for simultaneous analysis of multiple markers, for example, for a phenotyping project with five markers, this approach generates five distinct APPs, allowing independent evaluation of each marker.

Determination of object sizes for accurate phenotyping.

In theory, cellular phenotypes could be assigned to heatmap objects of any size, from larger than an entire cell down to a single pixel in the centre of the nucleus. Crucially, our method assigns cell phenotype probabilities based on deep learning interpretation of 512 x 512 pixel (256 x 256 µm) regions to single pixels, so that a single pixel contains information about the surrounding region. This transference of information to the centre explains the ability of nuclear pixels in the centre of the cell image in the heatmap to provide information about cytoplasmic or membranous signals.

The smaller the object used to represent each cell, the less the computational cost, but the less heatmap information will be used. Optimal object size will ensure that phenotype assignment is maximally accurate and efficient.

Therefore, to determine the minimum object size for accurate phenotype assignment, we compared 4 different nuclear outline sizes to a ground truth dataset. A nuclear detection APP was applied within a region of interest of size 29,409 µm2 to generate nuclear labels on a human LUAD image from mIF panel 2 (S1 Table). Each nuclear label was then outlined with an annotation to generate consistent nuclear X/Y coordinates for subsequent classification. The labelled image was assessed by a human pathologist to generate ground truth cytokeratin positivity, and 5 separate automated analyses were applied.

The five test groups were as follows:

  1. Group 1: Whole nuclear outlines generated by the APP.

  2. Group 2: Nuclear labels dilated by 3 pixels to include surrounding cytoplasm.

  3. Group 3: Nuclear labels dilated by 8 pixels to capture additional cytoplasmic regions.

  4. Group 4: Nuclear labels reduced to a single centre-of-mass pixel.

  5. Group 5: Nuclear labels reduced to a single pixel centre-of-mass pixel and dilated by 3 pixels.

Nuclear dilation was performed during the post-processing stage of the Visiopharm APP design. Using the ‘dilate’ tool in post-processing steps, we specified the number of pixels by which each nucleus would be expanded in order to create the corresponding test group labels. Labels in all 5 groups were then changed to a new positive label based on the pan-cytokeratin probability heatmap from PixlMap when the probability exceeded 0.5 as described above. Results for all methods were compared to ground truth.

Benchmarking.

To benchmark our method against the existing widely accepted phenotype methods, we compared PixlMap to the ‘gold standard’ manual assignment of phenotype identities by a pathologist, and to mean intensity-based phenotype generation which is a method built in to most commercially available platforms [20,21]. We applied a random region of interest (84,394 µm2) to 16 LUAD core images of a 6-plex immune phenotyping assay (S1 Table Panel 2). We then used a nuclear detection APP as described above to generate nuclear labels (total n=9409) within all regions. All 16 images, including their nuclear overlays, were copied into their corresponding comparison folders (ground truth, mean intensity, and PixlMap) for phenotype assignment. For the ground truth dataset, we manually assigned phenotypes to cells according to their lineage-defining markers (pan-cytokeratin, CD8, CD4, CD68 or SMA) which was further reviewed by our expert multi-disciplinary tissue imaging team and a pathologist. Next, for each nuclear label, under section’output variables’ in the APP design, we generated mean intensity measures for each marker and a pathologist assigned a positive staining threshold by manual assessment of images to create the mean intensity dataset. All cells below the appropriate threshold were classed as negative. We also generated mean probabilities using PixlMap heatmaps at 10%, 25%, 50%, 75%, and 90% confidences, to create the PixlMap dataset. X/Y coordinates for each nucleus were generated in all 3 datasets and used for performance metric analysis. Where PixlMap generated an incompatible phenotype, the cell was reclassified as ‘unclassifiable’. Weighted performance metrics (F1 score, accuracy, precision, recall, AUC-ROC) and confusion matrices were calculated using scikit-learn v1.4.0 [22] and displayed with seaborn v0.13.2 [23] in python v3.10. The weighted F1 score, accuracy, precision, and recall provide a balanced evaluation of classification performance, accounting for potential class imbalances. AUC-ROC further quantifies PixlMap’s discriminative power across phenotypes.

Results

We trained PixlMap on regional staining information for each cellular compartment, using compartment-specific staining patterns to generate training labels (Fig 2). The training set included manually annotated examples of cytoplasmic, membranous, and nuclear staining patterns. For membranous and cytoplasmic markers, entire regions containing positive cells were labelled, including nuclei no matter their expression intensity; in this way, the channel positivity status of the relevant extranuclear compartment was applied to the nuclear area as well as the rest of the cell. For nuclear markers, only the nuclei themselves were labelled. By focusing on nuclear outlines, PixlMap simplifies phenotype assignment by removing the requirement for complex whole-cell segmentation, while still capturing staining information relevant to cellular classification. This methodology bypasses the computational complexity and potential inaccuracies associated with traditional whole-cell segmentation, enabling more robust and efficient cellular phenotyping.

Fig 2. Training PixlMap on human lung adenocarcinoma cores.

Fig 2

Examples of the training labels used to train PixlMap with each compartmental stain: cytoplasmic (pan-cytokeratin), membranous (PD-L1), and nuclear pattern (FOXP3), respectively from left to right (top), shown next to their corresponding manually annotated labels (bottom). Each training label highlights positive regions, and unlabelled regions were classed as background. Training labels were generated from single channel images generated from mIF panels 1-4 (S1 Table). Scale bar = 50 µm.

Once PixlMap had shown stabilised deep learning convergence, we assessed classifier performance on test mIF images. PixlMap detected regions of positivity for all the markers within mIF images, providing accurate heatmaps which identify regions of positivity for cytoplasmic and membranous markers (Fig 3a and b, respectively) as well as positive nuclei (Fig 3c). Phenotypes are assigned by first applying a nuclear detection APP to generate nuclear labels before using the positivity heatmap from PixlMap to assign phenotypes to each label with specified confidence limits.

Fig 3. PixlMap extracts probability heatmaps from multiplex images to classify cells.

Fig 3

A. PixlMap identifies nuclei in cells with cytoplasmic staining. I.) A mIF image (panel 2 S1 Table) of human LUAD containing markers CD4 (cyan), CD8 (yellow), Ki-67 (magenta), CD68 (red), pan-cytokeratin (green), and smooth muscle actin (Orange), with a DAPI counterstain (blue). II.) Pan-cytokeratin channel of interest. III.) Pan-cytokeratin pixel probability heatmap (white, positive = 1, black, negative = 0). IV.) Nuclear labels. V.) Positive (magenta), and negative (blue) nuclear labels using PixlMap pan-cytokeratin heatmap, with a 50% confidence. Scale bar = 50 µm. B. PixlMap identifies nuclei in cells with positive membranous staining. I.) A mIF image (panel 7 in S1 Table) of human LUAD containing markers UBF1 (light blue), Calnexin (light pink), CD98/ATP1α cocktail (magenta), GLB1 (cyan), RPS6 (orange), GOLG1 (red), ATP5A (yellow), pan-cytokeratin (not shown), and a DAPI counterstain (blue). II.) CD98/ATP1α channel of interest. III.) CD98/ATP1α cocktail pixel probability heatmap (white, positive = 1, black, negative = 0). IV.) Nuclear labels. V.) Positive (magenta), and negative (blue) nuclear labels using PixlMap CD98/ATP1α heatmap, with a 50% confidence. Scale bar = 50 µm. C. PixlMap identifies positive nuclei. I.) A mIF image (panel 2 in S1 Table) of human LUAD containing markers CD4 (cyan), CD8 (yellow), Ki-67 (magenta), CD68 (red), pan-cytokeratin (green), and smooth muscle actin (Orange), with a DAPI counterstain. II.) Ki-67 channel of interest. III.) Ki-67pixel probability heatmap (white, positive = 1, black, negative = 0). IV.) Nuclear labels. V.) Positive (magenta), and negative (blue) nuclear labels using PixlMap Ki-67 heatmap, with a 50% confidence. Scale bar = 50 µm.

We set out to gauge the optimum label requirement size to correctly classify pan-cytokeratin positive cells by comparing 5 test groups to manually classify pan-cytokeratin positive cells (Fig 4). Test groups were as follows (total cells n = 401 per test group): Group 1 included nuclear-sized labels only; group 2 included nuclear labels + 3-pixel expansion; group 3 included nuclear labels +8-pixel expansion; group 4 included single-pixel in the centre of the nucleus, and group 5 included a 3-pixel-sized nucleus. We found that whole nuclear labelling was optimal for phenotype assignment (Fig 4E) and was comparable to the ground truth group with the most precision and accuracy, compared with the other groups (Fig 4F-J). Therefore, from this point on, we used whole nucleus labels for analysis.

Fig 4. PixlMap can correctly classify cells without the need for whole cell expansion.

Fig 4

A. A mIF image of panel 2 in S1 Table showing only pan-cytokeratin (green) and DAPI (blue) (left), and PixlMap’s pan-cytokeratin probability heatmap from the mIF image (right). Scale bar = 50 µm. B. A nuclear detection APP was applied in a default region of interest (29,409 µm2) Nuclei (blue) were outlined with an annotation (magenta) for X/Y calculations to ensure each X/Y coordinate stayed the same in all test groups. The image was then copied with its layer data into different folders for each test group (n = 5). Scale bar = 50 µm. C. Ground truth group: blue labels were manually assigned a pan-cytokeratin-positive label (yellow) by a pathologist (n = 121 positive, n = 280 negative). Scale bar = 50 µm. D. Comparison groups: blue labels in each comparison group (top) were changed to a pan-cytokeratin-positive label (yellow) (bottom) using the PixlMap pan-cytokeratin heatmap with 50% confidence (0.5): nuclear-size (left), nuclear-size + 3 pixels (middle left), nuclear-size + 8 pixels (middle), single pixel-sized (middle right), 3-pixel nucleus (far right). Scale bar = 50 µm. E-I. Confusion matrices comparing ground truth pan-cytokeratin assignment to PixlMap predictions by nuclear label size. Nuclear size only, nuclear label +3 pixels, nuclear label +8 pixels, 1-pixel label, and a 3-pixel label, respectively. Percentages normalised to predications. J. Bar plot showing weighted performance statistics for each label size (weighted F1 score, accuracy, precision, recall).

We next tested the ability of PixlMap to correctly assign phenotypes based on multiple markers expressed in the same compartment of different cell types (Fig 5a). PixlMap can extract probability heatmaps for CD4 and CD8 and use them to assign phenotypes, highlighting that PixlMap can distinguish between these T-cell types despite their identical morphologies. Furthermore, PixlMap can correctly classify cells with slightly different morphologies from those that were included in the training set. For example, CD68 membrane positivity in macrophages has a quite distinct pattern from all training cell types. Despite this, PixlMap was able to accurately phenotype CD68 positive cells (Fig 5b).

Fig 5. PixlMap can discriminate between phenotypes of various morphologies.

Fig 5

A. mIF panel 2 in S1 Table in human LUAD with CD4 (magenta), CD8 (yellow), and DAPI (blue) channels turned on only for visual (left). Middle shows the CD4 and CD8 pixel probability heatmaps extracted from the multiplex image by PixlMap (white, positive = 1, black, negative = 0). Right shows the nuclear overlays (blue) where nuclei were changed to a positive (yellow) label based on the corresponding CD4 (top) and CD8 (bottom) PixlMap probability heatmaps with 50% confidence (0.5). Scale bar = 50 µm. B. Highlights mIF panel 2 in S1 Table but with the pan-cytokeratin (green), CD68 (red), and DAPI (blue) channels turned on only for visual (left). Middle depicts the CD68 pixel probability heatmap extracted from the mIF image by PixlMap (white positive = 1, black, negative = 0). Left highlights the nuclear overlays (blue) were changed to a positive (yellow) label based on the corresponding CD68 probability heatmap with 50% confidence (0.5). Scale bar = 50 µm.

To further validate PixlMap against ground truth, we focussed on lineages defined by single markers. We compared manual annotations from a pathologist to PixlMap predictions of the following cell types: CD4-positive T-cells, CD8-positive T-cells, CD68-positive macrophages, pan-cytokeratin-positive epithelial cells, and SMA-positive myofibroblasts in human LUAD (S1 Fig 1) using panel 2 (S1 Table). We noted that PixlMap failed to accurately identify SMA positive cells, so we focussed on measuring performance on non-spindled cells for the remainder of this task, i.e., using the same markers in S1 Fig 1, but excluding SMA in the analysis (Fig 6). Our results further justify using PixlMap to assign class identities to labels of nuclear size, as it was highly comparable to manual phenotyping of these cells with multiple confidence limits (Fig 6A-F). Next, we compared PixlMap and mean intensity-based phenotyping and compared the respective outputs to the same ground truth labels (Fig 6H). Our results highlight that we can phenotype cells more accurately using PixlMap with nucleus segmentation only, than with the standard whole cell expansion-mean intensity method for phenotyping.

Fig 6. PixlMap phenotyping is comparable to pathologist ground truth.

Fig 6

A-E. Confusion matrices comparing ground truth assignment of pan-cytokeratin (CK), CD4, CD8, and CD68 to PixlMap predictions by confidence thresholds. Percentages normalised to predications. F. Bar plot showing weighted performance statistics for single lineage phenotyping for CK, CD4, CD8, and CD68 (weighted F1 score, accuracy, precision, recall). G. AUC-ROC comparing PixlMap to ground truth single predications for CK, CD4, CD8, and CD68. H. Confusion matrix comparing ground truth manual annotations provided by a pathologist and threshold-based phenotyping using mean pixel intensities of each marker. True = ground truth and predicted is thresholding. Percentages normalised to predications.

To evaluate whether our APP would work across different species and tissue types, we next tested PixlMap’s human tissue-trained phenotyping capabilities on murine mIF images (Fig 7). We ran a nuclear detection APP on images from an immunofluorescent panel (Fig 7A-D) (S1 Table, panel 5) of murine liver samples (n = 6), comprising of a total of 6208 cells. Nuclear labels (blue) were reclassified to a positive label (magenta) using PixlMap’s heatmap when the probability exceeded 0.5 (i.e., 50%) for the respective phenotypic input. Given PixlMap’s success in assigning single lineage phenotypes to cells in the human LUAD panel, we extended its application to murine liver samples to evaluate its performance in assigning single lineage phenotypes in this new context. Interestingly, despite being trained on human tumour tissue, we found that PixlMap can extract accurate positivity heatmaps for single lineage markers (CD4, CD8, and GS), and such heatmaps can be used to correctly assign class identities with high confidence compared to human ground truth assignment (Fig 7E-J).

Fig 7. PixlMap can extract probability heatmaps from murine liver samples.

Fig 7

A. PixlMap can extract probability heatmaps and classify cells in murine mIF images despite only being trained with human tissue images. The mIF panel (panel 5 in S1 Table) shown consists of GS (green), FOXP3 (not present), CD8 (red), Granzyme B (orange), CD45 (lilac), CD4 (cyan), c-Myc (magenta), β-Catenin (white), and a DAPI counterstain (blue). Scale bar = 50 µm. B. CD4 I.) Channel of interest II.) PixlMap heatmap III.) Nuclear labels IV.) CD4 positive (magenta), and CD4 negative (blue) based on a 50% PixlMap confidence. Scale bar = 50 µm. C. CD8 I.) Channel of interest II.) PixlMap heatmap III.) Nuclear labels IV.) CD8 positive (magenta), and CD8 negative (blue) based on a 50% PixlMap confidence. Scale bar = 50 µm. D. GS I.) Channel of interest II.) PixlMap heatmap III.) Nuclear labels IV.) GS positive (magenta), and GS negative (blue) based on a 50% PixlMap confidence. Scale bar = 50 µm. E-I. Confusion matrices comparing ground truth assignment of CD4, CD8, and GS to PixlMap predictions by confidence thresholds. Percentages normalised to predications. J. Bar plot showing weighted performance statistics for single lineage markers CD4, CD8, and GS (weighted F1 score, accuracy, precision, recall).

We revisited the spindled cell population to further explore PixlMap’s utility on cells with this morphology. We predict that PixlMap’s lack of ability to phenotype SMA positive cells is due to spindle-shaped cells being absent in the training data and therefore further explored this concept. We have shown that PixlMap has a limited capability to correctly extract probability heatmaps from fibroblast markers in a fibroblast-specific mIF panel (S1 Table Panel 6); HLA-DR, FAP, Podoplanin, and IL-6 (Fig 8C I-IV), and will need further training for SMA heatmap extraction (Fig 8C V). Therefore, additional training to include different fibroblast morphologies will be required for accurate fibroblast phenotype assignment from PixlMap.

Fig 8. PixlMap requires additional fibroblast morphology training for accurate fibroblast phenotyping.

Fig 8

A. mIF panel 6 in S1 Table with original mIF image of human LUAD with FAP (red), HLA-DR (cyan), IL6 (magenta), Podoplanin (yellow), SMA (orange), pan-cytokeratin (green), and DAPI (blue). Scale bar = 50 µm. B. I-V.) Original mIF channels showing FAP (red), HLA-DR (cyan), IL6 (magenta), Podoplanin (yellow), and SMA (orange), respectively, with DAPI (blue). Scale bar = 50 µm. C. PixlMap probability feature heatmap for the corresponding channels above in B as an input (white positive = 1, black, negative = 0). Scale bar = 50 µm.

In summary we have developed a method which can easily assign single lineage phenotype cell classes to circular cells. PixlMap offers a more efficient, accurate, and scalable method for the phenotyping of these cells, reducing the complexity of traditional segmentation approaches while maintaining robust performance.

Discussion

Accurate classification of cells in mIF images is essential to unlock the full wealth of biological information that they contain. Current phenotyping methods rely heavily on accurate cellular segmentation, and while methods to improve this continue to evolve, there is no universally trusted segmentation method. There are numerous factors affecting the quality of tissue, and, therefore, cellular segmentation, including cold ischaemic times, fixation times, and sample age. In addition, the unique biological signatures of each patient’s disease introduce variance between tumour tissues, making a standardised approach to cell classification challenging. Consequently, accurately segmenting these cells to gain knowledge of highly diverse and complex micro-environments remains a significant bottleneck. Cell expansion methods from nuclear masks are commonly used for mean intensity-based phenotyping methods. Without knowledge of the true cell boundaries, a method of this sort is necessary to ensure that extranuclear information is not wasted. While such approaches generally improve the ability to phenotype cells, they are vulnerable to the heterogeneity of cell shapes and sizes, which cannot be accommodated by a simple dilation of the nuclear envelope. Newer methods which include adding extra markers for cellular boundaries into mIF assays, singly or as cocktails, can aid cellular segmentation. However, for lower-plex analysis, this reduces the number of experimental proteins that can be used at once, there is no universal cell boundary cocktail for all cell types [5,6,24] and some element of ‘blind‘ nuclear expansion is always necessary.

Unlike other, more complex, deep learning methods which depend upon cellular segmentation [9], PixlMap requires limited hours spent annotating small datasets, making use of a highly accessible and well-characterised U-Net deep learning tool available in proprietary software [25]. Our simple training approach, using single channel information from mIF images to generate probability heatmaps generates an APP that is highly generalisable and can integrate information from multiple channels with multiple subcellular expression patterns. Crucially, we have shown that our method obviates the need for expansion of nuclear outlines to approximate cellular outlines, which is a significant source of error in classical approaches, and therefore does not require additional cell membrane markers for accurate phenotype assignment. This is achievable because our training data contains staining patterns of markers expressed in each major cellular compartment. By including nuclear areas in non-nuclear training annotations, the resulting heatmaps accurately predict cell phenotypes from nuclear labels even if the nuclei themselves are not stained. Our validation confirms that nuclear areas are sufficient to assign phenotypes to whole cells, and highly comparable to human assignment of phenotypes. Furthermore, we have shown that in our mIF datasets, increasing the object size for phenotype assignment (i.e., dilating from a nuclear label to generate an approximate whole cell label), is counterproductive, and even decreases the accuracy of phenotyping.

Cellular phenotyping based on mean pixel intensities of each marker is the current standard for phenotype assignment. However, our method is more nuanced, with training which includes real-world staining patterns. Such staining patterns can often be quite weak in absolute terms while still being convincingly positive to the human interpreter based on the pattern or staining quality; indeed, this subtle distinction is a key part of clinical immunohistochemical interpretation. This approach, therefore, should partly overcome the need for extensive normalisation methods in downstream analysis for large datasets with high staining variability. Our benchmarking results have shown superiority of PixlMap phenotype assignment over the standard method of generating mean intensities for cellular compartments. However, like all microscopy-based imaging methods, our method still requires, to an extent, good tissue quality to generate meaningful data to accurately label nuclei or objects of interest. Any method, however, that accurately labels the centre of a cell or nucleus can be deployed for PixlMap to accurately assign cell class identities.

PixlMap shows considerable robustness. Its value is not limited to human tissue images, despite only using human tissue annotations for training on LUAD. We have demonstrated that PixlMap correctly assigns class identities across species and unseen tissue types. We evaluated this in murine liver tissue, further proving that our method is highly generalisable. While we observed reduced performance in murine tissue compared to LUAD, we demonstrate the transferability of much of the performance across species, with considerable promise for future refinement of the approach with species-specific training data.

Regions of diverse lymphoid structures are densely packed with cells and typically represent problematic areas both for cell and nuclear segmentation. Therefore, it is imperative that training data are accurate, especially with high heterogeneity in cell populations with varying object sizes. We found that our algorithm can overcome this issue and phenotype T cells in lymphoid aggregate structures with high accuracy using only nuclear labelling, when compared to manual phenotyping.

Despite Visiopharm being used for all our image analysis in this study, our method is replicable in other software including open systems. Given that U-Net is a well understood and deployed network [26], our method should be replicable with U-Net in Python/ FOSS software with the same level of accuracy.

mIF images contain an abundance of useful information which have obvious potential in facilitating diagnostics and personalised medicine. New quantitative spatial assays are likely to be crucial as biomarkers for the correct positioning of new therapies, including anti-cancer immunotherapies and antibody-drug conjugates (ADCs) [27]. However, before these biomarkers can be implemented in the clinical setting, there are many obstacles to overcome such as assay standardisation, the choice of an appropriate implementation model, quality assurance, cost, data storage, and data complexity [28]. Clinical decision making may also require pathologists to undergo additional training to interpret such complex images [29]. Automated digital pathology workflows have potential to significantly streamline the process. We show that the PixlMap APP approach may overcome some major challenges in these digital pathology workflows, being highly robust in cellular phenotyping in the crowded tumour microenvironment, with the potential to improve patient stratification in a broadly applicable manner.

Despite the robustness of our APP in different tissue types and species, it still has some limitations to overcome. Due to extremely non-circular cells not being included in the training set, PixlMap will probably require an additional morphology category to assign phenotypes to fibroblasts and other elongated cells. Phenotyping fibroblasts in situ already presents significant biological and technical challenges. One major issue is the absence of a definitive, universal marker that uniquely identifies fibroblasts and distinguishes them from other cell types [30]. Therefore, relying on a single marker for mean intensity-based phenotyping may introduce inaccurate or misleading information. Additionally, fibroblasts exhibit substantial variability in their morphology—while they are often described as long and thin, not all fibroblasts have this feature, making morphological identification also unreliable in many cases [31]. Despite the lack of prior knowledge of fibroblast staining or morphology, PixlMap did perform better than expected at this challenging task. As a result, we believe, by training PixlMap on single channels containing fibroblast staining pattern information, we may be able to refine the detection of non-circular cells, as we need only a nuclear marker to assign class identities with PixlMap.

Similarly, macrophages do not have a consistent cellular shape, with their boundaries being unclear and discontinuous [32]. We have shown that PixlMap was able to phenotype macrophages in our dataset, however not as accurately as other cell types. We believe this is probably due to endogenous low-level CD4 signal in these cells that may have confused the algorithm. This could be ameliorated by the provision of additional macrophage-rich training images.

Furthermore, PixlMap will require further training on images generated by alternative staining and imaging platforms (Akoya PhenoCycler, Lunaphore COMET) due to the difference in autofluorescence removal in these imaging modalities [3335]. Each of these platforms also processes images differently, however, a cross-platform training approach could overcome this in future versions.

Conclusions and future work

PixlMap represents a paradigm shift in phenotyping approaches by eliminating the need for whole-cell segmentation, paving the way for simpler, more adaptable methods in tissue analysis. This work highlights that whole cell estimation by nuclear expansion is no longer necessary for accurate phenotype assignment using our method. With PixlMap, cellular phenotypes can be assigned using only nuclear segmentation. This approach simplifies the process of cellular phenotyping while maintaining high accuracy, leveraging our well-trained U-Net APP that assigns whole cell phenotypes directly to nuclei without the need to expand the entire cell structure. Moreover, the inclusion of confidence limits helps tailor the APPs predictions to different situations, depending on how much uncertainty is acceptable in a given context.

Future directions include further refining our APP to include examples of cells with different morphologies, e.g., fibroblasts. While others use known fibroblast markers to segment fibroblasts [36] and phenotype them based on their expression levels of specific fibroblast markers [37], there is scope for using our method for fibroblast phenotyping. We intend to further develop PixlMap as a cross-platform classifier by incorporating images from all of our mIF imaging platforms into our training data. Furthermore, we plan on testing the application of PixlMap to RNA/spot-based detection methods by including regions of cells with punctate signal in our training set until the desired stabilised convergence.

Supporting information

S1 Fig. PixlMap requires additional morphology training for fibroblast phenotyping.

A-E. Confusion matrices comparing ground truth assignment of pan-cytokeratin, CD4, CD8, CD68, and SMA to PixlMap predictions by confidence thresholds. Percentages normalised to predications. F. Bar plot showing weighted performance statistics (F1 score, accuracy, precision, recall).

(TIFF)

pone.0317865.s001.tiff (859.1KB, tiff)
S1 Table. mIF staining protocol parameters for the Ventana Discovery Ultra.

(DOCX)

pone.0317865.s002.docx (89.7KB, docx)
S1 Data. Script for single-channel image extraction from mIF images.

(TXT)

Acknowledgments

This paper was critically reviewed by Catherine Winchester (CRUK Scotland Institute).

We would like to thank Peter McHardy (CRUK Scotland Institute) for his support with Visiopharm infrastructure throughout this project. We would also like to thank Kara Luckett, Silvia Martinelli, and Cat Ficken from the Le Quesne lab at the CRUK Scotland Institute for their support with image acquisition and data curation throughout this project.

We acknowledge the support of NHS Research Scotland (NRS) Greater Glasgow and Clyde Biorepository for their ongoing management of the LATTICe TMA cohort.

Data Availability

Code used to generate the figures in this publication is available from DOI https://doi.org/10.5281/zenodo.15593393. The Visiopharm APPs used to generate all figures are available from DOI https://doi.org/10.5281/zenodo.15593785. Murine images and layer data used in this publication are available at https://figshare.com/s/b9c355bd9b978b2ca564 Raw data, including images, relating to the LATTICe TMA cohort used in this publication are subject to NHSGGC Biorepository ethical approval. To access this data a biorepository application and MTA are required, please contact the Lead Researcher or the NHSGGC Biorepository manager (clare.orange@nhs.scot) for advice on the application process.

Funding Statement

JLQ, LOJ, RLB, FB, LH, and IRP are funded by the Mazumdar Shaw Chair. KR is supported by a Pathological Society of Great Britain & Ireland and Jean Shanks Foundation clinical PhD fellowship (JSPS CPF 0422 04). TGB and AG were funded by an Accelerator Award from CRUK (Grant number: A26813) and CRUK Core funding (Grant numbers A17196, A31287 and CTRQQR-2021\100006). AG was funded by a CRUK HUNTER Studentship (Grant number: A28221). TGB, was funded by CRUK (Grant numbers: 23390 and DRCRPG-Nov22/100007). DM developed and maintains the Visiopharm software used in the study. DM was not involved in the data collection or data analysis parts of this study and declares no influence upon the results presented. The funders provided support in the form of salaries for all authors, but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section.

References

  • 1.Cousin S, Guégan J-P, Shitara K, Palmieri LJ, Metges JP, Pernot S, et al. Identification of microenvironment features associated with primary resistance to anti-PD-1/PD-L1 + antiangiogenesis in gastric cancer through spatial transcriptomics and plasma proteomics. Mol Cancer. 2024;23(1):197. doi: 10.1186/s12943-024-02092-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Kawamura I, Ohe R, Suzuki K, Kabasawa T, Kitaoka T, Takahara D, et al. Neighboring macrophage-induced alteration in the phenotype of colorectal cancer cells in the tumor budding area. Cancer Cell Int. 2024;24(1):107. doi: 10.1186/s12935-024-03292-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Baird R, Mason D. Multiplex: From Acquisition to Analysis. Artificial Intelligence Applications in Human Pathology. WORLD SCIENTIFIC (EUROPE). 2022. 99–135. doi: 10.1142/9781800611399_0006 [DOI] [Google Scholar]
  • 4.Bai Y, Zhu B, Rovira-Clave X, Chen H, Markovic M, Chan CN, et al. Adjacent Cell Marker Lateral Spillover Compensation and Reinforcement for Multiplexed Images. Front Immunol. 2021;12:652631. doi: 10.3389/fimmu.2021.652631 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Beechem JM, Danaher P, Sato H, Williams C, Rhodes M. Evaluating the Technical Performance of Single-Cell Spatial Molecular Imaging Technologies. nanostring.com. 2023. https://nanostring.com/wp-content/uploads/2023/09/WP_MK5986_CosMx_Data-Quality_r7.pdf [Google Scholar]
  • 6.6.10x Genomics. Xenium In Situ Multimodal Cell Segmentation: Workflow and Data Highlights. 10xgenomics.com. 2024. https://www.10xgenomics.com/support/in-situ-gene-expression/documentation/steps/assay/cell-segmentation-tech-note [Google Scholar]
  • 7.Graham S, Vu QD, Raza SEA, Azam A, Tsang YW, Kwak JT, et al. Hover-Net: Simultaneous segmentation and classification of nuclei in multi-tissue histology images. Med Image Anal. 2019;58:101563. doi: 10.1016/j.media.2019.101563 [DOI] [PubMed] [Google Scholar]
  • 8.Stevens M, Nanou A, Terstappen LWMM, Driemel C, Stoecklein NH, Coumans FAW. StarDist Image Segmentation Improves Circulating Tumor Cell Detection. Cancers (Basel). 2022;14(12):2916. doi: 10.3390/cancers14122916 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Stringer C, Wang T, Michaelos M, Pachitariu M. Cellpose: a generalist algorithm for cellular segmentation. Nat Methods. 2021;18(1):100–6. doi: 10.1038/s41592-020-01018-x [DOI] [PubMed] [Google Scholar]
  • 10.Moore DA, Sereno M, Das M, Baena Acevedo JD, Sinnadurai S, Smith C, et al. In situ growth in early lung adenocarcinoma may represent precursor growth or invasive clone outgrowth-a clinically relevant distinction. Mod Pathol. 2019;32(8):1095–105. doi: 10.1038/s41379-019-0257-1 [DOI] [PubMed] [Google Scholar]
  • 11.May S, Müller M, Livingstone CR, Skalka GL, Walsh PJ, Nixon C, et al. Absent expansion of AXIN2+ hepatocytes and altered physiology in Axin2CreERT2 mice challenges the role of pericentral hepatocytes in homeostatic liver regeneration. J Hepatol. 2023;78(5):1028–36. doi: 10.1016/j.jhep.2023.01.009 [DOI] [PubMed] [Google Scholar]
  • 12.Ntala C, Salji M, Salmond J, Officer L, Teodosio AV, Blomme A, et al. Analysis of Prostate Cancer Tumor Microenvironment Identifies Reduced Stromal CD4 Effector T-cell Infiltration in Tumors with Pelvic Nodal Metastasis. Eur Urol Open Sci. 2021;29:19–29. doi: 10.1016/j.euros.2021.05.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Xavier V, Martinelli S, Corbyn R, Pennie R, Rakovic K, Powley IR, et al. Mitochondrial double-stranded RNA homeostasis depends on cell-cycle progression. Life Sci Alliance. 2024;7(11):e202402764. doi: 10.26508/lsa.202402764 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Hatthakarnkul P, Ammar A, Pennel KAF, Officer-Jones L, Cusumano S, Quinn JA, et al. Protein expression of S100A2 reveals it association with patient prognosis and immune infiltration profile in colorectal cancer. J Cancer. 2023;14(10):1837–47. doi: 10.7150/jca.83910 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Officer LK, Andreou KE, Teodósio AV, He Z, Le Quesne JP. Automated Co-in Situ Hybridization and Immunofluorescence Using Archival Tumor Tissue. Methods Mol Biol. 2020;2148:245–56. doi: 10.1007/978-1-0716-0623-0_16 [DOI] [PubMed] [Google Scholar]
  • 16.Müller M, Bird TG, Nault J-C. The landscape of gene mutations in cirrhosis and hepatocellular carcinoma. J Hepatol. 2020;72(5):990–1002. doi: 10.1016/j.jhep.2020.01.019 [DOI] [PubMed] [Google Scholar]
  • 17.Harada N, Tamai Y, Ishikawa T, Sauer B, Takaku K, Oshima M, et al. Intestinal polyposis in mice with a dominant stable mutation of the beta-catenin gene. EMBO J. 1999;18(21):5931–42. doi: 10.1093/emboj/18.21.5931 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Kruspig B, Monteverde T, Neidler S, Hock A, Kerr E, Nixon C, et al. The ERBB network facilitates KRAS-driven lung tumorigenesis. Sci Transl Med. 2018;10(446):eaao2565. doi: 10.1126/scitranslmed.aao2565 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Müller M, May S, Hall H, Kendall TJ, McGarry L, Blukacz L, et al. Human-correlated genetic HCC models identify combination therapy for precision medicine. Springer Science and Business Media LLC. 2022. doi: 10.21203/rs.3.rs-1638504/v1 [DOI] [Google Scholar]
  • 20.Visiopharm. A complete workflow for all your multiplexed image analysis needs. Visiopharm.com. 2024. https://visiopharm.com/wp-content/uploads/guide-phenoplex-workflow.pdf [Google Scholar]
  • 21.Indica Labs Inc. HALO 3.2 Area Quantification FL Step-by-Step Guide. 2020. https://indicalab.com
  • 22.Abraham A, Pedregosa F, Eickenberg M, Gervais P, Mueller A, Kossaifi J, et al. Machine learning for neuroimaging with scikit-learn. Front Neuroinform. 2014;8:14. doi: 10.3389/fninf.2014.00014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Waskom M. seaborn: statistical data visualization. JOSS. 2021;6(60):3021. doi: 10.21105/joss.03021 [DOI] [Google Scholar]
  • 24.Cameron WD, Bennett AM, Bui CV, Chang HH, Rocheleau JV. Leveraging multimodal microscopy to optimize deep learning models for cell segmentation. APL Bioeng. 2021;5(1):016101. doi: 10.1063/5.0027993 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Ronneberger O, Fischer P, Brox T. U-Net: Convolutional networks for biomedical image segmentation. Medical image computing and computer-assisted intervention – MICCAI 2015. Cham: Springer. 2015. 234–41. [Google Scholar]
  • 26.Li M-Y, Zhu D-J, Xu W, Lin Y-J, Yung K-L, Ip AWH. Application of U-Net with Global Convolution Network Module in Computer-Aided Tongue Diagnosis. J Healthc Eng. 2021;2021:5853128. doi: 10.1155/2021/5853128 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Tan WCC, Nerurkar SN, Cai HY, Ng HHM, Wu D, Wee YTF, et al. Overview of multiplex immunohistochemistry/immunofluorescence techniques in the era of cancer immunotherapy. Cancer Commun (Lond). 2020;40(4):135–53. doi: 10.1002/cac2.12023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Mulholland EJ, Leedham SJ. Redefining clinical practice through spatial profiling: a revolution in tissue analysis. Ann R Coll Surg Engl. 2024;106(4):305–12. doi: 10.1308/rcsann.2023.0091 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Harms PW, Frankel TL, Moutafi M, Rao A, Rimm DL, Taube JM, et al. Multiplex Immunohistochemistry and Immunofluorescence: A Practical Update for Pathologists. Mod Pathol. 2023;36(7):100197. doi: 10.1016/j.modpat.2023.100197 [DOI] [PubMed] [Google Scholar]
  • 30.Gauthier V, Kyriazi M, Nefla M, Pucino V, Raza K, Buckley CD, et al. Fibroblast heterogeneity: Keystone of tissue homeostasis and pathology in inflammation and ageing. Front Immunol. 2023;14:1137659. doi: 10.3389/fimmu.2023.1137659 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Plikus MV, Wang X, Sinha S, Forte E, Thompson SM, Herzog EL, et al. Fibroblasts: Origins, definitions, and functions in health and disease. Cell. 2021;184(15):3852–72. doi: 10.1016/j.cell.2021.06.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Nikovics K, Favier A-L. Macrophage Identification In Situ. Biomedicines. 2021;9(10):1393. doi: 10.3390/biomedicines9101393 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Millian DE, Saldarriaga OA, Wanninger T, Burks JK, Rafati YN, Gosnell J, et al. Cutting-Edge Platforms for Analysis of Immune Cells in the Hepatic Microenvironment-Focus on Tumor-Associated Macrophages in Hepatocellular Carcinoma. Cancers (Basel). 2022;14(8):1861. doi: 10.3390/cancers14081861 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Schürch CM, Bhate SS, Barlow GL, Phillips DJ, Noti L, Zlobec I, et al. Coordinated Cellular Neighborhoods Orchestrate Antitumoral Immunity at the Colorectal Cancer Invasive Front. Cell. 2020;182(5):1341-1359.e19. doi: 10.1016/j.cell.2020.07.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Rivest F, Eroglu D, Pelz B, Kowal J, Kehren A, Navikas V, et al. Fully automated sequential immunofluorescence (seqIF) for hyperplex spatial proteomics. Sci Rep. 2023;13(1):16994. doi: 10.1038/s41598-023-43435-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Magness A, Colliver E, Enfield KSS, Lee C, Shimato M, Daly E, et al. Deep cell phenotyping and spatial analysis of multiplexed imaging with TRACERx-PHLEX. Nat Commun. 2024;15(1):5135. doi: 10.1038/s41467-024-48870-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Ogawa Y, Masugi Y, Abe T, Yamazaki K, Ueno A, Fujii-Nishimura Y, et al. Three Distinct Stroma Types in Human Pancreatic Cancer Identified by Image Analysis of Fibroblast Subpopulations and Collagen. Clin Cancer Res. 2021;27(1):107–19. doi: 10.1158/1078-0432.CCR-20-2298 [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Jordan Yaron

28 Apr 2025

Dear Dr. Le Quesne,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols .

We look forward to receiving your revised manuscript.

Kind regards,

Jordan Robin Yaron, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1.Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. Please note that funding information should not appear in any section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. Please remove any funding-related text from the manuscript.

4. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match.

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

5. Thank you for stating the following in the Competing Interests section:

“I have read the journal's policy and the authors of this manuscript have the following competing interests:DM is an employee of Visiopharm. All other authors declare no conflict of interest.”

We note that one or more of the authors are employed by a commercial company: Visiopharm

a. Please provide an amended Funding Statement declaring this commercial affiliation, as well as a statement regarding the Role of Funders in your study. If the funding organization did not play a role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript and only provided financial support in the form of authors' salaries and/or research materials, please review your statements relating to the author contributions, and ensure you have specifically and accurately indicated the role(s) that these authors had in your study. You can update author roles in the Author Contributions section of the online submission form.

Please also include the following statement within your amended Funding Statement.

“The funder provided support in the form of salaries for authors [insert relevant initials], but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section.”

If your commercial affiliation did play a role in your study, please state and explain this role within your updated Funding Statement.

b. Please also provide an updated Competing Interests Statement declaring this commercial affiliation along with any other relevant declarations relating to employment, consultancy, patents, products in development, or marketed products, etc. 

Within your Competing Interests Statement, please confirm that this commercial affiliation does not alter your adherence to all PLOS ONE policies on sharing data and materials by including the following statement: "This does not alter our adherence to  PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests) . If this adherence statement is not accurate and  there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please include both an updated Funding Statement and Competing Interests Statement in your cover letter. We will change the online submission form on your behalf.

6. We note that you have indicated that there are restrictions to data sharing for this study. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For more information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions.

Before we proceed with your manuscript, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., a Research Ethics Committee or Institutional Review Board, etc.). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of recommended repositories, please see

https://journals.plos.org/plosone/s/recommended-repositories. You also have the option of uploading the data as Supporting Information files, but we would recommend depositing data directly to a data repository if possible.

We will update your Data Availability statement on your behalf to reflect the information you provide.

7. Please include your full ethics statement in the ‘Methods’ section of your manuscript file. In your statement, please include the full name of the IRB or ethics committee who approved or waived your study, as well as whether or not you obtained informed written or verbal consent. If consent was waived for your study, please include this information in your statement as well.

8. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

  • The number of images as well as cells/image that were analyzed for each experiment.

  • Details on how mean intensity data were generated, thresholding approaches, expansion methods, and a representative mask from the images for each marker.

  • Direct comparison of actual cell number results from manual evaluation, PixlMap, and traditional methods for 2-3 markers, along with number of images and approximate number of cells.

  • Address, explain, or fix the issue with field/zoom representation in Figure 4A.

  • Address in greater detail potential issues with fibroblast detection.

  • Similarly, address potential accuracy issues in identifying CD68+ cells in Figure 5B.

  • Address whether surface markers are more accurately identified versus cytoplasmic markers.

  • Address weakness of metrics in liver tissue comparisons.

  • Clarify language throughout regarding manual versus threshold-based phenotyping.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

Reviewer #1: Partly

Reviewer #2: Partly

Reviewer #3: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously? -->?>

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: N/A

**********

3. Have the authors made all data underlying the findings in their manuscript fully available??>

The PLOS Data policy

Reviewer #1: Yes

Reviewer #2: No

Reviewer #3: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

Reviewer #1: While the paper is well presented, it significantly lacks technological novelty. Given the recent advances in generative models, and other advanced mathematical tricks (contrastive learning, different loss functions such as focal loss, ...), I do not entirely subscribe to the idea that you would need significant annotations for this task. In fact, a fairly simple CNN designed with the correct parameters (optimal receptive field) and loss functions could very well be suited here. The article does not talk about the technically novel aspects. I would expect large vision models (such as SAM) to atleast provide a initial ground truth for learning tasks. Unfortunately, I do not think this article is technically solid enough for being published. The methods used (basically a U-Net) are dated, and there's little innovation in terms of architectural design or modelling.

Reviewer #2: The number of images (and cells/image) analyzed should be provided for each experiment.

Since the performance of the approach is being directly compared to existing approaches (mean intensity), more details are needed on how the mean intensity data were generated (Figure 6H). What thresholding approach was employed? What expansion method was used? Show a representative mask from the images for each marker.

Provide actual results in terms of cell numbers of different lineages/markers as determined by manual evaluation, PixlMap, traditional analysis methods. Do this for at leash 2 or 3 markers and indicate the number of images and approximate number of cells.

In Figure 4A the PixlMap probability image does not seem to be the exactly the same field or zoom as the original image. This should be explained or corrected.

Reviewer #3: Pennie et al. generated a computational tool to classify biomarkers in the multiplexed immunofluorescence images. They note that this tool is inspired by the accuracy of human intuition in identifying positive/negative features. This tool performs classification of biomarkers without the need to define cell boundaries by expanding the nucleus – which is also the primary novelty of this work. The efficacy of this tool was validated in human lung adenocarcinoma samples and murine liver samples.

The manuscript contains sufficient background information for an uninformed reader. Methods are described in sufficient detail and adequate data is shown to draw conclusions. I recommend acceptance of this manuscript. I have the following questions and recommendations.

1. Is fibroblast detection using PixlMap reliable? The authors seem to suggest that it needs further training to be able to detect fibroblasts. I suggest moving Figure 8 to the Supporting Information section if the authors think it needs more work.

2. In Figure 5B, it appears that PixlMap fails to identify some CD68+ cells. What is the accuracy and precision in this case?

3. In Figure 6, about 32% of true unclassifiable features are predicted as CD4+ by PixlMap? I think it will be important to identify biomarker types that have higher accuracy with PixlMap. For example, authors already mention that spindle shaped cells are not yet accurately identified by PixlMap. Similarly, are surface markers more accurately identified as opposed to cytoplasmic ones?

4. Provide a comparison of the weighted metrics (F1 score, accuracy, precision, and recall) of existing models with your model.

5. Lines 50 and 51: The authors only talk about tumor microenvironment as a driver to quantify multiple biomarkers simultaneously. There is a need to perform similar analyses in many other non-cancer studies. I suggest changing the sentence to incorporate other applications.

6. In the murine liver tissue, the accuracy of the model is significantly lower than in human tissue. The F1 scores are also about 0.3 at all probability thresholds. It is understandable that the model is not trained on murine tissues. However, please justify the inclusion of this data with weak metrics in the main manuscript.

7. In Figure 6H, the caption mentions “Confusion matrix comparing PixlMap and threshold-based phenotyping using mean pixel intensities of each marker”. However, data is shown for Manual annotation vs. threshold-based phenotyping. Please clarify.

**********

what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org

PLoS One. 2025 Dec 3;20(12):e0317865. doi: 10.1371/journal.pone.0317865.r002

Author response to Decision Letter 1


26 Jun 2025

The Authors would like to take this opportunity to thank the Editor and Reviewers for the care and time taken in reviewing our work. We have responded to the feedback and suggestions received in detail in the attached document titled "Response to Reviewers".

Attachment

Submitted filename: Response to reviewers.pdf

pone.0317865.s005.pdf (102.9KB, pdf)

Decision Letter 1

Jordan Yaron

17 Aug 2025

PixlMap: A generalisable pixel classifier for cellular phenotyping in multiplex immunofluorescence images

PONE-D-25-00785R1

Dear Dr. Le Quesne,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager®  and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact billing support .

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Jordan Robin Yaron, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions??>

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously? -->?>

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available??>

The PLOS Data policy

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #3: Yes

**********

Reviewer #3: The authors have addressed all of my comments. I am satisfied with their response. I recommend acceptance of this manuscript in its revised version.

**********

what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy

Reviewer #3: No

**********

Acceptance letter

Jordan Yaron

PONE-D-25-00785R1

PLOS ONE

Dear Dr. Le Quesne,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Jordan Robin Yaron

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. PixlMap requires additional morphology training for fibroblast phenotyping.

    A-E. Confusion matrices comparing ground truth assignment of pan-cytokeratin, CD4, CD8, CD68, and SMA to PixlMap predictions by confidence thresholds. Percentages normalised to predications. F. Bar plot showing weighted performance statistics (F1 score, accuracy, precision, recall).

    (TIFF)

    pone.0317865.s001.tiff (859.1KB, tiff)
    S1 Table. mIF staining protocol parameters for the Ventana Discovery Ultra.

    (DOCX)

    pone.0317865.s002.docx (89.7KB, docx)
    S1 Data. Script for single-channel image extraction from mIF images.

    (TXT)

    Attachment

    Submitted filename: Response to reviewers.pdf

    pone.0317865.s005.pdf (102.9KB, pdf)

    Data Availability Statement

    Code used to generate the figures in this publication is available from DOI https://doi.org/10.5281/zenodo.15593393. The Visiopharm APPs used to generate all figures are available from DOI https://doi.org/10.5281/zenodo.15593785. Murine images and layer data used in this publication are available at https://figshare.com/s/b9c355bd9b978b2ca564 Raw data, including images, relating to the LATTICe TMA cohort used in this publication are subject to NHSGGC Biorepository ethical approval. To access this data a biorepository application and MTA are required, please contact the Lead Researcher or the NHSGGC Biorepository manager (clare.orange@nhs.scot) for advice on the application process.


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES