Skip to main content
PLOS One logoLink to PLOS One
. 2025 Mar 19;20(3):e0311242. doi: 10.1371/journal.pone.0311242

MCount: An automated colony counting tool for high-throughput microbiology

Sijie Chen 1,2, Po-Hsun Huang 1, Hyungseok Kim 1,2, Yuhe Cui 1, Cullen R Buie 1,*
Editor: Florian Rehfeldt3
PMCID: PMC11957731  PMID: 40106480

Abstract

Accurate colony counting is crucial for assessing microbial growth in high-throughput workflows. However, existing automated counting solutions struggle with the issue of merged colonies, a common occurrence in high-throughput plating. To overcome this limitation, we propose MCount, the only known solution that incorporates both contour information and regional algorithms for colony counting. By optimizing the pairing of contours with regional candidate circles, MCount can accurately infer the number of merged colonies. We evaluate MCount on a precisely labeled Escherichia coli dataset of 960 images (15,847 segments) and achieve an average error rate of 3.99%, significantly outperforming existing published solutions such as NICE (16.54%), AutoCellSeg (33.54%), and OpenCFU (50.31%). MCount is user-friendly as it only requires two hyperparameters. To further facilitate deployment in scenarios with limited labeled data, we propose statistical methods for selecting the hyperparameters using few labeled or even unlabeled data points, all of which guarantee consistently low error rates. MCount presents a promising solution for accurate and efficient colony counting in application workflows requiring high throughput, particularly in cases with merged colonies.

1. Introduction

Quantitative assessment of microorganisms is a critical procedure in the field of microbiology, and various methods have been developed to estimate microorganism population levels, including quantitative PCR [1], flow cytometry [2], spectrophotometry [3,4], and colony-forming units (CFU) counting [5,6]. Among these methods, CFU counting is the oldest and most widely used method, and its efficacy and reliability have been examined since the 1880s [6,7]. In addition, its simple operational protocol and minimal consumables cost have solidified CFU counting as the gold standard in microbiology.

However, traditional manual CFU counting is a time-consuming and labor-intensive process when dealing with many images (e.g., more than one hundred). Thus, researchers have devoted considerable effort to developing numerous solutions over the decades to realize an easier, faster, more accurate, and reliable counting method. Currently, popular solutions include commercial products such as SphereFlash (IUL Instruments) and ProtoCOL 3 (Synbiosis), as well as open-source tools based on various counting algorithms [818]. While commercial products are usually easy to operate, they are expensive and highly specialized for specific counting scenarios, and their proprietary programming nature makes them difficult to modify or share. In contrast, open-source tools are gaining more attention, especially those capable of batch processing [911]. NIST’s Integrated Colony Enumerator (NICE) is a tool that has been popular since 2009, which is based on the combination of extended minima function and thresholding algorithms [9]. It has a relatively short image processing time ( < 5 seconds per image) and a friendly user interface. OpenCFU, published in 2013, is a tool claiming to be faster, more accurate, and more robust than NICE [10]. It adopts the watershed algorithm along with a series of pre- and post-processing filters. AutoCellSeg was developed in 2017 [11] to reduce the hyperparameter selection effort during counting, based on a feedback-based watershed algorithm, and it has an interactive graphical user interface that is appealing to those not familiar with programming.

The existing colony-counting solutions work well on single petri dish plates of 60 cm2, where 25 ~  250 colonies are randomly distributed, making colony merging issues rare [7,19,20]. Most colony-plating protocols require careful selection of the serial dilution ratio to minimize the chance of merged colony. However, modern colony-counting requires higher throughput, and a typical example is plating samples from a 96-well plate on a single-well rectangular plate of 109 cm2 (Fig 1a). In this case, more than 10 colonies are plated on an area less than 1 cm2 for each sample, resulting in > 10 times higher density, which often leads to merged colonies. Furthermore, in such high-throughput workflows, the smaller area results in fewer image pixels per sample, reducing image quality. These difficulties lead to significant counting underestimation with current solutions (Fig 1b). Although NICE can recognize most single colonies, it tends to count merged colonies as one. AutoCellSeg typically counts each continuous region as one, regardless of the region’s shape, so it cannot handle scenarios involving colony merging. As for OpenCFU, it fails to recognize most colonies, including single colonies, due to the lower image quality for each sample of the rectangular plate. The failure of these solutions results from the fact that they rely on region-based algorithms and do not take contour information into account.

Fig 1. The need for a more powerful colony counting algorithm.

Fig 1

(a) Colonies of fluorescent E. coli NEB10-beta wereplated on agar in an 8 × 12 array format. As the area of each well is small, colony merging occurs frequently and the sub-image in the blue rectangle is an example. While it is challenging to determine the exact number of merged colonies in the yellow and cyan circles, a skilled person can count 3 and 2 colonies, respectively, which can be verified by the photograph taken a few hours earlier. (b) NICE correctly counts all single colonies but counts also merged colonies as one [9]; AutoCellSeg uses the watershed algorithm to count every connected region as one colony, regardless of the region’s shape [11]; OpenCFU applies sophisticated rules to count only perfect circle-shaped colonies but fails to recognize most single colonies [10]. All three algorithms underestimate the total number of colonies, which is 22.

Colony counting is a sub-domain of object counting that includes measurement of the number of biological cells [2123], pollen grains in air [24,25], and bubbles for analyzing two-phase systems [26,27]. Classical object counting approaches can be categorized into contour-based or region-based methods. Contour-based methods aim to recognize the shape only from the contour pixels and use methods such as the Hough-transform [28,29] and least squares fit for circular contour [26,28,30]. They also involve algorithms such as concave point detection methods to split the contour into segments [27,3033]. Contour-based methods can provide precise results given a high-definition image, but they typically require more computational resources. Region-based methods take all shape pixels into account, and classical methods include the extended minima function method [9,16,34], morphological operations [35], distance transformation [16,31,36], and the watershed algorithm [10,11,16,17,24,34]. Region-based methods can tolerate more noise and have faster processing speed, but their recognition accuracy is usually lower. To combine the advantages of both methods, some have made efforts to pair the extended minima method with contour segments [26,27,31,32]. In recent studies, optimization algorithms have been adopted [31,33], eliminating the computationally expensive exhausted pairing [26,27]. Surprisingly, although classical object counting algorithms are advancing rapidly in other sub-domains, colony counting lags behind with a reliance on region-based algorithms, leading to poor accuracy in high-throughput workflows.

In addition to classical object counting methods, neural networks have been applied across many fields due to their success in processing complex tasks [22,37,38]. However, when applied to colony counting, these methods encounter significant challenges, primarily due to the lack of standardized, large-scale, high-quality datasets. Most existing datasets are small, often only a few hundred images, which are insufficient for training deep learning models with numerous parameters. Additionally, neural networks often function as “black boxes”, which may provide final counts without visually delineating individual colonies, a drawback for biologists who need visual confirmation. While advanced models could include features like contouring, they require more robust datasets and computational resources. For these reasons, we believe classical algorithms remain a more suitable choice for colony counting.

To ensure user-friendliness and operational simplicity, counting tools typically emphasize a few intuitive hyperparameters. However, while these tools focus more on the physical meaning of different hyperparameters, they often lack discussion on properly tuning these hyperparameters, i.e., the hyperparameter optimization problem, which is an essential topic in the field of machine learning to ensure high quality performance with minimal human effort [39,40]. Consequently, deploying such solutions to various counting tasks, even with minor differences, can be challenging. Therefore, there is a need for not only a better colony counting algorithm but also a consistent method for hyperparameter optimization.

In this paper we propose a new solution, MCount (Merged-colony Counting), that can precisely infer the number of merged colonies from relatively small images, meeting the demands of high-throughput colony counting. MCount employs both region-based and contour-based algorithms, leading to much higher accuracy than existing tools. Since there is no standard benchmark for high-throughput colony counting tasks, we constructed a dataset for performance evaluation. In addition, we address the hyperparameter optimization problem. Given that the number of labeled data is often insufficient, we propose novel statistical methods for hyperparameter optimization, ensuring a low error rate even with minimal labeled data or even unlabeled data. Finally, we examine the statistical robustness of the proposed hyperparameter-tuning methods.

2. Methods

Fig 2 presents a detailed flowchart of MCount, depicting the process of foreground extraction, contour extraction, regional circle fitting, and optimization.

Fig 2. MCount uses a combination of contour-based and region-based information to accurately count merged colonies.

Fig 2

(a) The foreground and background are separated using Otsu thresholding [41], and colonies are segmented into disconnected segments, which are then smoothed to remove noise. (b) Contour-based algorithms are applied to each colony segment to identify turning points (colored in black), which divide the inner and outer contours into contour segments represented by different colors. Meanwhile, region-based algorithms generate candidate circles denoted in different colors that can fit the boundary well. Finally, proper circles are selected from candidate circles based on the shape of all contour segments. This selection process is converted into an optimization problem that is solved to correctly recognize merged colonies.

2.1. Algorithms adopted by MCount

2.1.1. Foreground extraction.

The first step is to obtain connected (overlapping) colony segments by using a series of filters. The original image is binarized using Otsu thresholding, which separates image pixels into two categories, i.e., foreground and background, by maximizing inter-class variance [41]. Next, the connected component labeling algorithm is used to separate disconnected segments. Finally, morphological operations (erosion operation followed by a dilation operation) are applied to each colony segment twice to remove white noise in the image.

2.1.2. Contour extraction.

The next step is to extract contours from each colony segment using a classic border following algorithm [42]. As each contour may correspond to the edges of several colonies, we need to split each contour into several segments, so that each piece only corresponds to one colony. We assume that any overlap between multiple colonies results in a shape with concave edge points that correspond to the intersections of the colony boundaries, which allows us to use the concave points to split the contour. To identify the concave points, we use an algorithm based on the polygon approximation algorithm [27] and the algorithm proposed by Zafari et al. [43], and the pseudo-code is provided in Algorithm 1.

In brief, as shown in lines 1-18, the contour is approximated using polynomial fitting, where all the vertices of the polygon are on the contour. Such vertices are called turning points, denoted as T , and hyperparameter d controls how densely the turning points are selected from the contour, as shown in S1 Fig. For example, when d increases, fewer turning points are selected. Next, all concave vertices of the polygon are further selected from the turning points by implementing code from lines 19-24. These concave points divide the contour into several segments, and each piece is denoted by a different color in Fig 2b.

Algorithm 1. Contour segmentation based on concave point detection. Given a contour C with a hyperparameter d that determines contour fineness, this algorithm first approximates C using polynomial fitting and selects turning points T from the polygon approximation. Then, it identifies concave points Conc from the turning points T to split the contour into several segments.

graphic file with name pone.0311242.g007.jpg

2.1.3. Regional circle fitting.

In this step, a set of candidate circles is generated from each colony segment. Two types of circles are considered. The first type of circle is identified using a region-based algorithm, where distance transformation is applied to the binary foreground image so that the center of circles is the local maxima of distance [44]. The other type of circle is obtained by a least square fit of each split contour piece [27]. In Fig 2b, all the candidate circles are denoted by different colors.

2.1.4. Optimization.

The final step of the algorithm is to pair the split contour segments with candidate circles, so that each contour will only match one circle. Denoting Ci as the i th contour piece and Oj as the j th circle (the edge of the circle rather than the center), the pairing problem that we solve is:

Xij1ifallocatingCitoOj0otherwise (1)
Yj=1ifanycontourisallocatedtoOj0otherwise (2)

Oj is a recognized as a colony only when Yj=1. To formulate this pairing problem, we define the distance Dij of allocating Oj to Ci as:

Dij=pminqpq¯,pCi,qOj (3)

where p and q are the pixels of contour piece and circle, respectively, and pq¯ represents the distance between pixel p and q.

The smaller Dij, the better the pairing of circle Oj with contour piece Cj. Thus, an intuitive way to pair all contour segments with proper circles is to match every contour piece to the closest circle, i.e., we are looking for a pairing that could lead to a minimized total distance, which could be converted into an optimization problem. However, simply minimizing the total distance without any constraints will lead to overfitting because different contour segments from the same colony might be fitted to multiple adjacent circles. Therefore, some constraints must be considered, and an example is to control the total number of recognized circles, jYj, which can be converted to a zero-one integer programming problem with objective function LXij,Yi:

Xij,Yj=argminXij,YiLXij,Yi=argminXij,Yii,jDijXij+λjYj (4)
s.t.jXij=1
0Xij1
YjiXijYj·M
0Yj1

where λ denotes the strength of the constraint and M denotes the total number of contour segments.

By solving equation (4), we can obtain all the recognized colonies Oj with Yj=1.

2.1.5. Reducing computational complexity of the optimization step.

We recognize that the most time-consuming part during the implementation of MCount is the optimization step because the zero-one integer programming problem is an NP-complete problem whose worst-case runtime grows exponentially. To mitigate this issue, the segment separation procedure is introduced into the foreground extraction step, which can dramatically reduce the computational complexity of the optimization step by reducing the total number of contour segments to a few hundred.

In addition, as there are several approximation algorithms available [45,46], which are already integrated into widely adopted linear programming modelers, e.g., PuLP in Python, we choose one (PuLP solver “COIN_CMD”) that yields a high recognition accuracy with relatively short computation time. The actual runtime of MCount is discussed in the Results and Discussion section.

2.2. Benchmark for performance evaluation

We first sought to evaluate the performance of MCount by using a benchmark dataset. As there is no standard benchmark available, existing colony counting tasks have relied on their own datasets [813]. However, we find that the datasets exhibited at least one of the following concerns when applied to MCount: (a) insufficient number of labeled images (typically less than 100); (b) too many colonies in each labeled image (more than 50), so that the counting number may no longer be the best criteria to evaluate recognition performance; (c) incorrect ground-truth labels, especially on highly merged colonies.

To address these issues, we have created our own benchmark dataset colonies of GFP fluorescent E. coli DH10-beta, a model organism designed for creating mutant libraries suitable for high-throughput purposes (S2 Fig). While our method is versatile and can effectively adapt to non-fluorescent imaging scenarios, the use of fluorescence imaging primarily facilitated the manual annotation process by providing clear and distinct colony boundaries, enabling more accurate ground-truth labeling essential for training and validating the model. The dataset includes 960 labeled images that are large enough to address the problem of an insufficiently large data set. These images were derived from 10 sets of 96-well plates, each with slightly different experimental conditions performed on different dates, ensuring that the dataset contains a certain degree of variation. These variations include differences in colony size, density, and the setup used during image collection, such as camera positioning and background lighting.

Additionally, we increased the size of our training set by decomposing these images into 15,847 colony segments using foreground extraction. Most colony segments have less than 5 colonies, which helps to address problem (b). To mitigate problem (c), we carefully labeled historical images in which the merged colonies were separated. The cell preparation procedure and more detailed discussions on problems (b) and (c) are included in the S1 File.

3. Results and Discussion

3.1. MCount only has two adjustable hyperparameters

Providing more hyperparameters can increase the flexibility of algorithm tuning, which often improves recognition accuracy. However, to prevent overfitting as well as to ensure user-friendliness, we aim to offer minimal adjustable hyperparameters by choosing the most important ones. In the case of MCount, we have identified two hyperparameters that significantly influence its accuracy: d and λ, which control the contour extraction and optimization steps, respectively.

Hyperparameter d represents the fineness of the polygon approximation fitting of the contour. When d is small, the contour will have more turning points, resulting in more contour segments being generated, as shown in Fig 3a. However, setting too small d should be avoided, because it will result in an unnecessarily large number of contour segments and candidate circles (each contour piece will generate one least-squares circle), leading to a longer runtime in the optimization step. This also reduces the algorithm’s robustness because each contour piece will be too short to accurately calculate the curvature. Conversely, if d is too large, very few turning points will be used to fit the contour, leading to an underestimation of the colony count. To minimize the average recognition error on segments, we implemented different values of d on 15,847 colony segments, as shown in Fig 3b, and found that the best value for d is 0.5 (when λ=26). The recognition error on a segment is defined as the difference between the number of colonies counted and the labeled number of colonies divided by the labeled number. Another way to define recognition error is to set 100% for a segment whenever the counted number differs from the label number. The plot for such a recognition error on segments is shown in S3 Fig (a).

Fig 3. MCount has two adjustable hyperparameters, d and λ, which control the contour fineness and constrain the circle number, respectively.

Fig 3

(a) The number of contour segments decreases as d increases, but excessively large d may lead to a failure to recognize merged colonies. The original image has two colonies. (b) The average error rate on the segment dataset versus d at λ=26, where the minimum average error rate of 3.41% is achieved when d=0.5. (c) A larger λ imposes a stronger constraint on the circle number. By tuning the λ value within the proper range, e.g., λ=10, MCount can correctly recognize merged colonies. The original image has four colonies. (d) The average error rate on the segment dataset versus λ when d=0.5, where the minimum average error rate of 3.16% is achieved at λ=38.

The other important hyperparameter is λ, which is defined in Equation (4) and reflects the trade-off between colony number constraints and pairing of contour segments to candidate circles. As shown in Fig 3c, when λ=0, there is no constraint on colony number, and almost every contour piece matches a unique candidate circle, even if some segments come from the same colony. When λ increases, the counted number monotonically increases, and λ=10 provides a reasonable recognition result. When λ is very large, such as λ=1000, the constraint is so strong that only one circle is drawn. We determined that the best value for λ to minimize the average recognition error on segments is 38 (when d=0.5) by testing colony segments. S3 Fig (b) shows the optimal values for λ using the alternative definition for recognition error.

Both hyperparameters d and λ are critical to achieve accurate colony-counting results with MCount. However, their optimum values are dependent upon various factors such as image resolution and colony size, thus there is no universal set of values that will work for all scenarios. The standard approach for determining the optimal hyperparameter values is to optimize metrics such as the average recognition error through grid search and k-fold cross-validation, as demonstrated in the following section.

3.2. MCount performance

We utilized grid search and 10-fold cross-validation to minimize the average error rate on 15,847 colony segments and 960 sub-images to systematically determine the appropriate values for the hyperparameters d,λ. The optimal hyperparameter values were found to be d,λ=0.5,38 and d,λ=0.5,26, respectively. In this section, where we focus on the performance on sub-images, we will assess the performance of MCount using d,λ=0.5,26.

Fig 4a shows a bubble plot of MCount’s result versus the ground-truth label on colony segments. The regression line of MCount is close to the ground-truth line, indicating that MCount provides a good estimate for colony recognition on the segment dataset. It is worth noting that the bubbles representing coordinates (1,1), (2,2), and (3,3) are large compared to other bubbles because segments with 1-, 2-, and 3-colonies account for a large portion (59.6%, 18.52%, and 7.93%, respectively) of the segment dataset, and most of them are correctly evaluated by MCount. Therefore, the regression line of MCount largely depends on segments with few colonies, even though the segment size in the dataset ranges from 1 to 27. MCount tends to slightly underestimate the number of colonies, but it is still accurate for most segments, as 92.26% of segments have no more than 5 colonies. In this case, the actual bias resulting from underestimation is zero because the counted number can only be natural numbers. The counted number monotonically decreases as λ increases, as shown in S4 Fig (a), and this underestimation issue can be addressed by choosing an appropriate value for λ. To better visualize the accuracy of MCount on colony segments, we used a rank error curve in Fig 4b, where all 15,847 segments are ranked according to their error rate. The x-axis represents the rank of each segment, and the y-axis represents the corresponding error rate. When the error rate is larger than zero, counting tends to overestimate, while counting underestimates when the error rate is smaller than zero. A zero error rate was achieved on 91.45% of the segments, demonstrating MCount's high accuracy on colony segments.

Fig 4. Comparison of MCount (λ=26 and d=0.5) and NICE performance in recognizing colony segments and sub-images.

Fig 4

(a) Bubble plot of MCount counting result versus ground truth label on 15,847 colony segments. The size of each bubble at location (x,y) is proportional to the number of segments that have x merged colonies but are recognized as y colonies by MCount. The blue line represents the regression line for MCount, while the green dashed line represents the regression line for a perfect algorithm that always gives the result as the label. MCount exhibits slight underestimation as the blue line is lower than the dashed green line. (b) Rank error curve of error rate values for 15,847 segments. MCount recognizes 91.45% of segments with zero error rate. (c) Bubble plot of Mcount, NICE, AutoCellSeg, and OpenCFU counting results versus ground truth count label on 960 colony sub-images. The blue line is much closer to the dashed green line compared to the orange line, indicating that MCount is more accurate in recognizing sub-images. (d) Comparison of MCount, NICE, AutoCellSeg, and OpenCFU performance in underestimation or overestimation of colony numbers. MCount has fewer underestimated results and less severe underestimation and overestimation compared to NICE. (e) Boxplot of error rate on 960 sub-images for MCount and NICE. MCount has a much lower average error rate and a smaller interquartile range, indicating more consistent performance.

The goal of the colony counting algorithm is to give an accurate count on sub-images rather than segments, therefore we further evaluated the accuracy of MCount on sub-images (Fig 4c-d) and compared it with NICE, AutoCellSeg, and OpenCFU (Fig 4e). Notably, many algorithms either lack accessible source code [17], are unavailable for download [23], or fail to install correctly [8,18]. Consequently, we selected the most popular and widely used tools — NICE [9], OpenCFU [10], AutoCellSeg [11] — for comparison.

Fig 4c is a bubble plot showing all algorithms versus the ground-truth label on sub-images. All algorithms have relatively uniform distributions indicated by the uniform bubble sizes. While MCount only slightly underestimates the number of colonies, which can be addressed by tuning the λ value as shown in S4 Fig (b), NICE, AutoCellSeg, and OpenCFU largely underestimate the number of colonies. Since AutoCellSeg tends to count every continuous region as a single colony regardless of its shape, it has the most severe underestimation issue; as OpenCFU struggles with low-resolution images, its fitted line cannot even penetrate the origin (x,y) = (0,0). Fig 4d shows the rank error curve of all algorithms, with MCount having better accuracy than the others. Fig 4e shows the boxplot of both algorithms’ recognition error on each of the 960 sub-images, with MCount having a 3.99% average error rate, substantially lower than the average error rates of NICE/ AutoCellSeg/ OpenCFU of 16.54%/ 33.54%/ 50.31% respectively.

To gain insights into the sensitivity of MCount to hyperparameters, we explored two key questions. First, we examined whether the optimal hyperparameters for each sub-dataset significantly differed from the global optimal parameters, noting that these 10 sub-datasets were generated under varying experimental conditions. The results, as shown in S1 Table, indicate that the variation in optimal hyperparameters across different sub-datasets is minimal, and their corresponding recognition error rates are all lower than the error rate observed when using the global optimal hyperparameters (3.99%). Additionally, we observed a pattern where the value of the hyperparameter d tends to increase as the mean colony number decreases. When culture time is consistent, sparser colony density typically leads to larger colony size, suggesting that d is positively related to colony size. The second question we explored was how well the globally determined optimal hyperparameters perform across different sub-datasets. The results, presented in S2 Table, show that the globally optimized hyperparameters applied to these 10 groups allow MCount to maintain robust performance across varied conditions, with average error rates that remain significantly lower than NICE.

The processing time of MCount was evaluated on the dataset by randomly selecting 100 sub-images. The average time per sub-image was found to be less than 1.7 seconds, indicating that MCount can meet the demand for high-throughput processing. However, the processing time can be further optimized by using a better optimizer solver. A comparison of the processing time for MCount and other solutions is presented in S3 Table.

3.3. Hyperparameter optimization

In previous sections, we utilized a standard method of grid search and cross-validation with the goal of minimizing the average error rate on the entire dataset to tune the hyperparameter set d,λ, which can be further improved by using techniques such as random search [40]. However, this can be computationally intensive and time-consuming when exploring the hyperparameter space in practice. Furthermore, this method requires a pre-labeled dataset with a large quantity of data points, e.g., 960 (as shown earlier), which is impractical when deploying MCount for other counting tasks. In this section, we explore alternative methods for determining the hyperparameter set d,λ that only require a few labeled or even unlabeled data points. Table 1 summarizes the hyperparameter optimization methods used in this work.

Table 1. Summary of hyperparameter optimization methods used in MCount. The table includes the method number, image type (labeled or unlabeled), number of images used for optimization, hyperparameter d optimization method, and hyperparameter λ optimization method. Method 1 uses grid search with cross-validation on 960 labeled images (Section 3.2 & 3.3). Method 2 and method 3 decoupled the optimization of d and λ and use the same empirical method for determining d (Section 3.3). Method 2 uses the average λ of 10 ~  20 labeled images (Section 3.3.1), while method 3 chooses the value of λ that leads to equidispersion on 40 ~  50 unlabeled images (Section 3.3.2).

Image type Number of images d optimization λ optimization Section
1 Labeled 960 Grid Search +  Cross Validation
2 Labeled 10 ~  20 Empirical Average of Samples 3.3.1
3 Unlabeled 40 ~  50 Empirical Equidispersion Assumption 3.3.2

Although d and λ are not completely independent, they can be considered separately since they control different aspects of the algorithm. Specifically, d governs the level of detail in the contour representation, while λ controls the number of circles used in the optimization step. In this way, we can decouple these two hyperparameters and investigate them separately.

We propose an empirical method to determine the appropriate value of d. MCount can recognize elliptical segments that are usually comprised of two colonies by dividing the contour of the elliptical colonies into multiple contour segments, as shown in Fig 3a. We began with d=100 and selected a few elliptical segments that appeared to have two colonies. We then visualized the contour of each selected elliptical segment and gradually decreased the value of d until the contour was divided into 3~10 segments. By applying this approach to the benchmark, we determined that d=0.5 is a suitable value. In the following sections, we focus on statistical methods for tuning λ and examine their consistency using statistical procedures.

3.3.1. The selection of λ by averaging a small number of labeled images.

Although labeling hundreds of colony images to calibrate MCount requires extensive effort, for other counting tasks labeling only a few dozen images may be feasible. Therefore, we are interested in statistical methods that require only a few labeled images to tune λ. Here we propose one such method:

  • (1)

    Obtain a small number of labeled images (e.g., n=10).

  • (2)

    For each labeled image, find the value of λi such that MCount gives the correct colony count. Note that the count monotonically decreases as λ increases, making it easy to find λi.

  • (3)

    Set the hyperparameter λ to the average value, λ=λ¯=1niλi.

In order to examine whether the proposed method can consistently lead to an appropriate λ value and determine how many images are needed in step (1), we carried out a simulation using the benchmark dataset of 960 labeled images. The simulation procedure is as follows:

  • (1)

    Randomly sample n images from the 960 labeled images.

  • (2)

    For each labeled image, find the value of λi such that MCount gives the same colony count as the label. Note that the count monotonically decreases as λ increases, making it easy to find λi.

  • (3)

    Calculate the average recognition error rate on the 960 images with λ=λ¯=1niλi.

  • (4)

    Repeat steps (1) to (3) 1000 times.

The simulation procedure allows us to investigate the distribution of the average recognition error rates from 1000 trials to see whether this hyperparameter-tuning method can consistently lead to low error rates. Fig 5 is a boxplot of simulation results under different values of n. Not surprisingly, a larger n leads to a narrower distribution and a lower mean/median. When n=1020, 985 (998) out of 1000 trials have a λ value leading to a low error rate (<8%) on the 960 images, which indicates consistently low recognition error rates. In practice, labeling ten to twenty colony images is a common practice in bio-labs, making it easy to calibrate MCount when applied to future counting tasks.

Fig 5. Distribution of average recognition error rates on the benchmark dataset when repeatedly implementing the hyperparameter-tuning method using n labeled images 1000 times.

Fig 5

In each replicate, a few (1, 3, 5, 10, 20, 30, or 50) images are randomly selected from the benchmark dataset, and λ is tuned for each image to a proper value so that MCount gives the same counting number as the label. Then, the average of λ is chosen for this replicate, and the average recognition error rate is calculated on the benchmark using this λ. By simulating this procedure for 1000 replicates, we can plot the distribution of average recognition error rates. As expected, increasing n results in a narrower distribution of average recognition error rates, leading to more consistent performance. Note that when n=10, all recognition errors fall in the range of 3.5% to 13% with a mean of 5.17% (median of 4.77%), much lower than the recognition error rate of NICE at 16.54% (15.79%).

3.3.2. The selection of λ based on equidispersion assumption on unlabeled images.

In many labs, there are often cases where acquiring a large number of images is feasible, but labeling even a single image is challenging. These situations require statistical methods that can tune λ using unlabeled images. Further, introducing some prior information could help make good use of the unlabeled data. One such prior is an assumption that the distribution of colony numbers features equidispersion, meaning that the mean is equal to the variance, a key property of the Poisson distribution. We discuss this prior information in more detail before introducing and examining our method in the following section. By utilizing this prior information/assumption, we hypothesize that we can obtain a reasonably accurate λ value using unlabeled images. By analyzing the simulation results we can determine the minimum number of unlabeled images required to obtain an accurate λ value for a specific task.

3.3.2.1. Hypothesis test for assessing equidispersion: It is assumed that many biological phenomena exhibit a Poisson distribution, and colony numbers are an example, where a fixed volume of liquid is independently and randomly sampled from the same source for every plating [47,48]. In this case, the equidispersion property is naturally satisfied.

However, it is important to note that the prior knowledge chosen in this method is the equidispersion property, not specifically the Poisson distribution. Equidispersion is a weaker requirement than the Poisson distribution, and it is possible that other distributions can also exhibit equidispersion. In S2 File and S4 Table, we further discuss this scenario and show that our benchmark is closer to a normal distribution than a Poisson distribution, although our equidispersion assumption still leads to a low recognition error.

We prefer assuming equidispersion to the Poisson distribution because it can be challenging to use statistical methods to confirm that a distribution follows Poisson, while equidispersion can be tested more easily. Here we propose a simple hypothesis test to assess equidispersion, known as the Poisson Dispersion test, which also serves as a likelihood test for the Poisson distribution:

H0:the data has equidispersion,i.e.,mean=variance
H1:meanvariance

The test statistic is:

D=i=1NXiX¯2X¯ (5)

Then we can calculate the p-value associated with the chi-square distribution (N-1 degrees of freedom). If the p-value is greater than the chosen level of significance, we fail to reject the null hypothesis H0 and can assume that equidispersion is satisfied.

3.3.2.2. Proposed method and its examination: Assuming that equidispersion is satisfied, we propose a statistical method that utilizes unlabeled images to tune λ as follows:

  • (1)

    Obtain a set of unlabeled images (e.g., n=40).

  • (2)

    Apply MCount to the images and obtain the counting results under different λ values.

  • (3)

    Calculate the mean and variance of the counting results under different λ values and find a value λ^ where the mean equals the variance.

  • (4)

    Set the hyperparameter λ=λ^.

To evaluate the effectiveness of the proposed method using unlabeled images, we conducted a simulation procedure as follows:

  • (1)

    Randomly select n images from a dataset comprising 96 labeled images with equidispersion. Erase the label of the selected images.

  • (2)

    Apply MCount to the n images and obtain counting results under different λ values.

  • (3)

    Calculate the mean and variance of the counting results under different λ values and find a value λ^ where the mean equals the variance.

  • (4)

    Calculate the average recognition error rate on 96 labeled images with λ=λ^.

  • (5)

    Repeat steps (1) to (4) 1000 times.

The boxplot in Fig 6 shows the distribution of the 1000 average recognition error rates for different values of n. Like the case for labeled images, larger n values lead to a narrower distribution and smaller mean. Interestingly, the median remains constant regardless of n. When n40, the maximum line of the boxplot is lower than 8%, indicating consistently low recognition error rates. However, the required n value is much larger than that for labeled images, as unlabeled images contain less information, requiring more training data for hyperparameter optimization.

Fig 6. Distribution of average recognition error rates on 96 images with equidispersion when repeatedly implementing the hyperparameter-tuning method using n unlabeled images 1000 times.

Fig 6

In each replicate, a few (3, 5, 10, 20, 30, 40, or 50) images with the label erased are randomly selected from 96 images, and λ is tuned so that the mean of the counting equals the variance. Then, the average recognition error rate is calculated on the 96 labeled images using this λ. By simulating this procedure for 1000 replicates, we can plot the distribution of average recognition error rates. As n increases, the distribution of average recognition error rates becomes narrower, leading to a more consistent distribution. Note that only when n>40, the maximum line of the box plot is lower than 8%, and the mean is equal or lower than 4.04%. The required n is much larger than the case for labeled images because unlabeled images have less information, requiring more training data for hyperparameter optimization. These results demonstrate that the proposed method is effective in achieving consistently low error rates using only a small number of unlabeled images for hyperparameter tuning.

Notably, the equidispersion criterion helps identify a range of reasonable lambda values, though it may not always yield a unique solution. In practice, we often observe that within a reasonable range—typically near the default values—a single optimal lambda value emerges. This is because variance is generally less sensitive to changes in lambda, while the mean is more responsive, aiding in narrowing down the solution. When multiple lambda values meet the criterion, they are applied to the colony counting process and manually inspected to determine the most accurate outcome, with clear indications when overestimation or underestimation occurs.

4. Conclusion

In this work, we propose a colony counting solution, MCount, that can recognize merged colonies that frequently occur in high-throughput workflows, which is beyond the capabilities of current solutions that adopt region-based algorithms. MCount extracts contour information and combines it with regional information using an optimization algorithm. To evaluate the performance of MCount, we prepared a GFP-fluorescent E. coli DH10-beta-based colony dataset, which is comprised of both sub-images and labeled segments. MCount maintains an average recognition error of 3.99% on the sub-image dataset (using grid search and 10-fold cross-validation to minimize the average error rate), which is much lower than current solutions like NICE of 16.54%.

Given that MCount only has two hyperparameters, it will be easy to deploy for other counting tasks. In addition to the standard hyperparameter optimization method, which requires several labeled images, we also proposed two methods that require a small number of labeled images or unlabeled images, respectively. To examine the statistical properties of the proposed methods, we conducted simulations and found that they all guarantee consistently low error rates compared to existing methods. The simulations showed that the method requiring labeled images achieved low error rates with as few as 10 labeled images, while the method suitable for unlabeled images required at least 40 images to achieve consistently low error rates.

Overall, the statistical evaluation of the proposed methods provides a strong basis for their potential deployment in various counting tasks. Future applications of this solution include colony classification for multiple strains on the same plate, which could be achieved by classification algorithms such as K-nearest neighbor based on colony color and size.

Supporting Information

S1 Fig. The Polygon Approximation Algorithm uses turning points on the contour to represent an inscribed polygon, where d controls how many turning points are generated.

(a) For consecutive three turning points Ti1, Ti, and Ti+1, d represents the distance from Ti to the line Ti1Ti+1¯. (b) The larger the value of d, the more turning points are generated to represent the contour. The turning points are represented as black dots, while contour segments divided by turning points are shown in different colors.

(TIF)

S2 Fig. A dataset that includes colony segments, labeled using historical photographs, is used to optimize and evaluate the performance of MCount.

(a) 960 sub-images are obtained by cropping 10 plates of fluorescent E. coli NEB10-beta and further divided into 15,847 segments using foreground extraction and segmentation. (b) The colony number distribution of sub-images shows that most sub-images have 10 ~  40 colonies. (c) The percentage of single-colony, two-colony, and three-colony segments is 59.6%, 18.52%, and 7.93%, respectively, which takes 86.05% in total. A well-performing algorithm is expected to correctly recognize almost all none and mildly merged colonies. The remaining percentage of segments, merged in a denser manner, requires the algorithm to infer sophisticated shapes. Note that 3.90% of segments are invalid because the merging of colonies is too severe to be labeled correctly, denoted as -1 in the left figure. (d) All segments are labeled according to their shape in a photograph taken about 4 hours ago to ensure labeling accuracy, including segments that are hard for humans to label.

(TIF)

pone.0311242.s002.tif (30.8MB, tif)
S3 Fig

The optimal values of d and λ are determined using a different definition of recognition error. (a) The average error rate on the segment dataset is plotted against d when λ=26. The minimum average error rate of 8.50% is achieved at d=0.5. (b) The average error rate on the segment dataset is plotted against λ when d=0.5. The minimum average error rate of 8.46% is achieved at λ=28. An error of each segment is defined as 1MCountLabel and the error was averaged across all segments, denoting the average error rate.

(TIF)

pone.0311242.s003.tif (1.5MB, tif)
S4 Fig

Increasing λ results in monotonously less counting leading to underestimation of colony number. The plot shows the MCount counting result with different λ values versus ground truth label on (a) 15,847 colony segments and (b) 960 sub-images, respectively. The green dashed line represents the regression line for a perfect colony estimator that always gives the result as the label, while the colored line represents the regression line for MCount. By tuning λ, it is possible to address the overestimation/underestimation issue. Increasing λ results in a monotonous decrease in the number of colonies counted, leading to an underestimation of colony number.

(TIF)

S1 Table. Variation of optimal hyperparameters across 10 sub-datasets generated under different experimental conditions.

Mean and variance of colony number, optimal hyperparameter values, and the corresponding recognition error rate for each sub-dataset. The results demonstrate minimal variation in hyperparameters across sub-datasets and consistently lower error rates compared to the global settings (3.99%).

(DOCX)

pone.0311242.s005.docx (16.2KB, docx)
S2 Table

Performance of MCount using globally optimized hyperparameters (λ=26 and d=0.5) across 10 sub-datasets. The table presents the average error rates for each sub-dataset when applying the global optimal hyperparameters. The results indicate that MCount maintains robust performance across varied experimental conditions, with error rates significantly lower than those achieved by NICE.

(DOCX)

pone.0311242.s006.docx (16.1KB, docx)
S3 Table. Processing time of different colony counting solutions for 100 randomly selected sub-images.

The optimization solver for MCount is ‘COIN_CMD’ in PuLP, and the solver can be changed to achieve faster processing times at the cost of sacrificing accuracy.

(DOCX)

pone.0311242.s007.docx (14.2KB, docx)
S4 Table. Mean, variance, and p-values of tests for 10 datasets.

The mean and variance values were calculated from the 96 colony counts of each dataset. The first column represents the dataset index. The second column represents the mean value of the 96 colony counts, and the third column represents the variance of the 96 colony counts. The 4th and 5th columns show the p-values for Kolmogorov–Smirnov (KS) tests for Normal and Poisson distributions, respectively, under the significant level of 0.05, where no rejection is made for any datasets. The 6th column shows the p-values for Poisson Dispersion tests, and datasets whose null hypotheses are rejected under the significant level of 0.05 are denoted with a star symbol * .

(DOCX)

pone.0311242.s008.docx (15.6KB, docx)
S1 File. Preparation and discussion on bacterial colony dataset.

Discussion, cultivation and preparation on E. coli Dataset used in this study.

(DOCX)

pone.0311242.s009.docx (16.4KB, docx)
S2 File. Discussion on the distribution of colony numbers.

(DOCX)

pone.0311242.s010.docx (14.2KB, docx)

Data Availability

The source code and integrated solution are freely available at https://github.com/hyu-kim/mcount. All methods implemented and data used are publicly available on Dryad at https://doi.org/10.5061/dryad.2280gb62f

Funding Statement

This work was supported by National Institutes of Health (NIH) through grant number RM1 GM135102 to C.R.B., S.C., P.-H.H., and H.K. received support from the Department of Energy’s Genome Sciences Program through grant SCW1039. URL1: https://grantome.com/grant/NIH/RM1-GM135102-01 URL2: https://genomicscience.energy.gov/llnl/ The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Smith CJ, Osborn AM. Advantages and limitations of quantitative PCR (Q-PCR)-based approaches in microbial ecology. FEMS Microbiol Ecol [Internet]. 2009 Jan 1 [cited 2023 Apr 2];67(1):6–20. Available from: https://academic.oup.com/femsec/article/67/1/6/518537 [DOI] [PubMed] [Google Scholar]
  • 2.Davey HM, Kell DB. Flow cytometry and cell sorting of heterogeneous microbial populations: the importance of single-cell analyses. Microbiol Rev. 1996;60(4):641–96. doi: 10.1128/mr.60.4.641-696.1996 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Beal J, Farny NG, Haddock-Angelli T, Selvarajah V, Baldwin GS, Buckley-Taylor R, et al. Robust estimation of bacterial cell count from optical density. Commun Biol. 2020;3(1):512. doi: 10.1038/s42003-020-01127-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Hazan R, Que Y-A, Maura D, Rahme LG. A method for high throughput determination of viable bacteria cell counts in 96-well plates. BMC Microbiol. 2012;12:259. doi: 10.1186/1471-2180-12-259 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Sieuwerts S, De Bok FAM, Mols E, De Vos WM, Van Hylckama Vlieg JET. A simple and fast method for determining colony forming units. Lett Appl Microbiol [Internet]. 2008 Oct 1 [cited 2023 Apr 2];47(4):275–8. Available from: https://academic.oup.com/lambio/article/47/4/275/6700914 [DOI] [PubMed] [Google Scholar]
  • 6.Willey JM, Sherwood Linda, Woolverton CJ, Prescott LM. Prescott, Harley, and Klein’s microbiology. 7th ed. McGraw-Hill Higher Education; 2008. 4–4 p.
  • 7.Breed RS, Dotterrer WD. The Number of Colonies Allowable on Satisfactory Agar Plates. J Bacteriol. 1916;1(3):321–31. doi: 10.1128/jb.1.3.321-331.1916 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Brugger S, Baumberger C, Jost M, Brugger J. Automated counting of bacterial colony forming units on agar plates. PLoS One. 2012;7(3):e33695. doi: 10.1371/journal.pone.0033695 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Clarke ML, Burton RL, Hill AN, Litorja M, Nahm MH, Hwang J. Low-cost, high-throughput, automated counting of bacterial colonies. Cytometry A. 2010;77(8):790–7. doi: 10.1002/cyto.a.20864 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Geissmann Q. OpenCFU, a new free and open-source software to count cell colonies and other circular objects. PLoS One. 2013;8(2):e54072. doi: 10.1371/journal.pone.0054072 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Khan AUM, Torelli A, Wolf I, Gretz N. AutoCellSeg: robust automatic colony forming unit (CFU)/cell analysis using adaptive image segmentation and easy-to-use post-editing techniques. Sci Rep. 2018;8(1):7302. doi: 10.1038/s41598-018-24916-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Choudhry P. High-Throughput Method for Automated Colony and Cell Counting by Digital Image Analysis Based on Edge Detection. PLoS One. 2016;11(2):e0148469. doi: 10.1371/journal.pone.0148469 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Cai Z, Chattopadhyay N, Liu WJ, Chan C, Pignol J-P, Reilly RM. Optimized digital counting colonies of clonogenic assays using ImageJ software and customized macros: comparison with manual counting. Int J Radiat Biol. 2011;87(11):1135–46. doi: 10.3109/09553002.2011.622033 [DOI] [PubMed] [Google Scholar]
  • 14.Sieuwerts S, de Bok FAM, Mols E, de vos WM, Vlieg JET van H. A simple and fast method for determining colony forming units. Lett Appl Microbiol. 2008;47(4):275–8. doi: 10.1111/j.1472-765X.2008.02417.x [DOI] [PubMed] [Google Scholar]
  • 15.Bray M-A, Vokes MS, Carpenter AE. Using CellProfiler for Automatic Identification and Measurement of Biological Objects in Images. Curr Protoc Mol Biol. 2015;10914.17.1-14.17.13. doi: 10.1002/0471142727.mb1417s109 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Bär J, Boumasmoud M, Kouyos RD, Zinkernagel AS, Vulin C. Efficient microbial colony growth dynamics quantification with ColTapp, an automated image analysis application. Sci Rep. 2020;10(1):16084. doi: 10.1038/s41598-020-72979-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Chen W-B, Zhang C. An automated bacterial colony counting and classification system. Inf Syst Front. 2009;11(4):349–68. doi: 10.1007/s10796-009-9149-0 [DOI] [Google Scholar]
  • 18.Zhang L. Machine learning for enumeration of cell colony forming units. Vis Comput Ind Biomed Art. 2022;5(1):26. doi: 10.1186/s42492-022-00122-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Sutton S. Counting colonies. Pharmaceutical Microbiology Forum Newsletter [Internet]. 2006 [cited 2023 Mar 28];12(9). Available from: www.remel.com/products/industry [Google Scholar]
  • 20.Tomasiewicz D, Hotchkiss D, Reinbold G, Read R, Hartman P. The most suitable number of colonies on plates for counting. Journal of Food Protection. 1980;43(4):282–6. [DOI] [PubMed] [Google Scholar]
  • 21.Xie W, Noble JA, Zisserman A. Microscopy cell counting and detection with fully convolutional regression networks. Comput. Methods Biomech. Biomed. Eng.: Imaging Vis. 2023;6(3):283–92. https://doi.org/10.1080/21681163.2016.1149104 [Google Scholar]
  • 22.Jiang N, Yu F. Multi-column network for cell counting. OSA Continuum. 2020;3(7):1834. [Google Scholar]
  • 23.Selinummi J, Seppälä J, Yli-Harja O, Puhakka JA. Software for quantification of labeled bacteria from digital microscope images by automated image analysis. Biotechniques. 2005;39(6):859–63. doi: 10.2144/000112018 [DOI] [PubMed] [Google Scholar]
  • 24.Costa CM, Yang S. Counting pollen grains using readily available, free image processing and analysis software. Ann Bot [Internet]. 2009 Oct 1 [cited 2023 Mar 28];104(5):1005–10. Available from: https://academic.oup.com/aob/article/104/5/1005/138988 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Monteiro FC. Pollen grain recognition through deep learning convolutional neural networks. AIP Conf Proc [Internet]. 2022 Apr 6 [cited 2023 Mar 28];2425(1):140001. Available from: https://aip.scitation.org/doi/abs/10.1063/5.0081614 [Google Scholar]
  • 26.de Langlard M, Al-Saddik H, Charton S, Debayle J, Lamadie F. An efficiency improved recognition algorithm for highly overlapping ellipses: Application to dense bubbly flows. Pattern Recognition Letters. 2018;101:88–95. [Google Scholar]
  • 27.Zhang W, Jiang X, Liu Y. A method for recognizing overlapping elliptical bubbles in bubble image. Pattern Recognition Letters. 2012;33(12):1543–8. [Google Scholar]
  • 28.Prasad D, Leung M, Cho S. Edge curvature and convexity based ellipse detection method. Pattern Recognition. 2012;45(9):3204–21. [Google Scholar]
  • 29.Thompson M, Duda RO, Hart PE. Pattern classification and scene analysis. Vol. 7, A Wiley-Interscience Publication; 1974. [Google Scholar]
  • 30.Zafari S, Eerola T, Sampo J, Kälviäinen H, Haario H. Segmentation of partially overlapping nanoparticles using concave points. In: International Symposium on Visual Computing [Internet]. Springer Verlag; 2015 [cited 2023 Mar 28]. Available from: https://link.springer.com/chapter/10.1007/978-3-319-27857-5_17 [Google Scholar]
  • 31.Zou T, Pan T, Taylor M, Stern H. Recognition of overlapping elliptical objects in a binary image. Pattern Anal Applic. 2021;24(3):1193–206. doi: 10.1007/s10044-020-00951-z [DOI] [Google Scholar]
  • 32.Guo X, Yu H, Rossetti B, Teodoro G, Brat D, Kong J. Clumped nuclei segmentation with adjacent point match and local shape-based intensity analysis in fluorescence microscopy images. Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS. 2018;2018:3410–3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Zafari S, Murashkina M, Eerola T, Sampo J, Kälviäinen H, Haario H. Resolving overlapping convex objects in silhouette images by concavity analysis and Gaussian process. J Vis Commun Image Represent. 2020;73102962. [Google Scholar]
  • 34.Jung C, Kim C. Segmenting clustered nuclei using H-minima transform-based marker extraction and contour parameterization. IEEE Trans Biomed Eng. 2010; 57(10):2600-2604. https://doi.org/10.1109/TBME.2010.2060336 [DOI] [PubMed] [Google Scholar]
  • 35.Park C, Huang J, Ji J, Ding Y. Segmentation, inference and classification of partially overlapping nanoparticles. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2012;35(3):1–1. [DOI] [PubMed] [Google Scholar]
  • 36.Paglieroni DW. Distance transforms: Properties and machine vision applications. CVGIP: Graphical Models and Image Processing. 1992;54(1):56–74. doi: 10.1016/1049-9652(92)90034-u [DOI] [Google Scholar]
  • 37.Zhang Y, Jiang H, Ye T, Juhas M. Deep Learning for Imaging and Detection of Microorganisms. Trends Microbiol. 2021;29(7):569–72. doi: 10.1016/j.tim.2021.01.006 [DOI] [PubMed] [Google Scholar]
  • 38.Zhang J, Li C, Rahaman M, Yao Y, Ma P, Zhang J. A comprehensive review of image analysis methods for microorganism counting: from classical image processing to deep learning approaches. Artificial Intelligence Review. 2022;55(4):2875–944. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Yang L, Shami A. On hyperparameter optimization of machine learning algorithms: Theory and practice. Neurocomputing. 2020;415295–316. [Google Scholar]
  • 40.Hutter F, Kotthoff L, Vanschoren J. Automated machine learning: Methods, systems, challenges [Internet]. Springer; 2019 [cited 2023 Apr 30]. Available from: http://www.springer.com/series/15602 [Google Scholar]
  • 41.Otsu N. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans Syst, Man, Cybern. 1979;9(1):62–6. doi: 10.1109/tsmc.1979.4310076 [DOI] [Google Scholar]
  • 42.Suzuki S, be KA. Topological structural analysis of digitized binary images by border following. Comput Vis Graph Image Process. 1985 Apr 1;30(1):32–46.
  • 43.Zafari S, Eerola T, Sampo J, Kälviäinen H, Haario H. Comparison of concave point detection methods for overlapping convex objects segmentation. In: Scandinavian Conference on Image Analysis [Internet]. Springer Verlag; 2017 [cited 2023 Mar 28]. Available from: https://link.springer.com/chapter/10.1007/978-3-319-59129-2_21
  • 44.Borgefors G. Distance transformations in digital images. Comput Vis Graph Image Process. 1986;34(3):344–71. [Google Scholar]
  • 45.Balas E. An Additive Algorithm for Solving Linear Programs with Zero-One Variables. Operations Research. 1965;13(4):517–46. doi: 10.1287/opre.13.4.517 [DOI] [Google Scholar]
  • 46.Glover F. A Multiphase-Dual Algorithm for the Zero-One Integer Programming Problem. Operations Research. 1965;13(6):879–919. doi: 10.1287/opre.13.6.879 [DOI] [Google Scholar]
  • 47.Pettigrew HM, Mohler WC. A rapid test for the Poisson distribution using the range. Biometrics. 1967;23(4):685–92. doi: 10.2307/2528422 [DOI] [PubMed] [Google Scholar]
  • 48.Fisher RA, Thornton HG, Mackenzie WA. The accuracy of the plating method of estimating the density of bacterial populations. Annals of Applied Biology [Internet]. 1922 Nov 1 [cited 2023 May 2];9(3–4):325–59. Available from: https://onlinelibrary.wiley.com/doi/full/10.1111/j.1744-7348.1922.tb05962.x [Google Scholar]

Decision Letter 0

Florian Rehfeldt

11 Mar 2024

PONE-D-24-05814MCount: An automated colony counting tool for high-throughput microbiologyPLOS ONE

Dear Dr. Buie,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Apr 25 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols .

We look forward to receiving your revised manuscript.

Kind regards,

Florian Rehfeldt

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. 

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

3. Thank you for stating the following in the Competing Interests section: 

"I have read the journal's policy and the authors of this manuscript have the following competing interests: C.R.B. is a Co-Founder and Advisor of Kytopen Corp.  The remaining authors declare no competing interests."

Please confirm that this does not alter your adherence to all PLOS ONE policies on sharing data and materials, by including the following statement: "This does not alter our adherence to  PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests).  If there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared. 

Please include your updated Competing Interests statement in your cover letter; we will change the online submission form on your behalf.

4. Thank you for uploading your study's underlying data set. Unfortunately, the repository you have noted in your Data Availability statement does not qualify as an acceptable data repository according to PLOS's standards.

At this time, please upload the minimal data set necessary to replicate your study's findings to a stable, public repository (such as figshare or Dryad) and provide us with the relevant URLs, DOIs, or accession numbers that may be used to access these data. For a list of recommended repositories and additional information on PLOS standards for data deposition, please see https://journals.plos.org/plosone/s/recommended-repositories.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: In their manuscript, Chen et al. present a sophisticated workflow for automated colony counting, which they name 'MCount'. The manuscript is well written, and in principle suitable for publication. Before publication, the authors may want to consider the following points:

1.) In the introduction, the authors discuss several existing tools for colony counting (NICE, AutoCellSeg, OpenCFU), but in the results section they only compare the performance of MCount to NICE. Since the authors went through the effort of creating a ground-truth dataset, it would be nice to see also the performance of the other methods on this dataset.

2.) MCount has two major tunable hyperparameters. As I understood, this is not the case for NICE. I acknowledge the effort the authors have put into guiding users in how to determine those hyperparameters. Still, to understand to what extent the improved performance of MCount is due to those tuned hyperparameters, it would be interesting to see how MCount would perform 'out-of-the-box' without parameter tuning, and how this would compare to NICE. To assess this, the authors could test how optimized hyperparameters tuned on the current dataset perform on an independent dataset, ideally obtained in an independent setting (e.g. by another lab).

3.) Along those lines, how sensitive is the optimum of the hyperparameters to various datasets? Especially datasets with similar colony density but obtained with a different imaging setup?

4.) To determine the hyperparameter lambda from unlabelled data, the authors present a strategy building on a equidispersion criteria. I did not understand, whether this criteria leads to a unique optimum for lambda? If not, what are the implications for this approach?

5.) The authors apply their method using a fluorescent strain. Is fluorescence imaging required for the good performance?

Reviewer #2: This work presents a classical image analysis workflow for counting CFUs. The method is described in all necessary detail, code and data are available, and the results are quantitatively analyzed. The main limitation of this manuscript is that it lacks a thorough comparison to existing literature and state-of-the-art. Apart from some qualitative comparison in the first figure and a few plots comparing their work to NICE in Fig. 4, no such comparison is performed. The literature cited is mostly from computer vision applications to other types of overlapping objects, but not microbiology or bioimaging.

The problem of counting objects is nowadays usually solved with object detection and/or segmentation neural networks, some of which are specifically taylored towards separating overlapping convex objects such as cell nuclei (e.g. Stardist, Cellpose). In the field of colony counting, some classical and ML-based literature not cited here (just as examples) include:

https://doi.org/10.2144%2F000112018

https://www.nature.com/articles/s41598-020-72979-4

https://link.springer.com/doi/10.1007/s10796-009-9149-0

http://arxiv.org/abs/2009.00926

https://doi.org/10.1364/OSAC.396603

https://doi.org/10.1016/j.tim.2021.01.006

https://vciba.springeropen.com/articles/10.1186/s42492-022-00122-3

The literature is for example reviewed in

https://doi.org/10.1016/j.tim.2021.01.006

https://link.springer.com/article/10.1007/s10462-021-10082-4

It is not necessary to cite all of this work in the manuscript, but the authors are encouraged to take the effort and spend some time studying the current literature, and then add a discussion of their work with respect to the current state of the art, and ideally also some more quantitative comparison.

Some technical remarks: it would be great if more documentation, explanation of the individual steps, and installation instructions could be added to the Jupter Notebooks. The figures appear in very low quality in the PDF, but this might be a production issue unrelated to the submitted material.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy .

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step.

PLoS One. 2025 Mar 19;20(3):e0311242. doi: 10.1371/journal.pone.0311242.r003

Author response to Decision Letter 0


10 Sep 2024

Response to Editor

[Comment]

Thank you for uploading your study's underlying data set. Unfortunately, the repository you have noted in your Data Availability statement does not qualify as an acceptable data repository according to PLOS's standards. At this time, please upload the minimal data set necessary to replicate your study's findings to a stable, public repository (such as figshare or Dryad) and provide us with the relevant URLs, DOIs, or accession numbers that may be used to access these data. For a list of recommended repositories and additional information on PLOS standards for data deposition, please see https://journals.plos.org/plosone/s/recommended-repositories.

[Response]

We appreciate the editor’s time reviewing our manuscript and Data Availability statement. We accept the suggestion and have uploaded the data set to Dryad. The relevant URL or DOI to access the data set is as follows: https://doi.org/10.5061/dryad.2280gb62f (forthcoming as of September 5, 2024; for temporary access in the meantime, please access via https://datadryad.org/stash/share/xh0Ek1hMtE-lihJ5wB590g-rnLcfL6G0irdWlaVxjhE). We have added the URL to the section “Data Availability” as follows:

“All methods implemented and data used are publicly available on Dryad (https://doi.org/10.5061/dryad.2280gb62).”

Response to Reviewer 1

[Comment 1-1]

In the introduction, the authors discuss several existing tools for colony counting (NICE, AutoCellSeg, OpenCFU), but in the results section they only compare the performance of MCount to NICE. Since the authors went through the effort of creating a ground-truth dataset, it would be nice to see also the performance of the other methods on this dataset.

[Response 1-1]

We appreciate the reviewer’s valuable comment. Indeed, we previously did not compare the results of MCount to AutoCellSeg and OpenCFU and instead outlined their limitations as described in Introduction and in Figure 1. Following reviewer’s comment, we have performed colony counting using AutoCellSeg and OpenCFU and have now added their results to Figure 4 for comparison.

As expected, we see that MCount has shown the lowest error rate compared to other methods such as NICE, AutoCellSeg, and OpenCFU. We have added our motivations and observations of the comparison study (lines 305-309 of main text)

[Comment 1-2]

MCount has two major tunable hyperparameters. As I understood, this is not the case for NICE. I acknowledge the effort the authors have put into guiding users in how to determine those hyperparameters. Still, to understand to what extent the improved performance of MCount is due to those tuned hyperparameters, it would be interesting to see how MCount would perform 'out-of-the-box' without parameter tuning, and how this would compare to NICE. To assess this, the authors could test how optimized hyperparameters tuned on the current dataset perform on an independent dataset, ideally obtained in an independent setting (e.g. by another lab)

[Response 1-2]

We appreciate your concerns on hyperparameter tuning and would like to address them as follows.

NICE Also Has Hyperparameters: It's important to note that NICE, like MCount, includes hyperparameters, such as sigma, i.e. the threshold of Otsu algorithm used in NICE. While many users might default to the standard values, these parameters significantly impact NICE's performance. However, in practice, many biologists prioritize experimental efficiency or believe that the error introduced by hyperparameters is smaller than the inherent variability in biological experiments. As a result, the adjustment process in NICE often involves providing a single example image and manually tuning the parameters based on subjective judgment, and in many cases, researchers simply use the default values without further calibration.

Necessity of Hyperparameters: We believe that having tunable hyperparameters is essential for a tool like MCount, as it allows the tool to be more adaptable to different scenarios and achieve optimal performance. From a data science perspective, the best approach to obtain optimal hyperparameters is to construct a small, well-labeled dataset that reflects the typical experimental setting, which can then be used to calibrate values like sigma. This approach ensures that MCount can maintain high accuracy across various experimental conditions and image qualities. However, the choice of hyperparameters presents a dilemma: on one hand, we want them to be sensitive enough to ensure adaptability across different experimental conditions; on the other hand, we want them to be less sensitive, making the tool easier to use for those who prefer simplicity. Given that MCount has the more advanced capability to differentiate overlapping colonies, its parameters are necessarily more sensitive compared to NICE.

Performance on Different Datasets: While we did not conduct experiments on a completely new and independent dataset, as our dataset, which consists of 10 groups of 96 images each, already exhibits sufficient variance. These 10 groups vary in terms of colony size, density, and even the setup used during image collection (e.g., variations in camera positioning, background lighting, etc.). This natural diversity in the dataset makes it suitable for evaluating the robustness of MCount's performance. We added corresponding clarification in the main text (lines 213-216). To address the reviewer's concerns, we applied the optimized hyperparameters to these 10 groups and assessed MCount's accuracy. Below is a summary of the results, which has now been added into Supplemental Information as S2 Table. These results demonstrate that MCount maintains strong performance across varied conditions within the original dataset, with an average error rate that remains significantly lower than NICE (16.54%). This suggests that the optimized hyperparameters are robust and generalizable across different experimental settings.

[Comment 1-3]

Along those lines, how sensitive is the optimum of the hyperparameters to various datasets? Especially datasets with similar colony density but obtained with a different imaging setup?

[Response 1-3]

The sensitivity of the optimal hyperparameters can indeed be characterized from two perspectives:

Effect of Hyperparameter Adjustment on Results: The first perspective is to observe how adjusting the hyperparameters affects the results. We previously addressed this and presented the result in S3 Fig, which illustrates how variations in hyperparameters impact the performance of MCount.

Variation of Optimum Across Different Datasets: The second perspective, as the reviewer suggested, concerns whether the optimum hyperparameters differ across datasets, particularly those with similar colony density but obtained using different imaging setups. To explore this, we calculated the optimal hyperparameters for each of the 10 sub-datasets in our study. Again, these 10 sets vary in terms of colony size, density, and the imaging setup. The results have now been summarized in S1 Table as follows:

These results indicate that while there is some variation in the optimal hyperparameters across different sub-datasets, MCount maintains robust performance. The differences in optimal hyperparameters are likely influenced by factors such as colony size, density, and variations in the imaging setup (e.g., camera positioning, background lighting). This analysis provides insights into the robustness and adaptability of MCount across different experimental conditions.

Additionally, we observed a pattern where the value of the hyperparameter 𝑑 tends to increase when the mean colony number decreases (considering a consistent culture time, which results in larger colonies). This correlation suggests that 𝑑 is positively related to colony size, which aligns with our expectations.

It's also worth noting that the recognition error rates for these individually optimized sub-datasets are significantly lower than the error rate observed when using a single set of optimal parameters for all 10 datasets combined (4%). This further underscores the benefit of tuning hyperparameters to suit specific experimental conditions.

Our assessment of MCount across different datasets has now been added to the main text (lines 320-331).

[Comment 1-4]

To determine the hyperparameter lambda from unlabelled data, the authors present a strategy building on a equidispersion criteria. I did not understand, whether this criteria leads to a unique optimum for lambda? If not, what are the implications for this approach?

[Response 1-4]

As the reviewer rightly pointed out, the equidispersion criterion does not guarantee a unique solution for the lambda parameter. Theoretically, there could be multiple lambda values that satisfy this criterion. However, this does not render the equidispersion criterion ineffective; rather, it serves as a tool to identify a range of lambda values that are relatively reasonable. We have added corresponding clarifications to the main text (lines 467-473).

Practical Observations: In our experimental experience with various datasets, we typically observe that within a reasonable range (usually around the default values of the parameters), we tend to obtain a single solution for lambda. This is because variance is relatively insensitive to changes in lambda, while the mean is more responsive to variations in lambda, helping to narrow down the solution.

Handling Multiple Lambda Values: In cases where multiple lambda values meet the equidispersion criterion, we can use these values to perform the colony counting and then manually inspect the results to determine which value provides the most reasonable outcomes. We have encountered situations where certain lambda values resulted in clearly overestimated or underestimated colony counts, making it evident which lambda should be chosen.

[Comment 1-5]

The authors apply their method using a fluorescent strain. Is fluorescence imaging required for the good performance?

[Response 1-5]

While fluorescence imaging can enhance performance, it is not strictly required. We have also successfully applied our method to non-fluorescent images in other, unpublished research. For instance, we used our approach to count colonies with different morphological characteristics in images containing mixtures of colonies with three different colors or sizes. These experiments demonstrate that our method is versatile and can effectively adapt to non-fluorescent imaging scenarios, provided the images are of sufficient quality to distinguish between colonies.

In this study, fluorescence imaging was primarily used to facilitate the manual annotation of the dataset. Fluorescence provides clear and distinct colony boundaries, making it easier to create accurate ground-truth labels, which are essential for training and validating the model.

We have added corresponding edits to the main text (lines 209-212).

Response to Reviewer 2

[Comment 2-1]

This work presents a classical image analysis workflow for counting CFUs. The method is described in all necessary detail, code and data are available, and the results are quantitatively analyzed. The main limitation of this manuscript is that it lacks a thorough comparison to existing literature and state-of-the-art. Apart from some qualitative comparison in the first figure and a few plots comparing their work to NICE in Fig. 4, no such comparison is performed. The literature cited is mostly from computer vision applications to other types of overlapping objects, but not microbiology or bioimaging. The problem of counting objects is nowadays usually solved with object detection and/or segmentation neural networks, some of which are specifically taylored towards separating overlapping convex objects such as cell nuclei (e.g. Stardist, Cellpose). In the field of colony counting, some classical and ML-based literature not cited here (just as examples) include:

https://doi.org/10.2144%2F000112018

https://www.nature.com/articles/s41598-020-72979-4

https://link.springer.com/doi/10.1007/s10796-009-9149-0

http://arxiv.org/abs/2009.00926

https://doi.org/10.1364/OSAC.396603

https://doi.org/10.1016/j.tim.2021.01.006

https://vciba.springeropen.com/articles/10.1186/s42492-022-00122-3

The literature is for example reviewed in

https://doi.org/10.1016/j.tim.2021.01.006

https://link.springer.com/article/10.1007/s10462-021-10082-4

It is not necessary to cite all of this work in the manuscript, but the authors are encouraged to take the effort and spend some time studying the current literature, and then add a discussion of their work with respect to the current state of the art, and ideally also some more quantitative comparison.

[Response 2-1]

We appreciate the reviewer for pointing out this limitation in our manuscript. We have taken the following steps to address the comment.

As per the reviewer’s suggestion, we have incorporated additional references to bioimaging and neural network (or deep learning) related literature in the manuscript. We also provided a paragraph discussing on why we chose not to adopt neural network (or deep learning) approaches. We would like to clarify that, broadly speaking, our method can also be classified as a machine learning (ML) method, which is why we avoid labeling other approaches as exclusively "ML methods." Perhaps it would be more appropriate to refer to our approach as a "classical method" and to the methods the reviewer mentioned as "neural network methods." Changes were made in throughout the Introduction of main text, more specifically at lines 79-87.

We have made an effort to include more quantitative comparisons in our revised manuscript. Specifically, we have modified Figure 4 to further include the counting performance of OpenCFU and AutoCellSeg. We believe this comparison is reasonable because, as the reviewer noted, one of the 2022 ML papers referred also uses OpenCFU and AutoCellSeg as benchmarks. We also attempted to include comparisons with newer tools, such as the ones referred above. However, we encountered several challenges: some of these tools do not provide source code, making replication difficult; others offer software that we were unable to download, or that did not function correctly after download. Please see the details of these challenges in the attached response letter.

[Comment 2-2]

Some technical remarks: it would be great if more documentation, explanation of the individual steps, and installation instructions could be added to the Jupter Notebooks. The figures appear in very low quality in the PDF, but this might be a production issue unrelated to the submitted material.

[Response 2-2]

We have updated the Jupyter Notebooks by adding additional documentation and comments to clarify the individual steps. We have also renamed some variables and functions to make the code more intuitive. We hope these changes improve the clarity and usability of the notebooks. The updated Jupyter Notebook is publicly available on Dryad (https://doi.org/10.5061/dryad.2280gb62f).

Regarding the figure quality issue, we would like to clarify that the version we submitted to PLOS ONE included the text and images uploaded separately. The text was provided in a DOCX format, and the images were submitted as high-resolution TIFF vector files. We did not upload any PDF versions of the images. Therefore, we guess the low image quality the reviewer mentioned here may be a production issue that occurred during the review process.

Attachment

Submitted filename: Response to Reviewers.docx

pone.0311242.s012.docx (387.7KB, docx)

Decision Letter 1

Florian Rehfeldt

17 Sep 2024

MCount: An automated colony counting tool for high-throughput microbiology

PONE-D-24-05814R1

Dear Dr. Buie,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager®  and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Florian Rehfeldt

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: (No Response)

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: (No Response)

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: (No Response)

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: (No Response)

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I thank the authors for their response. They have succesfully addressed all my concerns, and I support publication.

Reviewer #2: The authors now included a more detailed comparison with existing approaches and a discussion of recent literature in the field. The dataset on which the performance is demonstrated still is rather limited and not representative of a wide range of possible user scenarios, but since the tool and source code is available, readers can try and include the algorithm in their own workflows.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy .

Reviewer #1: No

Reviewer #2: No

**********

Acceptance letter

Florian Rehfeldt

PONE-D-24-05814R1

PLOS ONE

Dear Dr. Buie,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Florian Rehfeldt

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. The Polygon Approximation Algorithm uses turning points on the contour to represent an inscribed polygon, where d controls how many turning points are generated.

    (a) For consecutive three turning points Ti1, Ti, and Ti+1, d represents the distance from Ti to the line Ti1Ti+1¯. (b) The larger the value of d, the more turning points are generated to represent the contour. The turning points are represented as black dots, while contour segments divided by turning points are shown in different colors.

    (TIF)

    S2 Fig. A dataset that includes colony segments, labeled using historical photographs, is used to optimize and evaluate the performance of MCount.

    (a) 960 sub-images are obtained by cropping 10 plates of fluorescent E. coli NEB10-beta and further divided into 15,847 segments using foreground extraction and segmentation. (b) The colony number distribution of sub-images shows that most sub-images have 10 ~  40 colonies. (c) The percentage of single-colony, two-colony, and three-colony segments is 59.6%, 18.52%, and 7.93%, respectively, which takes 86.05% in total. A well-performing algorithm is expected to correctly recognize almost all none and mildly merged colonies. The remaining percentage of segments, merged in a denser manner, requires the algorithm to infer sophisticated shapes. Note that 3.90% of segments are invalid because the merging of colonies is too severe to be labeled correctly, denoted as -1 in the left figure. (d) All segments are labeled according to their shape in a photograph taken about 4 hours ago to ensure labeling accuracy, including segments that are hard for humans to label.

    (TIF)

    pone.0311242.s002.tif (30.8MB, tif)
    S3 Fig

    The optimal values of d and λ are determined using a different definition of recognition error. (a) The average error rate on the segment dataset is plotted against d when λ=26. The minimum average error rate of 8.50% is achieved at d=0.5. (b) The average error rate on the segment dataset is plotted against λ when d=0.5. The minimum average error rate of 8.46% is achieved at λ=28. An error of each segment is defined as 1MCountLabel and the error was averaged across all segments, denoting the average error rate.

    (TIF)

    pone.0311242.s003.tif (1.5MB, tif)
    S4 Fig

    Increasing λ results in monotonously less counting leading to underestimation of colony number. The plot shows the MCount counting result with different λ values versus ground truth label on (a) 15,847 colony segments and (b) 960 sub-images, respectively. The green dashed line represents the regression line for a perfect colony estimator that always gives the result as the label, while the colored line represents the regression line for MCount. By tuning λ, it is possible to address the overestimation/underestimation issue. Increasing λ results in a monotonous decrease in the number of colonies counted, leading to an underestimation of colony number.

    (TIF)

    S1 Table. Variation of optimal hyperparameters across 10 sub-datasets generated under different experimental conditions.

    Mean and variance of colony number, optimal hyperparameter values, and the corresponding recognition error rate for each sub-dataset. The results demonstrate minimal variation in hyperparameters across sub-datasets and consistently lower error rates compared to the global settings (3.99%).

    (DOCX)

    pone.0311242.s005.docx (16.2KB, docx)
    S2 Table

    Performance of MCount using globally optimized hyperparameters (λ=26 and d=0.5) across 10 sub-datasets. The table presents the average error rates for each sub-dataset when applying the global optimal hyperparameters. The results indicate that MCount maintains robust performance across varied experimental conditions, with error rates significantly lower than those achieved by NICE.

    (DOCX)

    pone.0311242.s006.docx (16.1KB, docx)
    S3 Table. Processing time of different colony counting solutions for 100 randomly selected sub-images.

    The optimization solver for MCount is ‘COIN_CMD’ in PuLP, and the solver can be changed to achieve faster processing times at the cost of sacrificing accuracy.

    (DOCX)

    pone.0311242.s007.docx (14.2KB, docx)
    S4 Table. Mean, variance, and p-values of tests for 10 datasets.

    The mean and variance values were calculated from the 96 colony counts of each dataset. The first column represents the dataset index. The second column represents the mean value of the 96 colony counts, and the third column represents the variance of the 96 colony counts. The 4th and 5th columns show the p-values for Kolmogorov–Smirnov (KS) tests for Normal and Poisson distributions, respectively, under the significant level of 0.05, where no rejection is made for any datasets. The 6th column shows the p-values for Poisson Dispersion tests, and datasets whose null hypotheses are rejected under the significant level of 0.05 are denoted with a star symbol * .

    (DOCX)

    pone.0311242.s008.docx (15.6KB, docx)
    S1 File. Preparation and discussion on bacterial colony dataset.

    Discussion, cultivation and preparation on E. coli Dataset used in this study.

    (DOCX)

    pone.0311242.s009.docx (16.4KB, docx)
    S2 File. Discussion on the distribution of colony numbers.

    (DOCX)

    pone.0311242.s010.docx (14.2KB, docx)
    Attachment

    Submitted filename: Response to Reviewers.docx

    pone.0311242.s012.docx (387.7KB, docx)

    Data Availability Statement

    The source code and integrated solution are freely available at https://github.com/hyu-kim/mcount. All methods implemented and data used are publicly available on Dryad at https://doi.org/10.5061/dryad.2280gb62f


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES