Skip to main content
PMC Canada Author Manuscripts logoLink to PMC Canada Author Manuscripts
. Author manuscript; available in PMC: 2015 Oct 1.
Published in final edited form as: J Biomol Screen. 2014 May 14;19(9):1314–1320. doi: 10.1177/1087057114534298

Rank ordering plate data facilitates data visualization and normalization in high throughput screening

Chand S Mangat 1, Amrita Bharat 1, Sebastian S Gehrke 1, Eric D Brown 1,2,*
PMCID: PMC4318693  CAMSID: CAMS4462  PMID: 24828052

Abstract

High-throughput screening (HTS) of chemical and microbial strain collections is an indispensable tool for modern chemical and systems biology; however, HTS datasets have inherent systematic and random error, which may lead to false positive or false negative results. Several methods of normalization of data exist; nevertheless, due to the limitations of each, no single method has been universally adopted. Here, we present a method of data visualization and normalization that is effective, intuitive and easy to implement in a spreadsheet program. For each plate, the data are ordered by ascending values and a plot thereof yields a curve that is a signature of the plate data. Curve shape characteristics provide intuitive visualization of the frequency and strength of inhibitors, activators and noise on the plate, allowing potentially problematic plates to be flagged. To reduce plate-to-plate variation, the data can be normalized by the mean of the middle 50% of ordered values, also called the interquartile mean (IQM) or the 50% trimmed mean of the plate. Positional effects due to bias in columns, rows or wells, can be corrected using the interquartile mean of each well position across all plates (IQMW) as a second level of normalization. We illustrate the utility of this method using datasets from biochemical and phenotypic screens.

Keywords: high-throughput screening, data normalization, rank ordering, interquartile mean, middle quartiles

Introduction

Once the exclusive domain of the pharmaceutical industry, high throughput screening (HTS) methodology has been widely adopted in academic settings over the last decade due to the accessibility of instrumentation and the power of high throughput approaches1,2. The parallel development of large chemical libraries and ordered genome-scale clone sets for model organisms has allowed significant advances in the fields of chemical and systems biology. HTS produces large datasets over a period of weeks or months. Data variability and quality can obscure bona fide chemical activities or genetic interactions. To be able to compare data across different days and different assay plates and assess data quality, various methods of normalization and data visualization are typically used, ranging from very simple to very sophisticated. Here, we present a simple and intuitive, but effective, approach to data visualization and normalization.

An important prerequisite for effective data normalization is to collect high quality data. The largest contributors to variation in cell-based assays are arguably the conditions of incubation and the cell growth period of the experiment, rather than setup or measurement. Such variations in growth conditions manifest, for example, as an edge effect where the outer wells behave differently than the inner wells, or as a stack effect, where plates behave differently depending on their positions within a stack. Edge effects can be significantly reduced with humidity-controlled incubators and protection of plates from direct exposure to the internal fans of the incubator to reduce evaporation. A practical solution to the stack effect, is to use well-designed racks in incubators, which leave an air space above and below each plate for even heating across the plate. These measures may reduce the impact of positional variation.

The data produced by a high-throughput screen will nevertheless have some level of variability, which necessitates data normalization so that hit selection is not biased towards or against a subset of the data. A common method of normalization of HTS data is to calculate the activity of the sample wells relative to high activity and low activity of on-plate controls. Further, the mean and variance of the controls are often used to normalize the data and set a threshold for hit selection. Chemical and microbial strain collections are often formatted such that controls are placed on the outer columns of the assay plate. Where the outer wells are often subject to significant edge effects and increased noise, control-based normalization can introduce unwanted bias. The edge effect problem can be remedied to an extent by placing controls throughout the plate; however, the format of the assay plates is limited by the format of the library being tested. Even when controls occupy almost 20% of the wells on a plate, the normalized data can display a large amount of variation across the screen due to outliers in the control data or differences in the preparation and stability of the control reagents compared to the samples. Further, ordered genomic libraries, such as the library of non-essential gene deletions in E. coli3, are often formatted in a manner that excludes positions for controls wells. For these reasons, normalization methods that are not based on control wells have particular advantage.

Several methods of normalization have been proposed that rely on the observation that the majority of samples in a screen are unperturbed4–9. One of the more sophisticated methods of normalization is the B-score, which corrects biases in plates, rows, columns, and batches4. The B-score method involves an iterative two-way median polish and a smoothing function. Implementation of the B-score is not trivial, as the data must be processed using sophisticated statistical software or a programming language, such as R or MatLab. Thus, while the B-score method offers a power tool for data processing, it also presents a formidable barrier to accessibility and understanding among non-experts. Indeed, where data normalization can be carried out with more accessible methodology, such as described herein, more users have the potential to benefit from hands-on data processing and its attendant benefits.

Another method of normalization of HTS data is the Z-score for plate-based normalization5. Like the B-score, the Z-score method assumes that the vast majority of compounds on a plate are inactive. To calculate the Z-score, the difference between the sample value and the plate mean is divided by the standard deviation of the plate. While this method is accessible, the Z-score has the disadvantage that both the mean and the standard deviation are influenced by the number of hits on the plate, which means that a hit might be missed if it is on a plate with many other hits. This would be less of an issue if compounds in chemical libraries were randomly arrayed; however, libraries tend to have related compounds on the same plate so hits tend to cluster on certain plates. Likewise, in ordered microbial clone sets, genes from the same pathway tend to be found on the same plate. A plate containing the pathway of interest will contain many hits, which may be masked after normalization by the Z-score, resulting in a false negative result.

In addition to the B-score and the Z-score, other statistical methods of non-control normalization have been proposed. In the background correction method, global trends and local fluctuations of the background surface are computed by polynomial least squares fit or by multiple linear regression6. Deviations of the background surface from the zero plane are considered to be due to systematic error and are subtracted from each well. The well-correction method of Makarenkov et al uses linear least-squares approximation or polynomial fit to look for a trend in each well position and either adds or subtracts the trend from the original values to bring the mean of each well position to zero7. This is followed by independent Z-score normalization of each well position. The R-score method is similar to the B-score but it estimates row and column effects using the robust linear model8. More recently, two new methods have been proposed: the matrix error amendment method, which estimates row and column effects using singular value decomposition to solve a system of linear equations and the partial mean polish, which is an extension of the Tukey median polish that is only applied to the rows and columns that show bias9. All of these methods have similar barriers to accessibility as the B-score.

Here, we propose an alternate method of normalization that is resistant to outliers and can be implemented with a simple spreadsheet program. Plots of the rank ordered data for each plate provide intuitive visual information on the spread of the data on the plate. The shape of each curve reflects the frequency and strength of inhibitors, activators and noise on that plate, allowing potentially problematic plates to be flagged. Quartiles are the three points that divide a rank ordered dataset into four equal parts. The interquartile region represents the central 50% of the values of a rank ordered dataset, which should exclude the inhibitors and activators on the plate. The interquartile mean is often also referred to as the 50% trimmed mean, where the highest 25% and lowest 25% of values in a dataset are trimmed before calculating the mean. In Excel, the IQM or 50% trimmed mean can be found with the function =TRIMMEAN(start:end,0.5), where start and end define the location of the data. Herein, we use biochemical and cell-based high-throughput datasets to show that normalization by the mean of the interquartile (middle 50%) of the data on each plate is an effective method of minimizing variability. Like the Z-score, the interquartile method avoids the potential problems of control-based normalization; however, since only the middle 50% of the data is used for normalization, the interquartile method has the advantage of not being influenced by the outliers on the plate.

Methods

Control-based normalization

The percent activity of each compound was calculated from the measured value using Eq. (1), where S represents the measured sample value and H and L represent the mean of the high activity and low activity controls, respectively, of on-plate controls.

%Activity=(S-LH-L)×100 Eq. (1)

B-score normalization4

The B-score normalization, as described by Brideau et al4, was carried out using the R-statistical package. Briefly, the Tukey median polish procedure was applied to a 16 × 24 matrix, representing a 384-well plate. The residual activity (rijp) for each sample in the ith row and jth column of the pth plate, was found according to rijp = S−(μp+Rip+Cjp), where S is the measured sample value, μp is the mean of the pth plate and Rip and Cip are the corresponding row and column effects. The median absolute deviation (MADp) of the pth plate was found according to MADp = median(|rijp−median(rijp)|). Finally, the B-score for each sample was calculated according to B-score = rijp/MADp. Parameters for the median polish function in R were as follows: eps = 1e-05, maxiter = 200, trace.iter = !TRUE, na.rm = TRUE. Parameters for the MAD function in R were as follows: constant = 1.4826, na.rm = FALSE, low = FALSE, high = FALSE.

Interquartile mean (IQM) normalization

Quartiles refer to the division of a rank-ordered dataset into four equal parts. The interquartile mean is the mean of the middle two quartiles or the middle 50% of rank ordered data. Sample data was normalized on a per plate basis by dividing every well of the plate by the interquartile mean of the plate according to Eq. (2).

IQMnormalizedactivity=(Sμiq) Eq. (2)

S represents the measured sample value and μiq represents the mean (μ) of the interquartile (iq) data of the plate. The value of μiq was found for each plate by sorting the data in ascending order using the =SORT function of Microsoft Excel software and then finding the mean of the interquartile (middle 50%) data. Sorting the data facilitated the creation of rank ordered plots. Alternatively, the interquartile mean can be found directly from unsorted data using the =TRIMMEAN function of Microsoft Excel. The trimmed mean of the middle 50% of the data, which is the same as the interquartile mean, can be found with the formula =TRIMMEAN(start:end,0.5) where start and end define the first and last cells of the data, respectively. As a quality control check for each plate, the interquartile mean was compared to the plate median and any plates where the two values differed by more than 15% were flagged as potentially problematic.

Interquartile mean well-based normalization (IQMW)

Data was first normalized according to Eq. (2) on a per plate basis then further normalized by the interquartile mean of each well position according to Eq. (3).

IQMWnormalizedactivity=(SIQMμiqw) Eq. (3)

SIQM represents the sample value after IQM-normalization, which was calculated by Eq. (2), and μiqw represents the mean (μ) of the interquartile values (iq) of each well position (w) across the entire screen.

Results and Discussion

Rank ordered plots for data visualization

When viewing a plot of an entire HTS dataset much of the data is obscured due to the density of information that is presented. The quality of HTS data can be evaluated more carefully by viewing data on a plate-by-plate basis. Rank ordering the data for each plate from lowest to highest value is a useful method of quickly visualizing inhibition (or activation) on the plate and identifying potentially problematic plates with, for example, many outliers. Figure 1A shows rank ordered plots of representative plates from a screen of the E. coli metabolic enzyme cystathionine β-lyase, which produces L-homocysteine and pyruvate10. The rank ordered data from a typical assay plate of the screen is represented by circles in Figure 1A; the data are symmetrically distributed, with inhibitors at the far left and activators at the far right. Examples of plates with atypical distributions are represented by triangles and squares; there are higher numbers of apparent inhibitors on the far left. When viewed as rank ordered plots, plates with a high number of outliers can be visually identified as potentially problematic or containing a high number of hits.

Figure 1. Plots of rank ordered data from representative assay plates of an enzymatic screen and comparison of interquartile mean to median.

Figure 1

The E. coli cystathionine β-lyase enzyme, which produces pyruvate and L-homocysteine, was previously screened against a chemical library10. (A) The data from four representative assay plates were rank ordered by ascending values and plotted. Vertical lines indicate the boundaries of the interquartile data that are marked by the high values of the first and third quartiles (Q1 and Q3), which represent the 25th and 75th percentile of the data. (B) Two plates were normalized by either the interquartile mean (closed symbols) or the median (open symbols). For the first plate, the interquartile mean and median were the same (circle) and for the second plate the IQM was 6% lower than the median (square). The symbol shapes are consistent with the rank ordered plots shown in panel A. (C) The fraction of IQM/median for each plate of the screen is shown. The solid line represents the average of all the IQM/median fractions and the hashed lines represent 3 standard deviations above and below the average.

A more quantitative and potentially automated method of flagging plates is comparing the interquartile mean to the median of the plate. A large discrepancy between the two values indicates a wide variation of the data, where more than half of the samples on the plate deviate from the mean. The value of the interquartile mean includes more of the plate data than the value of the median since the interquartile mean is calculated from the central 50% of the measured values whereas the median is represented by the central one or two values. Figure 1B shows the effect of normalizing two types of assay plates by either the median or the interquartile mean of the plate. For a typical screening plate with a relatively symmetric rank ordered plot, normalization by either method produced overlapping plots (Figure 1B, circles). For a plate with a left skew, the IQM and median differed by 6% and this produced a small difference in the normalization plots (Figure 1B, squares). Figure 1C shows the fraction of IQM divided by median for each plate of the screen. The average ratio of IQM/median was 0.995 because the two values are usually close to each other; however, out of the 625 plates screened, 7 plates had IQM/median values that were more than three standard deviations away from the average of 0.995. Special attention might be paid to these 7 plates, for example, creation of rank ordered plots to examine the distribution of hits and activators on the plate more closely. In this screen, the difference between IQM and median values was at most 6%, which had a small effect on normalization as described above in Figure 1B. For differences between IQM and median of more than 15%, we would recommend that screeners consider rescreening those plates to verify that the large number of outliers is not due to artifact.

Application of interquartile-based normalization to a biochemical screen

We applied interquartile mean normalization to a dataset from a screen of the essential E. coli GTPase EngA against a library of 30, 800 compounds11. EngA hydrolyzes GTP to GDP, producing phosphate, which was detected by the malachite green colorimetric assay. Each 384-well plate contained 32 high activity controls (uninhibited reaction) and 32 low activity controls (lacking enzyme) in the first two and last two columns. A plot of the controls from the entire screen showed that the high activity controls for this enzyme assay had slightly lower activity on the first day of the screen (Figure 2A). Figure 2B depicts the raw data from one replicate of the EngA screen as an index plot in chronological order. Individual days of screening are apparent from the periodic rise and fall of the data throughout the screen. This may be partially due to degradation of GTP to GDP and phosphate over the course of each day of screening. Normalization of the data by the controls, using Eq. (1), reduced the overall spread of the data and partially normalized the daily trend of increasing values (Figure 2C). Due to an idiosyncratic difference between the controls and the samples on the first day of the screen, the control-normalized activity is higher on this day.

Figure 2. Control-based or IQM normalization of an enzymatic screen.

Figure 2

A pilot screen of 30,8000 compounds was carried out in duplicate against the GTPase activity of EngA in 384-well density 11. Phosphate production was detected by the malachite green colorimetric assay by measuring absorbance at 600 nm. Shown are (A) the absorbance values of the high activity (black) and low activity (grey) controls, (B) the un-normalized absorbance values, (C) control normalized data represented as percent activity relative to high and low activity controls, calculated according to Eq. (1), or (D) IQM normalized data represented as percent activity relative to the interquartile mean, calculated according to Eq. (2). The solid line represents the mean of each dataset and the hashed lines represent 3 standard deviations above and below each mean.

IQM normalization was applied to the raw data, using Eq. (2). Compared to control-based normalization, the IQM method was more effective at equalizing the center of the data across the screen and narrowing the spread of data (Figure 2D). When using a global cut-off to select active molecules, effective normalization of the dataset reduces the tendency for plates with lower activity to be over-represented in hit selection (false positives) and plates with higher activity to be under-represented (false negatives). Outliers in the control data may skew the mean value of the controls and introduce bias to the data. The systematic increase in activity over the course of the day was normalized more effectively by interquartile normalization than by controls-based normalization, which is probably a reflection of slight but systematic differences between the controls and the samples.

Application of interquartile-based normalization to a whole cell phenotypic screen

We previously published a study describing the synergistic antibacterial effect of combining the antibiotic minocycline and the anti-diarrheal drug loperamide in Pseudomonas aeruginosa12. To investigate the mechanism of action of loperamide, the drug was screened against the Keio collection, which is an ordered E. coli gene deletion set of some 4,300 strains3. The sensitization of each deletion strain to treatment with loperamide and EDTA was calculated by dividing the cell density of the untreated samples by that of treated samples. Of note, the microplate format of the Keio library does not include empty wells that can be used for on-plate controls. Figure 3A depicts the response of each strain to treatment with loperamide and EDTA, when the fold sensitivity was calculated using the raw data. There was a high level of variation in the data and a trend towards decreased fold sensitization as the screen progressed, which may be due to differences in incubation time, incubation temperature and inoculum. Variation in each dataset was amplified by the division function used to calculate fold sensitization.

Figure 3. IQM normalization of a cell-based screen of a genetic library.

Figure 3

A library of strains containing deletions of non-essential genes in E. coli was screened for sensitivity to treatment with the drug loperamide in combination with EDTA12. Cultures were incubated in 96-well plates and growth was measured by optical density at 600 nm. (A) The fold sensitization to treatment with loperamide and EDTA was calculated by dividing the growth of untreated cultures by the growth of treated cultures. (B) The raw data from each condition, either untreated or treated with loperamide and EDTA, was first normalized by the IQM method before calculating fold sensitization.

The raw data from treated and untreated samples was normalized using the IQM method prior to calculating fold sensitization. IQM normalization resulted in a much narrower distribution of fold sensitization, which is expected, since most deletion strains should not show any genetic interactions with loperamide (Figure 3B). The mean sensitzation of all the plates was close to a value of 1 (no sensitization) and the trend towards decreasing sensitization in later plates was corrected.

Comparison of normalization by IQM and the B-score

We screened 142,000 compounds for inhibitors of E. coli growth in 384-well density (unpublished). When the mean value of each row and column was plotted for one replicate of the screen, the outer rows and outer columns had higher means than the inner rows and inner columns (Figure 4A and 4B). If hit selection was done on un-normalized data, more active molecules would be selected from the center wells because these wells generally have lower activity. Positional correction of HTS data should only be employed if biases in rows, columns or wells are present, otherwise the benefit of positional adjustment may be overshadowed by the potential to introduce error to the data.

Figure 4. B-score or IQM normalization of a cell-based screen of a chemical library.

Figure 4

A laboratory-derived strain of E. coli was screened for growth inhibition against a library of 142, 000 compounds at 384-density by measuring optical density at 600 nm (unpublished). Shown are the mean values of: un-normalized data in rows (A) and columns (B), B-score normalized data in rows (C), and columns (D), and IQMW normalized data in rows (E) and columns (F). Error bars represent one standard deviation. (G) IQMW normalized and B-score normalized data were plotted against each other. Data below the hashed lines represent the hits in the screen that were defined by a statistical cut-off of three standard deviations below the mean of each dataset, i.e., values lower than 0.725 and −6.06 for IQMW and B-score data, respectively.

To normalize this edge effect, we used either the B-score method or the interquartile mean of each well position. B-score normalization, which was performed using a script for the R statistical package, effectively normalized the mean values across all rows and columns (Figure 4C and 4D). Alternatively, normalizing the data using the interquartile mean of each plate (IQM) and then the interquartile mean of each well position (IQMW), which simultaneously takes row bias and column bias into account, was also effective at normalizing the mean values of all rows and columns (Figure 4E and 4F). Thus, both the B-score and well-based IQM approaches were effective at removing positional bias in this example.

To compare the difference in hit selection between IQM normalized and B-score normalized data, the two were plotted against each other (Figure 4G). A hit was defined as a compound with a value that was three standard deviations below the mean of all samples. Of the 1002 hits identified by the B-score normalization and the 979 hits identified by IQMW normalization, 862 hits were common to both methods, representing almost 90% of the hits. Of the 140 and 116 unique hits for the B-score and IQMW methods, respectively, most were only weakly potent and had values near the cutoff (Figure 4G). Similar confirmation rates were obtained for the two methods. In follow-up dose response curves, 69% of the B-score hits and 72% of the IQMW hits demonstrated minimum inhibitory concentrations (IC50) of < 100 μM. Although the sophisticated positional adjustment and smoothing functions that are carried out by the B-score are necessary in some screens, we conclude that IQMW normalization was also an effective method of normalizing positional effects in the screen described here and is highly accessible to a growing user base.

Here, we have illustrated the utility of interquartile-based normalization using rank ordered data from four HTS campaigns, including two biochemical screens and two cell-based screens. Since only the middle 50% of the data is used to calculate the interquartile mean, inhibitors and activators do not influence the normalization. Rank ordering data facilitates data visualization and the IQM approach to data normalization is intuitive, robust and easy to adopt. Further, while it has become standard to devote up to 20% of the wells of a screen to high and low activity controls, IQM normalization has the potential to increase the productivity of screens by dispensing with these control wells while still allowing effective normalization between plates or well positions.

Acknowledgments

We thank Gerry Wright for providing data from the screen of cystathione beta-lyase. E.D.B. acknowledges salary support from the Canada Research Chairs program and operating funds from the Canadian Institutes of Health Research (MOP-15496, MOP-64292, and MOP-81330) and Cystic Fibrosis Canada.

References

  • 1.The Academic Pursuit of Screening. Nat Chem Biol. 2007;3:433–433. doi: 10.1038/nchembio0807-433. [DOI] [PubMed] [Google Scholar]
  • 2.Brown ED. Screening in Academe: a Perspective on Implementation of University-Based Small Molecule Screening. J Biomol Screen. 2003;8:377–379. doi: 10.1177/1087057103257213. [DOI] [PubMed] [Google Scholar]
  • 3.Baba T, Ara T, Hasegawa Construction of Escherichia coli K-12 in-Frame, Single-Gene Knockout Mutants: the Keio Collection. Mol Syst Biol. 2006;2 doi: 10.1038/msb4100050. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Brideau C, Gunter B, Pikounis B, et al. Improved Statistical Methods for Hit Selection in High-Throughput Screening. J Biomol Screen. 2003;8:634–647. doi: 10.1177/1087057103258285. [DOI] [PubMed] [Google Scholar]
  • 5.Zhang JH, Chung TDY, Oldenburg KR. Confirmation of Primary Active Substances From High Throughput Screening of Chemical and Biological Populations: a Statistical Approach and Practical Considerations. J Comb Chem. 2000;2:258–265. doi: 10.1021/cc9900706. [DOI] [PubMed] [Google Scholar]
  • 6.Kevorkov D, Makarenkov V. Statistical Analysis of Systematic Errors in High- Throughput Screening. J Biomol Screen. 2005;10:557–567. doi: 10.1177/1087057105276989. [DOI] [PubMed] [Google Scholar]
  • 7.Makarenkov V, Zentilli P, Kevorkov D, et al. An Efficient Method for the Detection and Elimination of Systematic Error in High-Throughput Screening. Bioinformatics. 2007;23:1648–1657. doi: 10.1093/bioinformatics/btm145. [DOI] [PubMed] [Google Scholar]
  • 8.Wu Z, Liu D, Sui Y. Quantitative Assessment of Hit Detection and Confirmation in Single and Duplicate High-Throughput Screenings. J Biomol Screen. 2008;13:159–167. doi: 10.1177/1087057107312628. [DOI] [PubMed] [Google Scholar]
  • 9.Dragiev P, Nadon R, Makarenkov V. Two Effective Methods for Correcting Experimental High-Throughput Screening Data. Bioinformatics. 2012;28:1775–1782. doi: 10.1093/bioinformatics/bts262. [DOI] [PubMed] [Google Scholar]
  • 10.Ejim LJ, Blanchard JE, Koteva KP, et al. Inhibitors of Bacterial Cystathionine -Lyase: Leads for New Antimicrobial Agents and Probes of Enzyme Structure and Function. J Med Chem. 2007;50:755–764. doi: 10.1021/jm061132r. [DOI] [PubMed] [Google Scholar]
  • 11.Bharat A, Blanchard JE, Brown ED. A High-Throughput Screen of the GTPase Activity of Escherichia coli EngA to Find an Inhibitor of Bacterial Ribosome Biogenesis. J Biomol Screen. 2013;18:830–836. doi: 10.1177/1087057113486001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Ejim L, Farha MA, Falconer SB, et al. Combinations of Antibiotics and Nonantibiotic Drugs Enhance Antimicrobial Efficacy. Nat Chem Biol. 2011;7:348–350. doi: 10.1038/nchembio.559. [DOI] [PubMed] [Google Scholar]

RESOURCES