Skip to main content
Biostatistics (Oxford, England) logoLink to Biostatistics (Oxford, England)
. 2009 Feb 21;10(3):424–435. doi: 10.1093/biostatistics/kxp001

A simulation–approximation approach to sample size planning for high-dimensional classification studies

Perry de Valpine 1,2,3,4,*, Hans-Marcus Bitter 1,2,3,4, Michael P S Brown 1,2,3,4, Jonathan Heller 1,2,3,4
PMCID: PMC2697341  PMID: 19234308

Abstract

Classification studies with high-dimensional measurements and relatively small sample sizes are increasingly common. Prospective analysis of the role of sample sizes in the performance of such studies is important for study design and interpretation of results, but the complexity of typical pattern discovery methods makes this problem challenging. The approach developed here combines Monte Carlo methods and new approximations for linear discriminant analysis, assuming multivariate normal distributions. Monte Carlo methods are used to sample the distribution of which features are selected for a classifier and the mean and variance of features given that they are selected. Given selected features, the linear discriminant problem involves different distributions of training data and generalization data, for which 2 approximations are compared: one based on Taylor series approximation of the generalization error and the other on approximating the discriminant scores as normally distributed. Combining the Monte Carlo and approximation approaches to different aspects of the problem allows efficient estimation of expected generalization error without full simulations of the entire sampling and analysis process. To evaluate the method and investigate realistic study design questions, full simulations are used to ask how validation error rate depends on the strength and number of informative features, the number of noninformative features, the sample size, and the number of features allowed into the pattern. Both approximation methods perform well for most cases but only the normal discriminant score approximation performs well for cases of very many weakly informative or uninformative dimensions. The simulated cases show that many realistic study designs will typically estimate substantially suboptimal patterns and may have low probability of statistically significant validation results.

Keywords: Biomarker discovery, Experimental design, Generalization error, Genomic, Pattern recognition, Proteomic

1. INTRODUCTION

Recent years have seen an explosion of work on classification problems where the number of measured features per sample is vastly greater than the number of samples. For biological classification problems, such data arise from genomic DNA microarrays and proteomic mass spectrometry assays, from which investigators try to classify disease categories, tumor types, response to drugs, or other categories (Ludwig and Weinstein, 2005). Most of the efforts in method development have appropriately focused on what to do with real data sets (Wang and Shen, 2006, Adam and others, 2002). Generally speaking, various methods must select features (sometimes called biomarkers) to be used for classification and estimate a classifier without over-fitting to the many available data dimensions.

Because of the complexity of the algorithms involved, it is not straightforward to answer questions about study design. For example, if there are 10 informative and 5000 noninformative features and the best possible classification error rate is 5%, how many samples are necessary to have an 80% chance of estimating a classifier with less than 10% error rate for independent validation samples? Or, how many samples are necessary so that with probability 95%, the estimated classifier will perform statistically significantly better than a 50% error rate for independent validation samples, that is, conclude the study has at least found something nonrandom? Investigators planning studies have access to sound statistical principles but few specifics to serve as guideposts in evaluating sample sizes relative to hypothesized outcomes. Analysis of study design for high-dimensional classification studies has been identified as an important problem for genomics and proteomics because significant resources are required to execute such studies (Dobbin and Simon, 2007, Allison and others, 2006, Pusztai and Hess, 2004, Hwang and others, 2002).

Issues of sample size for genomic and proteomic pattern discovery studies are potentially quite important. Over 60 proteomics discovery studies have been published in recent years (Coombes and others, 2005, Baker, 2005). Many have sample sizes in the approximately 10–20 range; some notable cases with higher sample sizes (e.g. Adam and others, 2002, Petricoin, Ardekani, and others, 2002, Petricoin, Ornstein, and others, 2002, Zhang and others, 2004, Rogers and others, 2003) reveal that in broad terms, sample sizes of ∼50 per group are rare and of ∼100 per group are very rare. Implicit in some rationales for biomarker discovery studies is the possibility that multiple, individually weak biomarkers could combine to form a collectively strong diagnostic pattern. The observation that discovery studies often find nonspecific markers (Baker, 2005) also suggests that disease-specific patterns may require multiple, individually weak biomarkers. Detecting patterns of multiple weak biomarkers amid many noninformative data dimensions may require substantially greater sample sizes than detecting individually strong biomarkers.

In proteomics, early biomarker discovery and validation studies (Petricoin, Ardekani, and others, 2002, Petricoin, Ornstein, and others, 2002, Petricoin and Liotta, 2003, Rogers and others, 2003, Adam and others, 2002, Li and others, 2002, Adam and others, 2001) led to renewed attention toward potential pitfalls of design and analysis methods. These include low discovery and validation sample sizes, uncertainty about data preprocessing and statistical methods, low sample processing and measurement reproducibility within and between study sites, uncertainty about the biological nature and consistency of patterns, and lack of independent validation studies (Sorace and Zhan, 2003, Diamandis, 2004a, Diamandis, 2004b, Listgarten and Emili, 2005, Coombes and others, 2005, Ebert and others, 2006, Wilkins and others, 2006). Similar issues have been raised for genomic studies (e.g. Pusztai and Hess, 2004, Ludwig and Weinstein, 2005). Two important studies notable for their independent validation trials highlight the possibility—among many possible reasons for low validation success—that small sample sizes have been fundamentally limiting. Rogers and others (2003) saw sensitivity for renal cancer decline from ∼100% in discovery to ∼40% in validation, and Zhang and others (2004) saw specificity decline from ∼90% in discovery to ∼65% in validation.

For prospective analysis of pattern discovery study designs, purely simulation approaches quickly become cumbersome because there are many scenarios of interest, but purely analytical results are not easy to obtain. We take a middle road between simulations and approximations, with Monte Carlo methods for the feature-selection step and approximations for generalization error rates given each feature set. We use multivariate normal data and linear discriminant classification of features selected by univariate tests. While biologically simplistic, this framework captures the key impacts of both inaccurate feature selection and inaccurate classifier estimation. Related studies that use multivariate normal models include Pepe and others (2003), Hu and others (2005), Jung (2005), and Dobbin and Simon (2007), among others. Our approach gives order-of-magnitude faster estimation of generalization error compared to direct simulations, which are given for comparison. Both full simulation and simulation–approximation results are useful, but the latter can facilitate more practical exploration of study designs. Our approach also gives insight into which sources of variation are most important and suggests directions for future improvements.

We evaluate the simulation–approximation approach by comparing it to complete simulations that address meaningful study design questions (supplementary material available at Biostatistics online, http://www.biostatistics.oxfordjournals.org). We ask how validation error rate depends on the strength and number of informative features (and hence the minimum possible error rate), the number of noninformative features, the patient sample size, and the number of features allowed into the pattern. We find that typical sample sizes may perform poorly when there is a true pattern composed of many individually weak features. This result is not surprising based on general principles, but moving from principles to specific examples as guideposts is important for design of real studies.

We also give 2 approximations of the generalization (or test, or validation) error of a linear discriminant classifier when the training and validation samples do not follow the same distributions. The first is a delta approximation, from Taylor expansions of generalization error around the expected discriminant boundary. The second, and more successful, approximates the discriminant scores as normally distributed. Approximations of linear discriminant analysis with training and generalization samples from the same distributions have been reviewed by McLachlan (1992) and Wyman and others (1990). According to Wyman and others (1990) and Viollaz and others (1995), normal approximations of discriminant scores seem to be more accurate than other approaches, consistent with our results.

A related approach was given by Dobbin and Simon (2007), but ours appears to be more general and accurate (at the expense of being more computational). Theoretical bounds on generalization error from machine learning theory give another path of investigation (Hastie and others, 2001). For the related goal of identifying individually significant data dimensions (features), much study design work has built on feature-by-feature false discovery rate ideas (Benjamini and Hochberg, 1995, Storey, 2002, Efron, 2007). Feature-by-feature metrics of study design efficacy include the expected discovery rate (Gadbury and others 2004), anticipated average power (Pounds and Cheng, 2005), expected number of false discoveries (Tsai and others, 2005), and probability of informative features ranking highly (Pepe and others, 2003). Numerous recent studies give methods for feature selection or estimation of generalization error given real data, as opposed to prospective study design (e.g. Mukherjee and others (2003), Fu and others (2005), Wang and Shen (2006).

2. PROBLEM DEFINITION

Consider samples of size nj for each of J classes (j{1,,J}), with each sample having M dimensions. By a high-dimensional classification problem, we mean Mn, where n=j=1Jnj is the total sample size. For the training samples, from which the classifier will be estimated, let Inline graphic be the data vector for the ith sample of class j. Let Xj be all the data for class j and X be all the training data.

Let the number of dimensions of the data distributions that are truly informative (i.e. differ between classes) be MI and those that are truly uninformative be MU, with M = MI + MU. In the examples below, we will for simplicity use J = 2 and group means centered around 0 with all variances equal to 1. Let Δ be the vector of differences between class means for the informative dimensions, so the means from group 1 are (−0.5 Δ, 0MU) and the means from group 2 are (0.5 Δ, 0MU), where 0MU is a length MU vector of zeros. In this notation, a true pattern is defined by (Δ, MU) and a study design scenario is defined by (Δ, MU, n), where n = (n1, n2).

A classifier ψ(xG|X) predicts the class, j1,,J, of a new (generalization or validation) sample xG based on the training data, X. The generalization sample comes from one of the same distributions (for its unknown class) as the training samples. Define the conditional generalization error for class j as the expected fraction of incorrect classifications for a new sample, xGj, from class j given a training sample X,

2. (2.1)

where the expectation is over xGj sampled from true distribution j and the indicator function I(Inline graphic) is 1 if Inline graphic is true and 0 otherwise.

Define the conditional generalization error across all classes as

2. (2.2)

where P(j) is the probability that a new sample is from class j.

The generalization error for a new sample from group j is the conditional generalization error averaged over training samples:

2. (2.3)

where ET denotes expectation over training samples, X, with sample sizes n. Finally, the overall generalization error is

2. (2.4)

Given a generalization sample XG, with replicate data xGj from groups j=1,2, and a classification procedure ψ(xG|X), define the “pattern discovery power” as the expected probability of rejecting the null hypothesis that the predictions ψ(xGj|X) are independent of the true class labels, using an appropriate statistical test, with expectations over both the training and generalization samples. This is the probability that the independent validation step of an entire study concludes that the estimated classifier is at least better than random. This paper focuses on calculating generalization error rather than pattern discovery power, but the latter relates to one of the ultimate judgments about a study—whether something nonrandom has been independently validated—and is represented graphically with the simulation results.

3. SIMULATIONAPPROXIMATION OF GENERALIZATION ERROR

Next, we give a joint simulation and approximation approach to estimate efficiently the generalization error rates CGj and Gj for multivariate normal data analyzed with linear discriminant analysis. Define a partition of the space of X samples into R nonoverlapping regions, Ω1,,ΩR, that determine which dimensions of X are selected to estimate the classifier, that is, the feature selection. Define δ = (δ1, …, δM) to be a vector of 0s and 1s, with δk=1 if dimension k will be used for classification and 0 if not. For all XΩr, the same dimensions of X are used by the classifier (so R2M), so it makes sense to write δ as a function of Ωr: δrδr).

The generalization error for class j can be factored as

3. (3.1)

where P(·) is the probability indicated by its argument.

We develop approximations for ET[CGj(Δ,MU|X)|XΩr] based on the first 2 moments of P(X|XΩr), the probability density of training data sets given that they lead to feature selection δr. This is an expected generalization error given that the training and generalization samples do not come from the same distributions. We use Monte Carlo samples to estimate P(XΩr) and the first 2 moments of P(X|XΩr), which can be generated efficiently. In what follows, Ω{Ω1,,ΩR}.

In a real analysis, feature selection is intertwined with the problem of how many features to include, which is one type of regularization parameter that may be optimized over data-based estimates of generalization error, such as cross-validation. From the study design point of view, the goal is to provide insight into typical study outcomes under various scenarios. Instead of trying to include optimization of the number of features within each approximation, we calculate the approximation across a range of the feature-selection thresholds. This does not include variation or suboptimality in the feature-selection threshold in our estimates of generalization error distributions, but it does offer insight about the sensitivity of generalization error to the feature-selection threshold, which provides context and builds intuition for interpreting results with real data.

3.1. Monte Carlo approximation of feature selection

Next, we show how P(XΩ) and the mean and variance of P(X|XΩ) can be estimated with Monte Carlo methods. In the examples here, we assume feature selection is based on feature-by-feature univariate t-tests, which, when the data dimensions really are independent, makes the analysis optimistic because it “knows” this aspect of the “truth.” It is common to use feature-by-feature hypothesis tests to estimate false discovery rates as part of analyzing a high-dimensional study, so this simplification allows our results to stand side-by-side with expected false discovery rates and related ideas in considering study designs.

Consider a single data dimension, k, which may or may not be truly informative, for which δk will be 1 if the dimension is selected for the pattern and 0 if not. Let xijk be the kth dimension of sample i from class j. Let the n1 and n2 samples from groups j = 1 and j = 2, respectively, be normally distributed in dimension k: Inline graphic. Suppose the decision to include feature k in classification is based on the P-value of a t-test. One calculates Inline graphic for j=1,2; Inline graphic, where dfs=n1+n22 are the degrees of freedom of sk2; and Inline graphic. The feature is included if |tk| > t1Pc/2,dfs, where Pc is a threshold significance level for choosing δk=1 and t1Pc/2,dfs is the inverse cumulative t-density at 1Pc/2 with dfs degrees of freedom.

It is equivalent to consider the 2 independent random variables

3.1. (3.2)

and Inline graphic. Then,

3.1. (3.3)

and

3.1. (3.4)

Using Inline graphic or g(z,e2)=σ2e2/dfs in (3.4) gives an estimate of the mean difference between groups 1 and 2 or the within-group variance, respectively, given that the t-test is significant.

Working with the densities of z and e2 allows more efficient numerical methods to estimate (3.3) and (3.4) than if one worked with the densities of xijk directly. Next, 2 possible Monte Carlo implementations are given, but a variety of numerical methods could be used. For the case of a t-test, (3.3) is simply a cumulative density of a noncentral t-distribution with noncentrality parameter Inline graphic and dfs degrees of freedom. For a Monte Carlo estimate of (3.4), define {z(l),e2,(l)},l=1,,m, to be a simulated sample from Inline graphic, which can be generated efficiently with a Markov chain Monte Carlo (MCMC) algorithm. Then, a Monte Carlo estimate of (3.4) is

3.1. (3.5)

Even a small sample (by MCMC standards) of say m=100 can be reasonable for (3.5).

If one chose to extend the basic idea here for a test for which values of (3.3) are not as easily available as a noncentral t-distribution, then both (3.3) and (3.4) could be estimated by Monte Carlo. For that case, redefine {z(l),e2,(l)},l=1,,m, to be a Monte Carlo sample of size m from P(z,e2). Then, the natural estimates of (3.3) and (3.4) are

3.1. (3.6)

and

3.1. (3.7)

Extensions based on other Monte Carlo numerical integration techniques (such as importance sampling) are straightforward and not our focus here.

3.2. Approximations for generalization error

Let Inline graphic be the parameter vector of the classification function ψ, a linear discriminant function in the examples here. An estimated classifier ψ(xGj|X) is defined by estimated parameters Inline graphic = Inline graphic(X). For more concise notation, we view generalization error as a function of Inline graphic, that is, CGj(Δ, MU|X) = CGj(Inline graphic).

Delta approximation. A delta approximation for the class generalization error given XΩ is

3.2. (3.8)

where Inline graphic is the second derivative of CGj with respect to Inline graphicr and Inline graphics evaluated at Inline graphic is the covariance between the r and the s dimensions of Inline graphic|X ∈ Ω, and p is the number of features selected due to XΩ. The delta approximation is derived by Taylor series expansion of the expectation integral around Inline graphic. Note that although the dimensions (or features) are assumed to be independent for feature selection, after they are selected they are approximated as multivariate normal, so the covariances in (3.8) are not necessarily zero.

Normal score approximation. Classifiers typically involve a continuous score function, Inline graphic, with prediction of group 1, ψ(xGj|X)=1, if Inline graphic < 0 (by convention here) and prediction of group 2, ψ(xGj|X)=2, if Inline graphic > 0. The normal score approximation is to treat Inline graphic as normally distributed with mean E[Inline graphic] and variance V[Inline graphic]. Then,

3.2. (3.9)

where u1=+1, u2=1, and Φ(·) is the standard normal cumulative density function.

Relation to linear discriminant theory. For the case that ψ is a linear discriminant function, we need to calculate Inline graphic and Inline graphic for the delta approximation and E[Inline graphic] and V[Inline graphic] for the normal score approximation. Define xFij to be the selected training features (i.e. given XΩ) of the ith sample from class j. It is convenient to arrange the signs of the data in a consistent manner, so we assume (without loss of generality) that whenever dimension k is included in the classifier, Inline graphic (i.e. if Inline graphic, reverse the signs of the data). Then, define μFj and ΣF to be the mean vector and covariance matrix of xFij, respectively. The difference between means is ΔF=μF2μF1. By symmetry, μF1+μF2=0. The distributions of the xFij will not typically be normal because they are conditioned on a significant difference between normal sample means, but the approximation below uses exact expressions (see supplementary material available at Biostatistics online) for Inline graphic, and V[Inline graphic] under the assumption that the distributions are normal. The expressions use results of Siskind (1972) on the second moments of inverse Wishart distributions, which are related to the sampling distribution of Inline graphic. This allows full incorporation of multivariate sampling variability in estimating the linear discriminant classifier and uses the principle that second moment–based approximations derived from normal theory are often reasonable. Thus, there are really 2 approximations happening: an approximation of training features (given they have been selected) as multivariate normally distributed and either the delta approximation or normal score approximation of generalization error.

3.3. Linear discriminant analysis when the training and validation samples follow different distributions

As above, define a training sample of Inline graphic, i=1,,n1, from class 1 and Inline graphic, i=1,,n2, from class 2. Define ΔF=μF2μF1, with ΔF > 0 in every dimension. Define the “true” parameters of ψ in the linear discriminant case as Inline graphic=(w, a), where Inline graphic and a=0.5(μF1+μF2)=0. These are estimated by Inline graphic, where

3.3. (3.10)

is the pooled unbiased estimate of Inline graphic, and Inline graphic. This is the setup of standard linear discriminant analysis (McLachlan, 1992).

Define a validation sample from class j as Inline graphic. The discriminant score for a value xG is

3.3. (3.11)

with prediction of class 1 for Inline graphic < 0 and class 2 for Inline graphic > 0. If the training and validation samples came from the same distributions, then w and a would give the optimal discriminant function.

We maintain the generality of the prior log-odds ratio, log(P(1)/P(2)), in the derivations. In the simulations below, we assume P(1)=P(2). These values may be very different for a population screening test, where only a very small fraction is expected to have a disease condition, compared to a problem such as disease classification given disease presence. Consideration of P(1)P(2) is standard in balancing sensitivity and specificity of medical tests.

To use the delta approximation (3.8), we need the first 2 moments of Inline graphic and Inline graphic and the derivatives of the generalization error with respect to the elements of Inline graphic and Inline graphic. To use the normal score approximation (3.9), we need the first 2 moments of Inline graphic. These are given exactly in the supplementary material available at Biostatistics online for the approximation that the training samples are normally distributed given that the selected dimensions were individually significant.

3.4. Summation over feature spaces

It remains to complete the calculation (3.1) efficiently by combining the Monte Carlo estimates of (3.3) and (3.4) and the approximations (3.8) or (3.9). If the space of features that might be selected is relatively simple, then one might directly enumerate cases where P(XΩ) is appreciably greater than zero; this is not stated mathematically here. More generally, one can use a Monte Carlo sample from the space of selected features to approximate (3.1).

Let {Ω(l)},l=1,,m, be a sample from P(Ω)P(XΩ). Corresponding to each partition piece Ω, there is a distribution P(X|XΩ). Since this is characterized by ΔF and ΣF (estimated by (3.5)), we denote Inline graphic. Then, the Monte Carlo approximation of (3.1) is

3.4. (3.12)

For feature-by-feature selection as discussed above, the relationship δr=δ(Ωr),r=1,,R, is one-to-one, so we can identify P(δr)P(Ωr). Then, sampling from P(Ω) in practice amounts to simulating on a feature-by-feature basis whether each feature is selected.

3.5. Choice of feature-selection thresholds

The above simulation and approximation steps require a choice for the P-value cutoff, Pc, used for feature selection. In practice, one can consider a range of Pc-values based on heuristic considerations to encompass the value of Pc that minimizes the expected validation error. In the simulation results here (supplementary material available at Biostatistics online, summarized below), the following heuristics perform well. The lower bound PL of Pc is set to the value at which the probability of zero true discoveries is 30% because excluding most or all informative features will not lead to good patterns. The upper bound PU of Pc is the minimum of 2 values. The first is the Pc level at which the probability of including all informative features equals 80%, on the rationale that after including most or all informative features, error rates will only get worse as false features are added. The second is the Pc such that the expected number of uninformative features is N/2MI, that is, the expected total number of features if all truly informative features are included should not exceed N/2. In scenarios where the second bound was lower than the first, higher Pc would lead to worse validation error rates due to many uninformative features.

3.6. Summary of simulation–approximation method

In summary, the simulation–approximation procedure uses the following steps:

  • 1. Choose Δ, MU, and n to define a study scenario.

  • 2. Choose a useful range of feature-selection thresholds, Pc, which influence how many features are chosen in the feature-selection stage.

  • 3. For each (unique) dimension of Δ and each Pc, use the noncentral t-distribution and/or Monte Carlo methods to estimate
    • a) the probability that the feature will be selected,
    • b) the expected within-group variances and difference between group means given that the feature is selected.
  • 4. For the Monte Carlo approximation (3.12), generate a sample of training feature combinations, {Ω(l)}, for which the generalization error will be approximated.

  • 5. For each training feature combination, use the variances and mean differences given that the features are selected to approximate the generalization error using either (3.8) or (3.9) with the calculations in the supplementary material available at Biostatistics online.

  • 6. Sum the terms in (3.12).

4. SIMULATION STUDY

Results of simulations of 7 realistic study designs are detailed in the supplementary material available at Biostatistics online. The first 6 scenarios consider optimal (i.e. Bayes) error rates of 0.05, 0.10, and 0.20 with either 3 (few strong) or 12 (many weak) truly informative dimensions, while the seventh considers optimal error of 0.05 from 46 (very many, very weak) dimensions. All scenarios use equal discovery sample sizes for control and disease groups, n1=n2, with the same mean difference for all informative dimensions and 10 patients per group for validation power. The simulation–approximation is accurate with the normal score approximation in all scenarios and with the delta approximation in all scenarios except for very many, very weak true features. Both methods are most accurate when most of the variation in generalization error is due to variation in which features are selected rather than in discriminant parameters given the feature space. Much larger numbers of truly informative dimensions would render the approximations inaccurate, and, moreover, suggest methods beyond basic linear discriminant analysis (LDA), such as shrinkage methods to constrain high variances in estimated patterns.

Several realistic scenarios have limited statistical power for validation and lead to substantially suboptimal patterns. With 12 informative and 2000 uninformative features and optimal error rate of 20%, sample sizes of 20, 50, and 100 give median validation error rates around 48%, 40%, and 30–35%, respectively, with only sample sizes of 100 giving better than 50% power for validation. If the features give optimal error rate of 10%, then 50 patients per group give high validation power but with median error rates of roughly 18–22% for 1000–5000 uninformative dimensions. With an optimal error rate of 5%, 20 samples would give roughly 50–80% validation power at 5% significance for 1000–5000 uninformative dimensions.

For a given optimal error rate, it is much harder to find patterns from many weak than from few strong informative features. Given optimal error rate of 20%, 50 patients per group for 3 strong features give better results than 100 patients per group with 12 weak features. For optimal error rate of 10% or 5%, 20 patients per group for 3 strong features give roughly comparable performance to 50 patients per group for 12 weak features. In summary, by far the strongest factors in pattern discovery power are sample size and individual feature strength. Some of these results are sobering in light of sample sizes in typical studies. It is plausible that some real studies to discover diagnostic patterns from high-dimensional assays could have low power for independent validation and find patterns far from the best true pattern.

5. DISCUSSION

Prospective analysis of study design for high-dimensional pattern discovery is important to plan studies with reasonable expectations of success based on scientific guesswork about the types of real patterns that might exist. The complexity of feature selection and pattern analysis methods raises many challenges for prospective study design. Here, we have explored a middle road between simulation and approximation, with simulations to handle variability in the selected features and an approximation of linear discriminant analysis given that the selected features appear to be informative in training data.

One of the most complicated ways in which the scenarios here may be optimistic is their lack of multivariate patterns and pattern recognition methods. Multivariate patterns could include correlated features that appear to be individually weak but are collectively strong or even harder possibilities such as the classic “XOR” (checkerboard) problem, where each marginal distribution has no information and only more complicated models than LDA can represent the pattern. In such problems, the hazard of over-fitting is greater than for the simulations here and would likely produce less favorable results. Other directions for further exploration of the relationships between sample size, numbers of informative and noninformative features, true optimal error rate, and discovery and generalization error rates include the following: generation of data from distributions that are unknown to the learning method (i.e. non-normal), further development of the relationship between false discovery rates and pattern discovery power, and further theoretical development of accurate approximations and/or efficient simulations.

Supplementary Material

[Supplementary Material]
kxp001_index.html (756B, html)

Acknowledgments

This work was initiated while all authors were employed at Predicant Biosciences. We thank our colleagues at Predicant for insightful discussions and support. Conflict of Interest: None declared.

References

  1. Adam BL, Qu YS, Davis JW, Ward MD, Clements MA, Cazares LH, Semmes OJ, Schellhammer PF, Yasui Y, Feng ZD and others. Serum protein fingerprinting coupled with a pattern-matching algorithm distinguishes prostate cancer from benign prostate hyperplasia and healthy men. Cancer Research. 2002;62:3609–3614. [PubMed] [Google Scholar]
  2. Adam BL, Vlahou A, Semmes OJ, Wright GL. Proteomic approaches to biomarker discovery in prostate and bladder cancers. Proteomics. 2001;1:1264–1270. doi: 10.1002/1615-9861(200110)1:10<1264::AID-PROT1264>3.0.CO;2-R. [DOI] [PubMed] [Google Scholar]
  3. Allison DB, Cui XQ, Page GP, Sabripour M. Microarray data analysis: from disarray to consolidation and consensus. Nature Reviews Genetics. 2006;7:55–65. doi: 10.1038/nrg1749. [DOI] [PubMed] [Google Scholar]
  4. Baker M. In biomarkers we trust? Nature Biotechnology. 2005;23:297–304. doi: 10.1038/nbt0305-297. [DOI] [PubMed] [Google Scholar]
  5. Benjamini Y, Hochberg Y. Controlling the false discovery rate—a practical and powerful approach to multiple testing. Journal of the Royal Statistical Society Series B—Methodological. 1995;57:289–300. [Google Scholar]
  6. Coombes KR, Morris JRS, Hu JH, Edmonson SR, Baggerly KA. Serum proteomics profiling—a young technology begins to mature. Nature Biotechnology. 2005;23:291–292. doi: 10.1038/nbt0305-291. [DOI] [PubMed] [Google Scholar]
  7. Diamandis EP. Mass spectrometry as a diagnostic and a cancer biomarker discovery tool—opportunities and potential limitations. Molecular & Cellular Proteomics. 2004a;3:367–378. doi: 10.1074/mcp.R400007-MCP200. [DOI] [PubMed] [Google Scholar]
  8. Diamandis EP. Proteomic patterns to identify ovarian cancer: 3 years on. Expert Review of Molecular Diagnostics. 2004b;4:575–577. doi: 10.1586/14737159.4.5.575. [DOI] [PubMed] [Google Scholar]
  9. Dobbin KK, Simon RM. Sample size planning for developing classifiers using high-dimensional DNA microarray data. Biostatistics. 2007;8:101–117. doi: 10.1093/biostatistics/kxj036. [DOI] [PubMed] [Google Scholar]
  10. Ebert MPA, Korc M, Malfertheiner P, Rocken C. Advances, challenges, and limitations in serum-proteome-based cancer diagnosis. Journal of Proteome Research. 2006;5:19–25. doi: 10.1021/pr050271e. [DOI] [PubMed] [Google Scholar]
  11. Efron B. Correlation and large-scale simultaneous significance testing. Journal of the American Statistical Association. 2007;102:93–103. [Google Scholar]
  12. Fu WJ, Dougherty ER, Mallick B, Carroll RJ. How many samples are needed to build a classifier: a general sequential approach. Bioinformatics (Oxford) 2005;21:63–70. doi: 10.1093/bioinformatics/bth461. [DOI] [PubMed] [Google Scholar]
  13. Gadbury GL, Page GP, Edwards J, Kayo T, Prolla TA, Weindruch R, Permana PA, Mountz JD, Allison DB. Power and sample size estimation in high dimensional biology. Statistical Methods in Medical Research. 2004;13:325–338. [Google Scholar]
  14. Hastie T, Tibshirani R, Friedman J. The Elements of Statistical Learning. New York: Springer; 2001. [Google Scholar]
  15. Hu J, Zou F, Wright FA. Practical FDR-based sample size calculations in microarray experiments. Bioinformatics (Oxford) 2005;21:3264–3272. doi: 10.1093/bioinformatics/bti519. [DOI] [PubMed] [Google Scholar]
  16. Hwang DH, Schmitt WA, Stephanopoulos G, Stephanopoulos G. Determination of minimum sample size and discriminatory expression patterns in microarray data. Bioinformatics. 2002;18:1184–1193. doi: 10.1093/bioinformatics/18.9.1184. [DOI] [PubMed] [Google Scholar]
  17. Jung S-H. Sample size for FDR-control in microarray data analysis. Bioinformatics (Oxford) 2005;21:3097–3104. doi: 10.1093/bioinformatics/bti456. [DOI] [PubMed] [Google Scholar]
  18. Li JN, Zhang Z, Rosenzweig J, Wang YY, Chan DW. Proteomics and bioinformatics approaches for identification of serum biomarkers to detect breast cancer. Clinical Chemistry. 2002;48:1296–1304. [PubMed] [Google Scholar]
  19. Listgarten J, Emili A. Practical proteomic biomarker discovery: taking a step back to leap forward. Drug Discovery Today. 2005;10:1697–1702. doi: 10.1016/S1359-6446(05)03645-7. [DOI] [PubMed] [Google Scholar]
  20. Ludwig JA, Weinstein JN. Biomarkers in cancer staging, prognosis and treatment selection. Nature Reviews Cancer. 2005;5:845–856. doi: 10.1038/nrc1739. [DOI] [PubMed] [Google Scholar]
  21. McLachlan GJ. Discriminant Analysis and Statistical Pattern Recognition. New York: John Wiley & Sons; 1992. [Google Scholar]
  22. Mukherjee S, Tamayo P, Rogers S, Rifkin R, Engle A, Campbell C, Golub TR, Mesirov JP. Estimating dataset size requirements for classifying DNA microarray data. Journal of Computational Biology. 2003;10:119–142. doi: 10.1089/106652703321825928. [DOI] [PubMed] [Google Scholar]
  23. Pepe MS, Longton G, Anderson GL, Schummer M. Selecting differentially expressed genes from microarray experiments. Biometrics. 2003;59:133–142. doi: 10.1111/1541-0420.00016. [DOI] [PubMed] [Google Scholar]
  24. Petricoin EF, Ardekani AM, Hitt BA, Levine PJ, Fusaro VA, Steinberg SM, Mills GB, Simone C, Fishman DA, Kohn EC and others. Use of proteomic patterns in serum to identify ovarian cancer. Lancet. 2002;359:572–577. doi: 10.1016/S0140-6736(02)07746-2. [DOI] [PubMed] [Google Scholar]
  25. Petricoin EF, Liotta LA. Mass spectrometry-based diagnostics: the upcoming revolution in disease detection. Clinical Chemistry. 2003;49:533–534. doi: 10.1373/49.4.533. [DOI] [PubMed] [Google Scholar]
  26. Petricoin EF, Ornstein DK, Paweletz CP, Ardekani A, Hackett PS, Hitt BA, Velassco A, Trucco C, Wiegand L, Wood K and others. Serum proteomic patterns for detection of prostate cancer. Journal of the National Cancer Institute. 2002;94:1576–1578. doi: 10.1093/jnci/94.20.1576. [DOI] [PubMed] [Google Scholar]
  27. Pounds S, Cheng C. Sample size determination for the false discovery rate. Bioinformatics. 2005;21:4263–4271. doi: 10.1093/bioinformatics/bti699. [DOI] [PubMed] [Google Scholar]
  28. Pusztai L, Hess KR. Clinical trial design for microarray predictive marker discovery and assessment. Annals of Oncology. 2004;15:1731–1737. doi: 10.1093/annonc/mdh466. [DOI] [PubMed] [Google Scholar]
  29. Rogers MA, Clarke P, Noble J, Munro NP, Paul A, Selby PJ, Banks RE. Proteomic profiling of urinary proteins in renal cancer by surface enhanced laser desorption ionization and neural-network analysis: identification of key issues affecting potential clinical utility. Cancer Research. 2003;63:6971–6983. [PubMed] [Google Scholar]
  30. Siskind V. Second moments of inverse Wishart-matrix elements. Biometrika. 1972;59:690–691. [Google Scholar]
  31. Sorace JM, Zhan M. A data review and re-assessment of ovarian cancer serum proteomic profiling. BMC Bioinformatics. 2003;4:24. doi: 10.1186/1471-2105-4-24. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Storey JD. A direct approach to false discovery rates. Journal of the Royal Statistical Society Series B—Statistical Methodology. 2002;64:479–498. [Google Scholar]
  33. Tsai C-A, Wang S-J, Chen D-T, Chen JJ. Sample size for gene expression microarray experiments. Bioinformatics (Oxford) 2005;21:1502–1508. doi: 10.1093/bioinformatics/bti162. [DOI] [PubMed] [Google Scholar]
  34. Viollaz AJ, Sfer AM, Salvatierra SM. An approximation of the unconditional error rates of the sample linear discriminant function. Communications in Statistics—Theory and Methods. 1995;24:1941–1969. [Google Scholar]
  35. Wang JH, Shen XT. Estimation of generalization error: random and fixed inputs. Statistica Sinica. 2006;16:569–588. [Google Scholar]
  36. Wilkins MR, Appel RD, Van Eyk JE, Chung MCM, Gorg A, Hecker M, Huber LA, Langen H, Link AJ, Paik YK and others. Guidelines for the next 10 years of proteomics. Proteomics. 2006;6:4–8. doi: 10.1002/pmic.200500856. [DOI] [PubMed] [Google Scholar]
  37. Wyman FJ, Young DM, Turner DW. A comparison of asymptotic error rate expansions for the sample linear discriminant function. Pattern Recognition. 1990;23:775–783. [Google Scholar]
  38. Zhang Z, Bast RC, Yu YH, Li JN, Sokoll LJ, Rai AJ, Rosenzweig JM, Cameron B, Wang YY, Meng XY and others. Three biomarkers identified from serum proteomic analysis for the detection of early stage ovarian cancer. Cancer Research. 2004;64:5882–5890. doi: 10.1158/0008-5472.CAN-04-0746. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

[Supplementary Material]
kxp001_index.html (756B, html)
kxp001_1.pdf (1.1MB, pdf)

Articles from Biostatistics (Oxford, England) are provided here courtesy of Oxford University Press

RESOURCES