Abstract
The parameters of logistic regression models are usually obtained by the method of maximum likelihood (ML). However, in analyses of small data sets or data sets with unbalanced outcomes or exposures, ML parameter estimates may not exist. This situation has been termed ‘separation’ as the two outcome groups are separated by the values of a covariate or a linear combination of covariates. To overcome the problem of non-existing ML parameter estimates, applying Firth’s correction (FC) was proposed. In practice, however, a principal investigator might be advised to ‘bring more data’ in order to solve a separation issue. We illustrate the problem by means of examples from colorectal cancer screening and ornithology. It is unclear if such an increasing sample size (ISS) strategy that keeps sampling new observations until separation is removed improves estimation compared to applying FC to the original data set. We performed an extensive simulation study where the main focus was to estimate the cost-adjusted relative efficiency of ML combined with ISS compared to FC. FC yielded reasonably small root mean squared errors and proved to be the more efficient estimator. Given our findings, we propose not to adapt the sample size when separation is encountered but to use FC as the default method of analysis whenever the number of observations or outcome events is critically low.
Keywords: maximum likelihood estimation, logistic regression, Firth’s correction, separation, penalized likelihood, bias
1. Introduction
In medical research, logistic regression is a popular method used to study the relationship between a binary outcome and a set of covariates. Regression coefficients can be interpreted as log odds ratios and are usually estimated by the method of maximum likelihood (ML). Moreover, individualized prognosis can be obtained by estimating the probability of an outcome given the covariates, making logistic regression widely used in the era of personalized medicine.
Despite analytically attractive properties under some regulatory conditions as the sample size increases [1,2], in analyses of small or sparse data sets, the properties of ML estimator become questionable: ML coefficient estimates are biased away from zero and very unstable [3] or may even not exist [4]. The situation in which the ML estimate of at least one regression coefficient does not exist, i.e., diverges to plus or minus infinity, has been termed ‘separation’ as the two outcome groups are separated by the values of a covariate or a linear combination of covariates. Therefore, perfect predictions for some (quasi-complete separation) or for all observations (complete separation) of the data set the model is fitted on are obtained. The simplest case of separation arises when an odds ratio is estimated from a contingency table with one zero cell count. In addition to a small sample size and sparsity various other factors, such as a small relative frequency of one of two levels of the outcome variable (unbalanced outcome), rare exposures and strong associations with the outcome can give rise to separation [5,6]. A large number of strongly correlated covariates is another such factor that has induced a possibly unjustified promotion of the 10 events per variable rule in biomedical literature [7,8].
In order to solve the problem of non-existing ML coefficient estimates, applying Firth’s correction (FC) [9], originally intended for bias reduction in generalized linear models, was proposed [5]. FC was shown to be robust to the problem of separation, making it a default choice in situations where separation is likely to occur. Although separation could also arise as a feature of the data-generating process, we are interested in situations where separation occurs as a sampling artefact, which is a consequence of random sampling variation and can be removed by increasing sample size. Therefore, in such a case, a principal investigator is often advised to ‘bring more data’ in order to solve a separation issue. It is unclear if such a simple increasing sample size (ISS) strategy that keeps sampling new observations until separation is removed improves estimation compared to FC applied to the original data set.
The objective of this paper is therefore to investigate the performance of ML after separation has been removed by ISS compared to FC applied to the initial data. The paper is organized as follows. In the next section, we briefly summarize the logistic regression model together with the ML and the FC estimators. In addition, we provide a simple example illustrating the separation problem and describe the setup for a simulation study. Subsequently, we describe our findings evaluating the empirical performance of ML combined with ISS and FC. As illustrative examples, we consider preliminary and final analyses of two studies from colorectal cancer screening and from ornithology. Finally, we summarize our most important findings.
2. Materials and Methods
2.1. General
Assume that we have independent observations , where denotes a binary outcome with level 1 occurring with probability and , , denotes a vector of covariate values for the -th subject. With logistic regression, designed to provide individualized prognosis given by ensuring that [10], the logit (i.e., the log-odds) of is modeled as a linear combination of the covariate values:
(1) |
where is an intercept and , , are regression coefficients for the covariates . The regression coefficients for covariates can be interpreted as log odds ratios, corresponding to one-unit differences in . They are usually obtained by ML estimation, maximizing the log-likelihood function:
(2) |
In the simplest case with one binary covariate, observations can be classified according to and in a contingency table with observed frequencies
x | |||
0 | 1 | ||
y | 0 | ||
1 |
The ML estimate of is then given by:
(3) |
and predicted probabilities are equal to the proportions of events in the two groups.
Now suppose that at least one of the observed frequencies of the table, e.g., , is zero. Accordingly, the ML estimate of the predicted probability for is zero, while the ML regression coefficient is not defined. According to Equation (3), whenever one of the cell counts of a table is zero, is not defined. To obtain a finite regression coefficient, we have to assume . This is in accordance with our general assumption on separation being a sampling artefact, which can always be removed by increasing sample size. If so, a finite number of covariates can never be sufficient to perfectly predict an outcome in the underlying population.
In FC, the likelihood function is penalized by the Jeffreys invariant prior [9]. For tables, FC is equivalent to ML after adding a 0.5 to each of the observed frequencies [11]. Applying FC does not only remove the first-order bias from ML regression coefficient estimates in generalized linear models [9], which may be severe in small or sparse data sets. Moreover, FC is robust to the problem of separation [5], as predicted probabilities are pulled to 0.5 such that . Note that other methods imposing shrinkage on the regression coefficients, which are often based on weakly or strongly informative priors, can as well overcome the separation issue [11,12,13].
2.2. Simulation Study Setup
The simulation study setup is described as recommended by Morris [14].
Aims: We intended to systematically investigate the performance of ML after removing separation by ISS (ML+ISS) in comparison to FC applied to the original data, focusing on scenarios where separation is likely to occur. We also investigated the performance of combining FC with ISS (FC+ISS).
Data-generating mechanisms: To capture a context plausible in medical research, we considered a data-generating scheme as described in Binder et al. [15]. First, we generated data sets of size with a binary outcome variable and covariates of mixed types that were obtained by applying certain transformations to variables sampled from a standard multivariate normal distribution with correlation matrix , see Table 1. In this way, we obtained four binary covariates , two ordinal covariates, and , with three levels, and four continuous covariates (Table 1). In order to avoid extreme values, we truncated continuous covariates at the third quartile plus five times the interquartile distance of the corresponding distribution.
Table 1.
Correlation of | Type | E( ) | ||
---|---|---|---|---|
binary | 0.8 | |||
binary | 0.36 | |||
binary | 0.5 | |||
|
binary | 0.5 | ||
ordinal | 1.11 | |||
ordinal | 0.37 | |||
continuous | 54.5 | |||
continuous | 138.58 | |||
continuous | 106.97 | |||
- | continuous | 54.5 |
We considered a full factorial design, varying the number of covariates , the sample size the expected value of and the value of . This resulted in 72 possible combinations of simulation parameters. We simulated 1000 data sets with each of those combinations. We held the true regression coefficients of covariates constant across simulation scenarios, setting them to for binary covariates, and to for ordinal covariates. For continuous covariates, the effects were chosen such that the log odds ratio between the first and the fifth sextile of the corresponding distribution was 0.69. An intercept was determined for each simulation scenario such that the desired proportion of events was approximately obtained. Finally, we sampled the binary outcome for subject , , from a Bernoulli distribution as after calculating by . Whenever separation was encountered in the original data set of size , we added to these data new observations sampled from the same distribution to generate an increased data set of sample size in which separation was removed.
Methods: We analyzed each simulated data set by fitting a logistic regression model (1) and estimating the regression coefficients by:
ML after removing separation by ISS (ML+ISS);
FC applied to the original data; and
FC after removing separation by ISS (FC+ISS).
For ML, we estimated 95% confidence intervals (CI) by the Wald method, and for FC, by penalized likelihood profiles [5]. ML and FC estimation was performed using the logistf [16] package in R, version 3.5.0. We checked for the presence of separation by the algorithm [17] implemented in the brglm2 [18] package.
Estimands: The true regression coefficient was the estimand in our study as was considered the target covariate (e.g., exposure to a risk factor). was simulated as an unbalanced binary covariate with expected value 0.8.
Performance measures: As for prediction, separation might not necessarily be considered as a problem; we focused on estimation and evaluated bias ( and mean squared error (MSE; ) of regression coefficient estimates as well as the probability that a 95% CI excludes (type I error rate or power), the probability that it includes the true value of (coverage) and the width of the 95% CIs. To compare the performance of ML+ISS and FC, we defined the cost-adjusted relative efficiency (CARE) of ML + ISS relative to FC as:
(4) |
where is the increased sample size needed to remove separation averaged over 1000 simulation runs. This measure weighs the increased efficiency against the costs of increasing the sample size, compared to the efficiency of FC at the original sample size. By its definition, CARE suggests that FC is a more efficient estimator. A similar concept was introduced by Armstrong [19] for a different context.
3. Results
This section is divided into two parts: First, we report the results from the simulation study described in Section 2.2.; second, we illustrate the problem by two-real life data examples showing results from preliminary and final analyses. The examples are taken from colorectal cancer screening and from ornithology.
3.1. Results of Simulation Study
Before evaluating the performance of the methods, we describe the mean new sample size that was required to remove separation and its dependence on the prevalence of separation in the original data sets of size . These results are shown by means of a nested loop plot [20,21] in Figure 1. Clearly, the mean sample size was positively correlated with the prevalence of separation. Both and the prevalence of separation were generally higher in scenarios with smaller sample sizes, in scenarios where , and with larger effects of . Moreover, they were lower for scenarios with two covariates compared to scenarios with five covariates. Interestingly, many scenarios with ten covariates had the fewest separated data sets. This might seem counterintuitive since, for a fixed non-separated data set, omitting covariates can never induce separation. However, in our simulation study, scenarios with 5 and 10 covariates do not only differ in the number of covariates but also in the model used to generate the binary outcome.
The bias of is shown in Figure 2. For ML+ISS, positive bias was generally observed in scenarios where separation occurred in less than a third of simulated data sets. By requiring samples to be nonseparated, the ML+ISS strategy appeared to correct the small-sample bias of ML (away from zero), and in scenarios with high prevalence of separation, its corrective nature even resulted in some negative bias. Applying FC, a small-sample bias-reduction method indeed yielded almost unbiased regression coefficients in many scenarios; however, in extreme cases with small sample sizes and higher values of negative bias was non-negligible and could be severe, especially if, in addition, the expected outcome prevalence was only 10%. Generally, bias towards zero increased with higher values of . With the exception of some scenarios with strong effects, applying FC to the increased samples (FC+ISS) additionally increased the bias towards zero compared to FC estimation.
Figure 3 presents the MSE of and CARE of ML+ISS compared to FC. For most scenarios, there were no considerable differences in terms of MSE of between the methods. However, ML+ISS yielded in some cases much larger sample sizes with the average sample size up to 11.8 times higher than the original and a maximum difference between and of 6424 observations. Therefore, FC was a more efficient estimator in all but four scenarios, and in these four scenarios, CARE was very close to one. Interestingly, even with sample sizes of , FC usually performed slightly better than ML+ISS. In extreme situations with the highest values of and small sample sizes of , the MSE by FC was poor in absolute terms and could only slightly be improved by ML+ISS. The improvement came at the cost of much higher sample sizes, such that FC was substantially more efficient (CARE >> 1). FC often achieved reasonably small MSE even with large proportions of separation, but it failed considerably in small sample situations () with strongest effects of and , and in all small sample situations with and covariates included in the model. In the latter situations, ML+ISS as well yielded very unstable regression coefficient estimates with an extremely large MSE. As expected, by increasing the sample size, FC+ISS most often reduced the MSE of FC. However, it also occurred that the MSE was slightly increased despite the larger sample size: In some scenarios where bias towards zero was smaller for FC+ISS than FC, the latter yielded lower MSE, and, notably, in a few scenarios, FC outperformed FC+ISS with respect to both bias and the MSE.
Throughout scenarios with and , the type I error rate for FC was closest to the nominal level of 5%, while by ML+ISS and FC+ISS, the null hypothesis was on average rejected with slightly lower probability. The mean type I error rate over all these scenarios was 4.8% for FC, and 4.6% for ML+ISS and FL+ISS. In scenarios with all three approaches tended to be conservative; the type I error rate was on average 3.6% for FC and FC+ISS and 3.3% for ML+ISS. FC+ISS had the largest power throughout scenarios with a nonzero effect of . However, in scenarios with larger sample sizes of , all approaches resulted in similar power, ML+ISS performing slightly worse than FC and FC+ISS whenever . The power of FC was distinctively lower than the power of FC+ISS or ML+ISS in small sample situations with strong effects; throughout all scenarios with a sample size of and its power stayed close to the level of 5% irrespective of . The mean power to detect over all scenarios with was 49.7% for FC, 49.8% for ML+ISS, and 52% for FC+ISS, and for the mean power was 25.9% for FC, 33.7% for ML+ISS, and 38.2% for FC+ISS, respectively. See Figure S1 in Supplementary Materials for detailed results on power and type I error rate for over all simulated scenarios.
The empirical coverage levels and the width of the 95% CIs for are reported in Figures S2 and S3, respectively. In scenarios with the average coverage of the profile penalized likelihood CIs for FC and FC+ISS was close to nominal. For both approaches, there was only one out of 36 scenarios where the coverage was outside the ‘plausible’ interval , into which the empirical coverage rate of a method with perfect coverage of 0.95 falls with a probability of 0.99 when the coverage is estimated from the analysis of 1000 simulated data sets. The ML+ISS Wald CIs were, by contrast, slightly conservative and outside the plausible range in 7 out of 36 scenarios. The mean coverage (and the average width) over all scenarios with was 95.5% (3.2) for FC, 96% (3) for ML+ISS, and 95.5% (2.8) for FC+ISS. The estimates obtained in scenarios with were far less precise than for , with a mean coverage (and average width) of 96% (4.9) for FC, 96% (13.6 and 3.6 after excluding scenarios with and ) for ML+ISS, and 95.4% (3.5) for FC+ISS. Notably, the coverage of the CIs of FC and FC+ISS was out of the plausible range in 9 and the one of ML+ISS in 18 out of 36 scenarios. These CIs for all three methods were generally conservative; however, in two scenarios with and the coverage of the profile penalized likelihood CIs for FC+ISS was below 89%. Throughout all scenarios, the width increased with larger effect sizes of . FC’s CIs were on average the widest, followed by ML+ISS; FC+ISS resulted in the narrowest CIs.
3.2. Examples
3.2.1. Bowel Preparation Study
Our first example is a study comparing four different bowel purgatives (A, B, C, and D) and their effect on the quality of colonoscopy. The procedure was called a ‘success’ if the cecum was intubated by the endoscopist, or ‘failure’ otherwise (see [22] for more details). The data set used in a preliminary analysis consisted of 4132 patients and suffered from a separation issue as the cecum of patients that used purgative B, which was the smallest of the four groups, was always successfully intubated. The final data set consisted of 5000 patients, and there was now a single patient in group B whose cecum could not be intubated. Table 2 shows the regression coefficients estimated by ML and FC in the preliminary and the final analysis. All the analyses presented here were adjusted for age and sex of patients. In line with our simulation results, the point estimates for purgative B vs. A by FC decreased after increasing the sample size, and the confidence interval of FC in the final data set was narrower than that of ML.
Table 2.
Data Set | Bowel Purgative | [95% CI] | [95% CI] | |
---|---|---|---|---|
Preliminary version |
A | 2149 | reference | |
B | 239 | not available | 1.95 [−0.01, 6.8] | |
C | 596 | 0.98 [−0.21, 2.18] | 0.85 [−0.13, 2.16] | |
D | 1148 | −0.83 [−1.33, −0.34] | −0.83 [−1.32, −0.34] | |
Final version |
A | 2648 | reference | |
B | 267 | 1.4 [−0.59, 3.39] | 1.01 [−0.62, 2.64] | |
C | 799 | 0.83 [−0.11, 1.76] | 0.74 [−0.15, 1.64] | |
D | 1286 | −0.83 [−1.28, −0.39] | −0.83 [−1.27, −0.39] |
3.2.2. European Passerine Birds Study
The second example is a study investigating the influence of migration (migratory, nonmigratory) and type of diet (granivorous, insectivorous, and omnivorous) on the presence of intestinal parasites in the feces of European passerine birds (see [23] for more details). The data set used in the initial analysis consisted of 366 birds, and no intestinal parasites were present in any of the 17 granivorous birds. After the principal investigator was advised to ‘bring more data’ in order to solve the separation issue, the final data consisted of 385 birds, and intestinal parasites were present in 2 out of 30 granivorous birds. Table 3 shows the regression coefficient estimates, adjusted for migration, obtained by ML and FC estimation in the preliminary and the final analysis. Again, the FC estimates decreased dramatically after increasing the sample size.
Table 3.
Data Set | Diet | [95% CI] | [95% CI] | |
---|---|---|---|---|
Preliminary version |
Granivorous | 17 | reference | |
Insectivorous | 274 | not available | 1.53 [−0.7, 6.43] | |
Omnivorous | 75 | not available | 2.17 [−0.02, 7.06] | |
Final (ISS)version |
Granivorous | 32 | reference | |
Insectivorous | 276 | 0.75 [−0.82, 2.33] | 0.57 [−0.73, 2.26] | |
Omnivorous | 77 | 1.42 [−0.15, 2.98] | 1.24 [−0.05, 2.91] |
4. Discussion
In medical research, studies investigating a binary outcome often focus on estimating the effect of an independent variable adjusted for others by logistic regression, and ML is the most commonly used method to estimate such an effect. However, in the analysis of small or sparse data sets, ML coefficient estimates are biased away from zero and very unstable, or ML estimation is even impossible. This occurs when covariates can perfectly separate the observations into the groups defined by the levels of the outcome variable as in our examples. (In fact, this may also happen when there is no separation in the data set. Rareness of events together with numerical instability or inaccuracy may make fitted probabilities already indistinguishable from zero or one by software packages.) Researchers confronted with log-odds ratio estimates diverging to may question the plausibility of their analysis results, as it is not assumed that the effect of a variable can be truly ‘infinite’. Rather, a nonexisting estimate is the result of an extreme small sample bias and a consequence of ‘bad luck in sampling’ [6].
Software packages do not always warn a user of the nonconvergence of the iterative ML fitting procedure, and an alternative estimation method to plug-in as a remedy of the problem is usually not suggested. Although seldom reported, we are aware that especially in observational studies, increasing the sample size is a common practice as it offers the most straightforward and intuitive solution, albeit its properties have not been examined thoroughly yet and were perhaps rated overoptimistically. In this paper, we investigated the empirical statistical properties of such a strategy as opposed to FC, which has been described as an ‘ideal solution to the problem of separation’ [5], making it the most widely available penalty in software packages [12]. We compared the performance of both approaches in terms of MSE and bias of regression coefficients and also incorporated the cost of increased sample size when evaluating CARE of ML+ISS relative to FC applied on the original data set. To better understand the impact of ISS, we additionally included in our study a strategy where the first sample size is increased until separation is removed and then FC is applied.
For finite sample sizes, the sampling distribution of ML regression coefficients consists of a subdistribution of finite values and a proportion of diverging estimates that can be explained as a consequence of an extreme small sample bias. The ML+ISS analysis compensates the bias away from zero of ML estimation by the bias towards zero that is induced by sampling until the extreme bias is removed. This can be exemplified by assuming that the original data set and the added observations are analyzed separately. The FC estimate obtained from the model fitted on the added observations has a strong negative bias: In the colonoscopy study, e.g., for the added observations was equal to , while the estimate in the original data was . In parallel, FC estimation can be represented by ML estimation on augmented data consisting of the original and pseudo-observations [5,11]. The model fitted on the pseudo observations has . As such, the ML+ISS strategy indeed results in acceptable biases but sometimes comes at the cost of highly inflated sample sizes. FC applied to data sets of (much) smaller sample sizes, by contrast, generally yields almost unbiased estimates. While the bias away from zero in ML estimation is a result of poor applicability of large-sample properties in small samples, the ISS strategy actually adds bias towards zero and so the two biases may cancel each other out. If combined with the unbiased FC strategy, however, ISS leads to a bias towards zero. When the true effect size is very large, FC can be biased towards zero [5], and ISS can amplify this effect, as demonstrated in our simulation study.
In terms of cost-corrected efficiency, the ML+ISS strategy is clearly outperformed by FC applied to the original, pre-planned sample size. Therefore, whenever the event rate is low or exposures are rare, a good advice to a researcher planning a study would be to consider FC as the method of analysis. However, in studies where not only effect estimates but also predictions are of importance, e.g., if differences in outcomes attributable to different covariates should be described not only on a relative scale but also on the absolute scale of event probabilities, which is often more relevant, caution is advised—FC leads to predicted probabilities biased towards 0.5. This is a consequence of the unbiasedness of the linear predictor, which naturally cannot translate into unbiasedness of its nonlinear transformation on the probability scale. While pulling predicted probabilities towards 0.5 is not problematic when the outcome levels are approximately balanced, in situations of rare events, this bias becomes apparent. Recently, two methods to overcome this shortcoming—FLIC and FLAC—were proposed by Puhr et al. [11], which both yield average predicted probabilities equal to the observed event rate. Therefore, the bias in predicted probabilities is no longer of concern if one of those methods is applied. Alternatively, one can resort to Bayesian methods and use weakly informative priors centered around 0 (so-called shrinkage priors) to cope with separation, e.g., Cauchy priors [24] or log-F priors [12]. Other more heuristic corrections based on a redefinition of the outcome variable have also been proposed [13]. Despite the usefulness of these other procedures in some instances, unlike FC, these methods are mostly justified by their empirical behavior and less by theoretical considerations. Moreover, they are not invariant to linear transformations of the design matrix. For brevity, we have not included their evaluation in the present paper as head-to-head comparisons with FC were already performed [11]. In further simulations not included in this report, we also investigated the performance of FLAC, log-F, and the combined strategies FLAC+ISS and log-F+ISS. FLAC and log-F showed similar patterns of behavior as FC and can yield MSEs even smaller than FC as previously shown [11]. The addition of ISS to FLAC and log-F had the same impact as with FC.
In order to produce valid and informative research results, it is desirable, already at the design stage of a study, to not fully rely on approximations based on large-sample properties of ML methods. Moreover, the possibility of separation, or more generally, of effects of sparsity in discrete outcome data should be taken into consideration and a suitable analysis strategy chosen. Although FC can be seen as an advancement of the traditional ML analysis, it cannot solve all problems of data sparsity [25]. Caution is needed when applying FC in extreme situations where the sample sizes are (too) small and the outcomes unbalanced at the same time: It should be taken into account that in such situations, FC performs poorly and lacks power to detect even very strong effects. Finally, researchers should not believe that collecting more data could save a study after analysis has failed but are advised to consult and involve experienced biostatisticians already at the design stage of their studies. For example, a biostatistician may consider the use of simulation to estimate the precision with which an effect can be estimated at different sample sizes and how the result of the study may influence decision making [26]. In such simulations, meaningful assumptions on the distribution of exposure and confounders could be made. Furthermore, the possibility of separation is adequately considered and the superiority of FC vs. ML as the main analysis strategy could be demonstrated.
5. Conclusions
By means of a simulation study and the analysis of two real data examples, we compared two strategies for dealing with separation in logistic regression with respect to their empirical performance. We conclude that sampling observations until the problem of separation is removed have adequate performance in terms of precision and inference but are relatively inefficient compared to Firth’s correction applied to a data set of original size.
Acknowledgments
We thank Michael Kammer for programming the nested loop plot function which is publicly available [21]. The bowel preparation study data were provided by Elisabeth Waldmann and Monika Ferlitsch from the Medical University of Vienna, and the European passerine birds study data by Petra Bandelj and Rok Blagus from the University of Ljubljana. Open Access Funding by the Austrian Science Fund (FWF).
Supplementary Materials
The following are available online at https://www.mdpi.com/1660-4601/16/23/4658/s1, Figure S1: Type I error rate and power of . Figure S2: Coverage of the 95% confidence intervals for . Figure S3: Width of the 95% confidence intervals for .
Author Contributions
Conceptualization, G.H. and A.G.; methodology, H.Š., A.G., and G.H.; software, H.Š.; validation, H.Š.; formal analysis, H.Š.; investigation, H.Š.; resources, G.H.; data curation, H.Š.; writing—original draft preparation, H.Š.; writing—review and editing, A.G. and G.H.; visualization, H.Š.; supervision, G.H. and A.G.; project administration, G.H.; funding acquisition, G.H.
Funding
This research was funded by the Austrian Science Fund (FWF), grant number I 2276.
Conflicts of Interest
The authors declare no conflict of interest.
References
- 1.Cox D.R., Snell E.J. Analysis of binary data. 2nd ed. Chapman and Hall/CRC; London, UK: 1989. [Google Scholar]
- 2.Kosmidis I. Bias in parametric estimation: Reduction and useful side-effects. WIREs Comput. Stat. 2014;6:185–196. doi: 10.1002/wics.1296. [DOI] [Google Scholar]
- 3.King G., Zeng L. Logistic regression in rare events data. Polit. Anal. 2001;9:137–163. doi: 10.1093/oxfordjournals.pan.a004868. [DOI] [Google Scholar]
- 4.Albert A., Anderson J.A. On the existence of maximum likelihood estimates in logistic regression models. Biometrika. 1984;71:1–10. doi: 10.1093/biomet/71.1.1. [DOI] [Google Scholar]
- 5.Heinze G., Schemper M. A solution to the problem of separation in logistic regression. Stat. Med. 2002;21:2409–2419. doi: 10.1002/sim.1047. [DOI] [PubMed] [Google Scholar]
- 6.Mansournia M.A., Geroldinger A., Greenland S., Heinze G. Separation in logistic regression: Causes, consequences, and control. Am. J. Epidemiol. 2018;187:864–870. doi: 10.1093/aje/kwx299. [DOI] [PubMed] [Google Scholar]
- 7.Courvoisier D.S., Combescure C., Agoritsas T., Gayet-Ageron A., Perneger T.V. Performance of logistic regression modeling: Beyond the number of events per variable, the role of data structure. J. Clin. Epidemiol. 2011;64:993–1000. doi: 10.1016/j.jclinepi.2010.11.012. [DOI] [PubMed] [Google Scholar]
- 8.van Smeden M., de Groot J.A.H., Moons K.G.M., Collins G.S., Altman D.G., Eijkemans M.J.C., Reitsma J.B. No rationale for 1 variable per 10 events criterion for binary logistic regression analysis. BMC Med. Res. Methodol. 2016;16:163. doi: 10.1186/s12874-016-0267-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Firth D. Bias reduction of maximum likelihood estimates. Biometrika. 1993;80:27–38. doi: 10.1093/biomet/80.1.27. [DOI] [Google Scholar]
- 10.Hastie T., Tibshirani R., Friedman J. The Elements of Statistical Learning. 2nd ed. Springer New York Inc.; New York, NY, USA: 2001. pp. 119–128. [Google Scholar]
- 11.Puhr R., Heinze G., Nold M., Lusa L., Geroldinger A. Firth’s logistic regression with rare events: Accurate effect estimates and predictions? Stat. Med. 2017;36:2302–2317. doi: 10.1002/sim.7273. [DOI] [PubMed] [Google Scholar]
- 12.Greenland S., Mansournia M.A. Penalization, bias reduction, and default priors in logistic and related categorical and survival regressions. Stat. Med. 2015;34:3133–3143. doi: 10.1002/sim.6537. [DOI] [PubMed] [Google Scholar]
- 13.Rousseeuw P.J., Christmann A. Robustness against separation and outliers in logistic regression. Comput. Stat. Data Anal. 2003;43:315–332. doi: 10.1016/S0167-9473(02)00304-3. [DOI] [Google Scholar]
- 14.Morris T.P., White I.R., Crowther M.J. Using simulation studies to evaluate statistical methods. Stat. Med. 2019;38:2074–2102. doi: 10.1002/sim.8086. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Binder H., Sauerbrei W., Royston P. Multivariable Model-Building with Continuous Covariates: 1. Performance Measures and Simulation Design. University of Freiburg Germany; Breisgau, Germany: 2011. Technical Report FDM-Preprint 105. [Google Scholar]
- 16.Heinze G., Ploner M., Dunkler D., Southworth H. logistf: Firth’s Bias Reduced Logistic Regression. CRAN; Vienna, Austria: 2014. R package version 1.22. [Google Scholar]
- 17.Konis K. SafeBinaryRegression: Safe Binary Regression. CRAN; Vienna, Austria: 2013. R package version 0.1-3. [Google Scholar]
- 18.Kosmidis I. brglm2: Bias Reduction in Generalized Linear Models. CRAN; Vienna, Austria: 2018. R package version 0.1.8. [Google Scholar]
- 19.Armstrong B.G. Optimizing Power in Allocating Resources to Exposure Assessment in an Epidemiologic Study. Am. J. Epidemiol. 1996;144:192–197. doi: 10.1093/oxfordjournals.aje.a008908. [DOI] [PubMed] [Google Scholar]
- 20.Rücker G., Schwarzer G. Presenting simulation results in a nested loop plot. BMC Med. Res. Methodol. 2014;14:129. doi: 10.1186/1471-2288-14-129. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.LoopR: A R Package for Creating Nested Loop Plots. [(accessed on 1 October 2019)]; Available online: https://github.com/matherealize/loopR.
- 22.Waldmann E., Penz D., Majcher B., Zagata J., Šinkovec H., Heinze G., Dokladanska A., Szymanska A., Trauner M., Ferlitsch A., et al. Impact of high-volume, intermediate-volume and low-volume bowel preparation on colonoscopy quality and patient satisfaction: An observational study. United Eur. Gastroenterol. J. 2019;7:114–124. doi: 10.1177/2050640618809842. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Bandelj P., Blagus R., Trilar T., Vengust M., Vergles Rataj A. Influence of phylogeny, migration and type of diet on the presence of intestinal parasites in the faeces of European passerine birds (Passeriformes) Wildl. Biol. 2015;21:227–233. doi: 10.2981/wlb.00044. [DOI] [Google Scholar]
- 24.Gelman A., Jakulin A., Pittau M.G., Su Y.S. A weakly informative default prior distribution for logistic and other regression models. Ann. Appl. Stat. 2008;2:1360–1383. doi: 10.1214/08-AOAS191. [DOI] [Google Scholar]
- 25.Greenland S., Mansournia M.A., Altman D.G. Sparse data bias: A problem hiding in plain sight. BMJ. 2016;353:i1981. doi: 10.1136/bmj.i1981. [DOI] [PubMed] [Google Scholar]
- 26.Phillips C.V. The economics of more research is needed. Int. J. Epidemiol. 2001;30:771–776. doi: 10.1093/ije/30.4.771. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.