Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2017 Dec 20.
Published in final edited form as: Stat Med. 2016 Aug 16;35(29):5477–5494. doi: 10.1002/sim.7075

Estimating relative risk of a log-transformed exposure measured in pools

Emily M Mitchell a, Torie C Plowden b, Enrique F Schisterman a,*
PMCID: PMC5118194  NIHMSID: NIHMS808995  PMID: 27530506

Abstract

Pooling biospecimens prior to performing laboratory assays is a useful tool to reduce costs, achieve minimum volume requirements, and mitigate assay measurement error. When estimating the risk of a continuous, pooled exposure on a binary outcome, specialized statistical techniques are required. Current methods include a regression calibration approach, where the expectation of the individual-level exposure is calculated by adjusting the observed pooled measurement with additional covariate data. While this method employs a linear regression calibration model, we propose an alternative model that can accommodate log-linear relationships between the exposure and predictive covariates. The proposed model permits direct estimation of the relative risk associated with a log-transformation of an exposure measured in pools.

Keywords: Biomarkers, Design, Log-transformation, Pooled Specimens, Regression Calibration

1. Introduction

In epidemiological studies, pooling biospecimens prior to performing laboratory assays can help reduce lab costs, lessens assay measurement error in the observations, and achieve minimum volume requirements by combining only a small portion of each sample into a larger pool [15]. Pooling is currently used to screen blood for diseases such as HIV [610], biomonitor prevalence of environmental exposures [11, 12], and reduce measurement error in genetic microarray studies [13].

When analyzing pooled specimens in a regression setting, specialized statistical techniques are often needed. If the pooled samples are to be included as a predictor of a binary outcome, a set-based logistic regression model yields unbiased estimates when pools are matched on case status [5], or in matched survival studies [14]. If pools are not matched on the outcome, more complex methods may be required [15, 16]. This scenario could occur, for instance, if pools are formed and assayed before the outcome of interest is observed [15]. In such cases, Zhang and Albert [15] proposed a regression calibration (RC) approach to perform regression of a binary outcome on a pooled predictor. Regression calibration is typically applied in measurement error problems; in the pooling scenario, a pooled measurement is treated as a mis-measured value of each of the individual concentrations of the specimens comprising that pool. The existing regression calibration approach for pools uses ancillary information to adjust the observed measurement of the pooled sample, by estimating the linear relationship between the pooled variable and potential predictors. This method performs well when the pooled biomarker is approximately normally distributed.

If individual values of the biomarker are available (i.e. no pooling is performed), a regression model on a log-transformation of the exposure may be of interest, particularly when a biomarker is positive and right-skewed [17, 18]. When measured in pools, a log-transformation on the pooled measurements can induce computational complexity, due to the non-linearity of the log function. In light of recent developments for regression on a log-transformed, pooled outcome [19, 20], we propose an extension of the pooled regression calibration approach that will permit estimation of regression coefficients corresponding to a log-transformed biomarker, when only pooled measurements are available.

In Section 2, we set up the regression model and summarize the existing regression calibration method for an untransformed biomarker. In the following section we describe our proposed methods for extending this regression calibration approach to a log-transformed predictor variable that is logarithmically associated with the auxiliary predictive covariates. We then use simulation studies to compare the performance of the proposed approaches to existing methods in estimating relative risk, and we apply these strategies using data from the Effects of Aspirin in Gestation and Reproduction (EAGeR) trial to determine the association between prenatal serum leptin levels and the outcome of live birth.

2. Regression Calibration: Untransformed Predictor

In this section, we summarize the methods presented by Zhang and Albert [15] to apply a regression calibration approach to an untransformed exposure. While these methods were originally developed for a general link function, we specifically focus on the log link, to investigate the common scenario when the relative risk is of primary interest. Thus, for all of the methods that follow, we assume we are interested in estimating the relative risk associated with a biomarker of interest that is measured in pools. When all N specimens are measured individually, the relative risk associated with the untransformed exposure is commonly estimated under the model:

Pr(Yi=1Xi,Ci)=E(YiXi,Ci)=exp(α+βXi+γCi) (1)

for i = 1, …, N, where Yi is the binary outcome, Xi the exposure of interest that is only observed after laboratory analysis, Ci a vector of observed covariates, γ a vector of regression coefficients, and exp(β) is the relative risk associated with a 1-unit increase in X. Model (1) is commonly fit parametrically assuming that Y follows a binomial or Poisson distribution, often employing a sandwich estimator for robust standard error estimation [21, 22].

When the exposure is measured in pools instead of on individual specimens, the pooled measurement X¯i=gi-1j=1giXij is observed, assumed to be the average of the concentrations of specimens comprising that pool [23], where gi is the number of specimens in pool i, (i = 1, …, n), n denotes the total number of pools, and i=1ngi=N.

Zhang and Albert [15] showed that, through application of a first-order Taylor Series approximation, i could be related to the outcome approximately by:

Pr(Yij=1X¯i,Cij)exp{α+βE(XijX¯i,{Cij})+γCij}

where the ‘ij’ subscript denotes the jth subject in the ith pool. E(XijX¯i,{Cij}) is the expected value of the individual specimen concentration, given the observed measurement of pool i as well as additional covariate information { Cij:j=1,,gi}, which consists of the individual-level vector of covariates for each individual in pool i({Cij}), as well as potentially additional auxiliary variables. The quality of this Taylor Series approximation depends on the higher-order (2nd and higher) centered moments of the distribution f(XijX¯i,Cij), where the approximation works best when these values are small [15].

Under the linear regression calibration approach, a linear relationship between Xij and Cij is assumed, such that

Xij=ϕ0+ϕ1Cij+εij (2)

where the εij’s are assumed to be independent and identically-distributed with E(εij) = 0 and Var(εij) = σ2 for all j = 1, …, gi, i = 1, …, n. It follows that:

X¯i=ϕ0+ϕ1C¯i+εi (3)

where E(εi)=gi-1jE(εij)=0 and Var(εi)=gi-2jVar(εij)=gi-1σ2. C¯i is a matrix of covariates, where each column represents the average across the individual-level covariate values within pool i (i.e. C¯i=gi-1j=1giCij).

By combining equations (2) and (3), E(XijX¯i,{Cij}) can then be derived as follows:

Xij=(X¯i-X¯i)+ϕ0+ϕ1Cij+εij=X¯i-(ϕ0+ϕ1C¯i+εi)+(ϕ0+ϕ1Cij+εij)=X¯i+ϕ1(Cij-C¯i)+ηij (4)

where E(ηij) = E(εij) − E(εi) = 0 and thus

E(XijX¯i,{Cij})=X¯i+ϕ1(Cij-C¯i)=X¯i+E(XijCij)-E(X¯iC¯i).

Although the individual X’s are unobserved, ϕ̂ = (ϕ̂0, ϕ̂1) can be estimated by a weighted least squares on the pooled measurements [24]. Similar to the traditional regression calibration approach for measurement error, E(XijX¯i,{Cij}) acts as an adjustment on i, the observed ‘mis-measured’ value of Xij, given the information inherent in { Cij}. The estimated relative risk, exp(β̂), can then be obtained by substituting E^(XijX¯i,{Cij})=X¯i+ϕ^1(Cij-C¯i) for Xij in equation (1) and following with standard binomial or Poisson regression.

3. Regression Calibration: Log-transformed predictor

While the linear regression calibration approach described in Section 2 estimates the risk associated with an untransformed exposure, it is often of interest to estimate the risk based on a log-transformation of the exposure, particularly when that predictor is highly skewed. In this scenario, the regression model based on the individual-level data is:

Pr(Yij=1Xij,Cij)=exp{α+β(logXij)+γCij}. (5)

Here, exp(β) represents the risk associated with a 1-unit increase in log(X). To fit this regression model when X is measured in pools, we again employ a Taylor Series expansion, where

Pr(Yij=1X¯i,Cij)exp{α+βE(logXijX¯i,{Cij})+γCij}. (6)

To facilitate calibration of the log-transformed individual biomarker values, we propose two alternative methods to directly estimate E(logXijX¯i,{Cij}), which are particularly suited for capturing logarithmic relationships between the individual-level exposure and the covariates. First, we exploit the convenient summation properties of the gamma distribution to employ a parametric strategy, which will theoretically obtain maximum precision when the distributional assumptions are correctly specified. Next, we consider an alternative estimation procedure based on recently-developed quasi-likelihood methods for pooled measurements. This method is expected to be more flexible in accommodating alternate distributional assumptions, as it requires specification of the first two moments, as opposed to the full distribution, of the exposure conditional on the remaining covariates.

3.1. Parametric Calibration

Similar to the log-normal distribution, the gamma distribution can effectively characterize positive, right-skewed random variables. In addition, the convenient summation properties of the gamma distribution make it particularly amenable to analyzing biomarker data measured in pools [3, 23, 25, 26]. In this section, we exploit these convenient properties to obtain a closed form expression for E(logXijX¯i,{Cij}).

Suppose that ( XijCij) follows a gamma distribution with shape kij and scale θ, where logkij=ϕ0+ϕ1Cij and E(XijCij)=kijθ=θexp(ϕ0+ϕ1Cij). Letting Si = gii denote the sum of the individual specimens comprising pool i, it follows directly from the summation property of the gamma distribution that Si{Cij}~Gamma(jkij,θ) and (Si-Xij){Cij}~Gamma(jjkij,θ). Then (Xij/Si|Si, { Cij}) follows a beta distribution, since:

f(XijSi,Cij)=f(Xij,SiCij)f(SiCij)=f(XijCij)f(Si-XijCij)f(SiCij)=Γ(jkij)Γ(kij)Γ(jjkij)Xijkij-1(Si-Xij)(jjkij-1)Sijkij-1=Γ(jkij)SiΓ(kij)Γ(jjkij)(XijSi)(kij-1)(1-XijSi)(jjkij-1).

By applying a transformation of variables, we see that (Xij/SiSi,{Cij})~Beta(kij,jjkij).

We can then use the properties of the beta distribution to derive the conditional expectation of the individual concentrations given the measured average (or equivalently, the measured sum) of the pool as well as additional individual-level auxiliary variables:

E(logXijSi,{Cij})=log(Si)+E[log(XijSi)X¯i,{Cij}]=log(Si)+ψ(kij)-ψ(jkij)=log(Si)+E(logXijCij)-E(logSi{Cij})

where ψ is the digamma function [27] and the final step follows from the properties of the gamma distribution. Hence, similar to its linear analogue, the regression calibration step assuming a gamma distribution on X|C* involves adjusting the observed ‘mis-measured’ value log i with ancillary information based on the relationship between X and C*. To obtain the calibrated value of log Xij for each individual specimen, k^ij=exp(ϕ0+ϕ1Cij) is estimated for all i, j by optimizing the log-likelihood of the gamma distribution with constant scale parameter [20]. Calculation of E(logXijX¯i,{Cij}) is then straightforward. Once obtained, this value is plugged into the regression model (6). Example R code to obtain the maximum likelihood estimates from this gamma calibration approach is provided in the Appendix.

3.2. Quasi-Likelihood Calibration

While the parametric assumption on ( XijCij) will theoretically provide the most precise estimates of ϕ̂ when correctly specified, a less restrictive approach may be preferred if ( XijCij) does not follow a gamma distribution. For this method, we assume a linear relationship between log Xij and Cij, such that:

logXij=ϕ0+ϕ1Cij+εij (7)

where E(εij) = 0 and Var(εij) = σ2. Then, similar to the approach outlined in equation (4), we can re-write this expression as:

logXij=(logX¯i-logX¯i)+ϕ0+ϕ1Cij+εij=logX¯i-log{gi-1jexp(ϕ0+ϕ1Cij+εij)}+ϕ0+ϕ1Cij+εij=loggiX¯i-log{jexp(ϕ1Cij+εij)}+ϕ1Cij+εij.

Applying a first-order Taylor Series approximation to the second log function, the conditional expectation is then:

E(logXijX¯i,{Cij})loggiX¯i-log{jexp(ϕ1Cij)E(eεij)}+ϕ1Cij=loggiX¯i-logE(eε11)-log{jexp(ϕ1Cij)}+ϕ1Cij

which follows directly from the assumption that the εij’s are identically distributed. Estimates of ϕ1 are found using an extension of the quasi-likelihood method for pools [19]. Under equation (7), E(X¯i{Cij})=g-1jexp(ϕ0+ϕ1Cij), where ϕ0=ϕ0+logE{exp(ε11)}, and Var(X¯i{Cij})=νg-2jexp{2(ϕ0+ϕ1Cij)}, where ν = Var{exp(ε11)}/E{exp(ε11)}2. Then the following quasi-score function is solved:

i=1nX¯i-μiVi(ϕ)dμidϕ=0

where μi=gi-1jexp(ϕ0+ϕ1Cij) and Vi(ϕ)=gi-2jexp(ϕ0+ϕ1Cij). Note that ν is omitted from the function Vi since it does not affect estimation of the parameters of interest, ϕ. Normally, ν would be absorbed into the estimate of the dispersion parameter. However, since we only seek to identify a point estimate for E(logXijX¯i,{Cij}), a function of ϕ1, estimation of the dispersion parameter is unnecessary for our purposes. Furthermore, since ϕ0 is a function of E{exp(ε11)}, its estimator will not be an unbiased estimate of ϕ0. However, since the main value of interest does not depend on ϕ0, it is also treated as a nuisance parameter. After obtaining estimates of ϕ̂1 from the quasi-score function, E^(logXijX¯i,{Cij}) is substituted into (6). Since E(XijCij)=E(eε11)exp(ϕ0+ϕ1Cij) and E(X¯i{Cij})=g-1E(eε11)jexp(ϕ0+ϕ1Cij), we can write:

E(logXijX¯i,{Cij})α+logX¯i+logE(XijCij)-logE(X¯i{Cij})=α+log{X¯iE(XijCij)E(X¯i{Cij})} (8)

where α* = −log E(eε11). Thus, similar to the regression calibration goal of adjusting a mis-measured value based on auxiliary information, this approximation adjusts the ‘mis-measured’ value of i multiplicatively based on the additional information of the expected value of the individual-level biomarker concentration (Xij) given the individual-level observed covariates ( Cij). Furthermore, adding a constant (e.g. α*) to the approximation of E(logXijX¯i,{Cij}) will not change the estimate of β or γ in (6), since this constant will be absorbed into the intercept term. Thus, it is unnecessary to estimate the value of log E(eε11), and this term can also be treated as a nuisance parameter. Example code for each of these methods is provided in the Appendix.

4. Simulation Study

In this section we perform simulations to test the performance of each of these methods in estimating the relative risk of a log-transformed, pooled predictor on a binary outcome. 5000 simulations of 1000 observations were performed for various pool sizes to mimic the motivating study from the EAGeR trial, described in more detail in Section 6. Two confounders, C1 and C2, were generated to resemble Age and BMI, respectively. C1 was generated from a truncated normal distribution with a mean of 28 and standard deviation of 3.5, and with lower and upper limits of 18 and 42, to reflect the age distribution of the population of women in the EAGeR trial, specifically, pre-menopausal women attempting pregnancy. C2 was generated from a log-normal distribution with a geometric mean of E(log C2|C1) = 3.2 + 0.002(C1) and geometric variance of 0.05. To test these methods under various distributions of X|C, separate simulation studies were run, with X|C generated under a gamma distribution, a log-normal distribution, and a normal distribution.

For the first set of simulations, the exposure of interest was generated under a gamma distribution with scale parameter θ = 300 and:

E(XC1,C2)=exp{-1-0.02C1+3.2logC2}.

These parameters were chosen to mimic the distribution of leptin in the EAGeR study, as described in more detail in Section 6. Under this simulation, the parametric approach from Section 3.1 is correctly specified. Furthermore, the assumptions on the first and second moments of E(X|C1, C2) that are needed for the quasi-likelihood approach in (8) are satisfied when X|C is generated under a gamma distribution. Thus, the quasi-likelihood calibration procedure is also correctly-specified in this scenario.

For the second set of simulations, X was generated under a log-normal distribution with geometric standard deviation of 0.25 such that:

E(logXC1,C2)=-1.5-0.02C1+3.2logC2.

This distribution corresponds to the assumption of a linear association between log X and C, as in Section 3.2. Thus, the multiplicative adjustment to the pooled measurement using quasi-likelihood estimates of ϕ are correctly-specified in this scenario, while the assumption of the conditional gamma distribution needed for the parametric calibration method is misspecified. Results from this model will help illuminate potential consequences of misspecification of the distribution of X|C under the parametric regression calibration approach.

A skewed exposure distribution does not necessarily imply that a log-linear or gamma regression calibration model will provide the best predicted values of the individual-level X’s. Rather, the aptness of a log-linear or gamma model to predict individual-level exposure values depends more on the underlying link function than the skewness of the distribution [28]. For instance, if a covariate C is highly skewed, while X|C is normally-distributed, the marginal distribution of X may be skewed, even though a linear link function would more accurately capture the relationship between X and C. The advantage of the proposed regression calibration models is the ability to capture a logarithmic relationship between the exposure and some predictive covariates. Since the goal in fitting the calibration model is to ‘adjust’ the pooled values of the exposure back to the individual-level values, the model that most appropriately characterizes the relationship between X and C is preferred. Thus, we also conducted simulations where X|C is normally-distributed, in which case the linear regression calibration described in Section 2 would provide the best predictions of the individual-level X’s; expected values are incorporated into the outcome model as in Equation (8). In this scenario, X is generated with mean:

E(XC1,C2)=-100-0.02C1+3.2C22

and a standard deviation of 40. This model differs from the previous simulations since X is dependent on C22, rather than log C2, to generate a skewed marginal distribution for X while retaining an underlying linear link function.

After simulating the exposure of interest, the outcome Y for each simulation study was then generated under a binomial distribution with log link such that:

E(YX,C1,C2)=exp{0.8-0.1logX-0.015C1-0.01C2}

where C2=I(C225) is a binary variable designed to mimic an indication of whether a participant is overweight, where C2 represents BMI. The regression coefficients were chosen to emulate those observed in the EAGeR data analysis, where Y represents the outcome of live birth. Note that C2 and C2 are highly correlated, and C2 can be considered an auxiliary variable useful in helping predict the individual-level values of X. After generating the simulated data, pools with an equal number of specimens were artificially formed to have size 2, 4, and 8, to assess the potential impact of higher pool sizes on parameter estimation.

Models were fit under the complete data (“Full”), the naive model (“Naive”), the existing linear regression calibration approach (“Linear RC”), a regression calibration assuming a gamma distribution on X|C (“Gamma RC”), and a regression calibration approach assuming a log-linear relationship between X and C and estimating the corresponding parameters using quasi-likelihood (“Log-linear RC”). All models were fit assuming a linear relationship between Y and the log-transformed individual-level X’s. Poisson models with robust variance were fit according to (6), where the calibrated value for E(logXijX¯i,Cij) differed according to the assumptions underlying the relationship between X and C.

The full model, fit on the complete data, is considered the gold standard, since this is the model that would be available if all individual specimen concentrations were measured, assuming no measurement error. The naive model simply substitutes the pooled values for the individual X’s in (5), which is essentially equivalent to a linear regression calibration approach with no covariates (i.e. an intercept-only model). The linear regression calibration model assumes a linear relationship between X and C to obtain an estimate of the expected value of the individual level concentrations, given the observed covariates and pooled values, X^ij=E^(XijCij,X¯i). To fit this model, this calibrated value was substituted for X in (5).While this ad hoc substitution and log-transformation does not conform to the original proposed methods of Zhang and Albert [15], our goal in fitting this model was to determine the extent to which an inappropriate application of the regression calibration approach may result in invalid inference. Correct application of the regression calibration approach as described in Section 2 has been shown to provide unbiased and efficient estimates of β when only pooled measurements are available, and the goal is to estimate the risk associated with an untransformed exposure [15]. In addition, we demonstrate the utility of this method when X|C is normally-distributed in the third simulation scenario. Furthermore, while Zhang and Albert [15] tested several versions of the regression calibration approach, including imputation vs. augmentation, and a plug-in vs. normal distribution assumption, we only apply regression calibration under the imputation approach (as described in Section 2) and normal distribution assumption on X|C, since this method performed best, in general, in the simulation scenarios presented by Zhang and Albert [15]. In addition, while the Yij’s corresponding to the same pool are generally correlated with each other given i, we fit regression models under the correlation structure of working independence, since additional testing of alternate structures did not noticeably improve estimation of the parameters of interest or their standard errors [15].

In addition to fitting the full, naive, and linear regression calibration models, we also apply the regression calibration approach assuming a gamma distribution for X|C. Under the first simulation scenario, this model is correctly specified, so estimates of β should be consistent. While this approach is not correctly-specified under the second simulation study when X|C is generated from a log-normal distribution, the gamma distribution can adequately characterize skewed distributions in epidemiological studies [2931].

Estimation of the maximum likelihood estimates ϕ̂ under the gamma assumption was conducted using the optim function in R. To improve convergence, several starting values were tested. Finally, the regression calibration assuming a linear relationship between log X and C was fit, where R’s nleqslv function was applied to solve the non-linear system of quasi-likelihood estimating equations (see Appendix for details).

5. Simulation Results

Tables 1 and 2 provide the percent bias, empirical standard deviation, average estimated standard error, and 95% confidence interval coverage for β when X|C is generated under a gamma and lognormal distribution, respectively. Based on these simulation results, both the gamma and log-linear RC models give approximately unbiased estimates of the β coefficients with nominal 95% confidence interval coverage, when the true conditional exposure distribution is log-normal or gamma. These results hold regardless of pool size. Furthermore, the precision of these estimates under these proposed models is close to the full precision under the complete model, even for pools of size 8, when the total number of lab assays is reduced from 1000 to 125. This precision is likely a reflection of the predictive capacity of the model due to the high correlation between the simulated covariates and the exposure of interest.

Table 1.

Simulation results when X|C is generated under a gamma distribution. Relative bias, empirical standard deviation (SD), and average estimated standard error ( SE^) are multiplied by 100 to facilitate comparison between methods.

Pool Size Method β1 = −0.1
γ1 = −0.015
γ2 = −0.01
Bias x 100 SD x 100 SE^ x100 95% CI Coverage Bias x 100 SD x 100 SE^ x 100 95% CI Coverage Bias x 100 SD x 100 SE^ x 100 95% CI Coverage
2 Full −0.05 4.83 5.01 96.0 −0.02 0.74 0.75 95.5 0.03 7.75 7.85 95.5
Naive 4.71 5.19 5.37 85.8 0.12 0.73 0.74 95.3 −8.83 5.92 5.99 69.1
Linear RC 5.19 3.83 4.00 75.9 0.05 0.77 0.78 95.7 −5.07 7.05 7.08 89.1
Gamma RC −0.37 5.55 5.59 95.3 −0.02 0.74 0.75 95.4 0.10 8.23 8.24 95.2
Log-Linear RC −0.37 5.55 5.59 95.3 −0.02 0.74 0.75 95.5 0.11 8.23 8.24 95.2
4 Full −0.05 4.83 5.01 96.0 −0.02 0.74 0.75 95.5 0.03 7.75 7.85 95.5
Naive 5.61 6.75 6.83 86.8 0.15 0.73 0.74 95.1 −10.6 5.45 5.47 50.8
Linear RC 5.44 3.92 4.13 75.7 0.06 0.78 0.79 95.4 −5.48 7.01 7.03 88.0
Gamma RC −0.43 5.78 5.80 95.0 −0.02 0.74 0.75 95.4 0.12 8.43 8.40 95.5
Log-Linear RC −0.44 5.78 5.80 94.9 −0.02 0.74 0.75 95.3 0.14 8.43 8.41 95.4
8 Full −0.05 4.83 5.01 96.0 −0.02 0.74 0.75 95.5 0.03 7.75 7.85 95.5
Naive 5.94 9.17 9.22 90.6 0.16 0.73 0.74 95.1 −11.3 5.27 5.29 42.6
Linear RC 5.45 4.02 4.24 77.7 0.06 0.78 0.79 95.6 −5.54 6.98 7.05 87.7
Gamma RC −0.44 5.87 5.88 95.1 −0.02 0.74 0.75 95.3 0.14 8.52 8.48 95.1
Log-Linear RC −0.45 5.87 5.88 95.1 −0.02 0.74 0.75 95.2 0.14 8.52 8.49 95.1

Table 2.

Simulation results when X|C is generated under a log-normal distribution. Relative bias, empirical standard deviation (SD), and average estimated standard error ( SE^) are multiplied by 100 to facilitate comparison between methods.

Pool Size Method β1 = −0.1
γ1 = −0.015
γ2 = −0.01
Bias x 100 SD x 100 SE^ x100 95% CI Coverage Bias x 100 SD x 100 SE^ x 100 95% CI Coverage Bias x 100 SD x 100 SE^ x 100 95% CI Coverage
2 Full −0.01 4.74 4.84 95.5 0.00 0.69 0.70 95.2 0.04 7.30 7.36 95.3
Naive 4.73 4.86 4.96 84.2 0.13 0.68 0.70 94.8 −8.46 5.60 5.59 67.6
Linear RC 5.52 3.45 3.67 68.8 0.08 0.72 0.73 95.3 −5.16 6.43 6.59 88.4
Gamma RC 0.26 5.09 5.14 95.2 0.00 0.69 0.70 95.2 −0.41 7.56 7.57 95.4
Log-linear RC 0.32 5.06 5.11 95.1 0.00 0.69 0.70 95.3 −0.33 7.59 7.60 95.4
4 Full −0.01 4.74 4.84 95.5 0.00 0.69 0.70 95.2 0.04 7.30 7.36 95.3
Naive 5.68 6.11 6.22 85.1 0.15 0.68 0.70 94.8 −10.3 5.18 5.13 48.5
Linear RC 5.74 3.62 3.84 69.5 0.08 0.73 0.74 95.4 −5.50 6.45 6.58 87.0
Gamma RC 0.21 5.31 5.35 95.2 0.00 0.69 0.71 95.0 −0.32 7.77 7.76 94.9
Log-linear RC 0.28 5.27 5.31 95.2 0.00 0.69 0.71 95.0 −0.28 7.78 7.78 95.0
8 Full −0.01 4.74 4.84 95.5 0.00 0.69 0.70 95.2 0.04 7.30 7.36 95.3
Naive 5.95 8.30 8.34 88.8 0.16 0.68 0.70 94.6 −10.9 5.00 4.97 40.9
Linear RC 5.71 3.76 3.96 71.8 0.08 0.73 0.75 95.1 −5.50 6.50 6.62 87.5
Gamma RC 0.21 5.46 5.42 94.4 0.00 0.69 0.71 94.9 0.10 8.95 8.07 94.8
Log-linear RC 0.17 5.38 5.44 95.3 0.00 0.69 0.71 95.0 −0.14 7.94 7.90 94.9

On the other hand, regression coefficient estimates from the naive and linear regression calibration models (‘Linear RC’) are biased under the first two simulation scenarios. Although this bias is not excessive (~5 to 10% bias), the artificially low precision of these estimates results in sub-nominal 95% confidence interval coverage, despite accurate estimation of standard errors. These coverage rates, which drop below 70% for the linear regression calibration model and under 41% for the naive model, make these methods particularly susceptible to inflated Type I Error rates due to misspecification of the calibration model.

When the exposure is linearly related to the covariates, however, the linear regression calibration model performs best, with approximately unbiased coefficients, nominal 95% confidence interval coverage, and the lowest standard errors among the calibrated models (Table 3). Although the log-linear and gamma regression calibration models are misspecified, they continue to provide approximately unbiased coefficient estimates with close to nominal confidence interval coverage, demonstrating robustness to misspecification of the link function. Estimates under the naive model, on the other hand, remain biased, similar to those from the previous simulation scenarios, which subsequently corresponds to sub-nominal confidence interval coverage rates.

Table 3.

Simulation results when X|C is generated under a normal distribution. Relative bias, empirical standard deviation (SD), and average estimated standard error ( SE^) are multiplied by 100 to facilitate comparison between methods.

Pool Size Method β1 = −0.1
γ1 = −0.015
γ2 = −0.01
Bias x 100 SD x 100 SE^ x100 95% CI Coverage Bias x 100 SD x 100 SE^ x 100 95% CI Coverage Bias x 100 SD x 100 SE^ x 100 95% CI Coverage
2 Full 0.07 7.42 7.55 95.4 0.00 0.62 0.63 95.7 −0.05 7.27 7.19 95.0
Naive 4.94 7.57 7.63 89.4 −0.01 0.62 0.63 95.6 −5.68 5.25 5.20 79.8
Linear RC 0.07 7.44 7.57 95.5 0.00 0.62 0.63 95.7 −0.05 7.28 7.20 94.8
Gamma RC 0.34 7.58 7.56 95.1 0.00 0.62 0.63 95.7 −0.35 7.43 7.17 94.2
Log-Linear RC −0.21 7.75 7.80 95.3 0.00 0.62 0.63 95.6 0.08 7.39 7.28 94.8
4 Full 0.07 7.42 7.55 95.4 0.00 0.62 0.63 95.7 −0.05 7.27 7.19 95.0
Naive 5.80 9.73 9.68 90.3 −0.01 0.62 0.63 95.6 −6.79 4.76 4.70 69.1
Linear RC 0.06 7.46 7.58 95.4 0.00 0.62 0.63 95.7 −0.05 7.29 7.20 94.7
Gamma RC 0.18 7.82 7.69 94.7 0.01 0.63 0.63 95.5 −0.11 7.86 7.39 93.3
Log-Linear RC −0.43 7.97 7.98 94.9 0.00 0.62 0.63 95.6 0.17 7.46 7.35 94.6
8 Full 0.07 7.42 7.55 95.4 0.00 0.62 0.63 95.7 −0.05 7.27 7.19 95.0
Naive 6.03 13.18 13.14 92.4 −0.01 0.62 0.63 95.5 −7.20 4.56 4.52 64.2
Linear RC 0.06 7.46 7.58 95.3 0.00 0.62 0.63 95.6 −0.04 7.29 7.20 94.7
Gamma RC −0.16 7.96 7.84 94.7 0.01 0.63 0.63 95.3 0.11 7.96 7.43 93.7
Log-Linear RC −0.57 8.10 8.09 94.8 0.00 0.62 0.63 95.6 0.23 7.50 7.38 94.6

As mentioned previously and demonstrated in the simulation studies, the choice of calibration model is not dictated by the skewness of the marginal distribution of X. While the gamma and log-linear calibration models demonstrated robustness to misspecification in the final simulation study, the ability to choose the best model could improve estimate efficiency. While several methods exist for comparing model fit, the difficulty arises in the flexibility of these models to accommodate pooled, i.e. averaged, X values. Thus, in this scenario, we choose to use the root mean squared error (RMSE) to compare relative goodness of fit of the calibration models, due to its flexibility in assessing pooled values as well as its non-parametric underpinnings, which facilitate comparison across distributional assumptions. When applied to pooled values of X, the RMSE is defined as: RMSE=n-1i=1n(X¯i-X^i)2 and X^i=gi-1j=1giE^(XijCij) is the predicted pooled value based on the predicted individual-level values from the calibration model. While not a formal test of goodness of fit, comparing RMSE values across the calibration models can guide selection of the preferred model, where a lower RMSE value suggests a better fit. To demonstrate this, we provide average RMSE values for the calibration models under each simulation scenario in Table 4.

Table 4.

Average root mean squared error (RMSE) for the calibration models under each simulation scenario.

Pool Size Method Distribution of X|C
Gamma Log-normal Normal
2 Naive 5380 3546 793
Linear RC 2665 1992 28
Gamma RC 1161 1304 163
Log-linear RC 1167 1311 53
4 Naive 3799 2504 560
Linear RC 1876 1402 19
Gamma RC 817 916 107
Log-linear RC 821 922 31
8 Naive 2679 1765 394
Linear RC 1314 983 13
Gamma RC 571 645 61
Log-linear RC 575 646 20

When X|C is generated under the gamma or log-normal models, the average RMSE for the gamma and log-linear calibration models are quite similar, but both are considerably lower than the linear calibration model, indicating that both the gamma or log-linear calibration models fit the data better than the linear model. When X|C is generated under a normal model, however, a comparison of the RMSE values clearly indicates that a linear regression calibration model provides the best fit. In addition, the RMSE values from the naive model, which is equivalent to a linear calibration model with no covariates (the intercept-only model), has the highest RMSE in all scenarios. While the naive model is not a real contender for a calibration model in this scenario, RMSE values under this model can illustrate how much the other calibration models may be improving the predicted values of the X’s. If the RMSE for each calibration model were close to that from the naive model, the calibration step would be of little benefit to improving the fit of the final outcome model. Thus, comparing RMSE values can guide selection of the best calibration model, as well as provide an indication of the extent of improvement available over the naive model.

Due to the strong similarity in performance between gamma and log-linear regression calibration models, choice of which to fit could depend on the researcher’s preference concerning the assumptions and tools required to produce maximum likelihood or quasi-likelihood estimates, especially when the RMSE values fit under these models are similar. For instance, the gamma assumption requires full specification of the parametric distribution and involves maximization of the log-likelihood. The quasi-likelihood approach, while only requiring assumptions on the mean and variance of log X|C, relies on the performance of the Taylor Series approximation. If both procedures are employed but provide contradictory results, the discrepancy might indicate a more serious violation of the underlying model assumptions on the individual-level data.

6. Data Analysis

The Effects of Aspirin in Gestation and Reproduction (EAGeR) Trial was a prospective, multi-site, double-blinded randomized controlled trial, designed to evaluate the effects of preconception low-dose aspirin on gestation among women with a prior history of pregnancy loss. Women in the study were 18–40 years old, had reported 1 or 2 prior pregnancy losses and were actively attempting to conceive. These women had no prior gynecologic disease and had not been diagnosed with infertility. Baseline characteristics were recorded and blood draws were taken at randomization. Participants were followed for up to 6 cycles of attempting pregnancy; if they became pregnant, they were followed until the end of pregnancy [32].

Leptin is a hormone that plays a significant role in human physiology and reproduction, and recent studies have demonstrated an association between leptin in menstrual function and pregnancy-related conditions [3437]. The goal of this data analysis was to determine whether higher levels of leptin were associated with adverse pregnancy outcomes in the EAGeR trial after adjusting for potential confounders, where the outcome of interest is live birth. Due to the strong correlation between BMI and leptin [33], overweight status is considered a potential confounder of the relationship between leptin and pregnancy outcome. In this dataset, the median leptin concentration among normal weight women (BMI < 25) was 9.0 ng/mL, while the median concentration among overweight women was 34.3 ng/mL. Age was also considered as a potential confounder.

1052 observations with complete data on live birth outcome, leptin measurements, age, and BMI were available for analysis. Leptin concentration was measured for women at baseline using the Quantikine ELISA assay (R&D Systems, Minneapolis, MN), and BMI and age were assessed from the baseline questionnaire. The distribution of leptin concentration in the study sample was positive and right-skewed, with a median of 16.96 ng/mL and a mean of 23.42 ng/mL. A histogram of the individual-level concentrations as well as a normal Q-Q plot of log(leptin) are provided in Figure 1.

Figure 1.

Figure 1

Histogram and normal Q-Q plot of leptin and log(leptin) concentrations, respectively, measured on individual specimens obtained from participants in the EAGeR trial.

Since previous research has assessed the association between log-transformed leptin levels and health outcomes [3845], we apply a log-transformation to leptin to estimate the relative risk associated with a 1-unit increase of leptin on the log scale. Specifically, we seek to estimate the association between leptin and birth outcome after controlling for overweight status and age, based on the following regression model:

E(Livebirth)=exp{β0+β1log(leptin)+γ1Age+γ2Overweight} (9)

where ‘Live birth’ is the binary outcome indicating whether a participant had a live birth, ‘Age’ is measured in years, and ‘Overweight’ is a binary variable indicating a woman categorized as overweight or obese based on pre-pregnancy BMI (BMI ≥ 25). Since leptin concentrations were measured on individual specimens, we can directly compare results from analyses on artificially pooled samples to the gold standard based on individual-level measurements. To illustrate the proposed methods, pools were artificially formed first of size 2 (n = 526), then of size 4 (n = 263). For all models, Poisson distributions with log link and robust variance were fit to directly estimate β. Continuous BMI values were applied as ancillary information to help calibrate values of log-transformed leptin for analyses when only the pooled measurements are available.

Table 5 gives parameter estimates and standard errors under the full data as well as for each of the models discussed in Sections 2 and 3 for pool sizes of 2 and 4. In addition, RMSE values for the calibration models are provided as a relative goodness of fit criterion. Under the full data based on individual measurements of leptin, log(leptin) tris significantly negatively associated with live birth at a significance level of 0.05 (Relative Risk = 0.9). Consistent with previous knowledge, age was also negatively associated with live birth, an association that was preserved under all scenarios. When leptin is measured in pools, the naive and linear regression calibration (‘Linear RC’) models underestimate the magnitude of the association detected under the full data, and are unable to detect this significant association. Of note, the linear regression calibration model suffered from additional missing data (32 observations) due to a log-transformation on negative expected values in the regression calibration step.

Table 5.

Regression coefficient estimates and standard errors when regressing live birth on log(leptin), age, and overweight status in the EAGeR study. A ‘*’ indicates a statistically significant association at the 0.05 level. RMSE refers to the root mean squared error of the calibration model.

Pool Size Method log(Leptin) Age Overweight RMSE
1 Full Data −0.111 (0.040)* −0.018 (0.006)* −0.016 (0.078)
2 Naive −0.061 (0.051) −0.017 (0.006)* −0.124 (0.066) 13876
Linear RC −0.033 (0.038) −0.018 (0.006)* −0.109 (0.079) 7512
Gamma RC −0.120 (0.057)* −0.018 (0.006)* −0.014 (0.091) 7448
Log-linear RC −0.121 (0.057)* −0.018 (0.006)* −0.011 (0.091) 7453
4 Naive −0.035 (0.071) −0.017 (0.006)* −0.152 (0.061) 9814
Linear RC −0.059 (0.046) −0.019 (0.006)* −0.080 (0.088) 5258
Gamma RC −0.141 (0.074) −0.018 (0.006)* 0.011 (0.107) 5253
Log-linear RC −0.128 (0.069) −0.018 (0.006)* −0.003 (0.103) 5205

On the other hand, the regression calibration approaches assuming a gamma distribution or log-linear relationship give point estimates that are close to those under the full data. Furthermore, the standard errors of these estimates are precise enough to detect the significant association between log(leptin) and live birth with only 526 (as opposed to 1052) assays required. When pools are formed of size 4, none of the methods retain sufficient power to detect the significant association between log(leptin) and live birth. However, the gamma and log-linear RC models continue to calculate coefficient estimates that are similar to those under the full data.

The RMSE values provide a guide to choose the best calibration model in this analysis. When pool size is 2, the RMSE suggests that the gamma regression calibration model provides the best fit, with the log-linear calibration model performing similarly. In fact, these two models provide regression coefficient estimates similar to those from the full model, with the gamma model giving the closest estimates. When pool size is 4, the RMSE values suggest that the log-linear model is preferred and in fact, this model does slightly outperform the gamma and linear calibration models with respect to approximating estimates from the full data for log(leptin) and age. In real-world scenarios, when the full data is unavailable, the RMSE values can help select the best calibration model from among a set of pre-specified candidates.

This data analysis illustrates the potential benefits of applying a gamma or log-linear regression calibration approach to pooled data when the association between a log-transformed exposure and a binary outcome is of primary interest. Importantly, this analysis is for illustration purposes only and should not be interpreted as a complete and robust estimate of the association between leptin levels and live birth outcome, as this inference would require appropriate handling of additional statistical issues such as missing data, selection bias, measurement error, and other potential confounders.

7. Discussion

Pooling biospecimens prior to performing lab assays has been gaining traction as a cost-saving strategy in epidemiological studies in recent years with the development of statistical tools necessary for analysis. In this study, we have contributed to the toolkit of available analytical methods for pooled specimens by developing two new approaches to estimate relative risk when an exposure is measured in pools and the risk associated with the log-transformation of this exposure is of primary interest. We chose to focus on the logarithmic transformation of an exposure due to the popularity of this strategy in epidemiological studies. While it may be possible to extend these methods to additional monotonic transformations (e.g. square root), a thorough exploration of this topic is beyond the scope of this paper. In addition, while we only expounded upon a log link in the outcome model to specifically investigate the performance of these methods when estimating a relative risk, these methods could also be applied to general link functions, by combining the proposed calibration models for the expectation of a log-transformed exposure, with the original methodology for untransformed exposures as described in Zhang and Albert [15].

The proposed approaches were developed as an extension of an existing regression calibration method for pooled exposures, which was based on a linear adjustment to the pooled measurements, applying the measurement error approach of regression calibration to correct for the pools acting as ‘mis-measured’ values of the individual-level concentrations [15]. Since this method was not developed for a log-transformed exposure, our methods were proposed with the goal of direct calibration of the log-transformed individual-level biomarker values, particularly when the log-transformed exposure is linearly associated with the predictive covariates.

The first of the proposed methods exploited the convenient summation properties of the gamma distribution to derive a closed-form expression for the conditional expectation of the individual-level log-transformed exposures given the measured pooled values and supplementary auxiliary information. In contrast, the second approach assumes a linear relationship between the log-transformed exposure and the auxiliary variables. While the parametric methods assuming a gamma distribution should theoretically provide the most precise estimates when correctly-specified, our empirical simulations demonstrated negligible differences between the two methods. Implementation of each of these methods, however, does require different computational methods (e.g. maximizing the log-likelihood vs. solving a system of nonlinear equations), as demonstrated in the example code in the Appendix.

Since the regression calibration methods rely on a Taylor Series approximation, at least some amount of bias may be unavoidable. These methods, however, will often provide considerable improvement over the naive model with respect to estimating the regression coefficients, in which the pooled measurements are included as predictors in the regression model without any adjustment. The success of the regression calibration approach depends on the amount of relevant auxiliary variable information available, and the predictive value of these variables on the pooled exposure. If the biomarker of interest is uncorrelated with all of the auxiliary covariates, the regression calibration approach is indistinguishable from the naive method, which, as evidenced in the simulation studies, can result in bias and sub-nominal confidence interval coverage. On the other hand, a covariate that is perfectly predictive of the individual-level exposure concentrations will produce risk estimates that are equivalent to the full model. Thus, the goal of regression calibration is to shift estimates away from the biased estimates produced by the naive model, towards those from the full model. It may be difficult, however, to determine how close the adjusted estimates come to the gold standard of the individual-level data. While RMSE provides a useful tool to compare candidate calibration models, it does not provide an absolute measure of predictive capacity. Additional validation data containing measurements on individual-level exposure values may be required to assess the absolute fit of the calibration model.

Since the regression calibration approach can essentially be viewed as a single imputation approach, where observed relationships in the data are leveraged to predict expected values of the unobserved (individual-level) concentrations of the biomarker, a multiple imputation framework may improve this approach. This strategy could incorporate outcome information into the imputation procedure, and could be generalized to calibrate additional missing covariate information in the dataset. We are currently testing the feasibility of this method applied to pooled specimens.

As evidenced in the data analysis, pooling can impact the power to detect a significant association. Strategic pooling techniques, where pools are formed to be homogeneous with respect to important covariates, can help improve power when pooling is performed [4, 24]. In the data example, for instance, pooling specimens from subjects with similar BMI values could enhance prediction of the individual-level leptin values, subsequently improving precision of the regression coefficients in the outcome model. Similar methods to those presented here will still be applicable. Pooling on the outcome, however, creates additional complexities in the analysis, similar to those associated with outcome-dependent sampling [46, 47]. In a non-pooled scenario, logistic regression can accommodate outcome-dependent sampling (i.e. case-control sampling) directly, while log-linear regression requires additional adjustments such as propensity score weighting to mitigate selection bias. Similarly, when pools are formed to be homogeneous on the outcome, a set-based logistic regression model will provide unbiased regression coefficient estimates. However, it is unclear how to extend this model to accommodate a log-transformation on a pooled exposure [5]. Under log-linear regression with a pooled, log-transformed exposure, adjustments similar to propensity weighting may be necessary to counteract the bias induced by pooling on the outcome, as in the non-pooled scenario. While recent research has explored this topic in a slightly different scenario [46], additional efforts are needed to determine how a similar method could be extended to a log-transformed exposure. In summary, careful consideration of pooling design can help reduce cost, attain minimum volume requirements for lab assays, and reduce the number of measurements below the limit of detection. When appropriate analytical techniques are applied to these pooled measurements, reliable and efficient inference can be obtained.

Acknowledgments

Contract/grant sponsor: This research was supported by the Intramural Research Program of the Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, and by the Long-Range Research Initiative of the American Chemistry Council.

Appendix

Example R Code: Table A1 gives the first 12 observations from the example dataset on which subsequent code is based. Example code calculates estimates for ϕ̂, then E(log X|C*), and applies this value to the original log-linear regression model.

Table A1.

Example dataset of a binary outcome (Y), covariates (C1 and C2), and an exposure measured in pools (Pooled_X).

Y C1 C2 Pooled_X Pool_Number
1 0 1.818 0.549 1
0 0 2.865 0.549 1
1 0 1.319 0.549 1
1 0 2.458 0.549 1
0 0 1.538 0.575 2
0 1 2.209 0.575 2
0 0 1.757 0.575 2
0 1 1.817 0.575 2
0 0 2.359 1.096 3
0 0 1.765 1.096 3
1 1 1.750 1.096 3
0 0 1.623 1.096 3
## Data set-up:
# Calculate poolwise averages of additional covariates
     Pooled_C1 = ave(C1,Pool_Number)
     Pooled_C2 = ave(C2,Pool_Number)
# Calculate pool weights (i.e. pool size) for weighted least squares
     Pool_wts = tabulate(Pool_Number)[Pool_Number]
## Regression Calibration Step:
  # 1. Weighted Least Squares (Zhang and Albert, 2011)
         rc.ZA = glm(Pooled_X ∼ Pooled_C1+Pooled_C2,weights=Pool_wts)
         E.X = cbind(1,C1,C2)%*%rc.ZA$coef
         E.XP = ave(E.X,Pool_Number)
         ZA.Xhat = Pooled_X + E.X - E.XP
  # 2. Parametric Gamma assumption on X|C
       # Define log-likelihood for Gamma distribution:
          gamma.ll = function(theta,Pooled_X,C,pool=1:length(Pooled_X),pool.size){
              scale = theta[length(theta)] # scale parameter
              phi = theta[-length(theta)]
              mu.ij = exp(as.matrix(cbind(1,C))%*%phi)
              mu.i = ave(mu.ij,pool,FUN=sum)
              ll = sum(dgamma(Pooled_X,shape=mu.i/scale,
                            scale=scale/pool.size,log=T))
              return(-ll)
          }
          xfit.GA = optim(c(1,0,0,1),gamma.ll,Pooled_X=Pooled_X,
                                     # can try different starting values to improve convergence
                                     C=cbind(C1,C2),pool=Pool_Number,pool.size=Pool_wts)
          EX.GA = exp(cbind(1,C1,C2)%*%xfit.GA$par[1:3])
          ES.GA = ave(EX.GA,Pool_Number,FUN=sum)
          logGA.Xhat = log(Pooled_X*Pool_wts) + digamma(EX.GA) - digamma(ES.GA)
  # 3. Quasi-likelihood regression calibration
       # Define quasi-likelihood for alternate version:
          library(nleqslv)
          QL.htro = function(phi,Pooled_X,C,pool=1:length(Pooled_X),pool.size){
               mu.ij = exp(as.matrix(cbind(1,C))%*%phi)
               V.ij = mu.ijˆ2
               mu.i = ave(mu.ij,pool,FUN=mean)[!duplicated(pool)]
               V.i = (ave(V.ij,pool,FUN=sum)/pool.sizeˆ2)[!duplicated(pool)]
               dmu = apply(as.matrix(cbind(1,C))*c(mu.ij),2,ave,pool)[!duplicated(pool),]
               Q = t(dmu)%*%as.matrix((Pooled_X-mu.i)/V.i)
               return(Q)
          }
          xfit.QL = nleqslv(c(0,0,0),QL.htro, Pooled_X = Pooled_X[!duplicated(Pool_Number)],
                                       C=cbind(C1,C2),pool=Pool_Number,pool.size=Pool_wts)
          EX.QL = exp(cbind(1,C1,C2)%*%xfit.QL$x)
          EXP.QL = ave(EX.QL,Pool_Number)
          logQL.Xhat = log(Pooled_X)+log(EX.QL)-log(EXP.QL)
## Fit original regression model on calibrated X values
# Zhang and Albert (2011) version on log-transformed X:
   glm(Y∼C1+C2+log(ZA.Xhat),family=quasibinomial(link=log))
# Parametric Gamma Reg. Cal. on log(X):
   glm(Y∼C1+C2+logGA.Xhat,family=quasibinomial(link=log))
# Quasi-Likelihood Reg. Cal. on log(X):
   glm(Y∼C1+C2+logQL.Xhat,family=quasibinomial(link=log))

References

  • 1.Heffernan AL, Aylward LL, Toms LM, Sly PD, Macleod M, Mueller JF. Pooled biological specimens for human biomonitoring of environmental chemicals: opportunities and limitations. Journal of Exposure Science and Environmental Epidemiology. 2014;24:225–32. doi: 10.1038/jes.2013.76. [DOI] [PubMed] [Google Scholar]
  • 2.Mumford SL, Schisterman EF, Vexler A, Liu A. Pooling biospecimens and limits of detection: effects on ROC curve analysis. Biostatistics. 2006;7:585–98. doi: 10.1093/biostatistics/kxj027. [DOI] [PubMed] [Google Scholar]
  • 3.Schisterman EF, Vexler A, Ye A, Perkins NJ. A combined efficient design for biomarker data subject to a limit of detection due to measuring instrument sensitivity. Annals of Applied Statistics. 2011;5:2651–67. [Google Scholar]
  • 4.Vansteelandt S, Goetghebeur E, Verstraeten T. Regression models for disease prevalence with diagnostic tests on pools of serum samples. Biometrics. 2000;56:1126–33. doi: 10.1111/j.0006-341x.2000.01126.x. [DOI] [PubMed] [Google Scholar]
  • 5.Weinberg CR, Umbach DM. Using pooled exposure assessment to improve efficiency in case-control studies. Biometrics. 1999;55:718–26. doi: 10.1111/j.0006-341x.1999.00718.x. [DOI] [PubMed] [Google Scholar]
  • 6.Cardoso MS, Koerner K, Kubanek B. Mini-pool screening by nucleic acid testing for hepatitis B virus, hepatitis C virus, and HIV: preliminary results. Transfusion. 1998;38:905–7. doi: 10.1046/j.1537-2995.1998.381098440853.x. [DOI] [PubMed] [Google Scholar]
  • 7.Lan SJ, Hsieh CC, Yen YY. Pooling strategies for screening blood in areas with low prevalence of HIV. Biometrical Journal. 1993;35:553–65. [Google Scholar]
  • 8.Weinberg CR. HPV screening for cervical cancer in rural India. New England Journal of Medicine. 2009;361:305–6. [PMC free article] [PubMed] [Google Scholar]
  • 9.Center for Biologics Evaluation and Research. Guidance for Industry, Nucleic Acid Testing for Human Immunodeficiency Virus Type 1 and Hepatitis C Virus (HCV): Testing, Product Disposition, and Donor Deferral and Reentry. Silver Spring, MD: FDA; 2010. [Google Scholar]
  • 10.Emmanuel JC, Bassett MT, Smith HJ, Jacobs JA. Pooling of sera for human immunodeficiency virus (HIV) testing: an economical method for use in developing countries. Journal of Clinical Pathology. 1988;41:582–5. doi: 10.1136/jcp.41.5.582. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Calafat AM, Needham LL, Kuklenyik Z, Reidy JA, Tully JS, Aguilar-Villalobos M, Naeher LP. Perfluorinated chemicals in selected residents of the American continent. Chemosphere. 2006;63:490–6. doi: 10.1016/j.chemosphere.2005.08.028. [DOI] [PubMed] [Google Scholar]
  • 12.Bates MN, Buckland SJ, Garrett N, Ellis H, Needham LL, Patterson DG, Jr, Turner WE, Russell DG. Persistent organochlorines in the serum of the non-occupationally exposed New Zealand population. Chemosphere. 2004;54:1431–43. doi: 10.1016/j.chemosphere.2003.09.040. [DOI] [PubMed] [Google Scholar]
  • 13.Kendziorski C, Irizarry RA, Chen KS, Haag JD, Gould MN. On the utility of pooling biological samples in microarray experiments. Proceedings of the National Academy of Sciences of the United States of America. 2005;102:4252–7. doi: 10.1073/pnas.0500607102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Saha-Chaudhuri P, Weinberg CR. Specimen pooling for efficient use of biospecimens in studies of time to a common event. American Journal of Epidemiology. 2013;178:126–35. doi: 10.1093/aje/kws442. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Zhang Z, Albert PS. Binary regression analysis with pooled exposure measurements: a regression calibration approach. Biometrics. 2011;67:636–45. doi: 10.1111/j.1541-0420.2010.01464.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Lyles RH, Tang L, Lin J, Zhang Z, Mukherjee B. Likelihood-based methods for regression analysis with binary exposure status assessed by pooling. Statistics in Medicine. 2012;31:2485–97. doi: 10.1002/sim.4426. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Limpert E, Stahel WA, Abbt M. Log-normal distributions across the sciences: keys and clues. BioScience. 2001;51:341–52. [Google Scholar]
  • 18.McBean EA, Rovers FA. Statistical Procedures for Analysis of Environmental Monitoring Data & Risk Assessment. Prentice Hall; New Jersey: 1998. pp. 71–82. [Google Scholar]
  • 19.Mitchell EM, Lyles RH, Manatunga AK, Schisterman EF. Semiparametric regression models for a right-skewed outcome subject to pooling. American Journal of Epidemiology. 2015;181:541–8. doi: 10.1093/aje/kwu301. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Mitchell EM, Lyles RH, Schisterman EF. Positing, fitting, and selecting regression models for pooled biomarker data. Statistics in Medicine. 2015;34:2544–58. doi: 10.1002/sim.6496. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Zou G. A modified poisson regression approach to prospective studies with binary data. American Journal of Epidemiology. 2004;159:702–6. doi: 10.1093/aje/kwh090. [DOI] [PubMed] [Google Scholar]
  • 22.White H. Maximum likelihood estimation of misspecified models. Econometrica. 1982;50:1–25. [Google Scholar]
  • 23.Faraggi D, Reiser B, Schisterman EF. ROC curve analysis for biomarkers based on pooled assessments. Statistics in Medicine. 2003;22:2515–27. doi: 10.1002/sim.1418. [DOI] [PubMed] [Google Scholar]
  • 24.Mitchell EM, Lyles RH, Manatunga AK, Perkins NJ, Schisterman EF. A highly efficient design strategy for regression with outcome pooling. Statistics in Medicine. 2014;33:5028–40. doi: 10.1002/sim.6305. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Whitcomb BW, Perkins NJ, Zhang Z, Ye A, Lyles RH. Assessment of skewed exposure in case-control studies with pooling. Statistics in Medicine. 2012;31:2461–2472. doi: 10.1002/sim.5351. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Schisterman EF, Vexler A. To pool or not to pool, from whether to when: applications of pooling to biospecimens subject to a limit of detection. Paediatric and Perinatal Epidemiology. 2008;22:486–96. doi: 10.1111/j.1365-3016.2008.00956.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Johnson NL, Kotz S, Balakrishnan N. Continuous Univariate Distributions. Vol. 2. Wiley; New York: 1995. pp. 210–75. [Google Scholar]
  • 28.Draper NR, Smith H. Applied regression analysis. 3. Wiley; New York: 1998. [Google Scholar]
  • 29.Dick EJ. Beyond ‘lognormal versus gamma’: discrimination among error distributions for generalized linear models. Fisheries Research. 2004;70:351–66. [Google Scholar]
  • 30.Dennis B, Patil GP. The gamma distribution and weighted multimodal gamma distributions as models of population abundance. Mathematical Biosciences. 1984;68:187–212. [Google Scholar]
  • 31.Rehm J, Kehoe T, Gmel G, Stinson F, Grant B, Gmel G. Statistical modeling of volume of alcohol exposure for epidemiological studies of population health: the US example. Population Health Metrics. 2010;8:1–12. doi: 10.1186/1478-7954-8-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Schisterman EF, Silver RM, Perkins NJ, Mumford SL, Whitcomb BW, Stanford JB, Lesher LL, Faraggi D, Wactawski-Wende J, Browne RW, Townsend JM, White M, Lynch AM, Galai N. A randomised trial to evaluate the effects of low-dose aspirin in gestation and reproduction: design and baseline characteristics. Paediatric and Perinatal Epidemiology. 2013;27:598–609. doi: 10.1111/ppe.12088. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Considine RV, Sinha MK, Heiman ML, Kriauciunas A, Stephens TW, Nyce MR, Ohannesian JP, Marco CC, McKee LJ, Bauer TL, et al. Serum immunoreactive-leptin concentrations in normal-weight and obese humans. New England Journal of Medicine. 1996;334:292–5. doi: 10.1056/NEJM199602013340503. [DOI] [PubMed] [Google Scholar]
  • 34.Ahrens K, Mumford SL, Schliep KC, Kissell KA, Perkins NJ, Wactawski-Wende J, Schisterman EF. Serum leptin levels and reproductive function during the menstrual cycle. American Journal of Obstetrics and Gynecology. 2014;210:248e1–9. doi: 10.1016/j.ajog.2013.11.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Herrid M, Palanisamy SK, Ciller UA, Fan R, Moens P, Smart NA, McFarlane JR. An updated view of leptin on implantation and pregnancy: a review. Physiological Research. 2014;63:543–557. doi: 10.33549/physiolres.932674. [DOI] [PubMed] [Google Scholar]
  • 36.Kelesidis T, Kelesidis I, Chou S, Mantzoros CS. Narrative review: the role of leptin in human physiology: emerging clinical applications. Annals of Internal Medicine. 2010;152:93–100. doi: 10.1059/0003-4819-152-2-201001190-00008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Vázquez MJ, Romero-Ruiz A, Tena-Sempere M. Roles of leptin in reproduction, pregnancy and polycystic ovary syndrome: consensus knowledge and recent developments. Metabolism. 2015;64:79–91. doi: 10.1016/j.metabol.2014.10.013. [DOI] [PubMed] [Google Scholar]
  • 38.Falorni A, Bini V, Molinari D, Papi F, Celi F, Di Stefano G, Berioli MG, Bacosi ML, Contessa G. Leptin serum levels in normal weight and obese children and adolescents: relationship with age, sex, pubertal development, body mass index and insulin. International Journal of Obesity. 1997;21:881–890. doi: 10.1038/sj.ijo.0800485. [DOI] [PubMed] [Google Scholar]
  • 39.Matkovic V, Ilich JZ, Skugor M, Badenhop NE, Goel P, Clairmont A, Klisovic D, Nahhas RW, Landoll JD. Leptin Is Inversely Related to Age at Menarche in Human Females. The Journal of Clinical Endocrinology and Metabolism. 1997;82:3239–45. doi: 10.1210/jcem.82.10.4280. [DOI] [PubMed] [Google Scholar]
  • 40.Wannamethee SG, Tchernova J, Whincup P, Lowe GD, Kelly A, Rumley A, Wallace AM, Sattar N. Plasma leptin: associations with metabolic, inflammatory and haemostatic risk factors for cardiovascular disease. Atherosclerosis. 2007;191:418–26. doi: 10.1016/j.atherosclerosis.2006.04.012. [DOI] [PubMed] [Google Scholar]
  • 41.Asferg C, Mogelvang R, Flyvbjerg A, Frystyk J, Jensen JS, Marott JL, Appleyard M, Jensen GB, Jeppesen J. Leptin, not adiponectin, predicts hypertension in the Copenhagen City Heart Study. American journal of hypertension. 2010;23:327–33. doi: 10.1038/ajh.2009.244. [DOI] [PubMed] [Google Scholar]
  • 42.Zuo H, Shi Z, Yuan B, Dai Y, Wu G, Hussain A. Association between serum leptin concentrations and insulin resistance: a population-based study from China. PLoS One. 2013;8:e54615. doi: 10.1371/journal.pone.0054615. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Joung KE, Park KH, Zaichenko L, Sahin-Efe A, Thakkar B, Brinkoetter M, Usher N, Warner D, Davis CR, Crowell JA, Mantzoros CS. Early life adversity is associated with elevated levels of circulating leptin, irisin, and decreased levels of adiponectin in midlife adults. The Journal of Clinical Endocrinology and Metabolism. 2014;99:E1055–60. doi: 10.1210/jc.2013-3669. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Milaneschi Y, Sutin AR, Terracciano A, Canepa M, Gravenstein KS, Egan JM, Vogelzangs N, Guralnik JM, Bandinelli S, Penninx BW, Ferrucci L. The association between leptin and depressive symptoms is modulated by abdominal adiposity. Psychoneuroendocrinology. 2014;30:1–10. doi: 10.1016/j.psyneuen.2013.12.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Smith W, Schutte R, Huisman HW, Van Rooyen JM, Ware LJ, Fourie CM, Mels CM, Kruger R, McCarthy N, Schutte AE. Leptin is positively associated with blood pressure in african men with a low body mass index: the SAfrEIC study. Endocrine Care. 2015;47:145–51. doi: 10.1055/s-0034-1389926. [DOI] [PubMed] [Google Scholar]
  • 46.Lyles RH, Mitchell EM. On efficient use of logistic regression to analyze exposure assay data on pooled biospecimens (Technical Report 13-02) Department of Biostatistics and Bioinformatics, Rollins School of Public Health; Atlanta, GA: 2013. [Google Scholar]
  • 47.Zhou H, Chen J, Rissanen TH, Korrick SA, Hu H, Salonen JT, Longnecker MP. Outcome-dependent sampling: an efficient sampling and inference procedure for studies with a continuous outcome. Epidemiology. 2007;18:461–8. doi: 10.1097/EDE.0b013e31806462d3. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES