Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2016 Apr 6;43(4):996–1018. doi: 10.1111/sjos.12225

Linear Increments with Non‐monotone Missing Data and Measurement Error

Shaun R Seaman 1,, Daniel Farewell 2, Ian R White 1
PMCID: PMC5111617  PMID: 27867251

Abstract

Linear increments (LI) are used to analyse repeated outcome data with missing values. Previously, two LI methods have been proposed, one allowing non‐monotone missingness but not independent measurement error and one allowing independent measurement error but only monotone missingness. In both, it was suggested that the expected increment could depend on current outcome. We show that LI can allow non‐monotone missingness and either independent measurement error of unknown variance or dependence of expected increment on current outcome but not both. A popular alternative to LI is a multivariate normal model ignoring the missingness pattern. This gives consistent estimation when data are normally distributed and missing at random (MAR). We clarify the relation between MAR and the assumptions of LI and show that for continuous outcomes multivariate normal estimators are also consistent under (non‐MAR and non‐normal) assumptions not much stronger than those of LI. Moreover, when missingness is non‐monotone, they are typically more efficient.

Keywords: ignorability, imputation, missing not at random, mortal cohort inference, non‐ignorable missing data, partly conditional inference

1. Introduction

Many medical studies involve repeated measurement of an outcome over time on a set of patients. Such longitudinal studies include clinical trials and observational cohort studies. Missing outcome data are a common feature of these studies, typically arising because patients miss scheduled visits or drop out of the study. Statistical methods for analysing such incomplete datasets include inverse probability weighting (Seaman & White, 2013), multiple imputation (Little & Rubin, 2002), random‐effects models (Verbeke & Molenberghs, 2000), shared random‐effects models (Henderson et al., 2000), doubly robust estimation (Seaman & Copas, 2009) and the focus of this article: linear increments (LI).

Linear increments were introduced by Diggle et al. (2007) (henceforth ‘DF&H’), who brought ideas from survival analysis into the analysis of longitudinal data. This has consequent advantages such as the ability to use martingale central limit theorems. LI was later developed by Aalen & Gunnes (2010) (henceforth ‘A&G’). Pullenayegum & Feldman (2013) extended the LI approach to allow for irregular observation times. Gunnes et al. (2009a) & Gunnes et al. (2009b) and Kingsley et al. (2012) describe applications of LI to randomized trials and observational studies.

DF&H assumed that the change in the underlying outcome (the ‘increment’) of a patient between two consecutive measurement times t − 1 and t is the sum of a predictable component and a martingale increment, and that these underlying outcomes are then measured with mutually independent errors that are independent of all the underlying outcomes and whose variances are unknown (‘independent measurement error’). A linear model is specified for the dependence of the predictable component on covariates, and the parameters of this model are estimated using a series of linear regressions. The expected outcome at each time t is then estimated by summing the predictable components up to time t and averaging over the individuals, a method known as estimating the compensator. DF&H's basic model assumes that the increment between times t − 1 and t is independent of the underlying outcome at time t − 1. However, they suggested that this assumption could be weakened by including the measured outcome (or function thereof) at time t − 1 as a covariate in the model for the predictable component of the increment. DF&H's method was limited to monotone missing data, that is, data where if a patient's outcome is missing at time t then it is also missing at all later times.

A&G allowed for multivariate outcomes and non‐monotone missing data and explicitly modelled the increment between times t − 1 and t as a function of the outcome at time t − 1. They also treated the special situation where some missing outcomes result from patients dying, and, as an alternative to estimating the compensator, introduced an imputation method (which we call ‘LI‐LS imputation’), which is particularly suitable when some patients die. Unlike DF&H's model, A&G's model does not explicitly allow for independent measurement error.

Many of the commonly used methods for handling missing data in longitudinal studies assume data are missing at random (MAR), that is, the probability an outcome is observed depends only on observed data (see Seaman et al. (2013) for a formal definition). Among these methods is the multivariate normal (MVN) model fitted using maximum likelihood (Schafer, 1997). The LI approach was originally conceived as a computationally simple way to handle data that are missing not at random (MNAR), that is, where the probability an outcome is observed can depend on the unobserved data in a particular way. It involves the ‘discrete‐time independent censoring (DTIC)’ assumption that the expected increment between times t − 1 and t given the underlying outcomes up to time t − 1 and the missingness pattern in the increments up to time t does not depend on that missingness pattern. LI–LS imputation additionally requires another, stronger DTIC assumption and an ‘independent return’ assumption, that is, an assumption about the probability that a patient with missing outcome at time t will be observed again at a later time point. The relation between DTIC and MAR has not been investigated in depth.

In this article, we have several aims. In Section 2, we show how DF&H's LI model can be used with non‐monotone missing data. We demonstrate that, contrary to DF&H's suggestion, this model cannot simultaneously allow for independent measurement error and dependence of the expected increment on the previous outcome. DF&H allowed for independent measurement error; A&G allowed for dependence on the previous outcome. In Section 3, we clarify the relation between the DTIC, MAR and independent return assumptions. In Section 4, we prove that under specific (MNAR) assumptions, which for a continuous outcome are not much stronger than those required by LI–LS imputation, fitting the MVN model ignoring the missingness mechanism gives consistent estimation. Section 5 describes how the LI model can allow for dependence of the expected increment on outcomes prior to the previous one. In Section 6, we demonstrate using simulation studies that the MVN approach can be more efficient than estimating the compensator and LI–LS imputation when data are non‐monotone missing. In Section 7, we identify that using imputation to treat the situation where some patients die requires an assumption not mentioned by A&G. Finally, Section 8 describes an application of the various methods to data from the British household panel survey (BHPS).

2. The linear increments method

2.1. The LI model

Let Y it denotes a vector of length m representing an underlying outcome for individual i at time t (i = 1,…,N; t = 1,…,T). The corresponding measured outcome is Y it+e it, where e it is a measurement error. When the outcome is observed for individual i at time t, it is Y it+e it that is observed; Y it itself is never observed (except when e it=0). Let X i be a vector of p fully‐observed baseline covariates for individual i. The change (or ‘increment’), ΔY it=Y itY i,t − 1, in the underlying outcome between times t − 1 and t is assumed to be related to the underlying outcome at time t − 1 and the covariates by

ΔYit=αt+(βtI)Yi,t1+γtXi+εit(t=2,,T) (1)

where β t and γ t are, respectively, m × m and p × m matrices of parameters, I is the identity matrix, and ε it is a random vector of length m satisfying E(εitFi,t1)=0, with Fit={Xi,Yi1,,Yit} denoting the history of individual i's covariates and underlying outcomes up to time t. We assume ε it and e it have finite variance. An important special case of Equation (1) is where β t=I: then the expectation of ΔY it does not depend on Y i,t − 1. Equation (1) can be written equivalently as

Yit=αt+βtYi,t1+γtXi+εit(t=2,,T). (2)

Like DF&H, we assume that the measurement error process {e it:t = 1,…,T} is independent of all other processes (i.e. e i1,…,e iT are independent of X i,Y i1,ε i2,…,ε iT), that e it is independent of e is for all ts, and that E(e it) = 0. The underlying outcome processes, measurement error processes and missingness processes (described in Section 2.2) of different individuals are assumed independent (given X i,Y i1).

Some elements of β t and/or γ t may be constrained. For example, if baseline outcome Y i1+e i1 is included in X i, then β 2 can be constrained to equal I for identifiability. In much of this article, we shall consider two special cases of Equation (2). The first, which we call the case of ‘no independent measurement error’ (or just ‘no measurement error’) is where e it=0i,t. Strictly speaking, we mean by ‘no measurement error’ that there are no mutually independent measurement errors that are independent of X i, Y i1, ε i2, …, ε iT and whose variance is unknown (see Section 9 for case of known variance). The second case is where β t=It.

2.2. Missing data

Let R 0,it=1 if Y it+e it is observed and R 0,it=0 if Y it+e it is missing. Let R it=1 if R 0,it=R 0,it − 1=1 and R it=0 otherwise. Hence, R it indicates whether the measured increment (Y it+e it) − (Y i,t − 1+e i,t − 1) is observed, and R it=1 implies R 0,it=1. Let R0,it=(R0,i1,,R0,it) and Rit=(Ri1,,Rit) denote missingness histories up to and including time t. We assume R 0,i1=1∀i. The assumption that the measurement error process is independent of all other processes means that R 0,i1,…,R 0,iT are independent of e i1,…,e iT. We say that the ‘data are monotone missing’ if P(R 0,it + 1=1∣R 0,it=0) = 0 for all i (i.e. we mean that the data‐generating mechanism always generates monotone missing data). Let G it denote the vector consisting of X i and those elements of Y i1,…,Y it whose corresponding elements of R 0,it equal one. So, G it consists of the baseline covariates and the underlying outcomes at the times at which the measured (with error) outcome is observed. For example, if R0,i4=(1,1,0,1), then Gi4=(Yi1,Yi2,Yi4,Xi). Henceforth, we omit the i index, unless doing so causes ambiguity.

DF&H define DTIC as E(ΔYtFt1,e1,,et1,Rt)=E(ΔYtFt1,,e1,,et1) for all t. Because it is assumed that the measurement error process is independent of all other processes, this definition reduces to

E(ΔYtFt1,Rt)=E(ΔYtFt1)t. (3)

Note that Equation (3) can be equivalently written as E(YtFt1,Rt)=E(YtFt1)t. We discuss the relation between DTIC and MAR in Section 3. A&G give two versions of DTIC: Equation (3) and

E(YtFt1,R0,t)=E(YtFt1)t. (4)

For convenience, we call (3) and ((4)) ‘weak DTIC’ and ‘strong DTIC’, respectively. A&G point out that strong DTIC implies weak DTIC, and that when data are monotone missing, weak DTIC implies strong DTIC (i.e. they are equivalent).

2.3. Parameter estimation

Consider the model

E(Yt+et)=αtls+(βtls)(Yt1+et1)+(γtls)X. (5)

In general, αtls, βtls and γtls differ from the parameters of interest α t, β t and γ t in Equation (2). The former relates measured outcomes, the latter underlying outcomes. Let α^tls, β^tls and γ^tls denote the least‐squares estimators of αtls, βtls and γtls obtained by fitting model ((5)) to the set of patients with R t=1. DF&H offered a proof (Section 4.2) that α^tls, β^tls and γ^tls are unbiased consistent estimators of α t, β t and γ t in Equation (2) when weak DTIC holds, data are monotone missing and m = 1. However, this proof implicitly assumes that either (i) β t=1 and β^tls is fixed at 1 before calculating the least‐squares estimator of (αtls,γtls), or (ii) E{(Y t − 1+e t − 1)e t − 1}=0. Because (ii) is not true unless Var(e t − 1) = 0, this proof is not valid unless either (i) the expectation of the increment ΔY it does not depend on Y t − 1 (i.e. β t=1) and β^tls is fixed at the true value of β t or (ii) there is no measurement error. This lack of validity is demonstrated by Example 1 later. A&G also proved that α^tls, β^tls and γ^tls are unbiased estimators of α t, β t and γ t when weak DTIC holds and there is no measurement error. Their proof allows m > 1 and data to be non‐monotone missing.

The following theorem and corollary show that if weak DTIC holds and β^tls is fixed at the true value of β t, then (α^tls,γ^tls) is an unbiased consistent estimator of (α t,γ t) even if missingness is non‐monotone and/or there is measurement error. As explained later, this theorem is of most practical interest when the true value of β t is I. The theorem is followed by an example which illustrates that the least‐squares estimators may be neither unbiased nor consistent estimators of the parameters of interest when there is measurement error unless β^tls is fixed at the true value of β t.

Theorem 1

If the increments model of Equation (2) and the weak DTIC assumption of Equation (3) hold and β^tls is fixed at the true value of β t, then E{(α^tls,γ^tls)}=(αt,γt) and (α^tls,γ^tls)(αt,γt) as N.

Proofs of theorems are given in Appendix S1.

Example 1

m = 1, T = 2, Y 1∼Normal(0,1), e 1∼Normal(0,1), e 2=0 and Y 2=Y 1. Note that this is a special case of Equation (2) in which α 2=ε 2=0, β 2=1, and there is no covariate X. It can be shown that α2ls=0 and β2ls=0.5. Thus, even if there are no missing data, α^2ls and β^2ls are not unbiased or consistent estimators of α 2 and β 2. Moreover, when there is missing data, they may not even be unbiased or consistent estimators of α2ls and β2ls. For example, suppose that R 02=1 if and only if Y 1≥0. It can be shown that if Y 2+e 2=Y 2 is regressed on Y 1+e 1 using only those individuals with R 02=1, then α^2ls and β^2ls converge to 0.585 and 0.267, respectively, as N. Hence, β^2ls is a consistent estimator neither of β 2=1 nor of β2ls=0.5. Note that if β^2ls is fixed at β 2=1, then Theorem 1 implies that α^2ls converges to α 2=0.

In conclusion, the least‐squares estimators are unbiased for the parameters of the LI model if weak DTIC holds and either (i) β^tls is fixed at the true value of β t, or (ii) there is no measurement error. Otherwise, they may not be. Henceforth, the only value at which we consider fixing β^tls is I. This is the case of most practical interest: one could fix β^tls=I if one believed that expected increment E(ΔYtFt1) did not depend on Y t − 1 (i.e. β t=I). Otherwise, one could treat β t as unknown and estimate it. It seems unlikely one would wish to fix β t at a value other than I.

2.4. Estimating the compensator

From Equation (2), E(YtX,Y1)=αt+βtE(Yt1X,Y1)+γtX, and so

E(YtX,Y1)=αt+γtX+j=2t1k=j+1tβk(αj+γjX)+j=2tβjY1 (6)

where j=abβj means βbβb1βa. Thus, E(Y itX i,Y i1) can be estimated for each individual i in the dataset by replacing α t, β t and γ t in Equation (6) with α^tls, β^tls and γ^tls. Denote this estimate as Yitcpr. This is the estimating the compensator method. The overall mean E(Y t) can then be estimated as N1i=1NYitcpr. If (α^tls,β^tls,γ^tls) is an unbiased consistent estimator of (α t,β t,γ t), then this estimate of E(Y t) will also be unbiased and consistent.

Returning to Example 1, if α^2ls=0.585 and β^2ls=0.267 are used in Equation (6), then E(Y 2) is calculated to equal α^2ls=0.585, whereas its true value is zero.

2.5. LI–LS imputation

A&G propose an alternative to estimating the compensator when there is no measurement error. We call this method ‘LI‐LS imputation’, because it is based on the LI model of Equation (2) and the least‐squares (LS) estimators α^tls,β^tls,γ^tls of α t, β t and γ t. (Later, we shall introduce other LI imputation methods that differ from LI‐LS imputation only in that they use alternative estimators of α t, β t and γ t.) LI–LS imputation uses an actual outcome when it is observed and imputes a missing outcome as the most recently observed outcome updated by the expected increments. That is, letting Ytest denotes the value of Y t in the imputed dataset, we have Ytest=Yt if R 0t=1 and Ytest=α^tls+(β^tls)Yt1est+(γ^tls)X if R 0t=0.

A&G show that if Equation (2) holds, there is no measurement error, strong DTIC holds and

E{R0,t(Yt1estYt1)X,Y1}=0 (7)

then E(YtestY1,X)=E(YtY1,X). It follows that N1i=1NYitest is an unbiased estimator of E(Y t) and that if a linear regression model for Y with any or all of X and t as covariates is fitted to the imputed dataset {Yitest:i=1,,N;t=1,,T} using least squares, then the parameter estimators of this model are consistent.

The following theorem shows that LI–LS imputation can also be used when there is measurement error, provided that β t=I and β^tls is fixed at I.

Theorem 2

Suppose that the increments model of Equation (2), the strong DTIC assumption of Equation ((4)) and Equation (7) hold, that β t=I, and that β^tls is fixed at I. Then, E(YtestY1,X)=E(YtY1,X).

If the conditions of Theorem 2 are satisfied, then (i) N1i=1NYitest is an unbiased estimator of E(Y t) and (ii) if a linear regression model for Y with any or all of X and t as covariates is fitted to the imputed dataset using least squares, the parameter estimators are asymptotically unbiased.

Returning to Example 1, it can be shown that N1i=1NYi2est tends to 0.585 as N, and so is an inconsistent estimator of E(Y 2) = 0.

A&G say that Equation (7) will hold if

P(R0,t=1R0,k1,R0,k=1,R0,k+1=R0,k+2==R0,t1=0,X,Yk,Yk+1,,Yt)=P(R0,t=1R0,k1,R0,k=1,R0,k+1=R0,k+2==R0,t1=0,X,Yk) (8)

for all kt − 2, and imply that it is unlikely to hold otherwise. Equation (8) can be interpreted as meaning that the conditional probability of return at time t given dropout immediately after time k, no return between times k and t, the baseline covariates X, the most recently observed outcome Y k and subsequent outcomes does not depend on these subsequent outcomes (or, more informally, as ‘return after a dropout is independent of outcomes occurring since that dropout’). We shall call Equation (8) the ‘independent return’ assumption.

2.6. Estimating the compensator versus LI–LS imputation

A&G do not discuss the relative advantages and disadvantages of the two methods in detail but do say that LI–LS imputation can be used to give mortal‐cohort inference (as well as immortal‐cohort inference), whereas estimating the compensator gives immortal‐cohort inference only. We postpone consideration of mortal‐cohort inference until Section 7. The advantage of estimating the compensator is that it relies on fewer assumptions than LI–LS imputation. Notably, it does not make assumptions about the probability of return after dropout. The disadvantages are that it may be less efficient and less robust to violation of those assumptions than LI–LS imputation. LI–LS imputation may be more efficient because, unlike estimating the compensator, it uses Y t+e t values with R 0,t=1 but R 0,t − 1=0 (so that R t=0). It may be more robust to violation of the weak DTIC assumption because it relies on DTIC only to impute missing values. For example, if Y 3+e 3 is fully observed and Y 2+e 2 is missing whenever Y 2<0, then the estimate of E(Y 3) from LI–LS imputation will be unbiased, whereas that from estimating the compensator will be biased. Note that when data are monotone missing, the two estimators N1i=1NYitcpr and N1i=1NYitest of E(Y t) are equal (Aalen & Gunnes, 2010).

3. Relation between missingness assumptions

In this section, we consider the relation between DTIC and MAR and then the implications for the dropout and return processes of assuming DTIC.

Weak and strong DTIC are assumptions only about expectations. MAR, on the other hand, is an assumption about whole distributions. A slightly stronger assumption than strong DTIC, which is about whole distributions, is

f(YtFt1,R0,t)=f(YtFt1)t. (9)

We call Equation (9) ‘dDTIC’ (‘DTIC in distribution’). In practice, it seems likely that in most cases when strong DTIC holds, dDTIC will also hold. dDTIC can be written equivalently as

P(R0,t=1R0,t1,FT)=P(R0,t=1R0,t1,Ft1)t. (10)

(see Appendix S2). Thus, the hazards of dropout and return are not allowed to depend on future underlying outcomes.

Missing at random means that the probability the outcome is missing at time t can depend on covariates X and observed outcomes, including future observed outcomes, but not on missing outcomes (i.e. those at times t where R 0t=0). When there is measurement error, missingness depends on the outcomes observed with measurement error, not on the underlying outcomes. Conversely, the dDTIC assumption, combined with the assumed independence of the measurement error and missingness processes, allows the probability of missingness to depend on all past underlying outcomes (including those at times t where R 0t=0) but not on future underlying outcomes or the outcomes measured with error. When data are monotone missing and there is no measurement error, dDTIC is equivalent to MAR (and also to ‘sequential MAR’ — Hogan et al. (2004); Diggle et al. (2007)).

A simple example illustrates that estimating the compensator can give inconsistent estimation when there is measurement error and dropout depends on it. Suppose that m = 1 and T = 2, that Y 1=Y 2=0, that e 1 and e 2 are independently distributed Normal(0,1), that there are no baseline covariates X, that R 02=1 if and only if Y 1+e 1>0, and that β^2ls is constrained to equal 1. Whereas α 2=0, the least‐squares estimator α^2ls has negative expectation. Note that these data are MAR.

The following example illustrates that dDTIC and independent return can hold without MAR holding. In this example, the probability that an individual drops out at time t = 4 depends on the outcome at t = 2, which may be missing.

Example 2

m = 1 and T = 4, α t=0 and β t=1 for all t, there is no covariate X or measurement error, P(R 02=1∣Y 1,Y 2,Y 3,Y 4) = 0.5, R 03=1 always, and R 04=1 if and only if Y 2>0.

If, in addition to Equation (10), it is assumed that

P(R0,t=1R0,k1,R0,k=1,R0,k+1==R0,t1=0,Ft1)=P(R0,t=1R0,k1,R0,k=1,R0,k+1==R0,t1=0,Fk)k (11)

then the hazard of return after dropout cannot depend on any of the underlying outcomes after the most recent dropout. The following theorem shows that Equation (11) is sufficient but not necessary for independent return to hold.

Theorem 3

Let B k and C k be subvectors of (X,Y1,,Yk) such that both contain (X,Yk) and C k is a subvector of B k. If the increments model of Equation (2) and the dDTIC assumption of Equation (9) hold, then

P(R0,t=1R0,k1,R0,k=1,R0,k+1==R0,t1=0,Bk,Yk+1,,YT)=P(R0,t=1R0,k1,R0,k=1,R0,k+1==R0,t1=0,Bk),

implies

P(R0,t=1R0,k1,R0,k=1,R0,k+1==R0,t1=0,Ck,Yk+1,,YT)=P(R0,t=1R0,k1,R0,k=1,R0,k+1==R0,t1=0,Ck).

The converse is not true in general.

Theorem 3 also implies that the following assumption, which will be important in Section 4.1, is stronger than independent return but weaker than Equation (11):

P(R0,t=1R0,k1,R0,k=1,R0,k+1=R0,k+2==R0,t1=0,Gk,Yk+1,Yk+2,,Yt)=P(R0,t=1R0,k1,R0,k=1,R0,k+1=R0,k+2==R0,t1=0,Gk) (12)

for all kt − 2. We call Equation (12) the ‘strong independent return’ assumption. In the proof of Theorem 3, we give an example where independent return holds but strong independent return does not. However, in most real‐world situations, we believe it is unlikely that independent return would hold without strong independent return also holding. Note that the word ‘strong’ in ‘strong independent return’ is meant in a different sense from that in ‘strong DTIC’.

4. Use of MVN model assuming ignorability

A popular method for handling missing repeated outcome data is to assume that outcomes and covariates are normally distributed and calculate the maximum likelihood estimates of the mean and variance of this normal distribution ignoring the missingness mechanism and imposing no structure on the mean and variance (Schafer, 1997). We call this the ‘unstructured MVN’ method. These maximum likelihood estimates may be of interest in themselves, or they can be used to impute missing values. This approach yields consistent estimates when data are MAR and the MVN model is correctly specified (Seaman et al., 2013). (Indeed, under these conditions, the missingness mechanism is ‘ignorable’ in the following sense. The unstructured MVN method yields the same maximum likelihood estimates and imputed values as would be obtained if the MVN model were fitted jointly with any missingness model—that is, model for the missingness pattern given the outcomes and covariates—that assumes MAR and has parameters distinct from those of the MVN model (Seaman et al., 2013)). As we have noted, the assumptions used in estimating the compensator (weak DTIC) and LI–LS imputation (strong DTIC and Equation (7)) do not imply the data are MAR or normally distributed. Nevertheless, we now show (Section 4.1) that when there is no measurement error and slightly stronger assumptions than those of LI–LS imputation are satisfied, the unstructured MVN method yields consistent estimates even when MAR does not hold and the data are not normally distributed. We then show (Section 4.2) that a more efficient estimator can be obtained by constraining the variance matrix of the MVN distribution. We call this the ‘autoregressive MVN’ method. Finally, we show (Section 4.3) that if a further constraint is applied to the variance matrix, the resulting method provides consistent estimates even when there is measurement error, provided that β t=I for all t. We call this the ‘random‐walk MVN’ method. These MVN methods are typically more efficient than estimating the compensator and LI–LS imputation.

4.1. Unstructured MVN and no measurement error

We assume throughout Sections 4.1 and 4.2 that there is no measurement error, that is, e t=0. Let μ=E{(Y1,,YT,X)} and Σ=Var{(Y1,,YT,X)}. Let μ t denotes the subvector of μ corresponding to Y t (t = 1,…,T + 1), where Y T + 1 means X. Similarly, let Σ s,t denotes the submatrix of Σ corresponding to (Y s,Y t) (1≤s,tT + 1). So, for example, μ 3=E(Y 3), μ T + 1=E(X), Σ 3,T + 1=Cov(Y 3,X) and Σ T + 1,T + 1=Var(X).

Let μ^ and Σ^ denote the maximum likelihood estimates obtained by fitting the model (Y1,,YT,X)Normal(μ,Σ) with unstructured μ and Σ to the observed data and ignoring the missingness mechanism. That is, μ^ and Σ^ are the maximum likelihood estimates from the model GiTNormal(μR0,iT,ΣR0,iT,R0,iT)(i=1,,N), where μR0,iT and ΣR0,iT,R0,iT denote the subvector of μ and submatrix of Σ corresponding to G iT. This model can be fitted using an EM algorithm (e.g. using the norm package in R) (Schafer, 1997).

Theorem 4

If the increments model of Equation (2), the dDTIC assumption of Equation (9) and the strong independent return assumption of Equation (12) hold and Var(εtFt1)=Var(εt) for all t, then (μ^,Σ^) is a consistent estimator of (μ,Σ).

Theorem 4 may be surprising, because there is no assumption the data are MAR (e.g. Example 2 satisfies the conditions of Theorem 4, but the data are not MAR) nor that they are normally distributed. Note that Theorem 4 is not saying the missingness process is ‘ignorable’. In Example 2, missingness is not ignorable, because information about a missing value of Y 2 is gained by knowing whether Y 4 is observed. The intuition behind Theorem 4 is that although dDTIC and strong independent return allow dropout and return to depend on earlier unobserved outcomes, this dependence does not matter, because the autoregressive structure of the data means that the later outcomes are conditionally independent of these earlier outcomes given outcomes that are observed. Consider Example 2. Although whether or not Y 4 is observed depends on Y 2, which may be missing, Y 4 is independent of Y 2 given Y 3, which is observed. This results in observed outcomes being in an important sense representative of all the outcomes, as formalized in the following theorem.

Theorem 5

If the increments model of Equation (2), the dDTIC assumption of Equation (9) and the strong independent return assumption of Equation (12) hold, then for k < t,

a)f(YtR0,t2,R0,t1=R0,t=1,Gt1)=f(YtX,Yt1)b)f(YtR0,k1,R0,k=1,R0,k+1==R0,t1=0,R0,t=1,Gk) (13)
=f(YtX,Yk) (14)

If Equation (8) holds instead of Equation (12), then Equations ((13)) and ((14)) still hold but with G t − 1 and G k replaced by, respectively, (X,Y t − 1) and (X,Y k).

Equation ((13)) says that the conditional distribution of the outcome at time t given the previous observed outcomes in individuals who are observed at times t and t − 1 is the same as the distribution in the whole sample. Equation ((14)) says that the same is true of individuals returning after dropout. Because the joint distribution of the observed part of (Y1,,YT,X) can be factorized as the product of these conditional distributions, the parameters (μ,Σ) of the joint distribution can be consistently estimated (see proof of Theorem 4 for more details).

Having obtained estimates μ^ and Σ^, the corresponding estimates of α t, β t and γ t can be calculated as follows, and these used for LI imputation in place of the least‐squares estimates α^tls, β^tls, and γ^tls. We call this method ‘LI‐uMVN imputation’. Let

β^t=Σ^t1,t1Σ^t1,T+1Σ^T+1,T+11Σ^T+1,t11Σ^t1,tΣ^t1,T+1Σ^T+1,T+11Σ^T+1,t (15)
γ^t=Σ^T+1,T+11Σ^T+1,tΣ^T+1,T+11Σ^T+1,t1β^t (16)
α^t=μ^tβ^tμ^t1γ^μ^T+1 (17)

where μ^t denotes the tth element of μ^ and Σ^s,t denotes the (s,t)th element of Σ^(s,t=1,,T+1, with the (T + 1)th element corresponding to X).

Theorem 6

If μ^ and Σ^ are consistent estimators of μ and Σ, then α^t,β^t and γ^t are consistent estimators of α t,β t and γ t.

LI–uMVN imputation might be expected to be more efficient than estimating the compensator and LI–LS imputation, because the latter methods estimate α t, β t and γ t using only those observed outcomes Y t with R 0,t − 1=1, whereas the MVN method uses all observed outcomes.

If the motivation for using LI imputation is to enable a linear regression model for Y with any or all of X and t as covariates to be fitted to the imputed dataset, then an even more efficient alternative is available. It can be seen that the maximum likelihood estimates of the parameters in this linear regression are functions of μ^ and Σ^. A computationally convenient way to calculate these functions is to impute each missing Y t value as its conditional expectation given G iT (given by Equation ((18)) with μ and Σ replaced by μ^ and Σ^) and then fit the linear regression to the imputed data. We call this method ‘uMVN imputation’. More details and justification are given in Appendix S6. Unlike in LI imputation, where a missing value of Y t is imputed using only the individual's observed past (G t − 1), in uMVN imputation, missing values are imputed using his or her observed past and future (G T), and hence uMVN imputation may be more efficient when the data are non‐monotone missing. The conditional expectation of Y t given G iT is

μt+Σt,R0,iTΣR0,iT,R0,iT1GiTμR0,iT (18)

where Σt,R0,iT is the submatrix of Σ composed of the m rows corresponding to Y t and the columns corresponding to G iT.

So far, we have assumed there are baseline covariates in the model and that Y 1 is not treated as one of these covariates. If there are no baseline covariates, then γ t, X, Σ T + 1,t, Σ t,T + 1 and Σ T + 1,T + 1 should be omitted from all expressions in this article. If, on the other hand, Y 1 is included in X, then β 2 should be set equal to zero to ensure parameter identifiability.

Theorem 4 (and hence LI‐uMVN imputation and uMVN imputation) requires the assumption that the variance of ε t does not depend on the history Ft1. This is not required by the estimating the compensator and LI–LS imputation methods. When Y t is a continuous variable, one might be content to make this assumption. However, the LI approach can also be used for categorical outcomes. For example, A&G consider a Markov chain with q states. They define Y t to be a vector of q − 1 indicator variables for the state occupied at time t and show that α^tls, β^tls and γ^tls are closely related to the Aalen–Johansen estimator of the transition matrix. In this case, each element of Y tY t − 1 equals 0, 1 or −1, and the variance of ε t depends on Y t − 1. So, Var(εtFt1)=Var(εt) is not true. Even for a continuous outcome Y t, it may not be true if Y t is bounded above or below.

4.2. Autoregressive MVN and no measurement error

Although the unstructured MVN method may be expected often to be more efficient than estimating the compensator or LI–LS imputation, this is not guaranteed, because unlike those methods it does not exploit the autoregressive assumption in Equation (2). As shown in Appendix S3, Equation (2) implies the following constraint on Σ: for 1≤s<tT,

Σs,t=Σs,T+1ΣT+1,T+11ΣT+1,t+βtβt1βs+1Σs,sΣs,T+1ΣT+1,T+11ΣT+1,s (19)

The model defined by Y1,,YT,XNormal(μ,Σ) with the constraint of Equation (19) can be fitted by maximum likelihood to the observed data ignoring the missingness mechanism (an EM algorithm is described in Appendix S5). We call this the ‘autoregressive MVN’ method, and we call imputation using formula (18) with these estimates of μ and Σ ‘aMVN imputation’. The corresponding estimates of β t, γ t and α t are given by Equations (15)(17), and we call LI imputation using these estimates ‘LS ‐aMVN imputation’. These autoregressive MVN methods might be expected to be more efficient than the corresponding unstructured MVN methods.

Theorem 4 holds also for the autoregressive MVN method (proof in Appendix S1). Furthermore, the strong independent return assumption in that theorem can be replaced by the weaker independent return assumption. The following theorem shows a relation between autoregressive MVN and the methods of A&G.

Theorem 7

When data are monotone missing, the estimates of α t, β t and γ t obtained by using autoregressive MVN followed by Equations (15)(17) are equal to α^tls, β^tls and γ^tls. Furthermore, LI–LS imputation, LI–aMVN imputation and aMVN imputation yield identical imputed values.

4.3. Random‐walk MVN and measurement error

Suppose now that there is measurement error e t. Equation (2) implies that the underlying outcome Y t is conditionally independent of Y s for s < t − 1 given Y t − 1 and X, but it does not imply that the outcomes Y t+e t measured with error also have this autoregressive structure. Nevertheless, we show in Appendix S4 that when β t=I for all t, fitting the MVN model imposing the constraint of Equation (19) with β t=I (‘random‐walk MVN’ or ‘rMVN’) to the observed outcomes measured with error and ignoring the missingness mechanism gives consistent estimation of the parameters of Equation (2). The resulting LI–rMVN imputation and rMVN imputation methods can be more efficient than LI–LS imputation when data are non‐monotone missing.

5. Extension to higher‐order autoregression

Equation (2) implies that Y t is conditionally independent of Y 1,…,Y t − 2 given Y t − 1. So, if it is assumed that there is no measurement error (e t=0), then the measured outcomes are assumed to follow a first‐order autoregressive process. This can be quite a restrictive assumption. When measurement error is allowed, Equation (2) implies that the measured outcome Y t+e t can, in general, depend on all previous measured outcomes, thus allowing more flexibility. Unfortunately, as shown in Section 2.3, LI methods that allow for measurement error require the restrictive assumption that β t=I.

One way to allow an outcome measured at time t to depend on more than just the outcome measured at time t − 1, while still allowing β tI, is to define Y t to include the outcome (or outcomes) measured at both times t and t − 1 and set e t=0. This possibility was not explicitly mentioned by A&G but was mentioned in earlier work by the same authors (Gunnes et al., 2009a). There they considered only monotone missing data. We now describe how this approach could be used with monotone or non‐monotone missing data.

To avoid confusion, denote the outcome (or outcomes) of interest measured at time t as Z t, define Y t in Equation (2) as Yt=(Zt,Zt1) for t > 1 and Y 1=Z 1, and let e t=0 for all t. Equation (2) now allows for second‐order autoregression in Z t.

Note that, by definition, R 0,t=1 if and only if all elements of Y t are observed. (This was illustrated by A&G, who described an analysis where Y t included three quality‐of‐life outcomes all measured at time t; here R 0,t=1 only when all three were observed.) When Yt=(Zt,Zt1), R 0,t=1 requires that both Z t and Z t − 1 be observed. Because the two methods described by A&G only use values of Y it for which R 0i,t=1, this means that if Z it is observed but Z i,t − 1 and Z i,t + 1 are both missing, then Z it is treated as missing and individual i is regarded as not having returned at time t. This is also true of the uMVN and aMVN methods. Furthermore, R t=1 if and only if Z t, Z t − 1 and Z t − 2 are all observed. This means that if Z it and Z i,t + 1 are observed but neither Z i,t − 1 nor Z i,t + 2 are, then A&G's methods do not use Z it or Z i,t + 1 in the estimation of the parameters of Equation (2) (unlike the MVN methods).

Application of A&G's methods is now straightforward, except that it is not immediately obvious how LI–LS imputation should deal with an observed value of Z it when Z i,t − 1 is missing and Z i,t + 1 is observed. In this case, R 0i,t=0 and R 0i,t + 1=1. Because R 0i,t=0, LI–LS imputation involves imputing Y it (and hence Z it). However, because R 0i,t + 1=1, Y i,t + 1 (and hence Z it) is observed. Thus, we have both an observed and an imputed value of Z it. Fortunately, Theorem 5 implies that the conditional distributions of the observed and imputed values of Z it given the most recently observed value of Y ik (k < t) are the same. Therefore, it is valid to use either the observed or imputed value. We recommend using the former, for reasons of efficiency and robustness to possible misspecification of Equation (2).

Application of uMVN imputation is also straightforward. First, as with A&G's methods, any observed values of Z it for which Z i,t − 1 and Z i,t + 1 are both missing are deleted. One could then fit the MVN model to Y1,,YT,X. However, because the second element of Y t + 1 and first element of Y t are equal by definition (both equal Z t), there is no need to include both. Instead, one can just fit the model (Z1,,ZT,X)Normal(μZ,ΣZ). Missing values of Z t would then be imputed as their conditional expectations given the observed data, as in Equation ((18)). LI–uMVN imputation would instead involve calculating βt^, γt^ and αt^ from the maximum likelihood estimates of μ Z and Σ Z, as in Equations (15)(17), and then using these to impute in the same way as in LI–LS imputation. LI–aMVN imputation and aMVN imputation could be adapted to allow for second‐order autoregression by imposing a constraint analogous to Equation (19) on Σ.

Extension to third‐ (or higher‐) order autoregression is possible, by defining Y t=(Z t,Z t − 1,Z t − 2). However, the more elements Y t contains, the greater is the risk that one will be missing, causing more observed Z t values to be treated as missing.

6. Simulation studies

The following simple simulation studies are intended to demonstrate unbiasedness of the methods investigated in this paper when their assumptions are satisfied and biasedness when they are not, and to investigate their relative efficiencies. As indicated in Sections 2.6 and 4, estimating the compensator, LI–LS imputation, LI–aMVN imputation and aMVN imputation yield identical estimates when data are monotone missing. So, our simulation study scenarios were chosen to have high rates of dropout and return, in order to investigate how large the differences in efficiency can be when missingness is far from monotone. In most realistic situations, missingness would be non‐monotone but with lower rates of return, and so, the efficiency gains available there would be less than those demonstrated here. In the studies described in Sections 6.1 and 6.2, there is no measurement error. Appendix S8 includes a further study (Study 3) with measurement error and β t=I. Its results demonstrate that methods that estimate β t, rather than constraining it to equal its true value, are biased when there is measurement error, and also that, when data are non‐monotone missing, LI–rMVN imputation and rMVN imputation can be more efficient than LI–LS imputation with β t constrained as β t=I.

6.1. Study 1: first‐order autoregression

Data were generated from the following model. Let X ∼ Uniform(0,1), Y 1∼Normal(0.5X,1), Y t=0.4 + β t Y t − 1+0.5X + ε t(t = 2,…6), ε t∼Normal(1,1) with probability 0.5 and ε t∼Normal(−1,1) otherwise, R 1=1, logit{P(R0,t=1R0,t1=1,R0,t2,FT)}=ωt+X+(Yt1+Yt2)/2, and logit{P(R0,t=1R0,t1=0,R0,t2,FT)}=ϕt+X+(YLt+YLt1)/2, where β t=1.2 for all t, L t=argmaxj{j < t and R 0,j=1} (t≥2) denotes the most recent time at which the outcome was observed prior to time t, L 1=1 and Y 0=0. The values of ω t and ϕ t were chosen to make P(R 0,t=0∣R 0,t − 1=1) = 0.5 and P(R 0,t=1∣R 0,t − 1=0) = 0.5. This is a scenario in which dDTIC and strong independent return hold, and the data are MNAR. The (arguably unrealistic) bimodal distribution of ε t was chosen to illustrate that normality of (Y 1,…,Y 6,X) is not required by the MVN methods.

For each of 1000 simulated datasets, we estimated μ t=E(Y t) by estimating the compensator and by LI–LS, LI–uMVN, LI–aMVN, uMVN and aMVN imputation. Estimates were also calculated from the complete data (i.e. before imposing missingness) and from the complete cases (i.e. the mean of the outcomes with R 0t=1).

Table 1 shows the means and empirical SEs of the estimators. As expected, all but the complete‐cases estimator are approximately unbiased. LI imputation using the least‐squares estimates of α t, β t and γ t (LI–LS imputation) is considerably more efficient than estimating the compensator, and further efficiency is gained by replacing the least‐squares estimates by the corresponding parameter estimates from the unstructured or autoregressive MVN method (LI–uMVN or LI–aMVN imputation). Once α t, β t and γ t have been estimated, the LI imputation methods impute each individual's missing values using only his or her observed past. The uMVN and aMVN imputation methods, on the other hand, implicitly impute using the observed past and future. However, Table 1 shows that the efficiency gain from doing this is very small. It also shows that the difference in efficiency between the unstructured and autoregressive MVN methods is negligible.

Table 1.

Means and empirical SEs of estimated μ t in Simulation Study 1

Method μ 1 μ 2 μ 3 μ 4 μ 5 μ 6
True values 0.250 0.950 1.790 2.798 4.008 5.459
Means
Complete data 0.249 0.945 1.784 2.790 3.998 5.447
Complete cases 0.249 1.482 2.471 4.164 6.009 8.325
Estim. compens. 0.249 0.944 1.784 2.798 4.017 5.469
LI–LS impute 0.249 0.944 1.786 2.796 4.007 5.455
LI–uMVN impute 0.249 0.946 1.786 2.790 3.997 5.454
LI–aMVN impute 0.249 0.946 1.787 2.788 3.999 5.450
uMVN impute 0.249 0.947 1.784 2.791 3.998 5.455
aMVN impute 0.249 0.947 1.786 2.789 4.000 5.450
Empirical SEs
Complete data 0.045 0.084 0.120 0.160 0.204 0.255
Complete cases 0.045 0.115 0.172 0.224 0.278 0.319
Estim. compens. 0.045 0.113 0.227 0.343 0.477 0.615
LI–LS impute 0.045 0.113 0.179 0.238 0.296 0.354
LI–uMVN impute 0.045 0.102 0.140 0.186 0.234 0.292
LI–aMVN impute 0.045 0.100 0.139 0.186 0.233 0.291
uMVN impute 0.045 0.099 0.137 0.182 0.231 0.292
aMVN impute 0.045 0.097 0.136 0.182 0.230 0.291

LI, linear increment; LS, least‐square; MVN, multivariate normal; aMVN, autoregressive MVN.

We also fitted the linear regression model E(Y tX) = ψ 0+ψ 1 X + ψ 2 t + ψ 3 X t to datasets imputed by the five imputation methods. The results were in agreement with the results of Table 1. Again, all but the complete‐case estimator (i.e. the least‐squares estimator using outcomes with R 0t=1) are approximately unbiased; LI imputation using estimates of α t, β t and γ t from an MVN method is more efficient than LI imputation using the least‐squares estimates; and using the observed future to impute missing outcomes or constraining the variance matrix to be autoregressive provides little benefit (Table S1).

In Appendix S8, we present the results when the return mechanism P(R 0,t=1∣R 0,t − 1=1, R0,t2,FT) is modified so that the independent return assumption is violated. We demonstrate there that, as expected, estimating the compensator yields unbiased estimates of E(Y t) and E(Y tX), but the other methods (which assume independent return) are biased. Although the estimates from LI–LS imputation are biased, they are less biased than those from the MVN methods. This is because the least‐squares estimators of α t, β t and γ t are unbiased, whereas the corresponding MVN‐based estimators are biased.

6.2. Study 2: second‐order autoregression

Data were generated from the following model. Let X ∼ Uniform(0,1), Z 1∼Normal(0.5X,1), Z 2=0.4 + 1.5Z 1+0.5X + ε 2, and Z t=0.4 + 1.5Z t − 1−0.5Z t − 2+0.5X + ε t(3≤t≤7), where ε t∼Normal(1,1) with probability 0.5 and ε t∼Normal(−1,1) otherwise. Let R0,tZ=1 if Z t is observed, and R0,tZ=0 if Z t is missing. Z 1 is always observed.

Dropout is allowed at time 2 and when at least two consecutive outcomes have been observed. For times 4 and later, the probability of dropout depends on the rate of change in the outcome since time 2. Specifically, logit{P(R0,2Z=1FT)}=ω2+Z1+X, logit{P(R0,3Z=1R0,2Z=1,FT)}=ωt+W3+X and logit{P(R0,tZ=1R0,t1Z=R0,t2Z=1,R0,t3Z,FT)}=ωt+Wt+X (4≤t≤7), where W 3=Z 2Z 1 and W t=(Z t − 1Z 2)/(t − 3) (4≤t≤7). Return is possible at times 3, 4, 5 and 6. Specifically, logit{P(R0,tZ=1R0,t1Z=0,R0,t2Z,FT)}=ϕt+X+WLtZ (3≤t≤6), where LtZ=argmaxj{j<tandR0,jZ=1} denotes the most recent time at which the outcome Z t was observed prior to time t. The value of ω t (ϕ t) was chosen to make the overall probability of dropout (return) at time t among individuals at risk of dropout (return) at time t equal to 0.5. Note that dropout and return can depend on Z 2, which may be missing, and that dDTIC and the independent return assumption hold for the process Y t=(Z t,Z t − 1).

The results of applying the various methods with Y t defined as Y t=(Z t,Z t − 1) are shown in Tables 2 and S2. The autoregressive MVN methods have not been applied, as these were shown in Simulation Study 1 to be hardly more efficient than the unstructured MVN methods. As expected, all but the complete‐case estimator are approximately unbiased, estimating the compensator is least efficient, followed by LI–LS imputation, and uMVN imputation is only slightly more efficient than LI–uMVN imputation.

Table 2.

Means and empirical SEs of estimated μ t in Simulation Study 2

Method μ 1 μ 2 μ 3 μ 4 μ 5 μ 6 μ 7
True values 0.250 1.025 2.062 3.231 4.466 5.733 7.016
Means
Complete data 0.247 1.022 2.059 3.230 4.465 5.733 7.014
Complete cases 0.247 1.702 3.066 3.911 5.589 7.256 12.222
Estim. compens. 0.247 1.022 2.065 3.237 4.466 5.735 7.015
LI–LS impute 0.247 1.022 2.065 3.233 4.462 5.731 7.010
LI–uMVN impute 0.247 1.022 2.061 3.229 4.463 5.729 7.014
uMVN impute 0.247 1.023 2.061 3.227 4.463 5.729 7.014
Empirical SEs
Complete data 0.046 0.095 0.148 0.188 0.226 0.262 0.293
Complete cases 0.046 0.125 0.197 0.239 0.286 0.326 0.557
Estim. compens. 0.046 0.126 0.258 0.480 0.672 0.809 0.925
LI–LS impute 0.046 0.126 0.206 0.275 0.304 0.330 0.428
LI–uMVN impute 0.046 0.113 0.164 0.205 0.245 0.282 0.386
uMVN impute 0.046 0.110 0.159 0.200 0.241 0.282 0.387

LI, linear increment; LS, least‐square; MVN, multivariate normal.

7. Missingness due to death

A&G briefly discuss inference in the situation where a cause of missingness is death. They distinguish between ‘immortal‐cohort’ and ‘mortal‐cohort’ inference (also known, respectively, as ‘unconditional’ and ‘partly conditional’ inference). In this section, we elaborate on their brief discussion, in particular, establishing that an additional, unmentioned assumption is required for mortal‐cohort inference. Like A&G, we restrict ourselves to the case of no measurement error, although similar considerations would apply if there were measurement error. We begin by modifying earlier assumptions about the outcome, dropout and return processes so that they make sense when data can be missing due to death. Then, we explain what is meant by immortal‐cohort and mortal‐cohort inference. This is followed by an example which illustrates that the conditions stated by A&G as being sufficient for valid LI imputation are insufficient when interest is in mortal‐cohort inference. Finally, we prove that these conditions can be made sufficient by adding an additional assumption about the death process.

Let D i denote the time of individual's i last visit before death (D i=T if he or she does not die). Like A&G, we assume D i is known for all individuals. For individuals with D < t, the variable Y t describes an outcome that, in general, does not exist. In this setting, the meanings of the assumptions expressed by Equations (2), (3), (4), (8), (9) and (12) are unclear. It is therefore difficult to assess whether they are plausible for any particular given dataset. We now show that they can be replaced by Equations (20)(25), which are statements about only predeath outcomes, that is, outcomes that do exist.

Assume that

Yit=αt+βtYi,t1+γtXi+εit(t=2,,Di) (20)

with E(εitFi,t1,Dit)=0. This is a restriction of Equation (2) to predeath times, that is, to tD. Similarly, Equations (3), ((4)), ((8)), ((9)) and ((12)) are adapted to make them conditional on tD. Specifically, Equations (3), ((4)) and ((9)) are adapted to yield what we call the ‘mortal‐cohort weak DTIC’, ‘mortal‐cohort strong DTIC’ and ‘mortal‐cohort dDTIC’ assumptions, respectively:

E(ΔYtFt1,Rt,Dt)=E(ΔYtFt1,Dt)t, (21)
E(YtFt1,R0,t,Dt)=E(YtFt1,Dt)t,and (22)
f(YtFt1,R0,t,Dt)=f(YtFt1,Dt)t, (23)

while Equations ((8)) and ((12)) are adapted to yield the ‘mortal‐cohort independent return’ and ‘mortal‐cohort strong independent return’ assumptions, respectively:

P(R0,t=1R0,k1,R0,k=1,R0,k+1==R0,t1=0,X,Yk,Yk+1,Yt,Dt)=P(R0,t=1R0,k1,R0,k=1,R0,k+1==R0,t1=0,X,Yk,Dt) (24)

and

P(R0,t=1R0,k1,R0,k=1,R0,k+1==R0,t1=0,Gk,Yk+1,Yk+2,Yt,Dt)=PR0,t=1R0,k1,R0,k=1,R0,k+1==R0,t1=0,Gk,Dt. (25)

The stochastic process defined by Equation (20) terminates with Y D. In immortal‐cohort inference, no distinction is made between data missing due to death and data missing for other reasons, and both are imputed in the same way. Consequently, immortal‐cohort inference is about E(Y tY 1,X) in the following ‘supplemented’ outcome process. Define the ‘supplemented process’ as that defined by Equation (20) up to time D but with additional hypothetical postdeath outcomes Y D + 1,…,Y T for individuals with D < T that are assumed to obey

f(ΔYtFt1,R0,t,D<t)=f(ΔYtFt1,D<t)=f(ΔYtFt1,Dt)tD

It thus follows from, respectively, Equations (21), (22) and (23) that E{ΔYtFt1,Rt,I(Dt)}=E(ΔYtFt1), that E{YtFt1,R0t,I(Dt)}=E(YtFt1), and that

f{YtFt1,R0,t,I(Dt)}=f(YtFt1). (26)

Equations ((2)), (3), ((4)) and ((9)) therefore hold for the supplemented process. Hence, the least‐squares estimators α^tls, β^tls and γ^tls of α t, β t and γ t are unbiased and consistent for the parameters of Equation (2), and therefore also of Equation (20).

The proof in Section 3.3 of A&G establishes that LI–LS imputation respects E(YtestYtX,Y1)=0 for the supplemented process, provided that Equations ((2)) and ((7)) and strong DTIC hold, and hence one can validly use the imputed data to estimate E(Y tY 1,X), the expected outcome in the supplemented process.

In mortal‐cohort inference, on the other hand, the estimand is E(Y tY 1,X,Dt), the expected outcome at time t in individuals who are still alive at time t. A&G discuss immortal‐ versus mortal‐cohort inference. They (and Gunnes et al. (2009a)) defend immortal‐cohort inference, arguing that it may provide ‘a more fair comparison of treatments’ when one treatment improves survival in a way that means that patients with poor outcomes survive longer. However, Dufouil et al. (2004), Kurland & Heagerty (2005) and Kurland et al. (2009) generally favour mortal‐cohort inference, saying that for most purposes immortal‐cohort inference would be ‘inappropriate’, or that it is ‘generally inappropriate’ and ‘probably not of great scientific interest’ unless the death and outcome processes are independent.

A&G estimate E(Y tY 1,X,Dt) by using LI–LS imputation to impute only predeath missing outcomes, that is, calculating Yitest for individual i only if D it, and then fitting a linear regression model to the resulting imputed dataset {Yitest:i=1,,N;tDi} (rather than {Yitest:i=1,,N;t=1,,T}, as in Section 2.5). In order for this to be valid, it is necessary that E(YtestYtX,Y1,Dt)=0. However, the fact that E(YtestYtX,Y1)=0 for the supplemented process does not necessarily imply that E(YtestYtX,Y1,Dt)=0. A simple example illustrates this. Suppose T = 3, m = 1 and there are no baseline covariates X. Let P(Y 1=0) = 1 and P(Y 2=Y 3=0) = P(Y 2=Y 3=1) = 0.5. Suppose that the only dropout occurs between times 1 and 2, and its probability does not depend on Y 2, and that there is no return. Let P(D = 3∣R 0,2=1) = P(D = 3∣Y 2=0) = 1, so no one can die unless they drop out and have Y 2=1. Let P(D = 2∣R 0,2=0,Y 2=1) = P(D = 3∣R 0,2=0,Y 2=1) = 0.5, so half of those who drop out and have Y 2=1 die between times 2 and 3. When LI imputation or MVN imputation is applied to this population, the dropouts will all have their Y 3 imputed as 0.5. However, among the dropouts who are still alive at time 3, the mean value of Y 3 is actually 0.33.

Theorem 9 and its corollary, later, show that E(YtestYtX,Y1,Dt)=0 can be ensured by making two additional assumptions. We call these ‘independent death’:

P(D=tR0,k1,R0,k=1,R0,k+1==R0,t=0,X,Yk,,Yt,Dt)=P(D=tR0,k1,R0,k=1,R0,k+1==R0,t=0,X,Yk,Dt) (27)

and ‘strong independent death’:

P(D=tR0,k1,R0,k=1,R0,k+1==R0,t=0,Gk,Yk+1,Yk+2,,Yt,Dt)=P(D=tR0,k1,R0,k=1,R0,k+1==R0,t=0,Gk,Dt) (28)

Equation ((27)) means that the probability of dying between visits t and t + 1 for people who did not attend visit t does not depend on outcomes since their last observed outcome. This is analogous to the independent return assumption (Equation (8)), except that independent return is about the probability of returning after dropout, rather than dying after dropout. Just as independent death is analogous to independent return, strong independent death is analogous to strong independent return. The independent death assumption will often not be entirely plausible; it is more likely to hold approximately when long gaps between dropout and death are uncommon.

Theorem 8 shows the relation between the mortal‐cohort independent return assumptions and independent death assumptions, and between mortal‐cohort assumptions and independent return assumptions in the supplemented process.

Theorem 8

Suppose that the increments model of Equation (20) and the mortal‐cohort dDTIC assumption of Equation (23) hold. Then, (a) mortal‐cohort strong independent return and strong independent death together imply mortal‐cohort independent return and independent death; (b) mortal‐cohort independent return and independent death together imply independent return in the supplemented process; and (c) mortal‐cohort strong independent return and strong independent death together imply strong independent return in the supplemented process.

Theorem 9 establishes that, under the conditions of Theorem 8, the conditional distribution of a missing outcome given earlier observed outcomes is the same whether or not the missing outcome is after death.

Theorem 9

If the increments model of Equation (20), and the mortal‐cohort dDTIC, mortal‐cohort independent return and independent death assumptions (Equations (23), ((24)) and ((28))) hold, then for k < t,

f(YtR0,k1,R0,k=1,R0,k+1==R0,t=0,X,Yk,Dt)=f(YtX,Yk),

Under the conditions of Theorem 9, if LI imputation is carried out using the true values of α t, β t and γ t, then E(YtestYtX,Y1,Dt)=0.

In practice, the true values of α t, β t and γ t are unknown and must be estimated. Earlier in this section, we described when the least‐squares estimators are unbiased and consistent. Now, we consider the uMVN and aMVN estimators.

(of Theorems 4 and 8) If the increments model of Equation (20) and the mortal‐cohort dDTIC, mortal‐cohort strong independent return and strong independent death assumptions (Equations (23), ((25)) and ((28))) hold and Var(εtFt1,Dt)=Var(εtDt), then fitting the unstructured or autoregressive MVN model to the observed data ignoring the missingness mechanism yields consistent estimates of the mean and variance (μ,Σ) of the supplemented process, and hence of α t, β t and γ t. Furthermore, when fitting the autoregressive model, the mortal‐cohort strong independent return and strong independent death assumptions can be replaced by the mortal‐cohort independent return and independent death assumptions.

In Appendix S7, we discuss the validity of uMVN and aMVN imputation for mortal‐cohort inference. They are valid when ε t is normally distributed; otherwise, they are expected to have small bias. However, in view of the results in Section 6, where MVN imputation was only slightly more efficient than LI‐MVN imputation, the latter may be the better option.

8. Analysis of data from the BHPS

The BHPS began in 1991. The panel consisted of some 5500 households and 10,300 individuals drawn from many areas of Great Britain. Complex sampling weights were used to adjust for unequal selection probabilities and for non‐response. These weights were ignored in our analysis, which is intended to be illustrative rather than definitive. We examined the dependence of earnings (elicited for each year) on current age, adjusting for calendar time. We used waves 1–8 (1991–1998) (the period before booster samples was introduced) and restricted the sample to men aged 21–50 in 1991 who were observed to earn something during at least one of those 8 years. There were 3286 such men. Age was categorized as 21–25, 26–30, 31–40, 41–50 and 51–60. Mean earnings in each calendar year were estimated by estimating the compensator, LI–LS imputation and the MVN methods. Also, a linear regression model was fitted to the imputed dataset, regressing earnings in each year on current age group and calendar year. The baseline covariate X used in Equation (2) for all LI methods was age group in 1991. In order to handle missing values of earnings in 1991, we defined Y i2,…,Y i9 to be individual i's incomes in 1991–1998, respectively, defined Y i1=c for all i, where c is an arbitrary constant, and constrained β 2 in Equation (2) to equal zero (so the choice of c does not matter). SEs were estimated by bootstrapping.

Of the 3286 men, the percentage who had observed outcome at each wave varied from 55% (in 1983) to 64% (in 1987). 829 (15%) were observed at all eight waves, 658 (20%) had missing outcome at wave 1 but were observed at a later wave and did not drop out again, and 517 (16%) were observed at wave 1 but later dropped out and did not return. Of the remaining 1282 (39%) men, 976, 284 and 24 dropped out once, twice and thrice, respectively; and 1010, 258 and 14 returned once, twice and thrice, respectively. Of the 15,502 observed outcomes, 818 (5%) were immediately preceded and followed by missing outcomes; these are not used by the least‐squares estimators of the LI model and are completely ignored by LI methods that allow for autoregression of order two.

Table 3 shows the observed mean earnings in each year and the means estimated using seven LI methods. Estimates from LI–aMVN and LI–rMVN imputation (not shown) were very similar to those from LI–uMVN; estimates from aMVN and rMVN imputation (not shown) were very similar to those from uMVN imputation. All LI methods estimated the mean earnings in 1991 as lower than the observed mean. This is because the proportion of men reporting their earnings in 1991 increased with age group, and the mean earnings among those reporting also increased with age group, which suggests that the observed mean in 1991 is an overestimate of the mean in 1991 in the whole sample. Among the LI methods, the highest estimates came from estimating the compensator and the lowest from uMVN imputation. The difference ranged from 393 to 816 pounds. That these two methods differ most is not surprising, because the first makes no use of outcomes recorded immediately after return from dropout, whereas the second makes maximal use of these. Allowing for second‐order autoregression slightly increased estimates from LI–LS and LI–uMVN imputation but left those from estimating the compensator unchanged or slightly decreased.

Table 3.

Estimated mean earnings (and SEs) in calendar years 1991–1998 (1000's of pounds).

Year
Method 1991 1992 1993 1994 1995 1996 1997 1998
Estimated means
Observed data 14.35 14.55 15.28 15.60 16.85 17.25 17.96 18.63
Compensator 13.87 14.45 15.10 15.84 16.72 17.46 18.57 19.24
LI–LS 13.87 14.33 14.91 15.49 16.45 17.12 17.92 18.51
LI–uMVN 13.73 14.18 14.75 15.31 16.26 16.93 17.76 18.40
uMVN 13.48 13.96 14.49 15.14 15.99 16.69 17.75 18.46
Compensator(2) 13.87 14.45 15.12 15.89 16.72 17.41 18.41 19.11
LI–LS(2) 13.87 14.44 15.03 15.65 16.58 17.25 18.08 18.80
LI–uMVN(2) 13.77 14.33 14.88 15.49 16.44 17.14 18.02 18.74
Standard errors
Observed data 0.240 0.228 0.250 0.263 0.321 0.283 0.321 0.352
Compensator 0.233 0.234 0.260 0.281 0.293 0.298 0.349 0.403
LI–LS 0.233 0.213 0.228 0.239 0.265 0.258 0.282 0.333
LI–uMVN 0.220 0.202 0.216 0.230 0.265 0.253 0.285 0.328
uMVN 0.209 0.192 0.208 0.228 0.248 0.252 0.280 0.331
Compensator(2) 0.233 0.234 0.266 0.291 0.301 0.315 0.336 0.420
LI–LS(2) 0.233 0.220 0.240 0.254 0.276 0.273 0.287 0.362
LI–uMVN(2) 0.224 0.209 0.226 0.241 0.276 0.267 0.297 0.360

Methods are means of observed data, estimating the compensator, LI–LS imputation, LI–uMVN imputation and uMVN imputation.

All LI methods use a LI model with autoregression of order one, unless they are marked ‘(2)’, in which case they use a LI model with second‐order autoregression.

LS, least‐square; LI, linear increments; MVN, multivariate normal.

Estimated standard errors from LI–aMVN imputation and LI–rMVN imputation were very close to those from LI–uMVN imputation. Those from aMVN imputation were very close to those from uMVN imputation; those from rMVN were up to 10% greater. The largest standard errors came from estimating the compensator; these were even larger than those for the observed means. Standard errors from LI–LS imputation were smaller than those for the observed means. Those from LI–uMVN imputation tended to be slightly smaller than those from LI–LS imputation, and those from uMVN imputation slightly smaller still. This suggests that when there is substantial dropout and return, methods (like LI–uMVN and uMVN) that use all the data efficiently can give appreciably more precise estimates. Allowing for second‐order autoregression generally increased the standard errors, because this involves ignoring some of the observed outcomes.

Table 4 shows estimated coefficients when the linear regression of earnings on current age group and calendar year was fitted to observed data and to data imputed by the various LI methods. Mean earnings increased with current age, being 3000–4000, 6500–7500 and 7000–8000 pounds higher at ages 26–30, 31–40 and 41–50, respectively, than at 21–25. However, they were 1500–2500 pounds lower at ages 51–60 than at 41–50, possibly because of early retirement. The LI imputation methods all estimated that the differences between ages 26–50 and ages 21–25 were somewhat lower that the observed data suggest. Standard errors for LI methods were lower than those for the observed‐data analysis. Those for LI–uMVN and uMVN imputation were lower than those for LI–LS imputation. Allowing for second‐order autoregression increased standard errors, probably because some observed outcomes are then ignored.

Table 4.

Estimated coefficients (and standard errors) of linear regression of earnings (1000's of pounds) on current age group and year.

Current age Year
Method Intercept 26–30 31–40 41–50 51–60 –1991
Estimates
Observed data 8.322 4.018 7.386 7.948 5.472 0.475
LI–LS 8.517 3.879 7.109 7.774 5.792 0.457
LI–uMVN 8.557 3.614 6.855 7.519 5.496 0.468
uMVN 8.404 3.386 6.641 7.396 5.330 0.512
LI–LS(2) 8.504 3.783 7.236 7.907 6.127 0.473
LI–uMVN(2) 8.573 3.596 6.960 7.615 5.722 0.494
Standard errors
Observed data 0.322 0.377 0.459 0.525 0.799 0.049
LI–LS 0.298 0.335 0.436 0.520 0.713 0.046
LI–uMVN 0.284 0.315 0.422 0.496 0.691 0.045
uMVN 0.298 0.323 0.433 0.486 0.686 0.042
LI–LS(2) 0.300 0.338 0.447 0.541 0.764 0.047
LI.uMVN(2) 0.289 0.319 0.435 0.516 0.729 0.047

Methods are observed data, LI–LS imputation, LI–uMVN imputation and uMVN imputation. LI methods use a LI model with autoregression of order one, unless they are marked ‘(2)’, in which case they use a LI model with second‐order autoregression.

LI, linear increment; LS, least‐square; MVN, multivariate normal.

9. Discussion

Missing at random and the distributional form of DTIC (dDTIC) are equivalent assumptions when missingness is monotone and there is no independent measurement error. Otherwise, they differ. MAR allows dropout and return to depend on observed outcomes, including future outcomes but not on the underlying (error‐free) outcomes. DTIC allows dropout and return to depend on all past underlying outcomes but not on future outcomes or on the measurement error. Nevertheless, we have shown that some methods that ostensibly assume MAR can also be valid when data are MNAR but dDTIC holds and the underlying outcomes have an autoregressive structure. For example, in the absence of measurement error, fitting an unstructured MVN model and ignoring the missingness mechanism gives consistent estimation and can be more efficient than LI imputation using least squares (even when data are not normally distributed). In addition, it remains valid under the alternative assumption of MAR, even if the autoregressive structure does not apply. On the other hand, LI using least squares allows the variance of ε t to depend on the history Fi,t1, which may make this method more appealing than the MVN method when the outcome is categorical. Moreover, it is less prone to bias than the MVN method when the independent return assumption is violated. Estimating the compensator is the least efficient method but makes the fewest assumptions: it only requires weak DTIC and not independent return.

The MVN method can be made more efficient by constraining the variance matrix to have an autoregressive form. However, this was found in simulation studies to offer little benefit. Further investigation is required to determine whether it offers greater benefit with small sample sizes or when there are many time points.

As we have shown, when there is measurement error, none of the methods discussed in this paper allow the increment ΔY t to depend on underlying outcome Y t − 1 (i.e. for β tI), unless the magnitude of this dependence (i.e. the value of β t) is known, which is unlikely in practice. Just as measurement error can be handled when β t is known, it may be possible when β t is unknown but the measurement error variance, Var(e t), is known (Fuller, 1987).

Like A&G, we have focused on a first‐order autoregressive model. This is quite a restrictive assumption and may be too restrictive for some applications. So, we have shown in detail how the various methods can allow for autoregression of order s > 1 by including in the vector Y t not only the underlying outcome at time t but also those at times t − 1,…,ts + 1. A drawback of this approach is that because R 0,t=0 when Y it is not fully observed, some observed measurements may be ignored unless missingness is monotone. Alternative, more efficient extensions may be possible.

Software for applying some of the methods covered in this paper are described in Appendix S9. Like DF&H and A&G, we recommend using bootstrap to calculate standard errors.. However, it may be possible to derive expressions for sandwich variance estimators (Zeng & Lin, 2007; Farewell, 2010).

Note that we have treated the missingness processes of individuals as independent, whereas A&G's formulation is slightly more general, allowing for dropout and return of one individual to depend on the histories of other individuals.

A&G briefly discuss how LI could be used for causal inference about treatment effects in a non‐randomized longitudinal study with time‐dependent confounders, viewing the counterfactual untreated outcomes of treated individuals as missing data.

Apart from estimating the compensator, all methods discussed in this article require an assumption about the probability of return after dropout (the independent return assumption). Much of the published work on non‐monotone missing data in longitudinal studies assumes either MAR or that the probability of attending a visit at time t is independent of attendance at other times given the outcomes or given a random effect shared with the model for the outcomes, and so does not focus explicitly on return. Liao et al. (2012) give a recent review of some of this work. Articles in which an explicit model for the probability of return is used include Preisser et al. (2000), Lin et al. (2004) and Liao et al. (2012).

It would be interesting to understand better the connection between the random‐walk MVN method and the g‐inverse working singularity GEE method proposed by Farewell (2010). When β t=I and there are no covariates X, Equation (19) reduces to Σ s,t=Σ s,s for all s < t. A special case of this is Σ s,t=Σ s,s=s, which is the working covariance matrix used by Farewell (2010). Because this is the covariance matrix of a random walk where all increments have unit variance, it may be that the random‐walk MVN method is more efficient than the g‐inverse method when the variances of the increments differ at different times.

Finally, we note that we have not discussed the plausibility of the MAR and DTIC assumptions. This must be considered within the context of any given dataset. For example, medical data may be MAR when follow‐up of a patient is determined by the doctor on the basis of a measured health outcome, whereas DTIC may be more plausible when follow‐up is determined by the patient on the basis of his or her underlying health state. When data are monotone missing, MAR has a straightforward interpretation: dropout is independent of the present and future given the past. Interpretation when data are non‐monotone missing is more problematic, however (Robins & Gill, 1997). One advantage of the DTIC assumption may be that it is easily interpreted in both the monotone and non‐monotone case.

Supporting information

Supporting info item

Acknowledgements

We thank Lorenz Wernisch and Brian Tom for helpful discussions and two reviewers for very helpful and insightful comments.

S. R. S. and I. R. W. are funded by MRC Grants U1052 60558 and MC_US_A030_0015. D. F. is supported by an MRC Methodology Fellowship.

Seaman, S. R. , Farewell, D. , and White, I. R. (2016) Linear Increments with Non‐monotone Missing Data and Measurement Error. Scand J Statist, 43: 996–1018. doi: 10.1111/sjos.12225.

References

  1. Aalen, O. & Gunnes, N. (2010). A dynamic approach for reconstructing missing longitudinal data using the linear increments model. Biostat. 11, 453–472. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Diggle, P. , Farewell, D. & Henderson, R. (2007). Analysis of longitudinal data with drop‐out: objectives, assumptions and a proposal. Appl. Stat. 56, 499–529. [Google Scholar]
  3. Dufouil, C. , Brayne, C. & Clayton, D. (2004). Analysis of longitudinal studies with death and drop‐out: a case study. Stat. Med. 23, 2215–2226. [DOI] [PubMed] [Google Scholar]
  4. Farewell, D. M. (2010). Marginal analyses of longitudinal data with an informative pattern of observations. Biometrika 97, 65–78. [Google Scholar]
  5. Fuller, W. A. (1987). Measurement error models, John Wiley, New York. [Google Scholar]
  6. Gunnes, N. , Farewell, D. , Seierstad, T. & Aalen, O. (2009a). Analysis of censored discrete longitudinal data: estimation of mean response. Stat. Med. 28, 605–624. [DOI] [PubMed] [Google Scholar]
  7. Gunnes, N. , Seierstad, T. , Aamdal, S. , Brunsvig, P. , Jacobsen, A‐B , Sundstrom, S. & Aalen, O. (2009b). Assessing quality of life in a randomised clinical trial: correcting for missing data. BMC Med. Res. Method 9, 1–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Henderson, R. , Diggle, P. & Dobson, A. (2000). Joint modelling of longitudinal measures and event time data. Biostatistics 1, 465–480. [DOI] [PubMed] [Google Scholar]
  9. Hogan, J. W. , Roy, J. & Korkontzelou, C. (2004). Handling drop‐out in longitudinal studies. Stat. Med. 23, 1455–1497. [DOI] [PubMed] [Google Scholar]
  10. Kingsley, G. H. , Kowalczyk, A. , Taylor, H. , Ibrahim, F. , Packham, J. , McHugh, N. , Mulherin, D. M , Kitas, G. D. , Chakravarty, K. , Tom, B. D. M. , OŠKeeffe, A. G. , Maddison, P. J. & Scott, D. L. (2012). A randomized placebo‐controlled trial of methotrexate in psoriatic arthritis. Rheumatology 51, 1368–1377. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Kurland, B. & Heagerty, P. (2005). Directly parameterised regression conditioning on being alive: analysis of longitudinal data truncated by deaths. Biostat. 6, 241–258. [DOI] [PubMed] [Google Scholar]
  12. Kurland, B. , Johnson, L. , Egleston, B. & Diehr, P. (2009). Longitudinal data with follow‐up truncated by death: match the analysis method to research aims. Stat. Sci. 24, 211–222. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Liao, K. , Freres, D. & Troxel, A. (2012). A transition model for quality‐of‐life data with non‐ignorable non‐monotone missing data. Stat. Med. 31, 3444–3466. [DOI] [PubMed] [Google Scholar]
  14. Lin, H. , Scharfstein, D. & Rosenheck, R. (2004). Analysis of longitudinal data with irregular outcome‐dependent follow‐up. J. Roy. Stat. Soc. Ser. B 66, 791–813. [Google Scholar]
  15. Little, R. & Rubin, D. (2002). Statistical analysis with missing data, Wiley, New Jersey. [Google Scholar]
  16. Preisser, J. , Galecki, A. , Lohman, K. & Wagenknecht, L. (2000). Analysis of smoking trends with incomplete longitudinal binary responses. J. Amer. Stat. Assoc. 95, 1021–1031. [Google Scholar]
  17. Pullenayegum, E. & Feldman, B. (2013). Doubly robust estimation, optimally truncated inverse‐intensity weighting and increment‐based methods for the analysis of irregularly observed longitudinal data. Stat. Med. 32, 1054–1072. [DOI] [PubMed] [Google Scholar]
  18. Robins, J. & Gill, R. (1997). Non‐response models for the analysis of non‐monotone ignorable missing data. Stat. Med. 16, 39–56. [DOI] [PubMed] [Google Scholar]
  19. Schafer, J. (1997). Analysis of incomplete multivariate data, Chapman and Hall/CRC, Boca Raton. [Google Scholar]
  20. Seaman, S. & Copas, A. (2009). Doubly robust generalized estimating equations for longitudinal data. Stat. Med. 28, 937–955. [DOI] [PubMed] [Google Scholar]
  21. Seaman, S. , Galati, J. , Jackson, D. & Carlin, J. (2013). What is meant by ‘missing at random’?. Stat. Sci. 28, 257–268. [Google Scholar]
  22. Seaman, S. & White, I. (2013). Review of inverse probability weighting for dealing with missing data. Stat. Method Med. Res. 22, 278–295. [DOI] [PubMed] [Google Scholar]
  23. Verbeke, G. & Molenberghs, G. (2000). Linear mixed models for longitudinal data, Springer, New York. [Google Scholar]
  24. Zeng, D. & Lin, D. Y. (2007). Discussion of ‘analysis of longitudinal data with drop‐out: objectives, assumptions and a proposal’. Appl. Stat. 56, 544–545. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supporting info item


Articles from Scandinavian Journal of Statistics, Theory and Applications are provided here courtesy of Wiley

RESOURCES