Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2012 Aug 15.
Published in final edited form as: J Dev Orig Health Dis. 2011 Oct 12;2(6):322–329. doi: 10.1017/S2040174411000572

Analyzing growth trajectories

Ian W McKeague 1, Sara López-Pintado 2, Marc Hallin 3, Miroslav Šiman 4
PMCID: PMC3419544  NIHMSID: NIHMS394242  PMID: 22905314

Abstract

Growth trajectories play a central role in life course epidemiology, often providing fundamental indicators of prenatal or childhood development, as well as an array of potential determinants of adult health outcomes. Statistical methods for the analysis of growth trajectories have been widely studied, but many challenging problems remain. Repeated measurements of length, weight and head circumference, for example, may be available on most subjects in a study, but usually only sparse temporal sampling of such variables is feasible. It can thus be challenging to gain a detailed understanding of growth patterns, and smoothing techniques are inevitably needed. Moreover, the problem is exacerbated by the presence of large fluctuations in growth velocity during early infancy, and high variability between subjects. Existing approaches, however, can be inflexible due to a reliance on parametric models, require computationally intensive methods that are unsuitable for exploratory analyses, or are only capable of examining each variable separately. This article proposes some new nonparametric approaches to analyzing sparse data on growth trajectories, with flexibility and ease of implementation being key features. The methods are illustrated using data on participants in the Collaborative Perinatal Project.

Keywords: Growth curves, nonparametric Bayes, data depth contours

1 Introduction

There is a vast literature on the statistical analysis of human growth curves. The earliest work in this area concentrated on the formulation of parametric growth models, with Jenss, Bayley, Preece, Baines, Count and Gompertz being prominent contributors. These models are designed to capture known features of growth and development (such as the mid-childhood growth spurt) and have reached a high degree of sophistication with broad applications1,2. For example, such models have been used in searches for quantitative trait loci that control the key features of human growth trajectories3.

The purpose of the present article is to propose various new nonparametric modeling approaches that can bring greater flexibility, as well as ease of implementation, to the analysis of growth trajectories based on sparse data. Our emphasis is on methods that are suited for the study of prenatal or early childhood development, in which large fluctuations in growth velocity and high variability between subjects are not easily handled by parametric models. Despite a resurgent interest in the analysis of human growth trajectories, current statistical methods are limited by an over-reliance on parametric modeling and are only capable of examining each variable separately, or require computationally intensive methods that are unsuitable for exploratory analyses. Repeated measurements of length, weight, BMI and head circumference, for example, may be available on most subjects in a study, but usually only sparse temporal sampling of such variables is feasible. It can thus be challenging to gain a detailed understanding of growth patterns, and smoothing techniques are inevitably needed. Moreover, the problem is exacerbated by the presence of large fluctuations in growth velocity during early infancy, and high variability between subjects.

We propose a nonparametric Bayesian method4 for reconstructing growth velocity curves from sparse temporal data (or repeated measures) on a single variable. Figure 1 illustrates this method as applied to length measurements in a sample of 532 girls who participated in the Collaborative Perinatal Project (see Section 4 for further details). The left panel shows the reconstructed growth velocity curve (along with error bounds) of a specific individual, and the right panel replicates this for the whole sample. A key advantage of this method over existing approaches is that error bounds are included in the reconstruction. A version of data depth that is suitable for visualizing functional data5 is also discussed; the right panel of Fig. 1 highlights the deepest growth velocity curve in the sample and can be interpreted as a functional equivalent of the sample median.

Figure 1.

Figure 1

The left panel shows the reconstruction of an individual growth velocity curve (solid line) with error bounds (dashed lines); the right panel shows the reconstructed growth velocity curves for all individuals in the sample with the deepest curve highlighted.

In addition, we propose a method for visualizing patterns in the growth trajectories of multiple variables. Commonly used growth charts produce plots of univariate quantile curves, but such plots clearly omit all information related to dependencies between the various measurements under study. This is potentially misleading, as growth charts are often used as a diagnostic tool for detecting possible outliers, while a multivariate outlier clearly need not be an outlier from a marginal point of view, and vice versa. To address this problem, we introduce a method based on Tukey’s notion of halfspace data depth6, leading to the construction of flexible multiple-output growth charts, see for example Fig. 5.

Figure 5.

Figure 5

Estimated age-specific depth contours (cuts) in Example 2 for [left panel] head circumference (y2) and weight (y1), and [right panel] head circumference (y2) and length (y1), at equispaced ages t0 = 0 (birth; black), 1.75 (blue), 3.5 (green), 5.25 (cyan) and 7 (yellow) years. Observations are shown by dots (darker for younger subjects, lighter for older ones).

2 Growth velocities

Nonparametric frequentist approaches to the analysis of growth trajectories have been extensively studied in the setting of functional data analysis7,8. In particular, functional principal components analysis is used when it is of interest to estimate the “dominant modes of variation” of a sample of trajectories. Typically, however, a crucial first step is needed before such analyses are possible: the trajectories need to be reconstructed on a fine grid of equally spaced time points. Methods for reconstructing trajectories in this way have been studied using kernel smoothing7, smoothing splines8, local linear smoothing9, mixed effects models10,11, and principal components analysis through conditional expectations12,13.

In many settings involving functional data, the gradients of the trajectories (i.e., growth velocities) are of central interest, rather than the trajectories themselves, especially when dynamical effects are concerned. Difference quotients between observation times can be used to generate simple approximate gradients, but these estimates are piecewise constant and would not be suitable for use in functional data analysis unless the observation times are dense. In the case of regularly spaced observation times, spline smoothing to approximate the gradient of the trajectory over a fine grid is recommended8. More generally, methods of numerical differentiation, including spline smoothing, are an integral part of the extensive literature on ill-posed inverse problems for linear operator equations. In this literature, the observation times are usually viewed as becoming dense (for the purpose of showing convergence)14; in particular, the assumption of asymptotically dense observation times plays a key role in the study of penalized least squares estimation and cross-validation15,16.

Growth velocities can be reconstructed given sparse and irregularly spaced observation times (one observation time per trajectory is even enough) by borrowing strength between the trajectories in the data set. For such sparse observations, it has been shown that the best linear predictor of the gradient can be estimated in terms of estimated principal component scores, assuming Gaussian trajectories and that the pooled observation times become dense17. A disadvantage of this approach, however, is that data at the individual level plays a relatively minor role in the reconstruction, and its accuracy depends on how well each individual gradient can be represented in terms of a small number of estimated principal component functions (this in turn would require an accurate estimate of the covariance kernel of the trajectories, an unlikely scenario in the case of sparse observation times).

López-Pintado and McKeague (2011) 4,18 recently developed a flexible Bayesian approach to reconstructing growth velocities from sparse data, as outlined in the next section. Their approach is designed to adapt to observation times that are both sparse and irregularly spaced, and that can vary across subjects. The observation times are allowed to be arbitrary, as long as they include the endpoints of the time interval (so interpolation is possible). The prior distribution for the growth velocity is specified by a multivariate normal distribution at the observation times, and a tied-down Brownian motion between the observation times. This leads to an explicit representation of the posterior distribution in a way that exactly reproduces the data at the observation times. The empirical Bayes approach is then used to estimate the hyperparameters in the prior, borrowing strength between subjects, but in a simpler fashion than estimating principal component scores17. An important aspect of this approach is that reconstructed gradients can be computed rapidly over a fine grid, and then used directly as input into existing software, without the need for sophisticated smoothing techniques. Furthermore, a comparison of the results from repeated draws from the posterior distribution (multiple imputation) provides an easy way of assessing uncertainty in the conclusions (of standard functional data analyses) due to data sparsity.

The empirical Bayes approach is well developed for reconstructing individual growth velocity curves from parametric growth models19. A nonparametric Bayesian growth curve model has been developed for testing for differences in growth patterns between groups of individuals20. In addition, a nonparametric hierarchical-Bayesian growth curve model for reconstructing individual growth curves is available, but requires the use of computationally intensive Markov chain Monte Carlo methods21.

Bayesian reconstruction

We first consider how to reconstruct a growth velocity curve for a single subject. The observation times will typically vary slightly across the sample, but will be clustered around certain nominal ages (e.g., birth, 4 months, 8 months, 1 year, …). Let the observation times for the specific individual be 0 = t1 < t2 < … < tp = T, and assume that the endpoints of the time interval over which the reconstruction is needed are included. Letting the subject’s growth velocity at age t be X(t), the statistical problem is to estimate the growth velocity curve X = {X(t), 0 ≤ tT} from data on its integral over the gaps between the observation times. Reconstructing X based on such data is an ill-posed inverse problem in the sense that no unique solution exists, so some type of external information or constraint (i.e., regularization) is needed to produce a unique solution14.

The difference quotient estimate of X(t) in the interval between the ith and (i+1)th observation times is given by

yi=1Δititi+1X(s)ds,

where Δi is the length of the interval. Higher order difference estimates are produced by taking into account the proximity to neighboring observation times, say replacing yi by the weighted estimate ȳi = wiyi−1 + (1 − wi)yi, where wi = Δi/(Δi−1 + Δi) for i = 2, …, p − 1. Neither of these estimates borrow strength from other trajectories in the sample, but they provide the building blocks of empirical Bayes estimators that take advantage of the whole sample, as we now explain.

In the Bayesian approach to ill-posed inverse problems, regularization takes the form of specifying a prior distribution on X. It is desirable to make the prior flexible enough to cover a broad range of growth velocity patterns, yet simple enough that it is tractable to find the posterior distribution without the need for computationally intensive methods. López-Pintado and McKeague (2011)4 showed that this can be done using the following hierarchical prior: 1) at the observation times, X = (X(t1), …, X (tp))′ has a p-dimensional normal distribution with mean μ0 and non-singular covariance matrix Σ0, and 2) the conditional distribution of X given X is a tied-down Brownian motion with given infinitesimal variance σ2 > 0. Allowing an arbitrary (multivariate normal) prior at the observation times provides flexibility that would not be possible using a Brownian motion prior for the whole of X. In addition, the availability of data at these time points makes it possible to specify the hyperparameters in the multivariate normal (as we discuss below), which is crucial for practical implementation of our approach.

The posterior mean of X takes the computationally tractable form of a quadratic spline with knots at the observation times:

μ^(t)=μ^i+[μ^i+1-μ^i](t-ti)/Δi+6(t-ti)(ti+1-t)[yi-(μ^i+μ^i+1)/2]/Δi2

for t belonging to the interval between the ith and (i + 1)th observation times. Integration shows that μ̂(t) exactly reproduces the data. Here μ̂i is the ith component of the posterior mean of X, given by μ^=(0-1+Q)-1(0-1μ0+DY) where Y = (y1, ȳ2, , ȳp−1, yp−1)′,

Q=3σ2(1Δ11Δ1001Δ11Δ1+1Δ21Δ201Δ201Δp-1001Δp-11Δp-1),

and

D=6σ2diag(1Δ1,,1Δi-1+1Δi,,1Δp-1).

The posterior distribution is Gaussian, with a covariance kernel (not depending on Y) that takes a similarly tractable form as the mean.

The posterior mean μ̂(t) can be used for reconstructing the unobserved growth velocity X(t), provided various hyperparameters are specified in advance: the prior mean μ0 and prior precision matrix 0-1. This is done via a nonparametric empirical Bayes approach applied to the full sample of trajectories, initially treated as having identical sets of (nominal) observation times. The sample mean of Y is used to specify μ0. A constrained ℓ1 minimization method of sparse precision matrix estimation (clime)22 is applied to the (singular) sample covariance matrix of Y to specify 0-1. By restricting the resulting posterior covariance kernel and mean to the actual observation times for a given subject, we obtain suitable hyperparameters across the whole sample that adjust for any changes from the nominal observation times4.

The infinitesimal standard deviation σ is a smoothing parameter (playing the role of a time-scale), and can be selected using a type of cross-validation based on the prediction error from leaving-out an interior observation time4. We have found that μ̂(t) is relatively insensitive to σ. On the other hand, the width of credible intervals around μ̂(t) is roughly proportional to σ. In practice, insight into an appropriate choice of σ can also be gained through inspecting plots of μ̂(t), say for values σ in the range 1–3 for the height data (as suggested by cross-validation), and it is worthwhile to include pointwise 95% credible intervals around μσ(t) as a way of assessing the uncertainty in the reconstruction (see Fig. 3 for examples).

Figure 3.

Figure 3

Reconstructed growth velocity curves for two subjects in Example 1; posterior mean μ̂(t) (solid line), pointwise 95% credible intervals (dashed lines) based on σ = 1, 2, 3 in (a,d), (b,e) and (c,f), respectively; for one subject in (a,b,c), and a second subject in (d,e,f).

An R package “growthrate” implementing this reconstruction method has been developed by López-Pintado and McKeague (2011) 18. The package includes the data set and examples of the code used to compute the reconstructed growth velocity curves displayed in this article, and is available on the CRAN archive23.

Functional data depth

Given a sample of reconstructed growth velocity curves, it is of interest to look for “outlying” patterns of growth. One way to do this is to use the notion of functional data depth recently developed by López-Pintado and Romo (2009) 5 with the aim of introducing robust methods into functional data analysis. Robust methods are even more relevant in a functional setting than in multivariate problems because outliers can affect functional statistics in more ways and they can be more difficult to detect. For instance, a curve could be an “outlier” without having any unusually large value. This notion of depth is particularly convenient for identifying outliers because shape is also relevant in addition to magnitude. Direct generalization of multivariate depth (discussed in the next section) to functional data often leads to either depths that are computationally intractable or depths that do not take into account some natural properties of the functions, such as shape.

Let x1(t), …, xn(t) be a sample of real-valued functions defined on the time interval [0, T]. The band delimited by these curves is the set of points (t, y) such that xi(t) ≤ yxj (t) for some i, j = 1, …, n. An example for the case of n = 3 curves is provided in Figure 2. The band depth of a function x(t) is then defined as Dn,J (x) = p1 + + pJ, where J ≥ 2 is fixed, and pj is the proportion of bands that contain the graph of x among the bands derived from j curves in the sample. In the sequel we use band depth with J = 3, which is recommended5 for several reasons:1) when J is larger than 3 the index Dn,J can be computationally intensive, 2) bands corresponding to large values of J do not resemble the shape of any of the curves from the sample, 3) the band depth induced order is very stable in J, and 4) the band depth with J = 2 is the easiest to compute, but if two curves cross, the band delimited by them is degenerate at a point and it is unlikely that any other curve will be inside this band.

Figure 2.

Figure 2

A band determined by three curves (the shaded region), as used in the definition of functional data depth.

3 Growth charts and statistical depth

In this section we discuss the use of statistical depth for analyzing multiple growth variables (e.g., head circumference, weight and height) at fixed ages (in contrast to single variables at multiple ages, as studied in the previous section). Statistical depth was first considered for multivariate data to generalize order statistics, ranks and medians to higher dimensions. Given a probability distribution P on k-dimensional Euclidean space, the depth of a k-vector x represents the probability that a random draw from P is “more of an outlier” than x. Various definitions of multivariate depth have been proposed and analyzed24,25,26,27,28,29. The notion has been applied, for instance, as an attempt to extend rank tests to a multivariate context30, in control charts for multivariate processes31, confidence regions32, regression33, and for visualizing sample dispersion34.

Our discussion of multiple output growth charts involves Tukey’s notion of halfspace depth25, which is defined as follows. Consider all hyperplanes running through x: each Π divides ℝk into two closed halfspaces, with probabilities PΠ+ and PΠ-, respectively. Putting PΠ=min(PΠ-,PΠ+), select the hyperplane Π*, say, for which that probability PΠ reaches a minimum: that minimum is called the halfspace depth dP(x) = PΠ* = minΠ PΠ of x with respect to P.

The collection of all points x with given halfspace depth dP(x) is called a depth contour. An empirical version of this definition leads to the construction of empirical depth contours. Just as their population counterparts, empirical depth contours have the attractive geometric property that they enclose nested, convex sets. Besides, empirical depth contours are polytopes, each face of which runs through exactly k sample points (when generated from a continuous distribution P). For k = 1, depth contours reduce to pairs of quantiles of complementary order, τ and 1 − τ, where 0< τ < 1.

Quantile contours

The collection of empirical depth contours provides an interesting picture of the sample at hand, and a powerful data-analytical tool. Unfortunately, however, effective computation of depth contours was based, until recently, on algorithms with prohibitive complexity as k grows, and hardly implementable beyond k = 2 or 3 (although approximate methods are available35,36,37).

Hallin, Paindaveine and Šiman (2010)38 recently established a strong connection between half-space depth and regression quantiles. That connection has two important benefits: a quantile-based interpretation of depth contours, and, perhaps even more importantly, bringing the power of linear programming techniques to the practical computation of empirical contours. Moreover, that connection also opens the way to a tractable definition of (multiple-output) regression depth, and depth-based multiple-output growth charts.

First recall that the classical quantile of order τ, in a univariate sample X1, …, Xn, can be defined as a minimizer of i=1nρτ(Xi-a) over a ∈ ℝ, where ρτ(x) = x(τI [x < 0]) is the check function, and I is the indicator function; in the case τ = 1/2, note that ρτ(x) = |x|/2. This definition of quantiles naturally extends to a k-dimensional sample X1, …, Xn, with the empirical quantile hyperplane of order τ defined as a hyperplane Πτ={x:xk=bτ(x1,,xk-1)+aτ} that minimizes, over (a, b′) ∈ ℝk, the sum

i=1nρτ(Xi,k-a-b(Xi,1,,Xi,k-1))

of vertical weighted deviations, with the kth component representing the vertical direction. Now, choose an arbitrary unit vector uInline graphic, the unit sphere in ℝk, and consider it as the “vertical” direction: the “vertical” component of a vector X is then (u′X)u and, denoting by Γu a k × (k − 1) matrix of column unit vectors such that (u, Γu) constitutes an orthonormal basis of ℝk, we have X=u(uX)+Γu(ΓuX). Letting τ = τu, the directional empirical quantile hyperplane of order τ for direction u is obtained as in the above display, but with u characterizing the vertical direction, yielding a hyperplane Πτ={x:ux=bτΓux+aτ} minimizing, over (a, b′) ∈ ℝk, the sum

i=1nρτ(uXi-a-b(ΓuXi))

of weighted deviations along direction u, with weights (1 − τ) or τ according as Xi lies above or below the hyperplane. Fixed-τ collections of Πτu hyperplanes define polyhedral empirical quantile contours of order τ by means of the intersections of upper halfspaces corresponding to all the quantile hyperplanes of the same quantile level τ. Population versions are obtained in the same way, with sums replaced by mathematical expectations.

Quantile contours can be easily computed by parametric linear programming methods that can handle even samples up to size 500 and dimension k = 5; see Paindaveine and Šiman (2011) 39. The main finding in Hallin, Paindaveine and Šiman (2010) 38 is that halfspace depth contours and quantile contours actually coincide. As a consequence, quantile contours inherit the geometric features of depth contours mentioned earlier, benefit from the interpretation and the analytical features of quantiles, and allow linear programming numerical implementation. Another benefit is the possibility of reconstructing conditional depth/quantile contours via local methods—providing a convincing definition of (multiple output) regression depth contours and paving the way for the construction of multiple output growth charts, as explained in the next section.

Multiple output growth charts

Growth charts are expected to describe the distributions of selected body measurements in children, as a function of age. That description takes the form of a plot of quantiles against age. Existing methods are usually limited to producing marginal growth charts, that is, plots of univariate quantile curves. Such plots clearly omit all information related with dependencies between the various measurements under study. This is regrettable, as growth charts are often used as a diagnostic tool for detecting possible outliers, while a multivariate outlier clearly need not be an outlier from a marginal point of view, and vice versa. A semiparametric approach to multiple-output growth charts has been studied by Wei (2008)40.

The local methods described in a preprint of Hallin, Lu, Paindaveine and Šiman (2011) 6 allow for nonparametric multiple-output growth charts, hence a joint inspection of several measurements as a function of age. Let (ti, Xi), i = 1, …, n, be a random sample of k-dimensional growth measurements Xi, along with the age ti at which each observation was made. We are interested in using these data to infer the depth/quantile contours of X at a given age t0 (which may not be among the observation times). The local constant method consists in computing the weighted depth/quantile hyperplanes Πτt0={xux=bτt0Γux+aτt0} minimizing, over (a, b′) ∈ ℝk, the sum

i=1nwi(t0)ρτ(uXi-a-b(ΓuXi))

with u ranging over the unit sphere Inline graphic; the weights are of the type considered in traditional kernel methods: wi(t0) = K ((tit0)/h)/h for some univariate density K and bandwidth h > 0. For any given t0, this method yields a collection of nested “horizontal” cylinders (with respect to the t-axis), the intersection of which with the hyperplane t = t0 provides a reconstruction of the depth/quantile contours of X at age t0; such an intersection is called a t0-cut (see Fig. 5 for examples). These cuts can be obtained exactly by means of the algorithm and Matlab code presented in Paindaveine and Šiman (2011) 39.

4 Application to CPP data

In this section, we present some examples to illustrate the methods we have introduced. All the examples use data collected from participants in the Collaborative Perinatal Project (CPP) from examinations at the (nominal) ages of birth, four, eight, and twelve months, and three, four, and seven years. Here, by the “nominal” age we mean the targeted age of the measurement; the actual age of the measurement varies around the nominal age.

Example 1: Growth velocity curves

In our first example, we use the following inclusion criteria: female, birthweight 1800–4000 gms, gestational age 37–42 weeks, non-breast-fed, maternal age 20–40 years, the mother did not smoke during pregnancy, complete data on length and actual examination age, and increasing length measurements with age of examination (about 1% of the subjects were excluded under this criterion). This results in a data set of p = 7 height measurements on each of n = 532 subjects. As mentioned in Section 2, this data set is provided in the R package growthrate18, which also includes the code used to produce the growth velocity curves displayed below.

Figure 3 gives the reconstructed growth velocity curves (of length) for two subjects, and for three choices of σ. The choice σ = 1 produces very tight bands, which may be unrealistic because the growth rate is unlikely to have sharp bends at the observation times; the more conservative choices σ = 2 and 3 allow enough flexibility in this regard and appear to be more reasonable. Notice that the σ = 2 and σ = 3 bands bulge between observation times (and this is especially noticeable in the last observation time interval), which is a desirable feature since we would expect greater precision in the estimates close to the observation times.

Figure 4 is based on the notion of band depth defined at the end of Section 2, which allows the ordering of a sample of curves from the center outwards and consequently to define the middle 50% of curves, generalizing the notion of the classical boxplot to functional data. An R package “fbplot” for computing functional boxplots has been developed by Sun and Genton (2011)41. Such plots provide a useful diagnostic tool for detecting unusual patterns in the shape of individual growth velocity curves. In addition, the information provided by data depth could be used to create a variable describing the extent to which a subject has an unusual growth pattern, and used for predicting adult health outcomes. For example, regressing IQ at age seven on the indicator “not in the deepest 50%” and adjusting for various other covariates (birthweight, birthlength and gestational age), suggests that an unusual growth pattern is (negatively) associated with IQ (data not shown).

Figure 4.

Figure 4

Reconstructed growth velocity curves for the whole sample in Example 1 based on σ = 1, 2, 3 in (a,d), (b,e) and (c,f), respectively; the dark line in (a,b,c) is the deepest curve, and the dark bands in (d,e,f) are functional boxplots (representing the deepest 50% of the curves).

Example 2: Bivariate growth charts

This example is based on CPP data for 1775 girls from the Boston site, restricted to subjects having complete data on length, weight and head circumference. The monotonicity of length and head circumference as functions of age was violated (by more than 4 cm for length, and 3 cm for head circumference) by 12 individuals; those 12 highly suspicious observations were excluded, which still left n = 1268 complete records for the analysis.

Figure 5 displays the multiple output growth charts described earlier. The two plots show the bivariate t0-cuts of the growth trajectories of weight (kgs) and head circumference (cms), and length (cms) and head circumference (cms), at five equispaced ages between birth and 7 years. Head circumference is on the vertical axis in each plot. Clearly there is a much higher correlation between the pairs of variables at earlier ages than at later ages, especially in the left panel. These plots provide a useful diagnostic tool for detecting unusual patterns of growth in combinations of variables, and that might not be noticed in standard growth charts that examine each variable separately. For example, these pictures illustrate age dependence of both the correlation structure and the ratios of the plotted characteristics that could not be detected from the marginal univariate growth charts. The two plots also clearly show that marginal outliers need not be multivariate outliers and vice versa. Consequently, bivariate growth charts would rightly diagnose some children with small head circumference and small length or weight as normal even when the univariate growth charts indicated the contrary. Needless to say, depth contours could be constructed for any age, e.g., for the reference ages or for the age of a child under particular investigation.

5 Conclusion

We have proposed various new nonparametric methods for the analysis of growth trajectories, bringing greater flexibility as well as ease of implementation to existing approaches. For the CPP data set, these methods can lead to interesting findings about early childhood growth patterns. First we reconstructed growth velocity curves using an empirical Bayes technique that adapts to data sparsity and gives a way of assessing uncertainty in the reconstruction. Second, we discussed the use of functional data depth and functional boxplots which provide useful diagnostic tools for detecting unusual patterns in growth trajectories. Finally, using regression quantiles and Tukey’s notion of data depth, we proposed flexible and robust growth charts for multiple variables.

Acknowledgments

The research of Ian McKeague was supported in part by NIH grant R01 GM095722 (PI: McKeague) and NIH grant P01 AG 023028-01; he also gratefully acknowledges the help and encouragement of Ezra Susser. The research of Sara López-Pintado was supported by Spanish Ministry of Education and Science grant SEJ2007-67734. The research of Marc Hallin was supported by Sonderforschungsbereich “Statistical modelling of nonlinear dynamic processes” (SFB 823) of the Deutsche Forschungsgemeinschaft, and by a Discovery Grant of the Australian Research Council. The research of Miroslav Šiman was supported by Project 1M06047 of the Ministry of Education, Youth and Sports of the Czech Republic.

Footnotes

Statement of Interest: None declared.

References

  • 1.Dasgupta P, Hauspie R. Perspectives in Human Growth, Development and Maturation. Kluwer Academic Publishers; 2001. [Google Scholar]
  • 2.Sumiya T, Tashima T, Nakahara H, Shohoji T. Relationships between biological parameters of Japanese growth of height. Environmetrics. 2001;12:367–382. [Google Scholar]
  • 3.Li N, Das K, Wu R. Functional mapping of human growth trajectories. Journal of Theoretical Biology. 2009;261(1):33–42. doi: 10.1016/j.jtbi.2009.07.020. [DOI] [PubMed] [Google Scholar]
  • 4.López-Pintado S, McKeague IW. Recovering gradients from sparsely observed functional data. Biometrics. 2011 doi: 10.1111/biom.12011. under revision. http://www.columbia.edu/~im2131/ps/growthrate-package-reference.pdf. [DOI] [PMC free article] [PubMed]
  • 5.López-Pintado S, Romo J. On the concept of depth for functional data. Journal of the American Statistical Association. 2009;104(486):718–734. [Google Scholar]
  • 6.Hallin M, Lu Z, Paindaveine D, Šiman M. Local bilinear multiple-output quantile regression and regression depth. 2011 Preprint. [Google Scholar]
  • 7.Ferraty F, Vieu P. Nonparametric Functional Data Analysis. Springer; New York: 2006. [Google Scholar]
  • 8.Ramsay JO, Silverman BW. Functional Data Analysis. Springer; New York: 2005. [Google Scholar]
  • 9.Hall P, Müller HG, Wang JL. Properties of principal components methods for functional and longitudinal data analysis. Annals of Statistics. 2006;34:1493–1517. [Google Scholar]
  • 10.James G, Hastie TJ, Sugar CA. Principal component models for sparse functional data. Biometrika. 2000;87:587–602. [Google Scholar]
  • 11.Rice J, Wu C. Nonparametric mixed effects models for unequally sampled noisy curves. Biometrics. 2000;57:253–259. doi: 10.1111/j.0006-341x.2001.00253.x. [DOI] [PubMed] [Google Scholar]
  • 12.Yao F, Müller HG, Wang JL. Functional data analysis for sparse longitudinal data. Journal of the American Statistical Association. 2005;100:577–590. [Google Scholar]
  • 13.Yao F, Müller HG, Wang JL. Functional linear regression analysis for longitudinal data. Annals of Statistics. 2005;33:2873–2903. [Google Scholar]
  • 14.Kirsch A. Applied Mathematical Sciences. Springer-Verlag; New York: 1996. An Introduction to the Mathematical Theory of Inverse Problems, volume 120 of. [Google Scholar]
  • 15.Nashed MZ, Wahba G. Convergence rates of approximate least squares solutions of linear integral and operator equations of the first kind. Mathematics of Computation. 1974;28:69–80. [Google Scholar]
  • 16.Wahba G. Practical approximate solutions to linear operator equations when the data are noisy. SIAM Journal on Numerical Analysis. 1977;14(4):651–667. [Google Scholar]
  • 17.Liu B, Müller HG. Estimating derivatives for samples of sparsely observed functions, with application to online auction dynamics. Journal of the American Statistical Association. 2009;104:704–717. [Google Scholar]
  • 18.López-Pintado S, McKeague IW. Growthrate: Bayesian reconstruction of growth velocity. R package version 1.0. 2011 http://CRAN.R-project.org/package=growthrate.
  • 19.Shohoji T, Kanefuji K, Sumiya T, Qin T. A prediction of individual growth of height according to an empirical Bayesian approach. Annals of the Institute of Statistical Mathematics. 1991;43:607–619. [Google Scholar]
  • 20.Barry D. A Bayesian model for growth curve analysis. Biometrics. 1995;51(2):639–655. [PubMed] [Google Scholar]
  • 21.Arjas E, Liu L, Maglaperidze N. Prediction of growth: a hierarchical Bayesian approach. Biometrical Journal. 1997;39:741–759. [Google Scholar]
  • 22.Cai T, Liu W, Luo X. A constrained ℓ1 minimization approach to sparse precision matrix estimation. Journal of the American Statistical Association. 2011;106:594–607. [Google Scholar]
  • 23.R Development Core Team. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing; Vienna, Austria: 2011. http://www.R-project.org/ [Google Scholar]
  • 24.Mahalanobis PC. On the generalized distance in statistics. Proceedings of the Natural Academy of Science of India. 1936;13:1305–1320. [Google Scholar]
  • 25.Tukey JW. Mathematics and the picturing of data. Proceedings of the International Congress of Mathematicians; Vancouver, B C. 1974; 1975. pp. 523–531. Canad. Math. Congress, Montreal, Que. [Google Scholar]
  • 26.Oja H. Descriptive statistics for multivariate distributions. Statistics and Probability Letters. 1983;1:327–332. [Google Scholar]
  • 27.Liu R. On a notion of data depth based on random simplices. Annals of Statistics. 1990;18:405–414. [Google Scholar]
  • 28.Fraiman R, Meloche J. Multivariate L-estimation. Test. 1999;8:255–317. [Google Scholar]
  • 29.Zuo Y, Serfling RJ. General notions of statistical depth function. Annals of Statistics. 2000;28:461–482. [Google Scholar]
  • 30.Liu R, Singh K. A quality index based on data depth and multivariate rank test. Journal of the American Statistical Association. 1993;88:257–260. [Google Scholar]
  • 31.Liu R. Control charts for multivariate processes. Journal of the American Statistical Association. 1995;90:1380–1388. [Google Scholar]
  • 32.Yeh A, Singh K. Balanced confidence sets based on Tukey depth. Journal of the Royal Statistical Society Ser B. 1997;3:639–652. [Google Scholar]
  • 33.Rousseeuw P, Leroy AM. Robust Regression and Outlier Detection. Wiley; New York: 1987. [Google Scholar]
  • 34.Liu R, Parelius JM, Singh K. Multivariate analysis by data depth: Descriptive statistics, graphics and inference. Annals of Statistics. 1999;27:783–858. [Google Scholar]
  • 35.Zuo Y. Multidimensional trimming based on projection depth. Annals of Statistics. 2006;34(5):2211–2251. [Google Scholar]
  • 36.Cuesta-Albertos JA, Nieto-Reyes A. The random Tukey depth. Computational Statistics and Data Analysis. 2008;52(11):4979–4988. [Google Scholar]
  • 37.Kong L, Mizera I. Quantile tomography: using quantiles with multivariate data. 2010 Preprint, arXiv:0805.0056v1. [Google Scholar]
  • 38.Hallin M, Paindaveine D, Šiman M. Multivariate quantiles and multiple-output regression quantiles: from L1 optimization to halfspace depth. Annals of Statistics. 2010;38(2):635–669. [Google Scholar]
  • 39.Paindaveine D, Šiman M. Computing multiple-output regression quantile regions. Computational Statistics and Data Analysis. 2011 to appear. [Google Scholar]
  • 40.Wei Y. An approach to multivariate covariate-dependent quantile contours with application to bivariate conditional growth charts. Journal of the American Statistical Association. 2008;103(481):397–409. [Google Scholar]
  • 41.Sun Y, Genton MG. Functional boxplots. Journal of Computational and Graphical Statistics. 2011;20:316–334. [Google Scholar]

RESOURCES