Skip to main content
Entropy logoLink to Entropy
. 2019 Dec 31;22(1):55. doi: 10.3390/e22010055

Estimation of Dynamic Networks for High-Dimensional Nonstationary Time Series

Mengyu Xu 1, Xiaohui Chen 2, Wei Biao Wu 3,*
PMCID: PMC7516486  PMID: 33285830

Abstract

This paper is concerned with the estimation of time-varying networks for high-dimensional nonstationary time series. Two types of dynamic behaviors are considered: structural breaks (i.e., abrupt change points) and smooth changes. To simultaneously handle these two types of time-varying features, a two-step approach is proposed: multiple change point locations are first identified on the basis of comparing the difference between the localized averages on sample covariance matrices, and then graph supports are recovered on the basis of a kernelized time-varying constrained L1-minimization for inverse matrix estimation (CLIME) estimator on each segment. We derive the rates of convergence for estimating the change points and precision matrices under mild moment and dependence conditions. In particular, we show that this two-step approach is consistent in estimating the change points and the piecewise smooth precision matrix function, under a certain high-dimensional scaling limit. The method is applied to the analysis of network structure of the S&P 500 index between 2003 and 2008.

Keywords: high-dimensional time series, nonstationarity, network estimation, change points, kernel estimation

1. Introduction

Networks are useful tools to visualize the relational information among a large number of variables. An undirected graphical model belongs to a rich class of statistical network models that encodes conditional independence [1]. Canonically, Gaussian graphical models (or their normalized version partial correlations [2]) can be represented by the inverse covariance matrix (i.e., the precision matrix), where a zero entry is associated with a missing edge between two vertices in the graph. Specifically, two vertices are not connected if and only if they are conditionally independent, given the value of all other variables.

On one hand, there is a large volume of literature on estimating the (static) precision matrix for graphical models in the high-dimensional setting, where the sample size and the dimension are both large [3,4,5,6,7,8,9,10,11,12,13,14,15,16]. Most of the earlier work along this line assumes that the underlying network is time-invariant. This assumption is quite restrictive in practice and hardly plausible for many real-world applications, such as gene regulatory networks, social networks, and stocking market, where the underlying data generating mechanisms are often dynamic. On the other hand, dynamic random networks have been extensively studied from the perspective of large random graphs, such as community detection and edge probability estimation for dynamic stochastic block models (DSBMs) [17,18,19,20,21,22,23,24,25,26,27,28,29,30]. Such approaches do not model the sampling distributions of the error (or noise), since the “true” networks are connected with random edges sampled from certain probability models, such as the Erdős–Rényi graphs [31] and random geometric graphs [32].

In this paper, we view the (time-varying) networks of interests as non-random graphs. We adopt the graph signal processing approach for denoising the nonstationary time series and target on estimating the true unknown underlying graphs. Despite the recent attempts towards more flexible time-varying models [33,34,35,36,37,38,39,40], there are still a number of major limitations in the current high-dimensional literature. First, theoretical analysis was derived under the fundamental assumption that the observations are either temporally independent, or the temporal dependence has very specific forms, such as Gaussian processes or (linear) vector autoregression (VAR) [14,33,34,37,41,42,43]. Such dynamic structures are unduly demanding in view that many time series encountered in real applications have very complex nonlinear spatial-temporal dependency [44,45]. Second, most existing work assumes the data have time-varying distributions with sufficiently light tails, such as Gaussian graphical models and Ising models [33,34,36,41,42]. Third, in change point estimation problems for high-dimensional time series, piecewise constancy is widely used [41,42,46,47], which can be fragile in practice. For instance, financial data often appears to have time-dependent cross-volatility with structural breaks [48]. For resting-state fMRI signals, correlation analysis reveals both slowly varying and abruptly changing characteristics corresponding to modularities in brain functional networks [49,50].

Advances in analyzing high-dimensional (stationary) time series have been made recently to address the aforementioned nonlinear spatial-temporal dependency issue [14,37,43,51,52,53,54,55,56,57]. In [53,56,57], the authors considered the theoretical properties of regularized estimation of covariance and precision matrices, based on various dependence measures of high-dimensional time series. Reference [38] considered the non-paranormal graphs that evolve with a random variable. Reference [37] discussed the joint estimation of Gaussian graphical models based on a stationary VAR(1) model with special coefficient matrices, which may also depend on certain covariates. The authors applied a constrained L1-minimization for inverse matrix estimation (CLIME) estimator with a kernel estimator of covariance matrix and developed consistency in the graph recovery at a given time point. Reference [14] studied the recovery of the Granger causality across time and nodes assuming a stationary Gaussian VAR model with unknown order.

In this paper, we focus on the recovery of time-varying undirected graphs on the basis of the regularized estimation of the precision matrices for a general class of nonstationary time series. We simultaneously model two types of dynamics: abrupt changes with an unknown number of change points and the smooth evolution between the change points. In particular, we study a class of high-dimensional piecewise locally stationary processes in a general nonlinear temporal dependency framework, where the observations are allowed to have a finite polynomial moment.

More specifically, there are two main goals of this paper: first, to estimate the change point locations, as well as the number of change points, and second, to estimate the smooth precision matrix functions between the change points. Accordingly, our proposed method contains two steps. In the first step, the maximum norm of the local difference matrix is computed at each time point and the jumps in the covariance matrices are detected at the location where the maximum norms are above a certain threshold. In the second step, the precision matrices before and after the jump are estimated by a regularized kernel smoothing estimator. These two steps are recursively performed until a stopping criterion is met. Moreover, a boundary correction procedure based on data reflection is considered to reduce the bias near the change point.

We provide an asymptotic theory to justify the proposed method in high dimensions: point-wise and uniform rates of convergence are derived for the change point estimation and graph recovery under mild and interpretable conditions. The convergence rates are determined via subtle interplay among the sample size, dimensionality, temporal dependence, moment condition, and the choice of bandwidth in the kernel estimator. Our results are significantly more involved than problems for sub-Gaussian tails and independent samples. We highlight that uniform consistency in terms of time-varying network structure recovery is much more challenging and difficult than pointwise consistency. For the multiple change point detection problem, we also characterize the threshold of the difference statistic that gives a consistent selection of the number of change points.

We fix some notations: Positive, finite, and non-random constants, independent of the sample size n and dimension p, are denoted by C,C1,C2,, whose values may differ from line to line. For the sequence of real numbers, an and bn, we write an=O(bn) or anbn if limsupn(an/bn)C for some constant C< and an=o(bn) if limn(an/bn)=0. We say anbn if an=O(bn) and bn=O(an). For a sequence of random variables Yn and a corresponding set of constants an, denote Yn=OP(an) if for any ε>0 there is a constant C>0 such that P(|Yn|/an>C)<ε for all n. For a vector xRp, we write |x|=(j=1pxj2)1/2. For a matrix Σ, |Σ|1=j,k|σjk|, |Σ|=maxj,k|σjk|, |Σ|L1=maxkj|σjk|, |Σ|F=(j,kσjk2)1/2 and ρ(Σ)=max{|Σx|:|x|=1}. For a random vector zRp, write zLa, a>0, if za=:[E(|z|a)]1/a<. Let z=z2. Denote ab=min(a,b) and ab=max(a,b).

The rest of the paper is organized as follows: Section 2 presents the time series model, as well as the main assumptions, which can simultaneously capture the smooth and abrupt changes. In Section 3, we introduce the two-step method that first segments the time series based on the difference between the localized averages on sample covariance matrices and then recovers the graph support based on a kernelized CLIME estimator. In Section 4, we state the main theoretical results for the change point estimation and support recovery. Simulation examples are presented in Section 5 and a real data application is given in Section 6. Proof of main results can be found in Section 7.

2. Time Series Model

We first introduce a class of causal vector stochastic processes. Next, we state the assumptions to derive an asymptotic theory in Section 4 and explain their implications. Let εiRp,iZ be independent and identically distributed (i.i.d.) random vectors and Fi=(,εi1,εi) be a shift process. Let Xi(t)=(Xi1(t),,Xip(t)) be a p-dimensional nonstationary time series generated by

Xi(t)=H(Fi;t), (1)

where H(·;·)=(H1(·;·),,Hp(·;·)) is an Rp-valued jointly measurable function. Suppose we observe the data points Xi=Xi,n=Xi(ti) at the evenly spaced time intervals ti=i/n,i=1,2,,n,

Xi,n=H(Fi;i/n). (2)

We drop the subscription n in Xi,n in the rest of this section. Since our focus is to study the second-order properties, the data is assumed to have a mean of zero.

Model (1) is first introduced in [58]. The stochastic process Xi(t)iZ,t[0,1) can be thought of as a triangular array system, double indexed by i and t, while the observations (Xi)i=1n are sampled from the diagonal of the array. On one hand, when fixing the time index t, the (vertical) process Xi(t)iZ is stationary. On the other hand, since H(Fi;ti) is allowed to vary with ti, the diagonal process (2) is able to capture nonstationarity.

The process (Xi)iZ is causal or non-anticipative as Xi is an output of the past innovations (εj)ji and does not depend on future innovations. In fact, it covers a broad range of linear and nonlinear, stationary and non-stationary processes, such as vector auto-regressive moving average processes, locally stationary processes, Markov chains, and nonlinear functional processes  [53,58,59,60,61].

Motivated by real applications where nonstationary time series data can involve both abrupt breaks and smooth varies between the breaks, we model the underlying processes as piecewise locally stationary with a finite number of structural breaks.

Definition 1 (Piecewise locally stationary time series model).

Define PLSι([0,1],L) as the collection of mean-zero piecewise locally stationary processes on [0,1], if for each (X(t))0t1PLSι([0,1],L), there is a nonnegative integer ι such that X(t) is piecewise stochastic Lipschitz continuous in t with Lipschitz constant L on the interval [t(l),t(l+1)),l=0,,ι, where 0=t(0)<t(1)<t(ι)<t(ι+1)=1. A vector stochastic process (X(t))0t1PLSι([0,1],L) if all coordinates belong to PLSι([0,1],L). For the process (X0(t))0t1 defined in (1), this means that there exists a non-negative integer ι and a constant L>0, such that

max1jpHj(F0;t)Hj(F0;t)L|tt|for allt(l)t,t<t(l+1),0lι.

Remark 1.

If we assume (Xi(t))0t1PLSι([0,1],L),iZ, then it follows that for each i=ik,,i+k, where k/n0, and that t(l)i,i<t(l+1) for some 0lι, we have

max1jpHj(Fi;i/n)Hj(Fi;i/n)Lk/n=o(1).

In other words, within a locally stationary time period, in a local window of i, (Xij)ikii+k can be approximated by the stationary process (Xij(i/n))ikii+k for each j=1,,p. This justifies the terminology of local stationarity.

The covariance matrix function of the underlying process is Σ(t)=σjk(t)1j,kp, t[0,1], where σjk(t)=E(Hj(F0;t)Hk(F0;t)), and the precision matrix function is Ω(t)=Σ(t)1=ωjk(t)1j,kp. The graph at time t is denoted by G(t)=(V,E(t)), where V is the vertex set and E(t)={(j,k):ωjk(t)0}. Note that (Xi(t))tPLSι([0,1],L),iZ implies piecewise Lipschitz continuity in Σ(t) except at the breaks t(1),,t(ι). In particular, if sup0t1max1jpHj(F0;t)C for some constant C>0, then

|Σ(s)Σ(t)|2CL|st|,s,t[t(l),t(l+1)),l=0,,ι. (3)

The reverse direction is not necessarily true, i.e., (3) does not indicate (Xi(t))tPLSι([0,1],L), iZ in general. As a trivial example, let εij=21/2 with probability 2/3 and 2 with probability 1/3 i.i.d for all i,j. At time tk=k/n, let Xij(tk)=(1)ktkεij. Then for any k and k such that k+k is odd, |Σ(tk)Σ(tk)|=|tktk|, while X01(tk)X01(tk)2=tk+tk.

Assumption 1 (Piecewise smoothness).

(i) Assume (Xi(t))0t1PLSι([0,1],L) for each iZ, where L>0 and ι0 are constants independent of n and p. (ii) For each l=0,,ι, and 1j,kp, we have σjk(t)C2[t(l),t(l+1)).

Now we introduce the temporal dependence measure. We quantify the dependence of Xi(t)iZ by the dependence adjusted norm (DAN) (cf. [62]). Let εi be an independent copy of εi and Fi,{m}=(,εim1,εim,εim+1,,εi). Denote Xi,{m}(t)=Xi1,{m}(t),,Xip,{m}(t), where Xij,{m}(t)=Hj(Fi,{m};t), 1jp. Here Xi,{m}(t) is a coupled version of Xi(t), with the same generating mechanism and input, except that εim is replaced by an independent copy εim.

Definition 2 (Dependence adjusted norm (DAN)).

Let constants a1,A>0. Assume sup0t1X1j(t)a<,j=1,,p. Define the uniform functional dependence measure for the sequences (Xij(t))iZ,t[0,1] of form (1) as

θm,a,j=sup0t1Xij(t)Xij,{m}(t)a,j=1,,p,

and Θm,a,j=i=mθi,a,j. The dependence adjusted norm of (Xij(t))iZ,t[0,1] is defined as

X·,ja,A=supm0(m+1)AΘm,a,j,

whenever X·,ja,A<.

Intuitively, the physical dependence measure quantifies the adjusted stochastic difference between the random variable and its coupled version by replacing past innovations. Indeed, θm,a,j measures the impact on Xij(t) uniform over t by replacing εim while freezing all the other inputs, while Θm,a,j quantifies the cumulative influence of replacing εm on (Xij(t))i0 uniform over t. Then X·,ja,A controls the uniform polynomial decay in the lag of the cumulative physical dependence, where a depends on the the tail of marginal distributions of X1,j(t) and A quantifies the polynomial decay power and thus the temporal dependence strength. It is clear that X·,ja,A is a semi-norm, i.e., it is subaddative and absolutely homogeneous.

Assumption 2 (Dependence and moment conditions).

Let Xi(t) be defined in (1) and Xi in (2). There exist q>2 and A>0 such that

ν2q:=supt[0,1]max1jpE|Xj(t)|2q<andNX,2q:=max1jpX·,j2q,A<. (4)

We let MX,q:=1jpX·,j2q,Aq1/q and write NX=NX,4, MX=MX,2. The quantities MX,q and NX,2q measure the Lq-norm aggregated effect and the largest effect of the element-wise DANs respectively. Both quantities play a role in the convergence rates of our estimator.

Obviously, we have XijXij,{m}aθm,a,j and max1jpE|Xij|2qν2q for all 1in. In contrast to other works in a high-dimensional covariance matrix and network estimation, where sub-Gaussian tails and independence are the keys to ensure consistent estimation. Assumption 2 only requires that the time series have a finite polynomial moment, and it allows linear and nonlinear processes with short memory in the time domain.

Example 1 (Vector linear process).

Consider the following vector linear process model

H(Fi;t)=m=0Am(t)εim,

where εi=(ε1,,εp) and εij are i.i.d. with mean 0 and variance 1, and εijqCq for each iZ and 1jp with some constants q>2 and Cq>0. The vector linear process is commonly seen in literature and application [63]. It includes the time-varying VAR model where Am(t)=A(t)m as a special example.

Suppose that the coefficient matrices Am(t)=(am,jk(t))1j,kp,m=0,1, satisfy the following condition.

  • (A1) 

    For each 1j,kp, am,jk(t)C2[0,1].

  • (A2) 

    For each 1jp, there is a constant CA,j>0 such that for each t[0,1], k=1pam,jk(t)2CA,j(m+1)2(A+1) for all m0.

  • (A3) 

    For any t,t[0,1], m=0k=1p[am,jk(t)am,jk(t)]2L2|tt|2 for each j=1,,p.

Note that

σjk(t)=m0Am,j·(t)Am,k·(t),Θm,q,j2Cqq1m=0(Am,j·Am,j·)1/2,Xij(t)Xij(t)2=m=0Am,j·k=1p[am,jk(t)am,jk(t)]2,

where Am,j·(t) is the jth row of Am(t). Under conditions (A1)–(A3), one can easily verify that for each 1j,kp, the process satisfies: (1) σjk(t)C2[0,1]; (2) X·,jq,ACq(q1)CA,j (due to Burkholder’s inequality, cf. [64]); (3) Hj(F0;t)Hj(F0;t)L|tt|.

Conditions (A1)–(A3) implicitly impose smoothness in each entry of the coefficient matrices, sparseness in each column of the entry and evolution, and polynomial decay rate in the lag m of each entry and its derivative.

For 1lι, let δjk(t(l)):=σjk(t(l))σjk(t(l)) and Δ(t(l))=δjk(t(l))1j,kp, where σjk(t(l))=limtt(l)σjk(t) is well-defined in view of (3). We assume that the change points are separated and sizeable.

Assumption 3 (Separability and sizeability of change points).

There exist positive constants c1(0,1) and c2>0 independent of n and p such that max0lι(t(l+1)t(l))c1 and δ(tl):=|Δ(tl)|c2.

In the high-dimensional context, we assume that the inverse covariance matrices are sparse in the sense of their L1 norms.

Assumption 4 (Sparsity of precision matrices).

The precision matrix |Ω(t)|L1κp for each t[0,1], where κp is allowed to grow with p.

If we further assume that the eigenvalues of the covariance matrices are bounded from below and above, i.e., there exists a constant 0<c<1, such that cinft[0,1]|Σ(t)|2supt[0,1]|Σ(t)|2c1, then the covariance matrices and precision matrices are well-conditioned. In particular, as |Ω(t)Ω(t)|c2|Σ(t)Σ(t)|, a small perturbation in the covariance matrix would guarantee a small change of the same order in the precision matrix under the spectral norm.

3. Method: Change Point Estimation and Support Recovery

In graphical models (such as the Gaussian graphical model or partial correlation graph), network structures relevant to correlations or partial correlations are second-order characteristics of the data distributions. Specifically, the existence of edges coincides with non-zero entries of the inverse covariance matrix. We consider the dynamics of time series with both structural breaks and smooth changes. The piecewise stochastic Lipschitz continuity in Definition 1 allows the time series to have discontinuity in the covariance matrix function at time points t(l),l=1,,ι (i.e., change points), while only smooth changes (i.e., twice continuous differentiability of the covariance matrix function in Assumptions 1) can occur between the change points.

In the presence of change points, we must first remove the change points before applying any smoothing procedures since |Ω(t)Ω(t)||Σ(t)|L11|Σ(t)|L11|Δ(t)|, i.e., a non-negligible abrupt change in the covariance matrix will result in a substantial change of the graph structure for sparse and smooth covariance matrices. Thus our proposed graph recovery method consists of two steps: change point detection and support recovery.

Let hhn>0 be a bandwidth parameter such that h=o(1) and n1=o(h), and Dh(0)={h,h+1/n,,1h} be a search grid in (0,1). Define

D(s)=n1i=0hn1XnsiXnsii=1hnXns+iXns+i,sDh(0). (5)

To estimate the change points, compute

s^1=argmaxsDh(0)|D(s)|. (6)

The following steps are performed recursively. For l=1,2,, let

Dh(l)=Dh(l1){s^l2h,,s^l+2h}c, (7)
s^l+1=argmaxsDh(l)|D(s)|, (8)

until the following criterion is attained:

maxsDh(l)|D(s)|<ν, (9)

where ν is an early stopping threshold. The value of ν is determined in Section 4, which depends on the dimension and sample size, as well as the serial dependence level, tail condition, and local smoothness. Since our method only utilizes data in the localized neighborhood, multiple change points can be estimated and ranked in a single pass, which offers some computational advantage than the binary segmentation algorithm [41,46].

Once the change points are claimed, in the second step, we consider recovering the networks from the locally stationary time series before and after the structural breaks. In [11], where Xi,i=1,,n are assumed with an identical covariance matrix, the precision matrix Ω^ is estimated as,

Ω^λ=argminΩRp×p|Ω|1s.t.|Σ^ΩIdp|λ, (10)

where Σ^ is the sample covariance matrix. Inspired by (10), we apply a kernelized time-varying (tv-) CLIME estimator for the covariance matrix functions of the multiple pieces of locally stationary processes before and after the structural breaks. Let

Σ^(t)=i=1nw(t,ti)XiXi, (11)

where

w(t,i)=Kb(ti,t)i=1nKb(ti,t) (12)

and Kb(u,v)=K(|uv|/b)/b. The bandwidth parameter b satisfies that b=o(1) and n1=o(b). Denote Bn=nb. The kernel function K(·) is chosen to have properties as follows.

Assumption 5 (Regularity of kernel function).

The kernel function K(·) is non-negative, symmetric, and Lipschitz continuous with bounded support in [1,1], and that 11K(u)du=1.

Assumption 5 is a common requirement on the kernel functions and can be fulfilled by a range of kernel functions, such as the uniform kernel, triangular kernel, and the Epanechnikov kernel. Now the tv-CLIME estimator of the precision matrix Ω(t) is defined by Ω˜(t)=ω˜jk(t)1j,kp, where ω˜jk(t)=min(ω^jk(t),ω^kj(t)), and Ω^(t)Ω^λ(t)=(ω^jk(t))1j,kp,

Ω^λ(t)=argminΩRp×p|Ω|1s.t.|Σ^(t)ΩIdp|λ. (13)

Similar hybridized kernel smoothing and the CLIME method for estimating the sparse and smooth transition matrices in high-dimensional VAR model has been considered in [65], where change point is not considered. Thus in the current setting we need to carefully control effect of (consistently) removing the change points before smoothing.

Then, the network is estimated by the “effective support” defined as follows.

G^(t;u)=(g^jk(t;u))1j,kp,whereg^jk(t;u)=I|ω˜jk(t)|u. (14)

It should be noted that the (vanilla) kernel smoothing estimator (11) of the covariance matrix does not adjust for the boundary effect due to the change points in the covariance matrice function. Thus, in the neighborhood of the change points, a larger bias can be induced in estimating Σ(t) by Σ^(t). As a remedy, we apply the following reflection procedure for boundary correction. Suppose tT^b+h2(j) for 1jι, Denote T^d(j):=[s^jd,s^j+d) for d(0,1). We replace (11) by

Σ^(t)=i=1nw(t,ti)x˘ix˘i,

and then apply the rest of the tv-CLIME approach. Here

x˘i=xiif (is^jn)(ts^jn)0;x2s^jniotherwise. (15)

4. Theoretical Results

In this section, we derive the theoretical guarantees for the change point estimation and graph support recovery. Roughly speaking, Proposition 1 and 2 below show that under appropriate conditions, if each element of the covariance matrix varies smoothly in time, one can obtain an accurate snapshot estimation of the precision matrices as well as the time-varying graphs with high probability via the proposed kernel smoothed constrained l1 minimization approach.

Define Jq,A(n,p)=MX,q(pϖq,A(n))1/q, where ϖq,A(n)=n,n(logn)1+2q,nq/2Aq if A>1/21/q, A=1/21/q, and 0<A<1/21/q, respectively.

Proposition 1 (Rate of convergence for estimating precision matrices: pointwise and uniform).

Suppose Assumptions 2, 4, and 5 hold with ι=0. Let Bn=bn for n1=o(b) and b=o(1).

  • (i) 
    Pointwise. Choose the parameter λCκp(b2+Bn1Jq,A(Bn,p)+NX(logp/Bn)1/2) in the tv-CLIME estimator Ω^λ(t) in (13), where C is a sufficiently large constant independent of n and p. Then for any t[b,1b], we have
    |Ω^λ(t)Ω(t)|=OP(κpλ). (16)
  • (ii) 
    Uniform. Choose λCκpb2+Bn1Jq,A(n,p)+NXBn1(nlog(p))1/2 in the tv-CLIME estimator Ω^λ(t) in (13), where C is a sufficiently large constant independent of n and p. Then we have
    supt[b,1b]|Ω^λ(t)Ω(t)|=OP(κpλ). (17)

The optimal order of the bandwidth parameter b=b in (17) is the solution to the following equation:

b2=Bn1max(Jq,A(n,p),NX(nlog(p2))1/2),

which implies that the closed-form expression for b is given by

b=C1n1Jq,A(n,p)1/3+C2NX1/3n1/6log(p)1/6

for some constants C1 and C2 that are independent of n and p.

Given a finite sample, to distinguish the small entries in the precision matrix from the noise is challenging. Since a smaller magnitude of a certain element of the precision matrix implies a weaker connection of the edge in the graphical model, we instead consider the estimation of significant edges in the graph. Define the set of significant edges at level u as E*(t;u)=(j,k):gjk*(t;u)0, where

gjk*(t;u)=I|ωjk(t)|>u.

Then, as a consequence of (17), we have the following support recovery consistency result.

Proposition 2 (Consistency of support recovery: significant edges).

Choose u as u=C0κp2b2, where C0 is taken as a sufficiently large constant independent of n and p. Suppose that u=o(1) as n,p. Then under conditions of Proposition 1, we have that as n,p,

Psupt[b,1b](j,k)Ec(t)Ig^jk(t;u)000, (18)
Psupt[b,1b](j,k)E*(t;2u)Ig^jk(t;u)=000. (19)

Proposition 2 shows that the pattern of significant edges in the time-varying true graphs G(t),t[b,1b], can be correctly recovered with high probability. However, it is still an open question to what extent the edges with magnitude below u can be consistently estimated, which can be naturally studied in the multiple hypothesis testing framework. Nonetheless, hypothesis testing for graphical models on the nonstationary high-dimensional time series is rather challenging. We leave it as a future problem.

Propositions 1 and 2 together yield that the consistent estimation of the precision matrices and the graphs can be achieved before and after the change points. Now, we provide the theoretical result of the change point estimation. Theorem 1 below shows that if the change points are separated and sizable, then we can consistently identify them via the single pass segmentation approach under suitable conditions. Denote

h=C1n1Jq,A(n,p)1/3+C2NX1/3n1/6log(p)1/6,

where C1 and C2 are constants independent of n and p.

Theorem 1 (Consistency of change point estimation).

Assume XiRp admits the form (2). Suppose that Assumptions 2 to 3 are satisfied. Choose the bandwidth h=h, and ν=(1+L)h2 in (5) and (9) respectively. Assume that h=o(1) as n,p. We find that there exist constants C1,C2,C3 independent of n and p, such that

P(|ι^ι|>0)C1pϖq,A(n)MX,qqν2qqnqc2q1/3+C2p2expC3(nlog2(p)NX2)1/3. (20)

Furthermore, in the event {ι=ι^}, the ordered change-point estimator (s^(1)<s^(2)<<s^(ι^)) defined in (7) satisfies

max1jι|s^(j)t(j)|=OP(h2). (21)

Proposition 2 and Theorem 1 together indicate the consistency in the snapshot estimation of the time-varying graphs before and after the change points. In a close neighborhood of the change points, we have the following result for the recovery of the time-varying network. Denote S:=b,1b](1jι^T^h2+bc(j) as the time intervals between the estimated change points, and N:=[0,b)1jι^(T^h2+bT^h2c)(1b,1] as the recoverable neighborhood of the jump.

Theorem 2.

Let Assumptions 2 to 5 be satisfied. We have the following results as n,p.

  • (i) 
    Between change points. For tS, take b=b and u=u, where b and u are defined in Proposition 2. Suppose u=o(1). We have
    suptSmaxj,k|σ^j,k(t)σj,k(t)|=OP(b2). (22)
    Choose the penalty parameter as λ:=C1κpb2, where C1 is a constant independent of n and p. Then
    suptS|Ω^λ(t)Ω(t)|=OP(κp2b2).
    Moreover,
    PsuptS(j,k)Ec(t)Ig^j,k(t;u)0=01, (23)
    PsuptS(j,k)E*(t;2u)Ig^jk(t;u)=0=01. (24)
  • (ii) 
    Around change points. For sN, take b=b:=C1n1Jq,A(n,p)1/2+C2NX1/2n1/4log(p)1/4, and u=u:=C0κp2b, where C0, C1 and C2 are constants independent of n and p. Suppose u=o(1). We have
    suptNmaxj,k|σ^j,k(t)σj,k(t)|=OP(b).
    Choose the penalty parameter as λ:=C1κpb, where C1 is a constant independent of n and p. Then
    suptN|Ω^λ(t)Ω(t)|=OP(κp2b). (25)
    Moreover,
    PsuptN(j,k)Ec(t)Ig^j,k(t;u)0=01, (26)
    PsuptN(j,k)E*(t;2u)Ig^j,k(t;u)=0=01. (27)

Note that the convergence rates for the covariance matrix entries and precision matrix entries in case (ii) around the jump locations are slower than those for points well separated from the jump locations in case (i). This is because on the boundary due to the reflection, the smooth condition may no longer hold true. Indeed, we only take advantage of the Lipschitz continuous property of the covariance matrix function. Thus, we lose one degree of regularity in the covariance matrix function, and the bias term b2 in the convergence rate of the between-jump area becomes b around the jumps. We also note that around the smaller neighborhood of the jump J:=1jι^T^h2, due to the larger error in the change point estimation, consistent recovery of the graphs is not achievable.

5. A Simulation Study

We simulate data from the following multivariate time series model:

Xi=m=0100Am(i)ϵim,i=1,,n,

where Am(i)Rp×p,1m100,1in, and ϵim=(ϵim,1,,ϵim,p), with ϵm,k, mZ, j=1,,p generated as i.i.d. standardized T(8) random variables. In the simulation, we fix n=1000 and vary p=50 and p=100. For each m=1,,100, the coefficient matrices Am(i)=(1+m)βBm(i), where β=1, and Bm(1) is an Rp×p block diagonal matrix. The 5×5 diagonal blocks in Bm(i) are fixed with i.i.d. N(0,1) entries and all the other entries are 0.

We consider the number of abrupt changes is ι=2 and (nt(1),nt(2))=(300,650). The matrix A0(i) is set to be a zero matrix for i=1,2,,299, while A0(i)=A0(299)+αα, i=300,301,,649, and A0(i)=A0(649)αα, i=650,651,,1000, where the first 20 entries in α are taken to be a constant δ0 and the others are 0.

We let the coefficient matrices A1(i)={am,jk(i)}1j,kp evolve at each time point, such that two entries are soft-thresholded and another two elements increase. Specifically, at time i, we randomly select two elements from the support of A1(i), which are denoted as {a1,jlkl(i)},l=1,2 and that a1,jk(i)0, and set them to a1,jlkl(i)=sign(a1,jlkl(i))(|a1,jlkl(i)0.05|). We also randomly select two elements from A1(i) and increase their values by 0.03.

Figure 1 and Figure 2 show the support of the true covariance matrices at i=100,200,,900.

Figure 1.

Figure 1

Support of the true covariance matrices, p=50.

Figure 2.

Figure 2

Support of the true covariance matrices, p=100.

In detecting the change points, the cutoff value ν of detection is chosen as follows. After removing the neighborhood of detected change points, we obtain Dh(l) by ordering Dh(l),Dh(l), where l is obtained from (9) with ν=0. For l=1,2,,l1, compute

Rh(l)=Dh(l)Dh(l+1).

We let ι^=argmax0ll1Rh(l) and set ν=Dh(ι^).

We report the number of estimated jumps and the average absolute estimation error, where the average absolute estimation error is the mean of the distance between the estimated change points and the true change points. As is shown in Table 1 and Table 2, there is an apparent improvement in the estimation accuracy as the jump magnitude increases and dimension decreases. The detection is relatively robust to the choice of bandwidth.

Table 1.

Average distance.

Bandwidth 0.14 0.16 0.18 0.2 0.22 0.24
p=50 δ0=1 23.4 21.0 17.47 16.6 14.7 16.5
δ0=2 7.4 6.9 8.3 8.1 7.2 6.3
p=100 δ0=1 37.2 30.1 26.4 25.5 21.2 21.3
δ0=2 7.8 8.2 9.9 6.9 8.9 7.6

Table 2.

Number of estimated change points.

Bandwidth 0.14 0.16 0.18 0.2 0.22 0.24
p=50 δ0=1 2.38 2.16 1.99 2.00 2.00 2.00
δ0=2 2.46 2.31 2.00 2.00 2.00 2.00
p=100 δ0=1 2.25 2.09 1.99 1.99 2.00 2.00
δ0=2 2.38 2.19 2.00 2.00 2.00 2.00

We evaluate the support recovery performance of the time-varying CLIME at the lattice 100,200,,900 with λ=0.02,0.06,0.1. We take the uniform kernel function and the bandwidth is fixed as 0.2. At each time point t0, two quantities are computed: sensitivity and specificity, which are defined as:

sensitivity=1j,kpI{g^jk(t0;u)0,gjk(t0;u)0}1j,kpI{gjk(t0;u)0},specificity=1j,kpI{g^jk(t0;u)=0,gjk(t0;u)=0}1j,kpI{gjk(t0;u)=0}.

We plot the Receiver Operating Characteristic (ROC) curve, that is, sensitivity against 1-specificity. From Figure 3 and Figure 4 we observe that, due to a screening step, the support recovery is robust to the choice of λ, except at the change points, where a non-negligible estimation error of the covariance matrix is induced and the overall estimation is less accurate. As the effective dimension of the network remains the same at p=50 and p=100 by the construction of the coefficient matrix Am(i), there is no significant difference in the ROC curves at different dimensions.

Figure 3.

Figure 3

ROC curve of the time-varying CLIME, p=50.

Figure 4.

Figure 4

ROC curve of the time-varying CLIME, p=100.

6. A Real Data Application

Understanding the interconnection among financial entities and how they vary over time provides investors and policy makers with insights into risk control and decision making. Reference [66] presents a comprehensive study of the applications of network theory in financial systems. In this section, we apply our method to a real financial dataset from Yahoo! Finance (finance.yahoo.com). The data matrix contains daily closing prices of 420 stocks that are always in the S&P 500 index between 2 January 2002 through 30 December 2011. In total, there are n=2519 time points. We select 100 stocks with the largest volatility and consider their log-returns; that is, for j=1,,100,

Xij=logpi+1,j/pij,

where pij is the daily closing price of the stock j at time point i. We first compute the statistic (5) and (6) for the change point detection. We look at the top three statistics for different bandwidths. For bandwidth k=n1/5=0.21, we rank the test statistic and find that the location for the top change point is: 7 February 2008 (ns^1=1536), which is shown in Figure 5. The detected change point is quite robust to a variety of choices of bandwidth. Our result is partially consistent with the change point detection method in [48]. In particular, the two breaks in 2006 and 2007 were also found in [48] and it is conjectured that the 2007 break may be associated to the U.S. house market collapse. Meanwhile, it is interesting to observe the increased volatility before the 2008 financial crisis.

Figure 5.

Figure 5

Break size |Ds|. From 4 February 2004, to 30 November 2009.

Next, we estimate the time-varying networks before and after the change point at 26 May 2006 with the largest jump size. Specifically, we look at four time points at: 813, 828, 888, and 903, corresponding to 23 March 2006, 13 April 2006, 11 July 2006, and 1 August 2006. We use tv-CLIME (13) with the Epanechnikov kernel with the same bandwidth as in the change point detection to estimate the networks at the four points. Optimal tuning parameter λ is automatically selected according to the stability approach [67]. The following matrix shows the number of different edges at those four time points. It is observed that the time of the first two time points (813 and 828) and the last two (888 and 903) has a higher similarity than across the change point at time 858. The estimated networks are shown in Figure 6. Networks in the first and second row are estimated before and after the estimated change point at time 858, respectively. It is observed that at each time point the companies in the same section tend to be clustered together such as companies in the Energy section: OXY, NOV, TSO, MRO, and DO (highlighted in cyan). In addition, the distance matrix of estimated networks is estimated as

0332350396332039442835039402343964282340.

Figure 6.

Figure 6

Estimated networks at time points 813, 828, 888, and 903, corresponding to 23 March 2006, 13 April 2006, 11 July 2006, and 1 August 2006. Colors correspond to the nine sections in the S&P dataset.

7. Proof of Main Results

7.1. Preliminary Lemmas

Lemma 1.

Let (Yi)iZ be a sequence that admits (2). Assume YiLq for i=1,2,, and the dependence adjusted norm (DAN) of the corresponding underlying array (Yi(t)) satisfies Y·q,A< for q>2 and A>0. Let (ω(t,ti))i=1n be defined in (12) and suppose that the kernel function K(·) satisfies Assumption 5. Denote ϖq,A(n)=n,n(logn)1+2q,nq/2Aq if A>1/21/q, A=1/21/q, and 0<A<1/21/q, respectively. Then there exist constants C1,C2 and C3 independent of n, such that for all x>0,

supt(0,1)Pi=1nw(t,ti)YiE(Yi)>xC1ϖq,A(Bn)Y·q,AqBnqxq+C2expC3Bnx2Y·2,A2. (28)
Psupt(0,1)i=1nw(t,ti)YiE(Yi)>xC1ϖq,A(n)Y·q,AqBnqxq+C2expC3Bn2x2nY·2,A2. (29)

Proof. 

Let Si=j=1iYiE(Yi). Note that

supt(0,1)i=1nw(t,ti)Yi=supt(0,1)i=1nw(t,ti)(SiSi1)supti=1n1w(t,ti)w(t,ti+1)Si+suptw(t,1)SnBn1max1in|Si|,

where the last inequality follows from the fact that supti=1n|w(t,ti)w(tti+1)|Bn1, due to Assumption 5.

To see (29), it suffices to show

Pmax1in|Si|>xC1ϖq,A(n)Y·q,Aqxq+C2expC3x2nY·2,A2. (30)

Now, we develop a probability deviation inequality for max1in|j=1iαjYj|, where αj0, 1jn are constants such that 1jnαj=1. Denote P0(Yi)=E(Yi|εi)E(Yi) and

Pk(Yi)=E(Yi|εik,,εi)E(Yi|εik+1,,εi).

Then we can write

max1in|j=1iαjYj|max1in|j=1iαjP0(Yj)|+max1in|k=1nj=1iαjPk(Yj)|+max1in|k=n+1j=1iαjPk(Yj)|. (31)

Note that (P0(Yj))jZ is an independent sequence. By Nagaev’s inequality and Ottaviani’s inequality, we have that

P(max1in|j=1iαjP0(Yj)|x)j=1nαjqP0(Yj)qqxq+expC3x2j=1nαj2P0(Yj)22j=1nαjqxqYjq+expC3x2j=1nαj2, (32)

where the last inequality holds because P0(Yj)q2Yjq by Jensen’s inequality. Since j=i+1αjPk(Yj) is a martingale difference sequence with respect to σ(εi+1k,εi+2k,), we have that |k=1+nj=i+1nαjPk(Yj)| is a non-negative sub-martingale. Then by Doob’s inequality and Burkholder’s inequality, we have

Pmax1in|k=n+1j=1iαjPk(Yj)|xP|k=n+1j=1nαjPk(Yj)|x2+Pmax1in|k=n+1j=1+inαjPk(Yj)|x2k=1+nj=1nαjPk(Yj)qqxq(j=1nαj2)q/2Θn,qqxqΘn,qqnq/21j=1nαjqxq. (33)

Now, we deal with the term max1in|k=1nj=1iαjPk(Yj)|. Define am=min(2m,n) and Mn=logn/log2. Then

max1in|k=1nj=1iαjPk(Yj)|m=1Mnmax1in|l=1i/amj=1+(l1)ammin(lam,i)k=1+am1amαjPk(Yj)|. (34)

Let Aodd={1li/am,l is odd} and Aeven={1li/am,l is even}. We have

Pmax1in|l=1i/amZl,m,i|xPmax1in|AoddZl,m,i|x/2+Pmax1in|AevenZl,m,i|x/2,

where we have that Zl,m,i:=j=1+(l1)ammin(lam,i)αjPam1am(Yj) is independent of Zl+2,m,i for 1li/am,1mMn,1in, as Pam1am(Yj):=k=1+am1amPk(Yj) is am-dependent. Therefore, we can apply Ottaviani’s inequality and Nagaev’s inequality for independent variables. As a consequence,

Pmax1in|l=1i/amZl,m,i|x1ln/amZl,m,nqqxq+expC3x21ln/amZl,m,n22.

Again, by Burkholder’s inequality, we have that for q2,

Zl,m,nqk=1+am1amj=1+(l1)ammin(lam,n)αjPk(Yj)q(j=1+(l1)ammin(lam,n)αj2)1/2(Θam1Θam).

Note j=1+(l1)ammin(lam,n)αj2am(q2)/q(j=1+(l1)ammin(lam,n)αjq)2/q. Let τm=m2/m=1Mnm2, and we have τmm2 as 1m=1Mnm2π2/6. In respect to (34), we have that

Pmax1in|k=1nj=1iPk(Yj)|xm=1MnPmax1in|l=1i/amZl,m,i|τmxi=1nαjqxqY·q,Aqm=1Mnτmqam(1/2A)q1+m=1MnexpC3x2τm2am2Aj=1nαj2Y·2,A2. (35)

Note m=1Mnτmqam(1/2A)q1n1ϖq,A(n), and

m=1MnexpC3x2τm2am2Aj=1nαj2Y·2,A2expC3x2j=1nαj2Y·2,A2.

Combining (31), (32), (33) and (35), we obtain

Pmax1in|j=1iαjYjE(Yj)|>xC1ϖq,A(n)j=1nαjqY·q,Aqnxq+C2expC3x2j=1nαj2Y2,A2. (36)

Now, we have (30) by taking αj=n1 for j=1,,n. Note that since K(·) has bounded support, for any given t[b,1b], we have

P|i=1nw(t,ti)(YiEYi)|>xP|i=BnBnw(t,ttn+i)(Ytn+iEYtn+i)|>xC1ϖq,A(Bn)i=BnBnw(t,ttn+i)qY·q,AqBnxq+C2expC3x2i=BnBnw(t,ttn+i)2Y·2,A2.

Therefore (28) follows from (36) by taking αj=w(t,tn+j), and note that for any t[b,1b], i=BnBnw(t,ttn+i)βBn1β for a constant β2. □

Lemma 2.

Suppose (Xij)iZ,1jp satisfys Assumption 2. Furthermore, let Assumption 5 hold. Let ϖq,A(n) be defined as in Lemma 1. Then there exist constants C1,C2, and C3 independent of n and p, such that for all x>0, we have

supt(0,1)P|i=1nω(t,ti)XiXiE(XiXi)|xC1ν2qqpϖq,A(Bn)MX,qqBnqxq+C2p2expC3Bnx2ν42NX2, (37)

and

Psupt(0,1)|i=1nw(t,ti)XiXiE(XiXi)|xC1ν2qqpϖq,A(n)MX,qqBnqxq+C2p2expC3Bn2x2nν42NX2. (38)

Proof. 

For 1j,kp, let Yi,jk=XijXik. We now check the conditions in Lemma 1 for (Yi,jk)1in. Denote Yi,jk,{m}=Xij,{m}Xik,{m}. Then the uniform functional dependence measure of (Yi,jk)i is

θm,q,jkY=supiYi,jkYi,jk,{m}q=supiXijXikXij,{m}Xik,{m}qsupiXij(XikXik,{m})q+supiXik,{m}(XijXij,{m})q.

Thus the DAN of the process Y·,jk satisfies that

Y·,jkq,AsupiXij2qX·,k2q,A+supiXik2qX·,j2q,Aνq(X·,k2q,A+X·,j2q,A).

The result follows immediately from Lemma 1 and the Bonferroni inequality. □

Lemma 3.

We adopt the notation in Lemma 2. Suppose Assumptions 2, 1, and 5 hold with ι=0. Recall Bn=nb, where b0 and Bn/n as n. Then there exists a constant C independent of n and p such that Σ^(t) in (11) satisfies that for any t[c,1c],

|Σ^(t)Σ(t)|=OPb2+MX,qν2qBn1(pϖq,A(Bn))1/q+ν4NX(logp/Bn)1/2. (39)

Furthermore,

supt[c,1c]|Σ^(t)Σ(t)|=OPb2+MX,qν2qBn1(pϖq,A(n))1/q+ν4NXBn1[nlogp]1/2. (40)

Proof. 

First, we have

Eσ^jk(t)σjk(t)=i=1nw(t,ti)[σjk(ti)σjk(t)].

Approximating the discrete summation with integral, we obtain for all 1j,kp,

supt[b,1b]Eσ^jk(t)σjk(t)11K(u)[σjk(ub+t)σjk(t)]du=OBn1.

By Assumption 1, we have

σjk(ub+t)σjk(t)=ubσjk(t)+12u2b2σjk(t)+o(b2u2).

Thus we have supt[c,1c]|Eσ^(t)σ(t)|=OBn1+b2, in view of Assumption 5. By Lemma 2, we have

supt(0,1)PΣ^(t)EΣ^(t)xC1pνqqMX,qqϖq,A(Bn)Bnqxq+C2p2expC3Bnx2NX2.

Denote u=C4MX,qν2qBn1(pϖq,A(Bn))1/q+ν4NX(logp/Bn)1/2 for a large enough constant C4, then for any t(0,1),

Σ^(t)EΣ^(t)=OP(u).

Thus (39) is proved. The result (40) can be obtained similarly. □

7.2. Proof of Main Results

Proof of Proposition 1. 

Given (39) and (40), the proof of (16) is standard. (See, e.g., Theorem 6 of [11]). For λ and λ* given in Proposition 1, by Lemma 3, we have that, respectively,

λsuptEκp|Σ^(t)Σ(t)|, (41)
λEκpsupt|Σ^(t)Σ(t)|. (42)

Then note that for any t[0,1], for any λ>0,

|Ω^λ(t)Ω(t)||Ω(t)|L1|Σ(t)Ω^λ(t)Idp||Ω(t)|L1|Σ^(t)Ω^λ(t)Idp|+|(Σ(t)Σ^(t))Ω(t)|+|Ω^λ(t)Ω(t)|L1|Σ^(t)Σ(t)|

where by construction, we have |Σ^(t)Ω^λ(t)Idp|λ and |Ω^λ(t)Ω(t)|L12κp. Consequently,

|Ω^λ(t)Ω(t)|κpλ+3κp|Σ^(t)Σ(t)|. (43)

Then (16) and (17) follow from (41) to (43). □

Proof of Proposition 2. 

Theorem 2 is an immediate result of (17). □

Proof of Theorem 1. 

Denote rj,1jι as the time point(s) of the time of jump ordered decreasingly in the sense of the infinite norm of covariance matrices, i.e., |Δ(r1)||Δ(r2)||Δ(rι)||Δ(s)| for s(0,1){r1,,rι}c. (Temporal order is applied if there is a tie.) Let Th(j)=[rjh,rj+h). For h=o(1), as a result of Assumption 3, Th(j)Th(i)= if ij for n sufficiently large. That is to say, each time point s(0,1) is in the neighborhood of, at most, one change point.

For any s[t(j),t(j+1)), j=0,1,,ι, denote D(s)=E[D(s)] and

D(s)=(hs+t(j))Δ(t(j)),t(j)s<t(j)+h0,t(j)+hs<t(j+1)h(h+sr)Δ(t(j+1)),t(j+1)hst(j+1). (44)

Then, for s1jι[t(j)+h,<t(j+1)h), by (3), we have

|Σ(s+t)Σ(s)|Lt,|t|h,

we can easily verify that

sups[0,1]|D(s)D(s)|Lh2. (45)

Note that |D(s)| is maximized at s=r1 and |D(r1)|=h|Δ(r1)|. By the triangle inequalities, we have that for some positive constant C, for any s[0,1],

|D(r1)||D(s)|hc2|D(r1)D(r1)||D(s)||D(s)D(s)|hc2|D(s)|2Lh2c2(|sr1|h)2Lh2. (46)

On the other hand, since |D(r1)||D(s^1)|, we have

|D(r1)||D(s^1)||D(r1)||D(s^1)|+|D(r1)D(r1)|+|D(s^1)D(s^1)||D(r1)D(r1)|+|D(s^1)D(s^1)|. (47)

Denote the event A:={sups[h,1h]|D(s)D(s)|h2} and let Yi=(Yi,jk)1j,kp, Yi,jk=XijXikσi,jk. Note that

|Djk(s)Djk(s)|=1ni=1hnYns+1i,jki=1hnYns+i,jk. (48)

By Lemma 2, we have for any x>0,

Psups[h,1h]|D(s)D(s)|xC1pϖq,A(n)MX,qqν2qqnqxq+C2p2expC3nx2NX2. (49)

It follows that

|D(r1)||D(s^1)|=OPh1Jq,A(n,p)+NXh1(n1log(p))1/2.

Taking h=h, we have

|s^1r1|=OP(h2).

Furthermore, we have

P(A)1C1pϖq,A(n)MX,qqν2qqnqc2q1/3C2p2expC3(nlog2(p)NX2)1/3.

Let Ak:={max1jk|s^jrj|c212(L+1)h2} for some 1kι. Assume AkA. Under Ak we have that [rjh,rj+h)T^2h(j)=:[s^j2h,s^j+2h) for 1jk and rk+11jkT^2h(j) as a consequence of Assumption 3. According to (46) and (47), we have if A is true, |s^k+1rk+1|c212(L+1)h2, which implies Ak+1A. The result (21) follows from deduction.

Suppose A holds. By the choice of ν, as a consequence of (45) and (49), and that νh, we have that

sups[0,1]|D(s)D(s)|ν.

As a result,

min1jι|D(rj)|c2hνν,

i.e., ι^ι. On the other hand, since 1jιT^2h(j) is excluded from the searching region for sι+1, we have

sups1jιT^2h(j)c|D(s)|ν.

In other words, {ι^=ι}A. Thus (20) is proved. □

Proof of Theorem 2. 

We adopt the notations in the proof of Theorem 1 and assume that E holds. Similar to Lemma 3, we have that by Lemma 2, for any t(0,1),

Σ^(t)EΣ^(t)=OP(u),

where u=C4MX,qν2qBn1(pϖq,A(Bn))1/q+ν4NX(logp/Bn)1/2 for a large enough constant C4.

Since under E, Tb(j)T^b+h2(j). For t1jιT^b+h2(j)c[b,1b], we have that for all 1j,kp,

Eσ^jk(t)σjk(t)=11K(u)[σjk(ub+t)σjk(t)]du+OBn1=bσjk(t)11uK(u)du+12b2σjk(t)+o(b2)11u2K(u)du+OBn1=O(b2+Bn1).

On the other hand, for t1jιT^b+h2(j)Th2c(j)[0,b][1b,1], due to reflection, we no longer have that differentiability. As a result of the Lipschitz continuity, we get

Eσ^jk(t)σjk(t)=11K(u)[σjk(ub+t)σjk(t)]du+OBn1=O(b+Bn1).

The result (22) follows by the choices of b. The rest of the proof are similar to that of Proposition 1 and Theorem 2. □

Acknowledgments

X.C. acknowledges that part of this work was carried out at the MIT Institute for Data, System, and Society (IDSS).

Author Contributions

Methodology, M.X., X.C., W.B.W.; writing—original draft preparation, M.X., X.C., W.B.W.; writing—review and editing, M.X., X.C., W.B.W., software, M.X. All authors have read and agreed to the published version of the manuscript.

Funding

X.C.’s research is supported in part by NSF CAREER Award DMS-1752614 and UIUC Research Board Award RB18099. W.B.W.’s research is supported in part by NSF DMS-1405410.

Conflicts of Interest

The authors declare no conflict of interest.

References

  • 1.Lauritzen S. Graphical Models. Clarendon Press; Oxford, UK: 1996. [Google Scholar]
  • 2.Peng J., Wang P., Zhou N., Zhu J. Partial Correlation Estimation by Joint Sparse Regression Models. J. Am. Stat. Assoc. 2009;104:735–746. doi: 10.1198/jasa.2009.0126. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Meinshausen N., Bühlmann P. High-dimensional graphs and variable selection with the lasso. Ann. Stat. 2006;34:1436–1462. doi: 10.1214/009053606000000281. [DOI] [Google Scholar]
  • 4.Friedman J., Hastie T., Tibshirani R. Sparse Inverse Covariance Estimation with the Graphical Lasso. Biostatistics. 2008;9:432–441. doi: 10.1093/biostatistics/kxm045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Banerjee O., El Ghaoui L., d’Aspremont A. Model selection through sparse maximum likelihood estimation for multivariate Gaussian or binary data. J. Mach. Learn. Res. 2008;9:485–516. [Google Scholar]
  • 6.Rothman A.J., Bickel P.J., Levina E., Zhu J. Sparse Permutation Invariant Covariance Estimation. Electron. J. Stat. 2008;2:494–515. doi: 10.1214/08-EJS176. [DOI] [Google Scholar]
  • 7.Yuan M. High Dimensional Inverse Covariance Matrix Estimation via Linear Programming. J. Mach. Learn. Res. 2010;11:2261–2286. [Google Scholar]
  • 8.Yuan M., Lin Y. Model selection and estimation in the Gaussian graphical model. Biometrika. 2007;94:19–35. doi: 10.1093/biomet/asm018. [DOI] [Google Scholar]
  • 9.Ravikumar P., Wainwright M.J., Raskutti G., Yu B. High-dimensional covariance estimation by minimizing ℓ1-penalized log-determinant divergence. Electron. J. Stat. 2011;5:935–980. doi: 10.1214/11-EJS631. [DOI] [Google Scholar]
  • 10.Candès E., Tao T. Rejoinder: “The Dantzig selector: Statistical estimation when p is much larger than n”. Ann. Stat. 2007;35:2392–2404. doi: 10.1214/009053607000000532. [DOI] [Google Scholar]
  • 11.Cai T., Liu W., Luo X. A constrained ℓ1 minimization approach to sparse precision matrix estimation. J. Am. Stat. Assoc. 2011;106:594–607. doi: 10.1198/jasa.2011.tm10155. [DOI] [Google Scholar]
  • 12.Cai T., Liu W. Adaptive thresholding for sparse covariance matrix estimation. J. Am. Stat. Assoc. 2011;106:672–684. doi: 10.1198/jasa.2011.tm10560. [DOI] [Google Scholar]
  • 13.Fan J., Feng Y., Wu Y. Network Exploration via the Adaptive Lasso and SCAD penalties. Ann. Appl. Stat. 2009;3:521–541. doi: 10.1214/08-AOAS215. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Basu S., Shojaie A., Michailidis G. Network Granger causality with inherent grouping structure. J. Mach. Learn. Res. 2015;16:417–453. [PMC free article] [PubMed] [Google Scholar]
  • 15.Loh P.L., Bühlmann P. High-dimensional learning of linear causal networks via inverse covariance estimation. J. Mach. Learn. Res. 2014;15:3065–3105. [Google Scholar]
  • 16.Loh P.L., Wainwright M.J. Structure estimation for discrete graphical models: Generalized covariance matrices and their inverses. Ann. Stat. 2013;41:3022–3049. doi: 10.1214/13-AOS1162. [DOI] [Google Scholar]
  • 17.Lèbre S., Becq J., Devaux F., Stumpf M.P., Lelandais G. Statistical inference of the time-varying structure of gene-regulation networks. BMC Syst. Biol. 2010;4:1–16. doi: 10.1186/1752-0509-4-130. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Przytycka T.M., Singh M., Slonim D.K. Toward the dynamic interactome: It’s Toward the dynamic interactome: it’s about time. Brief. Bioinform. 2010;11:15–29. doi: 10.1093/bib/bbp057. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Khandani A.E., Lo A.W. What happened to the quants in August 2007? Evidence from factors and transactions data. J. Financ. Mark. 2011;14:1–46. doi: 10.1016/j.finmar.2010.07.005. [DOI] [Google Scholar]
  • 20.Chi K.T., Liu J., Lau F.C. A network perspective of the stock market. J. Empir. Financ. 2010;17:659–667. [Google Scholar]
  • 21.Durante D., Dunson D.B., Vogelstein J.T. Nonparametric Bayes modeling of populations of networks. J. Am. Stat. Assoc. 2017;112:1516–1530. doi: 10.1080/01621459.2016.1219260. [DOI] [Google Scholar]
  • 22.Durante D., Dunson D.B. Locally adaptive dynamic networks. Ann. Appl. Stat. 2016;10:2203–2232. doi: 10.1214/16-AOAS971. [DOI] [Google Scholar]
  • 23.Han Q., Xu K., Airoldi E. Consistent estimation of dynamic and multi-layer block models; Proceedings of the 32nd International Conference on Machine Learning; Lille, France. 6–11 July 2015; pp. 1511–1520. [Google Scholar]
  • 24.Danaher P., Wang P., Witten D.M. The joint graphical lasso for inverse covariance estimation across multiple classes. J. R. Stat. Soc. Ser. B Stat. Methodol. 2014;76:373–397. doi: 10.1111/rssb.12033. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Dondelinger F., Lèbre S., Husmeier D. Non-homogeneous dynamic Bayesian networks with Bayesian regularization for inferring gene regulatory networks with gradually time-varying structure. Mach. Learn. 2013;90:191–230. doi: 10.1007/s10994-012-5311-x. [DOI] [Google Scholar]
  • 26.Pensky M. Dynamic network models and graphon estimation. Ann. Stat. 2019;47:2378–2403. doi: 10.1214/18-AOS1751. [DOI] [Google Scholar]
  • 27.Pensky M., Zhang T. Spectral clustering in the dynamic stochastic block model. Electron. J. Stat. 2019;13:678–709. doi: 10.1214/19-EJS1533. [DOI] [Google Scholar]
  • 28.Bhattacharjee M., Banerjee M., Michailidis G. Change Point Estimation in a Dynamic Stochastic Block Model. arXiv. 20181812.03090 [Google Scholar]
  • 29.Bartlett T.E., Kosmidis I., Silva R. Two-way sparsity for time-varying networks, with applications in genomics. arXiv. 20181802.08114 [Google Scholar]
  • 30.Gaucher S., Klopp O. Maximum likelihood estimation of sparse networks with missing observations. arXiv. 20191902.10605 [Google Scholar]
  • 31.Erdös P., Rényi A. On Random Graphs I. Publ. Math. Debr. 1959;6:290–297. [Google Scholar]
  • 32.Penrose M. Random Geometric Graphs. Oxford University Press; Oxford, UK: 2003. [Google Scholar]
  • 33.Zhou S., Lafferty J., Wasserman L. Time Varying Undirected Graphs. Mach. Learn. 2010;80:295–319. doi: 10.1007/s10994-010-5180-0. [DOI] [Google Scholar]
  • 34.Kolar M., Xing E. On time varying undirected graphs; Proceedings of the 14th International Conference on Artificial Intelligence and Statistics (AISTATS 2011); Ft. Lauderdale, FL, USA. 11–13 April 2011. [Google Scholar]
  • 35.Kolar M., Song L., Xing E. Estimating time-varying networks. Ann. Appl. Stat. 2010;4:94–123. doi: 10.1214/09-AOAS308. [DOI] [Google Scholar]
  • 36.Kolar M., Xing E.P. Sparsistent Estimation Of Time-Varying Markov Sparsistent Estimation Of Time-Varying Markov Random Fields. arXiv. 20090907.2337 [Google Scholar]
  • 37.Qiu H., Han F., Liu H., Caffo B. Joint estimation of multiple graphical models from high dimensional time series. J. R. Stat. Soc. Ser. B Stat. Methodol. 2015;78:487–504. doi: 10.1111/rssb.12123. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Lu J., Kolar M., Liu H. Post-regularization Inference for Dynamic Nonparanormal Graphical Models. arXiv. 20151512.08298 [Google Scholar]
  • 39.Ahmed A., Xing E.P. Recovering time-varying networks of dependencies Recovering time-varying networks of dependencies in social and biological studies. Proc. Natl. Acad. Sci. USA. 2009;106:11878–11883. doi: 10.1073/pnas.0901910106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Tibshirani R., Saunders M., Rosset S., Zhu J., Knight K. Sparsity and smoothness via the fused lasso. J. R. Stat. Soc. Ser. B. 2005;67:91–108. doi: 10.1111/j.1467-9868.2005.00490.x. [DOI] [Google Scholar]
  • 41.Cho H., Fryzlewicz P. Multiple-change-point detection for high dimensional time series via sparsified binary segmentation. J. R. Stat. Soc. Ser. B Stat. Methodol. 2015;77:475–507. doi: 10.1111/rssb.12079. [DOI] [Google Scholar]
  • 42.Roy S., Atchadè Y., Michailidis G. Change-point estimation in high-dimensional Markov random field models. J. R. Stat. Soc. Ser. B Stat. Methodol. 2017;79:1187–1206. doi: 10.1111/rssb.12205. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Zhou S. Gemini: Graph estimation with matrix variate normal instances. Ann. Stat. 2014;42:532–562. doi: 10.1214/13-AOS1187. [DOI] [Google Scholar]
  • 44.Tong H. Non-Linear Time Series: A Dynamical System Approach. Oxford University Press; Oxford, UK: 1993. [Google Scholar]
  • 45.Fan J., Yao Q. Nonlinear Time Series: Nonparmatric and Parametric Methods. Springer; Berlin/Heidelberg, Germany: 2003. [Google Scholar]
  • 46.Fryzlewicz P. Wild Binary Segmentation for multiple change-point detection. Ann. Stat. 2014;42:2243–2281. doi: 10.1214/14-AOS1245. [DOI] [Google Scholar]
  • 47.Kokoszka P., Leipus R. Change-point estimation in ARCH models. Bernoulli. 2000;6:513–539. doi: 10.2307/3318673. [DOI] [Google Scholar]
  • 48.Aue A., Hörmann S., Horváth L., Reimherr M. Break detection in the covariance structure of multivariate time series models. Ann. Stat. 2009;37:4046–4087. doi: 10.1214/09-AOS707. [DOI] [Google Scholar]
  • 49.Chang C., Glover G.H. Time-frequency dynamics of resting-state brain connectivity measured with fMRI. NeuroImage. 2010;50:81–98. doi: 10.1016/j.neuroimage.2009.12.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Hutchison M., Womelsdorf T., Gati J., Everling S., Menon R. Resting-state networks show dynamic functional connectivity in awake humans and anesthetized macaques. Hum. Brain Mapp. 2013;34:2154–2177. doi: 10.1002/hbm.22058. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Wiesel A., Bibi O., Globerson A. Time varying autoregressive moving average models for covariance estimation. IEEE Trans. Signal Process. 2013;61:2791–2801. doi: 10.1109/TSP.2013.2256900. [DOI] [Google Scholar]
  • 52.Qiu H., Han F., Liu H., Caffo B. Robust Portfolio Optimization under High Dimensional Heavy-Tailed Time Series. Johns Hopkins University; Baltimore, MD, USA: 2014. Technical Report. [Google Scholar]
  • 53.Chen X., Xu M., Wu W.B. Covariance and precision matrix estimation for high-dimensional time series. Ann. Stat. 2013;41:2994–3021. doi: 10.1214/13-AOS1182. [DOI] [Google Scholar]
  • 54.Chen X., Xu M., Wu W.B. Regularized Estimation of Linear Functionals of Precision Matrices for High-Dimensional Time Series. IEEE Trans. Signal Process. 2016;64:6459–6470. doi: 10.1109/TSP.2016.2605079. [DOI] [Google Scholar]
  • 55.Basu S., Michailidis G. Regularized estimation in sparse high-dimensional time series models. Ann. Stat. 2015;43:1535–1567. doi: 10.1214/15-AOS1315. [DOI] [Google Scholar]
  • 56.Bhattacharjee M., Bose A. Consistency of large dimensional sample covariance matrix under weak dependence. Stat. Methodol. 2014;20:11–26. doi: 10.1016/j.stamet.2013.08.005. [DOI] [Google Scholar]
  • 57.Shu H., Nan B. Estimation of Large Covariance and Precision Matrices from Temporally Dependent Observations. arXiv. 2014 doi: 10.1214/18-AOS1716.1412.5059 [DOI] [Google Scholar]
  • 58.Draghicescu D., Guillas S., Wu W.B. Quantile curve estimation and visualization for nonstationary time series. J. Comput. Graph. Stat. 2009;18:1–20. doi: 10.1198/jcgs.2009.0001. [DOI] [Google Scholar]
  • 59.Wu W.B. Nonlinear system theory: Another look at dependence. Proc. Natl. Acad. Sci. USA. 2005;102:14150–14154. doi: 10.1073/pnas.0506715102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Zhou Z., Wu W.B. Local linear quantile estimation for nonstationary time series. Ann. Stat. 2009;37:2696–2729. doi: 10.1214/08-AOS636. [DOI] [Google Scholar]
  • 61.Zhou Z., Wu W.B. Simultaneous inference of linear models with time varying coefficients. J. R. Stat. Soc. 2010;72:513–531. doi: 10.1111/j.1467-9868.2010.00743.x. [DOI] [Google Scholar]
  • 62.Wu W.B., Wu Y.N. Performance bounds for parameter estimates of high-dimensional linear models with correlated errors. Electron. J. Stat. 2016;10:352–379. doi: 10.1214/16-EJS1108. [DOI] [Google Scholar]
  • 63.Ltkepohl H. New Introduction to Multiple Time Series Analysis. Springer; Berlin/Heidelberg, Germany: 2007. [Google Scholar]
  • 64.Chow Y., Teicher H. Probability Theory: Independence, Interchangeability, Martingales. Springer; New York, NY, USA: 1997. p. 414. [Google Scholar]
  • 65.Ding X., Qiu Z., Chen X. Sparse transition matrix estimation for high-dimensional and locally stationary vector autoregressive models. Electron. J. Stat. 2017;11:3871–3902. doi: 10.1214/17-EJS1325. [DOI] [Google Scholar]
  • 66.Allen F., Babus A. The Network Challenge: Strategy, Profit, and Risk in an Interlinked World. FT Press; Hoboken, NJ, USA: 2009. Networks in Finance. [Google Scholar]
  • 67.Liu H., Roeder K., Wasserman L. Stability Approach to Regularization Selection (StARS) for High-Dim Graphical Models; Proceedings of the 23rd International Conference on Neural Information Processing Systems (NIPS’10); Vancouver, BC, Canada. 6–9 December 2010; [PMC free article] [PubMed] [Google Scholar]

Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES