Abstract
This paper studies simultaneous inference for factor loadings in the approximate factor model. We propose a test statistic based on the maximum discrepancy measure. Taking advantage of the fact that the test statistic can be approximated by the sum of the independent random variables, we develop a multiplier bootstrap procedure to calculate the critical value, and demonstrate the asymptotic size and power of the test. Finally, we apply our result to multiple testing problems by controlling the family-wise error rate (FWER). The conclusions are confirmed by simulations and real data analysis.
Keywords: high-dimensional factor model, multiple testing, multiplier bootstrap, simultaneous inference
1. Introduction
The high-dimensional factor model is becoming more and more important in different scientific areas including finance and macroeconomics. For example, the data in the World Bank contain two-hundred countries over forty years and the number of stocks can be in the thousands which is larger than or of the same order of the sample size for portfolio allocation. Due to its broad applications, much efforts have been devoted to analyzing factor model in different aspects. Examples include estimation of factors and loadings for latent factor model [1,2], covariance matrix estimation [3,4,5,6], and simultaneous inference for factor loadings of dynamic factor model [7,8], among others.
This work focuses on the simultaneous inference for the loading matrix with observed factors, which is an important issue in the analysis of approximate factor models. For example, in the study of gene expression genomics, it is commonly assumed that each gene is associated with only a few factors. For example, the authors of [9] showed that several oncogenes are related to Rb/E2F pathway rather than any other pathways. The authors of [10] also considered sparse loading matrix for gene expression data. Therefore, it is necessary to test the sparsity of the factor loadings. In the literature, some inference procedures have been proposed for latent factor models. For example, in low-dimensional setting, the authors of [11] considered testing the homogeneity assumption, i.e., the loadings associated to a factor are identical for all variables. The same testing problem has been considered by the authors of [12] in high-dimensional situation. As for observed factors, to the best of our knowledge, very limited work has been conducted.
Inference for the factor loadings with observed factors is not trivial. The approaches for latent factors cannot be directly applied to observed factors. The major difficulty is due to the high dimensionality, which poses significant challenges in deriving the asymptotic null limiting distribution of the test statistic. We propose a test statistic based on the maximum discrepancy measure. The distribution of this statistic is attractive in high-dimensional statistical inference such as model selection, simultaneous inference, and multiple testing. Examples include the works in [13,14,15,16,17], among others.
We use the multiplier bootstrap procedure to obtain the critical value of our test statistic. Based on the fact that the test statistic can be approximated by the sum of the independent random variables, we show that the proposed multiplier bootstrap method consistently approximates the null limiting distribution of the test statistic, and thus the testing procedure achieves the prespecified significance level asymptotically. There are some related works applying multiplier bootstrap method to high-dimensional inference; see in [16,18,19], among others. However, their procedures require sparsity assumption on the parameters and cannot be directly applied to factor model. Compared with the works with latent factors, we do not require homogeneity constraints or sparsity on the model and our procedure is adaptive to high-dimensional regime.
Another application of our procedure is the multiple testing problem. Combining the multiplier bootstrap method with step-down procedure proposed by [17], we show that our procedure has a strong control of the family-wise error rate (FWER). Our method is asymptotically non-conservative as compared to the Bonferroni–Holm procedure since the correlation among the test statistics has been taken into account. We also want to point out that any procedure controlling the FWER will also control the false discovery rate [20] when there exist some true discoveries.
The rest of the paper is organized as follows. In Section 2.1, we develop the multiplier bootstrap procedure for simultaneous test of parameters for a single factor and demonstrate its asymptotic level and power. In Section 2.2, we give the result of simultaneous test of parameters for multiple factors. Section 3 discusses the multiple testing problem by combining the multiplier bootstrap procedure with the step-down method proposed by [17]. Section 4 investigates the numerical performance of the proposed test by simulations. We also conduct real data analysis on portfolio risk of S&P stocks via Fama–French model in Section 5. The proofs of the main results are given in Appendix A.
Finally, we introduce some notation. For set S, let denote the cardinality of S. Let be the vector of zeros. For matrix , denote by and the minimum and maximum eigenvalues of , respectively. The matrix element-wise maximum norm and norm are defined as and , respectively. For and , denote by and . Let be the ith column of the identity matrix. We write if is smaller than or equal to up to a universal positive constant. For , we write . For two sets, A and B, denotes their symmetric difference, that is, .
2. Methodology
2.1. Simultaneous Test for a Single Factor
We consider the factor model defined as follows,
| (1) |
where is the observed response for the ith variable at time t, is the unknown vector of factor loadings, is the observed vector of common factors, and is the latent error. Here, K is a fixed integer denoting the number of factors, p is the number of variables, and T denotes the sample size. Model (1) is commonly used in finance and macroeconomics, see, e.g., in [3,4,21], among others.
Denote by , and , then model (1) can be re-expressed as
| (2) |
We first focus on testing the coefficients corresponding to a single factor, i.e., the kth factor. Specifically, we consider the following simultaneous testing problem that given ,
| (3) |
where G is a subset of and are prespecified values. For example, if are 0, then the hypotheses are able to test whether the variables with indices in G are significantly associated with the kth factor. Throughout the paper, is allowed to grow as fast as p, which may have exponential growth with T as in Assumption 3.
The ordinary least squares (OLS) estimator is applied to estimate , where and . Therefore,
| (4) |
We propose the following test statistic for ,
where . For each , the asymptotic normality of is straightforward due to the central limit theorem. However, when diverges with p, it is very challenging to demonstrate the existence of the limiting distribution of . In order to approximate the asymptotic distribution of , we will use the multiplier bootstrap method. From (4), we know
| (5) |
where and .
In order to apply the multiplier bootstrap procedure, we need to approximate by sum of independent random variables. As is consistent for , we can replace the former with the latter in , and define . Then, for each , are i.i.d. and well approximates .
We then apply the multiplier bootstrap procedure to approximate the distribution of . Denote by the covariance matrix of , and hence , where . We know that is -consistent for . To estimate , we first calculate the residuals
Denote by , then the error covariance matrix is estimated by
Let , a sequence of i.i.d. independent of , be the multiplier random variables. Then the multiplier bootstrap statistic is defined as
Conditioning on , the covariance of and is , which can sufficiently approximate the covariance between and . Then, the bootstrap critical value can be obtained via
is calculated by generating repeatedly. In our simulations and real data, we conduct bootstrap 500 times. We now present some technical assumptions.
Assumption 1.
- (i)
are with and .
- (ii)
There exist constants such that .
- (iii)
and are independent.
Assumption 2.
There exist positive constants , such that for any , , and ,
The “” condition in Assumption 1 is commonly considered in the literature for high-dimensional inference, see, e.g., in [16]. Assumption 1 (ii) allows the bounded eigenvalue of the error covariance matrix. As noted in [22], such assumption is satisfied by two situations: (1) , where is a stationary ergodic process with spectral density f, and (2) where , is a stationary process as above and is a noise process independent of . In Example 1 in [22], they demonstrated that ARMA() process satisfies Assumption 1 (ii). Furthermore, this assumption is commonly considered in the literature, see, e.g., in [4,15].
Assumption 2 allows the application of the large deviation theory to and . In this paper, we assume that and have exponential-type tails. Let and .
Assumption 3.
Suppose , and there exists a constant , such that , where .
Assumption 4.
There exists a constant such that .
Assumption 3 is needed in Bernstein-type inequality [23] and commonly assumed in the literature for Gaussian approximation theory. Assumption 4 is also reasonable by bounding the eigenvalues of .
Theorem 1.
Under Assumptions 1–4, we have
Theorem 1 demonstrates that the multiplier bootstrap critical value well approximates the quantile of the test statistic. It is worth mentioning that our method does not require any sparsity assumption on either or .
The proof of Theorem 1 depends on the two results: (1) is sufficiently close to and (2) the covariances of and are well approximated by the bootstrap version. The first result is demonstrated in Lemma A7 that there exist and such that
where and . The second result is shown in Lemma A6 that
i.e., the maximum discrepancy between the empirical and population covariance matrices converges to zero.
Based on Theorem 1, for a given significance level , we define the test by
| (6) |
The hypothesis is rejected whenever .
Bootstrap is a commonly used resampling method and full theories about it can be found in [24]. There are many versions of bootstrap, for example, wild bootstrap, empirical bootstrap, and multiplier bootstrap, among others. As discussed in [25], other exchangeable bootstrap methods are asymptotically equivalent to the multiplier bootstrap. As our test statistic can be approximated by the maximum of sum of independent random vectors, we adopt the multiplier bootstrap method in [25] based on Gaussian approximation.
Alternatively, we propose the studentized statistic for , where . Similarly to Section 2.1, we define the multiplier bootstrap statistic as
where are independent of . Then, the bootstrap critical value can be obtained via
Theorem 2 below justifies the validity of the bootstrap procedure for the studentized statistic.
Theorem 2.
Under the assumptions in Theorem 1, we have
Based on this result, for a given significance level , we define the test by
The hypothesis is rejected whenever .
For the studentized statistic, we can derive its asymptotic distribution. By Lemma 6 of the work in [15], for any and as , we have
However, the above alternative testing procedure may not perform well in practice, because it requires diverging p, and the convergence rate is typically slow.
In contrast to the extreme value approach, our testing procedure explicitly accounts for the effect of in the sense that the bootstrap critical value depends on G. Therefore, our approach is more robust to the change in .
Next, we turn to the (asymptotic) power analysis of the above procedure. Denote by the kth column of . We focus on the case where as below. Define the separation set
| (7) |
where . Let with , which is the correlation matrix of .
Assumption 5.
Suppose for some constant .
Theorem 3.
Under Assumptions 1–5, for any , we have
As long as one entry of has a magnitude larger than , our bootstrap-assisted test can reject the null correctly. Therefore, with being sparse, our procedure performs well in detecting non-sparse alternatives. According to Section 3.2 of [26], the separation rate is minimax optimal under suitable assumptions.
2.2. Simultaneous Test for Multiple Factors
In this section, we test the elements of the loading matrix corresponding to different factors. The testing problem can be stated as follows,
where is a subset of . Define
We propose the studentized test statistic
From the linear expansion in (5), the multiplier bootstrap statistic is defined as
where are independent of . Then, the bootstrap critical value can be obtained via
Let , and for a constant .
Theorem 4.
Suppose , under Assumptions 1,2 and 4, we have
Based on Theorem 4, for a given significance level , we define the test by
The hypothesis is rejected whenever .
Now we turn to the power analysis of the test . Similar to Section 2.1, we focus on the case where as and define the separation set
Let . We consider the following condition.
Assumption 6.
Suppose for some constant .
The asymptotic power of the testing procedure is given as follows.
Theorem 5.
Under the assumptions in Theorem 4 and Assumption 6 , for any , we have
3. Multiple Testing with Strong FWER Control
In this section, we study the following multiple testing problem,
For simplicity, we set and let j be fixed. We combine the bootstrap-assisted procedure with the step-down method proposed by [17]. Our method can be seen as a special case in Section 5 of [25]. Note that this framework can cover the case of testing equalities because equalities can be rewritten as pairs of inequalities.
We briefly illustrate the control of the FWER. Full details and theory can be found in [25]. Let be the space for all data generating processes, and be the true process. Each null hypothesis is equivalent to for some . For any , denote by with . The strong control of the FWER means that
| (8) |
where denotes the probability distribution under the data-generating process .
For , denote . For a subset , let be the bootstrapped estimate for the -quantile of . The step-down procedure in [17] is described as follows. Define at the first step and reject all satisfying . If no is rejected, then stop the procedure. If some are rejected, let be the set of indices for those hypotheses not being rejected at the first step. On step , let be the subset of hypotheses that were not rejected at step . Reject all hypotheses for satisfying that . If no hypothesis is rejected, then stop the procedure. Proceed in this way until the algorithm stops.
Romano and Wolf [17] proved the following result:
| (9) |
| (10) |
Therefore, we can show that the step-down method together with the multiplier bootstrap provide strong control of the FWER by verifying (9) and (10). The theoretical results are given in the proposition below. The proofs are similar to those of Theorem 1, which are omitted here.
Proposition 1.
Under the assumptions in Theorem 1, the step-down procedure with the bootstrap critical value satisfies (8).
Our multiple testing method has the following two important features: (i) It can be applied to models with an increasing dimension; (ii) It takes into account the correlation amongst statistics and hence is asymptotically non-conservative.
In the simulation, we also consider Benjamini–Hochberg procedure [20] to control the false discovery rate (FDR), which is summarized as follows. For each of , we calculate the p-values based on the studentized test statistic. Let be the ordered p-values, and denote by the null hypothesis corresponding to . Let , and then reject all for .
4. Simulation Study
This section examines the performance of the proposed testing procedure by a simulation study. We fix the number of factors , the sample size , and let the dimensionality p increase from 50 to 600. Throughout the simulation, we consider testing the first column of and repeat multiplier bootstrap 500 times.
Each row of is generated independently from , where is identity matrix. Let with . Here, we consider two models for the covariance structure .
-
(a)
Model 1 (sparse): where , for , where and otherwise. .
-
(b)
Model 2 (non-sparse): where and for .
Under each model, and are generated independently from and , respectively.
We calculate the empirical sizes of test for each column of under each model by considering hypothesis (3) with and being the true value of . The results are summarized in Table 1. Here “NST”, “ST” denote the non-studentized, studentized Bootstrap-based test, respectively, and “EX” denotes the test using extreme value distribution. The estimated sizes of the three tests are reasonably close to the nominal level 0.05 for the values of p ranging from 50 to 600.
Table 1.
Empirical sizes of tests, , , and 500 replications.
| p = 50 | p = 100 | p = 200 | p = 400 | p = 600 | |
|---|---|---|---|---|---|
| Model 1 | |||||
| NST | 0.076 | 0.060 | 0.058 | 0.050 | 0.058 |
| ST | 0.074 | 0.064 | 0.064 | 0.058 | 0.078 |
| EX | 0.046 | 0.046 | 0.038 | 0.046 | 0.058 |
| Model 2 | |||||
| NST | 0.050 | 0.052 | 0.056 | 0.060 | 0.038 |
| ST | 0.070 | 0.058 | 0.064 | 0.068 | 0.048 |
| EX | 0.038 | 0.030 | 0.024 | 0.018 | 0.016 |
For all , by varying with and , we plot the empirical powers of and in Figure 1. For ease of presentation, we only consider . The results for other dimensionality are similar in spirit, and are not presented here. For all tests, the significance level is fixed at . From Figure 1, we can tell that the empirical rejection rate grows from the nominal level to one as c deviates away from zero. The difference between NST test and ST test is slight. For small p, the EX test does not perform well because this approach requires diverging p. Furthermore, for non-sparse error covariance matrix, our method performs better than the EX method. These numerical results confirm our theoretical analysis.
Figure 1.
Empirical powers of the NST, ST, and EX methods. The figures in the left panels are based on Model 1, while those in the right panels are for Model 2. The red solid line corresponds to the nominal level. (a) , (b) , (c) , (d) , (e) , (f) .
Next, we study the numerical performance of the step-down method in Section 3 and compare it with the Bonferroni–Holm procedure. Consider the following two-sided multiple testing problem; among all with . For Models 1 and 2, the first entries of are and , respectively, and the rest are equal to . We set and .
We employ both the step-down method based on the studentized/non-studentized test statistic, and the Bonferroni–Holm procedure (based on the studentized test statistic) to control the FWER. We denote these three procedures by NST-FWER, ST-FWER, and BH-FWER, respectively. For comparison, we also consider using Benjamini–Hochberg procedure to control FDR. We denote this procedure by BH-FDR. Based on 500 replications, we calculate the average empirical FWER
for methods NST-FWER, ST-FWER, and BH-FWER, the average empirical FDR
for method BH-FDR, and the average empirical power
for all the four methods, where and . Under each model, we consider the case and . Table 2 and Table 3 report the empirical FWER, FDR, and the average power. From Table 2 and Table 3, the proposed and Bonferroni–Holm procedures provide similar control on the FWER, and Benjamini–Hochberg procedure can control FDR. The empirical powers of the step-down method and Benjamini–Hochberg procedure are higher than that of the Bonferroni–Holm procedure. It is also seen that controlling the FDR is more powerful than controlling the FWER.
Table 2.
Empirical family-wise error rate (FWER) and false discovery rate (FDR) with power in the brackets of multiple testing based on Model 1, , and 500 replications.
| T | Method | p = 50 | p = 200 | p = 500 | p = 600 | |
|---|---|---|---|---|---|---|
| 200 | 3 | NST-FWER | 0.058 (0.551) | 0.062 (0.405) | 0.048 (0.309) | 0.056 (0.291) |
| ST-FWER | 0.074 (0.554) | 0.082 (0.431) | 0.086 (0.337) | 0.090 (0.324) | ||
| BH-FWER | 0.054 (0.528) | 0.070 (0.409) | 0.074 (0.319) | 0.068 (0.300) | ||
| BH-FDR | 0.061 (0.635) | 0.064 (0.470) | 0.086 (0.380) | 0.069 (0.353) | ||
| 15 | NST-FWER | 0.056 (0.569) | 0.050 (0.412) | 0.040 (0.306) | 0.046 (0.303) | |
| ST-FWER | 0.066 (0.583) | 0.086 (0.430) | 0.074 (0.334) | 0.084 (0.327) | ||
| BH-FWER | 0.060 (0.561) | 0.066 (0.410) | 0.056 (0.310) | 0.068 (0.309) | ||
| BH-FDR | 0.043 (0.810) | 0.064 (0.655) | 0.06 (0.518) | 0.061 (0.509) | ||
| 400 | 3 | NST-FWER | 0.050 (0.935) | 0.058 (0.889) | 0.062 (0.839) | 0.058 (0.808) |
| ST-FWER | 0.070 (0.937) | 0.062 (0.885) | 0.078 (0.842) | 0.066 (0.813) | ||
| BH-FWER | 0.052 (0.931) | 0.054 (0.873) | 0.062 (0.834) | 0.052 (0.795) | ||
| BH-FDR | 0.057 (0.957) | 0.056 (0.924) | 0.064 (0.889) | 0.068 (0.863) | ||
| 15 | NST-FWER | 0.058 (0.947) | 0.054 (0.881) | 0.040 (0.819) | 0.070 (0.815) | |
| ST-FWER | 0.052 (0.946) | 0.066 (0.881) | 0.058 (0.825) | 0.084 (0.882) | ||
| BH-FWER | 0.050 (0.942) | 0.056 (0.871) | 0.050 (0.809) | 0.060 (0.806) | ||
| BH-FDR | 0.035 (0.989) | 0.052 (0.968) | 0.056 (0.946) | 0.059 (0.941) |
Table 3.
Empirical FWER and FDR with power in the brackets of multiple testing based on Model 2, , and 500 replications.
| T | Method | p = 50 | p = 200 | p = 500 | p = 600 | |
|---|---|---|---|---|---|---|
| 200 | 3 | NST-FWER | 0.044 (0.805) | 0.052 (0.692) | 0.066 (0.622) | 0.056 (0.609) |
| ST-FWER | 0.058 (0.807) | 0.066 (0.701) | 0.084 (0.638) | 0.066 (0.621) | ||
| BH-FWER | 0.030 (0.759) | 0.042 (0.620) | 0.024 (0.517) | 0.024 (0.505) | ||
| BH-FDR | 0.039 (0.819) | 0.046 (0.691) | 0.038 (0.592) | 0.030 (0.570) | ||
| 15 | NST-FWER | 0.042 (0.805) | 0.060 (0.697) | 0.058 (0.626) | 0.048 (0.618) | |
| ST-FWER | 0.050 (0.809) | 0.080 (0.708) | 0.080 (0.637) | 0.072 (0.630) | ||
| BH-FWER | 0.028 (0.757) | 0.042 (0.621) | 0.034 (0.530) | 0.038 (0.519) | ||
| BH-FDR | 0.035 (0.922) | 0.044 (0.822) | 0.046 (0.746) | 0.040 (0.717) | ||
| 400 | 3 | NST-FWER | 0.060 (0.989) | 0.052 (0.985) | 0.052 (0.971) | 0.050 (0.970) |
| ST-FWER | 0.064 (0.989) | 0.054 (0.985) | 0.068 (0.970) | 0.072 (0.973) | ||
| BH-FWER | 0.046 (0.983) | 0.022 (0.975) | 0.026 (0.951) | 0.024 (0.945) | ||
| BH-FDR | 0.045 (0.995) | 0.034 (0.990) | 0.040 (0.972) | 0.035 (0.973) | ||
| 15 | NST-FWER | 0.066 (0.992) | 0.072 (0.986) | 0.056 (0.975) | 0.056 (0.975) | |
| ST-FWER | 0.072 (0.992) | 0.076 (0.986) | 0.056 (0.975) | 0.064 (0.975) | ||
| BH-FWER | 0.046 (0.988) | 0.036 (0.973) | 0.024 (0.952) | 0.022 (0.950) | ||
| BH-FDR | 0.043 (0.999) | 0.042 (0.998) | 0.031 (0.993) | 0.044 (0.992) |
5. Real Data Analysis
This section conducts hypothesis testing for financial data from 1 January 2017 to 14 March 2018. The dataset consists of daily returns of 491 stocks from S&P 500 index. In addition, we collected Fama–French three factors [21] in the same period. In summary, the panel matrix is a 300 by 491 matrix , in addition to a factor matrix of size 300 by 3. Here, 300 is the number of days and 491 is the number of stocks.
We first centralize and standardize the factor matrix and is centralized as well. We consider testing the sparsity of each column of and repeat the multiplier bootstrap 500 times. Simultaneous test of parameters corresponding to multiple factors is also considered. The hypotheses are
where , with or and . The results are depicted in Table 4. For the first column of , it is therefore not reasonable to assume . However, we can claim that the last two columns of are sparse. Hence, a sufficiently large number of stocks are not influenced by the last two factors.
Table 4.
Results of sparse testing.
| = 10 | = 30 | = 50 | = 70 | = 90 | |
|---|---|---|---|---|---|
| 1st loading | R | R | R | R | R |
| 2nd loading | A | A | A | A | A |
| 3rd loading | A | A | A | A | R |
| 2nd and 3rd loading | A | A | A | R | R |
Note: “A” means accepting the null hypothesis; “R” denotes rejecting the null hypothesis.
Acknowledgments
We thank the three referees for insightful comments and suggestions.
Appendix A. Technical Details
We prove the main results in this section. First, we introduce some notations. Throughout this section, we denote by constants that do not depend on and may vary from place to place. Define and . Let be a sequence of vectors. Define with and with . Denote by with . We begin by presenting some useful lemmas that will be used in the proofs of the main results.
Lemma A1.
Suppose that the random variables both satisfy the exponential-type tail condition: There exist and such that ,
Then, for some and , and any ,
- (i)
, where with ,
- (ii)
, where .
Proof of Lemma A1.
The proof of the first claim can be found in the proof of Lemma A.2 in [4], thus we prove the second claim. For any , we have
where . Pick up an , and ; then, it can be shown that is increasing when . Therefore, when , which implies when ,
When ,
Then the proof is complete. ☐
Lemma A2.
Under Assumptions 1–4, we have
- (i)
.
- (ii)
.
- (iii)
.
Proof of Lemma A2.
For a proof, see the proof of Lemma A.3 and Lemma B.1 in [4]. ☐
Lemma A3.
If a random variable X satisfies exponential-type tail: there exist and such that , . Then .
Proof of Lemma A3.
Note that
It is not hard to check when , . When , , where and . Then, the proof is complete. ☐
Lemma A4.
Under the assumptions in Theorem 1, there exist constants such that
Proof of Lemma A4.
Implied by Assumption 3, we have for some constants . We then apply Corollary 2.1 of [25] to the sequence . What we should check is its Condition (E.2), that is, uniformly over i,
where and B is some large enough constant. By Lemmas A1 and A3 we have uniformly for . This implies . Uniformly for , we have
By Lemma A1, we know and is exponential-type tail. Then by Lemma A3 we have =O(1) and . Thus, we can find large enough B such that the above condition is satisfied. Then, the proof is complete. ☐
Lemma A5.
Under the assumptions in Theorem 1, there exists a sequence of positive numbers such that and .
Proof of Lemma A5.
By Lemma A2 (iii), we have . Since and , we have
On the other hand,
Choosing , by Assumption 3, the proof is complete. ☐
Lemma A6.
Under the assumptions in Theorem 1, we have for every and ,
Proof of Lemma A6.
For , let with . Recall that . As uniformly for , by Lemma A2 (i), we have
By Lemma A5 and Assumption 3, choosing , we have . By Lemma 3.1 of [25], on the event , we have for all , and so on this event
implying the first claim. The second claim follows similarly. ☐
Lemma A7.
Under the assumptions in Theorem 1, there exist such that
where , .
Proof of Lemma A7.
The arguments in the proof of Lemma A5 imply that
By Lemma A2 (ii), uniformly for , we have
Choosing , we have
Note that
then the proof is complete. ☐
Lemma A8.
Under the assumptions in Theorem 4, we have
- (i)
,
- (ii)
,
Proof of Lemma A8.
(i) By Assumption 1 and Lemma A1, satisfies the exponential tail condition, with parameter as shown in Lemma A1. Thus, satisfies the exponential tail condition, with parameter . It follows from that . Therefore, by the Bernstein’s inequality [23], there exist constants , , for any
(A1) Using Bonferroni’s method, we have
Let for some . It is not hard to check that when (by assumption), for large enough C,
and
As , this proves (i).
(ii) By Assumption 1 and Lemma A1, satisfies the exponential tail condition for the tail parameter . Therefore, again by the Bernstein’s inequality and the Bonferroni method on similar to (A1) with the parameter , it follows from that . Thus, when for large enough C, as K is fixed, the term
and the rest terms on the right-hand side of the inequality, multiplied by are of order . Hence when (by assumption), we have
which completes the proof. ☐
Lemma A9.
Under the assumptions in Theorem 4, we have
- (i)
- (ii)
Proof of Lemma A9.
(i) By the triangular inequality, we have
For I, we have
By Lemma 3.1 of [4], we have . It is straightforward to see that
then we have,
By Lemma A8 (ii), we have , which implies that
Part is similar to , thus we have
Then the proof is complete.
(ii) By the triangular inequality, we have
which proves the result. ☐
Proof of Theorem 1.
Without loss of generality, we set . First, we prove the following fact,
(A2) For , let with . In addition, Let and . For every , note that
where (1) follows from Lemmas A6 and A7, (2) follows from Lemma A4, and (3) follows from Lemma 2.1 in [25] and the fact that has no point masses. Then, by the definition of in Lemma A4, we have
where . The right-hand side of the above inequality is , which has proved (A2). Since , similar arguments imply that
which completes the proof. ☐
Proof of Theorem 2.
From the arguments in the proof of Lemma A6, we have
which implies that . We then have
(A3) Define and . Note that
where .
On the event for all ,
On the other hand,
Therefore, on the above event, . By Lemma A5, we can find such that and . Thus by Lemma A7 and (A3), we have
for and .
Let . Note that
On the event for all , we have
which implies that
Choosing , we can show that . The rest of the proofs are similar to those in the proof of Theorem 1. We skip the details. ☐
Proof of Theorem 3.
Let . Following the arguments in the proof of Theorem 2, we can show that the distribution of can be approximated by . Under Assumption 5, by Lemma 6 of [15], we have for any and as
It implies that
(A4) The bootstrap consistency result implies that
(A5) where is the th quantile of . Consider any such that . Using the inequality for any , we have
(A6) where as i is fixed and grows. From the proof of Theorem 2, we know the difference between and is asymptotically negligible. Thus, by (A6) and the fact that , we have
(A7) The conclusion thus follows from (A7) and (A5) provided that is small enough. ☐
Proof of Theorem 4.
Without loss of generality, we set . Define
Let
denote the maximum discrepancy between the empirical and population covariance matrices. By the triangular inequality, we have
Note that , by Lemma A9 (ii), we have
By Lemma A2 (iii) and , we have
Since , we have
The above results hold uniformly for , thus we have
The rest of the proofs is similar to those in the proof of Theorem 2. We skip the details. ☐
Proof of Theorem 5.
For a proof, see the proof of Theorem 3. ☐
Author Contributions
X.G. conceived and designed the experiments; Y.W. performed the experiments; Y.W. analyzed the data; X.G. contributed to analysis tools; X.G. and Y.W. wrote the paper. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the National Natural Science Foundation of China, grants 12071452, 11601500, 11671374 and 11771418 and the Fundamental Research Funds for the Central Universities.
Conflicts of Interest
The authors declare no conflict of interest.
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Bai J., Liao Y. Efficient estimation of approximate factor models via penalized maximum likelihood. J. Econ. 2016;191:1–18. [Google Scholar]
- 2.Heinemann A. Efficient estimation of factor models with time and cross-sectional dependence. J. Appl. Econ. 2017;32:1107–1122. [Google Scholar]
- 3.Fan J., Fan Y., Lv J. High dimensional covariance matrix estimation using a factor model. J. Econ. 2008;147:186–197. [Google Scholar]
- 4.Fan J., Liao Y., Mincheva M. High-dimensional covariance matrix estimation in approximate factor models1. Ann. Stat. 2011;39:3320–3356. doi: 10.1214/11-AOS944. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Fan J., Liao Y., Mincheva M. Large covariance estimation by thresholding principal orthogonal complements. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 2013;75:603–680. doi: 10.1111/rssb.12016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Fan J., Wang W., Zhong Y. Robust covariance estimation for approximate factor models. J. Econom. 2019;208:5–22. doi: 10.1016/j.jeconom.2018.09.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Dickhaus T., Pauly M. Time Series Analysis and Forecasting. Springer; Berlin, Germany: 2016. Simultaneous statistical inference in dynamic factor models; pp. 27–45. [Google Scholar]
- 8.Dickhaus T., Sirotko-Sibirskaya N. Simultaneous statistical inference in dynamic factor models: Chi-square approximation and model-based bootstrap. Comput. Stat. Data Anal. 2019;129:30–46. [Google Scholar]
- 9.Lucas J., Carvalho C., Wang Q., Bild A., Nevins J.R., West M. Sparse statistical modelling in gene expression genomics. Bayesian Inference Gene Expr. Proteom. 2006;1:155–176. [Google Scholar]
- 10.Carvalho C.M., Chang J., Lucas J.E., Nevins J.R., Wang Q., West M. High-dimensional sparse factor modeling: applications in gene expression genomics. J. Am. Stat. Assoc. 2008;103:1438–1456. doi: 10.1198/016214508000000869. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Reis R., Watson M.W. Relative goods’ prices, pure inflation, and the Phillips correlation. Am. Econ. J. Macroecon. 2010;2:128–157. [Google Scholar]
- 12.Amengual D., Repetto L. Testing a Large Number of Hypotheses in Approximate Factor Models. CEMFI; Madrid, Spain: 2014. Technical Report. [Google Scholar]
- 13.Candès E., Tao T. The Dantzig selector: Statistical estimation when p is much larger than n. Ann. Stat. 2007;35:2313–2351. [Google Scholar]
- 14.Cai T., Liu W., Xia Y. Two-sample covariance matrix testing and support recovery in high-dimensional and sparse settings. J. Am. Stat. Assoc. 2013;108:265–277. doi: 10.1080/01621459.2012.758041. [DOI] [Google Scholar]
- 15.Cai T.T., Liu W., Xia Y. Two-sample test of high dimensional means under dependence. J. R. Stat. Soc. Ser. (Stat. Methodol.) 2014;76:349–372. [Google Scholar]
- 16.Zhang X., Cheng G. Simultaneous inference for high-dimensional linear models. J. Am. Stat. Assoc. 2017;112:757–768. doi: 10.1080/01621459.2016.1166114. [DOI] [Google Scholar]
- 17.Romano J.P., Wolf M. Exact and approximate stepdown methods for multiple hypothesis testing. J. Am. Stat. Assoc. 2005;100:94–108. [Google Scholar]
- 18.Zhu Y., Yu Z., Cheng G. High dimensional inference in partially linear models; Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics (AISTATS); Naha, Okinawa, Japan. 16–18 April 2019; pp. 2760–2769. [Google Scholar]
- 19.Zhang X., Cheng G. Bootstrapping high dimensional time series. arXiv. 20141406.1037 [Google Scholar]
- 20.Benjamini Y., Hochberg Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. J. R. Stat. Soc. Ser. (Methodol.) 1995;57:289–300. doi: 10.1111/j.2517-6161.1995.tb02031.x. [DOI] [Google Scholar]
- 21.Fama E.F., French K.R. The cross-section of expected stock returns. J. Financ. 1992;47:427–465. doi: 10.1111/j.1540-6261.1992.tb04398.x. [DOI] [Google Scholar]
- 22.Bickel P., Levina E. Some theory for Fisher’s linear discriminant function, “naive Bayes”, and some alternatives when there are many more variables than observations. Bernoulli. 2004;10:989–1010. doi: 10.3150/bj/1106314847. [DOI] [Google Scholar]
- 23.Merlevède F., Peligrad M., Rio E. A Bernstein type inequality and moderate deviations for weakly dependent sequences. Probab. Theory Relat. Fields. 2011;151:435–474. doi: 10.1007/s00440-010-0304-9. [DOI] [Google Scholar]
- 24.Kosorok M.R. Introduction to Empirical Processes and Semiparametric Inference. Springer; New York, NY, USA: 2008. [Google Scholar]
- 25.Chernozhukov V., Chetverikov D., Kato K. Gaussian approximations and multiplier bootstrap for maxima of sums of high-dimensional random vectors. Ann. Stat. 2013;41:2786–2819. [Google Scholar]
- 26.Ingster Y.I., Tsybakov A.B., Verzelen N. Detection boundary in sparse regression. Electron. J. Stat. 2010;4:1476–1526. doi: 10.1214/10-EJS589. [DOI] [Google Scholar]

