Skip to main content
Journal of Applied Statistics logoLink to Journal of Applied Statistics
. 2020 Oct 26;49(3):709–721. doi: 10.1080/02664763.2020.1837084

On a generalization of the test of endogeneity in a two stage least squares estimation

Ayyub Sheikhi a,CONTACT, Fatemeh Bahador a, Mohammad Arashi b,*
PMCID: PMC9042057  PMID: 35706770

Abstract

In situations that the predictors are correlated with the error term, we propose a bridge estimator in the two-stage least squares estimation. We apply this estimator to overcome the multicollinearity and sparsity of the explanatory variables, when the endogeneity problem is present.The proposed estimator was applied to modify the Durbin-Wu-Hausman (DWH) test of endogeneity in the presence of multicollinearity. To compare our modified test with the existing DWH for detection of an endogenous problem in multi-collinear data, some numerical assessments are carried out. The numerical results showed that the proposed estimators and the suggested test perform better for the multi-collinear data. Finally, a genetical data set is applied for illustration the our results by estimating the coefficients parameters in the presence of endogeneity and multicollinearity.

Keywords: Bridge estimator, Durbin-Wu-Hausman test, Endogeneity, Instrumental variable, multicollinearity, sparsity

1. Introduction

Under some certain regularity conditions, the Ordinary Least Squares (OLS) estimator produces the unbiased and consistent estimators. An important assumption is that the predictors in the model are uncorrelated with the error term. When this condition is not satisfied, then the OLS estimator yields biased and inconsistent estimates and referred to as the endogeneity problem (Ebbes [10]; Bowden and Turkington [3]). An approach to overcome the endogeneity problem is applying some instrumental variable estimators or two-stage least squares which assumes that there is a secondary predictor, termed as instrumental variable, that is correlated to the predictor but not with the error term.

Another important assumption in the classical regression model is that the predictors are uncorrelated. When this assumption is violated, the explanatory variables are nearly dependent which refers as multicollinearity problem and the OLS yields poor estimates. In order to combat this problem, several approaches have been considered and these include the ridge regression estimator introduced by Hoerl and Kennard [17], as one. They considered a shrinkage method to overcome the problem of multicollinearity in the estimation of regression parameters; see ([12,21,28,32] and [22]).

Many authors described the connection between endogeneity and multicollinearity in the literature. Recently, Philips and Evans [26] discussed on approximating and reducing bias in an estimator method called Two Stage Least Squares (2SLS) estimation while Hansen and Kuzbur [15] proposed a jackknife instrument variable estimator with regularization at each jackknife iteration that helps alleviate the bias. Interested readers may refer to [5,6,8] for other developments in this regard.

On the other hand, in the context of variable selection, it is well known that the least absolute shrinkage and selection operator (LASSO) of Tibshirani [29] produces sparse solution. In the presence of endogenous regressors, the direct using of the Lasso selector fails as sparsity of coefficients in regression model does not correspond to sparsity of linear projection coefficients [2].

This paper uses the bridge regression method to combat multicollinearity and sparsity in estimation of the instrumental variable models.

It is known that finding valid instrumentals is not easy. Moreover, using instrumental variables estimators for eliminating the inconsistency of the OLS causes the loss of efficiency. Thus, researchers must find a good way of determining when they must apply the instrument variable estimator, and when they do not have to. Wu [31], Hausman [16] and Durbin [9] proposed some tests for detecting endogenous regressors in the instrumental analysis. The secondary purpose of this paper is to modify the Hausman test statistic in multi-collinear models. The main and preliminary results are given in the next section. In section 3, we propose our test statistic that is used for tesing of endogeneity under the multicollinearity situation and provide some asymptotic results. Section 4 is devoted to a simulation study to investigate the performance of the introduced test statistic. Finally, an application of our results in the analysis of a genetic data is presented in Section 5.

2. Main results

Consider the multiple linear regression model

y=Xβ+ϵ (1)
xi=Zηi+ui,i=1,2,,p, (2)

where y=(y1,,yn) is the vector of response variables, X=(x1,,xp) is an n×p matrix of endogenous explicative variables, β=(β1,,βp) is the vector of parameters, and ϵ=(ϵ1,,ϵn) is the vector of random disturbances with E(ϵ)=0 and E(ϵϵ)=σε2In. It is further assumed that ϵ follows the n-variate normal distribution. Moreover, Z is an n×k matrix of the exogenous instrumental variables, ηi is its k×1 corresponding coefficient vector, and ui,i=1,2,,n, is an n×1 vector of N(0,σu2In) errors. It is also assumed that the covariance matrix of (εi,ui) is Σ.It is known that the two stage least squares (2SLS) estimator of β is obtained in two stages as follows: in the first stage, we have to estimate the regression coefficient ηi using the OLS approach as the below:

η^i=(ZZ)1Zxi.

Then, using the predicted values x^i=Zη^i in the second stage, y is regressed on X^=(x^1,,x^p) and the estimator of coefficient vector β is of the form

β^2SLS=(X^X^)1X^y=(XPX)1XPy, (3)

where P=Z(ZZ)1Z is a symmetric and idempotent matrix.

Suppose now in the model (2) there are some ill-conditioned in the matrix XX. The most ill-conditioned situations are multicollinearity and sparsity. In such situation, we use the bridge penalty in the second stage of 2SLS method to deal with these problems. The first stage estimation is the same as the traditional 2SLS, but in the second stage of estimation we apply the bridge estimator in the regression equation y=PXβ+ε. Then by shrinking the estimation under j=1p|βj|γ<c for some γ,c>0, the bridge 2SLS estimator of β after solving the following optimization problem:

β^2SLSB=argminβRp{(yX^β)(yX^β)+λn1p|β|γ},

where λn0 is the tuning parameter, |β|γ=(|β1|γ,,|βp|γ), and 1p is a p-tuple of 1's. Using local quadratic approximation (LQA) of Fan and Li [11] the penalty term can be approximated around a local vector βo and the bridge estimator has the following closed form

β^2SLSB=[C+Σλ(βo)]1X^y,C=X^X^, (4)

where Σλ(βo)=λγ2Diag(|β1o|γ2,,|βpo|γ2). For more details one may refer to [24],[33], and [25].Obviously, when γ=1, the LASSO of [30] can be approximated by

β^2SLSL=argminβRp{(yX^β)(yX^β)+λn1p|β|}=[C+λ2Diag(|β1o|1,,|βpo|1)]1X^y.

However, the ridge estimator is of the following form for γ=2, without approximation

β^2SLSR=argminβRp{(yX^β)(yX^β)+λn1pT|β|2} (5)
=(XPX+kI)1XPy. (6)

In order to study the properties of these estimators, The following regularity conditions are assumed:

(A.1) The p×p and k×k matrices Σxx=E(xixi) and Σzz=E(zizi) are nonsingular and finite.

(A.2) E(xiεi)0, E(ziεi)=0 and the k×p matrix Σxz=E(zixi) is finite and of full rank.

(A.3) By the law of large numbers, there exist nonnegative definite matrices Σxx, Σzz and Σxz such that

1ni=1nxixiΣxx,1ni=1nziziΣzz,1ni=1nxiziΣxz.

The following theorem discusses on the asymptotic behavior of the 2SLS bridge estimator β^2SLSB.

Theorem 1

Under the assumptions (A.1)–(A.3), if γ1, λnnλ00, then

n(β^2SLSBβ)DargminV(u),

where

V(u)=2uΣxzΣzz1W+uΣxzΣzz1Σzxu+λ0j=1pujsgn(βj)βjγ1ifγ>1,V(u)=2uΣxzΣzz1W+uΣxzΣzz1Σzxu+λ0j=1p[ujsgn(βj)I(βj0)+|uj|I(βj=0)],ifγ=1,

and W follows the N(0,σ2Σzz) distribution.

Proof.

Define

Vn(u)=i=1n[(εiux^in)2εi2]+λnj=1p[|βj+ujn|γ|βj|γ],

where u=(u1,,up) and x^i=Zη^+ui,i=1,2,,n. Here, the convergence of the first term of Vn(u) is proven and the convergence of the second term can be similarly found in Knight and Wenjiang [19]. Note that

i=1n[(εiux^in)2εi2]=ui=1nxi^xi^n2ui=1nxi^εin=ux^x^n+x^εn. (7)

Using the central limit theorem and (A.1)-(A.3), the RHS of (7) converges in distribution to 2uΣzxΣzz1W+uΣxzΣzz1Σzxu which completes the proof.

Corollary 1

β^2SLSR has asymptotically normal distribution with mean λ0Σzx1Σzz

Σxz1β and variance-covariance matrix σ2Σzx1ΣzzΣxz1.

Proof.

Setting γ=2 and β^2SLSR=(XPX+kI)1XPy in Theorem 1, implies that

V(u)=2uΣxzΣzz1W+uΣxxΣzz1Σzxu+λ0uβ,

so,

argminV(u)=(ΣxzΣzz1Σzx)1ΣxzΣzz1Wλ0(ΣxzΣzz1Σzx)1β=Σzx1Wλ0Σzx1ΣzzΣxz1βN(λ0Σzx1ΣzzΣxz1β,σ2Σzx1ΣzzΣxz1).

The proof is complete.

3. The proposed test

In this section, we propose a test statistic to diagnose endogeneity when the multicollinearity problem exists in the explanatory variables. Suppose that the regression model between the outcome vector y and the predictor matrix X is expressed as y=Xβ+ϵ.

Then, for the endogeneity testing hypotheses

H0:cov(X,ϵ)=0H1:cov(X,ϵ)0. (8)

Hausman[20] introduced the following test statistic

H=(β^2SLSβ^OLS)[Var(β^2SLS)Var(β^OLS)]1(β^2SLSβ^OLS) (9)

Under the null hypothesis H0, it follows a central chi-square distribution with p degrees of freedom (χ2(p)) distribution where p is the number of unknown parameters.

Many fields of sciences, such as marketing, genomics, and basic sciences have studied detection of multicollinearity see e.g. Batterham et al. [1]. An important consequence of a high degree of multicollinearity is the high estimated variance of β^OLS. Therefore, in this case, the Hausman test statistic tends to zero and always leads to accepting the null hypothesis. Therefore, the Hausman test in the presence of the multicollinearity problem in the model may lead to incorrect conclusions about regressors' endogeneity. Guo et al. [14] introduced a test statistic in this context without considering the problem of multicollinearity.

We propose a modified version of the Hausman test statistic as follows:

H=(β^2SLSBβ^R)[Var(β^2SLSB)Var(β^R)]1(β^2SLSBβ^R), (10)

where β^R=(XX+kI)1Xy is the ordinary ridge estimation in the model (2) and k>0 is the ridge parameter. It is expected that the null hypothesis of endogeneity is rejected if the statistic H significantly differs from zero.

To perform the hypothesis test, we need to find the null distribution of the statistic H. To this end, we take the following assumptions

(B.1) X and Z are stochastic and nonsingular matrices.

(B.2) The p×p matrices M1=plim(XXn) and M2=plim(XPXn) exist and are non-singular.

Theorem 2

Under (B.1)-(B.2), (A.3) and the null hypothesis (8), the statistic H in (10) has asymptotically (χ2(p)).

Proof.

Following the approach of Hausman [20], it suffices to prove the following three assertions:

1. Under the null hypothesis (8), β^R and β^2SLSB are consistent estimators of β.

2. β^R is an efficient estimator under the null hypothesis (8).

3. β^R and β^2SLSB have asymptotically normal distributions.

Under (B.1)–(B.2) we have

plimβ^R=plim[(XX+kI)1Xy=plim[(XX+kI)1X(Xβ+ε)]=plim[(XX+kIn)1XXβn]+plim[(XX+kIn)1Xεn]=M11M1β+0=β,

which proves the consistency of β^R. Also, the assumption (A.3) imply that Σ^xz=1nXZΣxz and therefore around a local vector βo we have

plimβ^2SLSB=lim[(XPX+Σλ(βo))1XP(Xβ+ε)]=(M2+0)1M2β+plim[(XZn)(ZZn)1(ZXn)+Σλ(βo)n]1[(XZn)(ZZn)1(Zεn)]=β+[ΣxzΣzz1Σzx]1[ΣxzΣzz1Σzε]=β.

To establish the efficiency of β^R, the Cramer-Rao inequality implies that

Cov(β^R)(βEβ^R)In1(β)(βEβ^R)

where In(β)=σ2XX is the Fisher's information matrix. Given that E(β^R)=(XX+kI)1(XX)β, we conclude that:

Cov(β^R)σ2(XX+kI)1(XX)(XX+kI)1,

and also

Var(β^R)=σ2(XX+kI)1(XX)(XX+kI)1,

which means that β^R is an efficient estimator.

The asymptotic normality of β^R is verified by Knight and Wenjiang [19] which they have shown that β^R asymptotically follows a normal distribution with mean λ0Σxx1β and variance-covariance matrix σ2Σxx1. Finally, the asymptotic normality of β^2SLSB follows from Theorem (1), and this completes the proof.

4. Simulation study

In this section, we carry out some numerical analyses of the two most applied special cases of the bridge estimator i.e. β^2SLSR and β^2SLSL. First we compare the performance of β^2SLSR and β^2SLSL with the traditional estimators β^OLS, β^2SLS, β^R and β^lasso based on their absolute relative bias (R.Bias) and estimated mean square errors (EMSE) under different degrees of endogeneity. Next, the our proposed test statistic is compared to Durbin- Wu- Hausman test in terms of the estimated power.

textbfWe generate data in the form of the vector W=(X,Z,ϵ) from a multivariate normal distribution with mean zero and a specific variance covariance matrix. In order to impose multicollinearity to the data, the variance covariance matrix is adopted such that the correlation between the covariates is equal to the constant value 0.9 and there is a slight correlation (0.4) between all instruments and the predictor variables, meaning that all instrument variables are relevant. Moreover, for endogeneity, different levels of ρ{0,0.1,0.2,0.3,0.4,0.5,0.6,0.7,0.8} are considered. Using the approach of principal component analysis, we take the largest eigenvector of XX as the vector of coefficients β ([18]). Then, the response vector generated following y=Xβ+ε. Specifically, in this experiment, we take the number of covariates and the number of instruments as px=2 and pz=3, respectively. Then by generating new error terms, the experiment is replicated 5000 times. The absolute R.Bias and EMSE for the standardized values of any estimator β^ say, are calculated respectively as follows:

R.Bias(β^)=15000l=15000j=1pxβ^jlβjβj,EMSE=15000l=15000j=1px(β^jlβj)2

where β^jl is the estimation of β^j in the lth replication of the simulation. The value of the ridge parameter k is obtained via the generalized cross-validation technique introduced by Golub et al. [13] to obtain a good estimator. See also Liu and Jiang [20] and references therein. The results are summarized in Table 1, for the values n = 50 and ρ=0,0.5,0.81.

Table 1. Estimated parameter values, absolute R.Bias and EMSE for n=50 and different values of ρ.

ρ=0
Coefficients β βols^ β2sls^ βridge^ β2slsr^
β1 0.7 0.658 0.671 0.678 0.624
β2 0.5 0.509 0.576 0.502 0.5548
R.bias   0.402 0.879 0.486 0.647
EMSE   0.126 1.019 0.116 0.343
ρ=0.5
Coefficients β βols^ β2sls^ βridge^ β2slsr^
β1 0.7 1.400 0.740 1.366 0.737
β2 0.5 0.044 0.484 0.066 0.476
R.bias   1.158 0.825 1.103 0.636
EMSE   0.812 0.777 0.706 0.327
ρ=0.8
Coefficients β βols^ β2sls^ βridge^ β2slsr^
β1 0.7 1.838 0.774 1.815 0.757
β2 0.5 −0.252 0.459 0.236 0.468
R.bias   1.891 0.805 1.851 0.691
EMSE   1.909 0.853 1.831 0.419

Table 1 shows that in the case ρ=0 where we only have the multicollinearity problem, the OLS estimator has the lowest R.bias value and the ridge estimator has the lowest EMSE value. While in the presence of a problem of the endogeneity, the EMSE and the absolute R.Bias values of the 2SLS-R estimator are lesser than the other three competitors.

Figure 1 depicts the absolute R.bias and EMSE values of the three estimators for different values of ρ and n.

Figure 2.

Figure 2.

Criteria plot of Bias and EMSE for n = 20, 100.

Figure 1.

Figure 1.

Plot of absolute R.Bias and EMSE for n = 20, 100.

In view of Figure 1, for enough large sample size, if the endogeneity level is almost greater than 0.2, then the new estimator is always better than the other three estimators.

In order to evaluate the performance of β^2SLSL, the sparse and endogenous samples are generated. For this purpose, the vector W=(X,Z,ϵ) with the multivariate normal distribution with mean zero and the variance covariance matrix is considered such that all instruments are relevant and for the different levels of the endogeneity corresponding to the values ρ=0,.,1. The vector β is also taken as the form β=(0.6,0.001) to create sparsity conditions. Here again, we set the number of covariates and the number of instruments as px=2 and pz=3 and we compare the performance of new 2SLS-L estimator with the three estimators OLS, 2SLS and Lasso. The results are summarized in Table 2 and plotted in Figure 3.

Table 2. Estimated parameter values, absolute R.Bias and EMSE for n=50 and different values of ρ.

ρ=0
Coefficients β β^ols β^2sls β^lasso β^2slslasso
β1 0.6 0.604 0.610 0.572 0.349
β2 0.001 −0.0004 0.0001 0.0002 0.0004
R.bias   0.227 0.507 0.217 0.502
EMSE   0.041 0.354 0.040 0.308
ρ=0.6
Coefficients β β^ols β^2sls β^lasso β^2slslasso
β1 0.6 1.196 0.656 1.137 0.371
β2 0.001 0.003 −0.0001 0.002 0.0003
R.bias   0.684 0.511 0.598 0.467
EMSE   0.386 1.096 0.319 0.261
ρ=1
Coefficients β β^ols β^2sls β^lasso β^2slslasso
β1 0.6 1.589 0.643 1.567 0.524
β2 0.001 0.0003 −0.002 0.0001 0.0002
R.bias   1.042 0.524 0.996 0.454
EMSE   1.005 0.997 0.934 0.266

Table 3. Estimated parameter values and MSPEs for the OLS, the 2SLS, the Ridge and 2SLS-R methods.

Estimation β^ols β^2sls β^ridge β^2slsr
Intercept 0.0055 0.0064 0.0055 0.0061
β1^ −0.1747 −5.868 −0.1707 −3.5670
β2^ 0.2384 1.5041 0.2280 0.8182
β3^ 0.0565 4.1146 0.0502 2.6057
Estimated MSPE 0.9914 0.9882 0.9901 0.9832

As we can see in Table 2, almost having an endogeneity level greater than 0.6, the 2SLS-L is superior to the other three estimators.

At the end of this section, the power of our proposed test of endogeneity for a set of data with multicollinearity and endogeneity ill-conditioned is studied. We compare its power with the power of the existing DWH test. For this purpose we generate samples from model (1) by simulating vector W=(X,Z,ϵ) which has a multivariate normal distribution with mean zero and consider the variance covariance matrix in such a way that the correlation between the covariates is equal to 0.9 and all instruments are relevant. Also, the number of instruments and covariates are assumed as pz=3 and px=3, respectively. Further, we consider the beta coefficient vector as β=(2,3,2). We have considered 9 different values of the endogeneity levels as ρ=0,0.1,0.2,0.3,0.4,0.5,0.6,0.7,0.8 and used the exact critical value based on the distribution of statistic at 0.05 significance level test. For each value of the endogeneity level, the power of the proposed test is estimated by N = 1000 simulated samples of size n = 100 from the alternative hypothesis. As shown in Figure 3, these two test statistics have almost the same power under the null hypothesis while under the alternative hypothesis our modified test behaves better, than the DWH test in the presence of the multicollinearity, in terms of the power criterion.

Figure 3.

Figure 3.

Comparison of the DWH test statistic and the new modified DWH test statistic H.(solid line: DWH test statistic; dashed line: new modified DWH test statistic).

5. Genetic data analysis

In this section, we apply our results to carry out a regression analysis of a genetic data set that suffers from multicollinearity and endogeneity. Several authors have reported the presence or absence of a relationship between the genetic variant(s) and the outcome as the primary analysis result, rather than an instrumental analysis, ( Didelez and Sheehan [7], Burgess and Dylan [4]). We use the our method to overcome endogeneity and multicollinearity in the data set of Health and Retirement Study (HRS), which is supported by the National Institute on Aging (NIA U01AG009740, RC2 AG036495, RC4 AG039029). This data contain genotype and phenotype variants which may effect on some outcomes such as educational attainment. see e. g. Rietveld et al. [27] and Okbay et al. [23].

We are involved in running multiple regression to predict educational attainment (EA) using systolic blood pressure (SBP), main arterial pressure (MAP) and pulse pressure (PP) as predictor variables. These predictors are extremely correlated (with ρ^=0.922 between MAP and PP and VIF=60.79 in the regression equation between these exposure variables). On the other hand, the dependent variable EA and these three predictors are correlated in which correlation between residual of regression equation and EA is about 0.9997, so, we choose another three variables major depressive disorder (MDD), schizophrenia polygenic score (SPS) and subjective well-being (SWB) as instrumental variables since their correlation with EA is 0.030, 0.015 and 0.012, respectively. We first assessed the fitted value at the first stage and observed that the previous multicollinearity in the matrix XPX still exists and consequently a cross-validation technique incorporated to obtain the ridge parameter which is equal to k = 1. Hence, using (6) we can estimate the parameters on the model. We also calculate the Mean Squared Prediction Error (MSPE) criterion to evaluate the performance of the estimators. To end this, we divide the real data into two subgroups, training and testing, such that 80 percent of the data set to the training set and the rest in the test group. Table 3 presents the estimated values of the coefficients of SBP, MAP, and PP using the four methods OLS, 2SLS, ridge and regularized 2SLS, along with the MSPE values for the test group.

According to the results of this paper, the proposed new estimator offers smaller estimated MSPE compared to the others. Furthermore, the value of the proposed test statistic H, is equal to 5.401 while for the DWH test, it is lesser and equals to 0.144. Note that both of these values of tests statistics provide evidences in favor of H0.

6. Conclusion

In this study we used the bridge method to combat the multicollinearity and sparsity in a two-stage least squares estimation. In particular, we studied a ridge-2SLS as well as a Lasso-2SLS estimation procedures. Some large sample properties of the proposed estimator investigated and compared with some other estimators using a simulation study. Furthermore, we proposed a modification of the DWH test of endogeneity. Using a simulation study we observed that the our new test has higher power than the power of the existing DWH test in multi-collinear data.

Acknowledgments

The authors would like to thank the anonymous reviewers for the insightful comments which led to an improvement of this paper. The work of M. Arashi was based upon research supported in part by the National Research Foundation (NRF) of South Africa SARChI Research Chair UID: 71199. Opinions expressed and conclusions arrived at in this study are those of the author and are not necessarily to be attributed to the NRF.

Note

1

We use the R software and all codes are available upon request.

Disclosure statement

No potential conflict of interest was reported by the author(s).

References

  • 1.Batterham A.M., Tolfrey K., and George K.P., Nevill's explanation of Kleiber's 0.75 mass exponent: an artifact of collinearity problems in least squares models?, J. Appl. Physiology 82 (1997), pp. 693–697. doi: 10.1152/jappl.1997.82.2.693 [DOI] [PubMed] [Google Scholar]
  • 2.Belloni A. and Chernozhukov V., L1-Penalized quantile regression in high-Dimensional sparse models, The Ann. Stat. 39 (2011a), pp. 82–130. doi: 10.1214/10-AOS827 [DOI] [Google Scholar]
  • 3.Bowden R. and Turkington D., Instrumental Variables, Cambridge University Press, NY., 1984. [Google Scholar]
  • 4.Burgess S. and Dylan S., Predicting the direction of causal effect based on an instrumental variable analysis: A cautionary tale, J. Cau. Inf. 4 (2016), pp. 49–59. doi: 10.1515/jci-2015-0024 [DOI] [Google Scholar]
  • 5.Carrasco M., A regularization approach to the many instruments problem, J. Econometrics 170 (2012), pp. 383–398. doi: 10.1016/j.jeconom.2012.05.012 [DOI] [Google Scholar]
  • 6.Carrasco M. and Tchuente G., Regularized LIML for many instruments, J. Econometrics 186 (2015), pp. 427–442. doi: 10.1016/j.jeconom.2015.02.018 [DOI] [Google Scholar]
  • 7.Didelez V. and Sheehan N.A., Mendelian randomization as an instrumental variable approach to causal inference, Statist. Meth. Med. Res. 16 (2007), pp. 309–330. doi: 10.1177/0962280206077743 [DOI] [PubMed] [Google Scholar]
  • 8.Dixon P., Davey Smith G., and von Hinke S., Pharmaco Economics, 34, 2016, 1075. Available at 10.1007/s40273-016-0432-x [DOI] [PMC free article] [PubMed]
  • 9.Durbin J., Errors in variables, Revue De L'institut International De Statistique 22 (1954), pp. 23–32. doi: 10.2307/1401917 [DOI] [Google Scholar]
  • 10.Ebbes P., Latent instrumental variables – a new approach to solve for endogeneity, Ph.D. thesis, University of Groningen Economics and Business. 2004.
  • 11.Fan J. and Li R., Variable selection via non concave penalized likelihood and its oracle properties, J. Amer. Statist. Assoc. 96 (2001), pp. 1348–1360. doi: 10.1198/016214501753382273 [DOI] [Google Scholar]
  • 12.Garciá C.B., Garciá J., Loṕez Martiń M.M., and Salmeroń R., Collinearity: revisiting the variance inflation factor in ridge regression, J. Appl. Stat. 42 (2015), pp. 648–661. doi: 10.1080/02664763.2014.980789 [DOI] [Google Scholar]
  • 13.Golub G.H., Heath M., and Wahba G., Generalized cross-validation as a method for choosing a good ridge parameter, Technometrics 21 (1979), pp. 215–223. doi: 10.1080/00401706.1979.10489751 [DOI] [Google Scholar]
  • 14.Guo Z., Kang H., Cai T.T., and Small D.S., Testing endogeneity with possibly invalid instruments and high dimensional covariates, arXiv preprint arXiv:1609.06713, 2016
  • 15.Hansen C. and Kozbur D., Instrumental variables estimation with manyWeak instruments using regularized JIVE, J. Econ. 182 (2014), pp. 290–308. doi: 10.1016/j.jeconom.2014.04.022 [DOI] [Google Scholar]
  • 16.Hausman J., Specification tests in econometrics, Econometrica 46 (1978), pp. 1251–1271. doi: 10.2307/1913827 [DOI] [Google Scholar]
  • 17.Hoerl A.E. and Kennard R.W., Ridge regression: biased estimation for nonorthogonal problems, Technometrics 12 (1970), pp. 55–67. doi: 10.1080/00401706.1970.10488634 [DOI] [Google Scholar]
  • 18.Jolliffe I.T., Principal Component Analysis, Springer series in statistics, Vol. 29, 2002. [Google Scholar]
  • 19.Fu W. and Knight K., Asymptotic for lasso-type estimators, Ann. Statist. 28 (2000), pp. 1356–1378. doi: 10.1214/aos/1015957397 [DOI] [Google Scholar]
  • 20.Liu X.Q. and Jiang H.Y., Optimal generalized ridge estimator under the generalized cross-validation criterion in linear regression, Linear Alg. App. 436 (2012), pp. 1238–1245. doi: 10.1016/j.laa.2011.08.032 [DOI] [Google Scholar]
  • 21.Lukman A.F., Ayinde K., Binuomote S., and Onate A.C., Modified ridge-type estimator to combat multicollinearity: application to chemical data, J. Chemom. 2019), pp. e3125. 10.1002/cem.3125. [DOI] [Google Scholar]
  • 22.Lukman A.F., Ayinde K., Sek S.K., and Adewuyi E., A modified new two-parameter estimator in a linear regression model, Modelling and Simul. Eng. (2019), pp. 10.1155/2019/6342702. [DOI] [Google Scholar]
  • 23.Okbay A., Beauchamp J., Fontana M., Lee J., Pers T., Rietveld C., Turley P., Chen G., Emilsson V., Meddens S., and Oskarsson S., Genome-wide association study identifies 74 loci associated with educational attainment, Nature 533 (2016), pp. 539–542. doi: 10.1038/nature17671 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Park C. and Yoon Y., Bridge regression: adaptivity and group selection, J. Statist. Plann. Inf. 141 (2011), pp. 0. [Google Scholar]
  • 25.Park C. and Yoon Y.J., Bridge regression: adaptivity and group selection, J. Stat. Plann. Inf. 141 (2011), pp. 3506–3519. doi: 10.1016/j.jspi.2011.05.004 [DOI] [Google Scholar]
  • 26.Phillips G. and Evans G, Approximating and reducing bias in 2SLS estimation of dynamic simultaneous equation models, Comp. Statist. Data Anal. 100 (2016), pp. 734–762. doi: 10.1016/j.csda.2015.11.011 [DOI] [Google Scholar]
  • 27.Rietveld C., Medland S., Derringer J., Yang J., Esko T., Martin N., Westra H., Shakhbazov K., Abdellaoui A., Agrawal A., and Albrecht E., GWAS of 126,559 individuals identifies genetic variants associated with educational attainment, Science 340 (2013), pp. 1467–1471. doi: 10.1126/science.1235488 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Saleh A.K., Ehsanes Md., Arashi M., and Kibria B.M.G., Theory of Ridge Regression Estimation with Applications, John Wiley, USA, 2019. [Google Scholar]
  • 29.Tibshirani R., Regression shrinkage and selection via the lasso, J. Roy. Statist. Soc.: Ser. B 58 (1996), pp. 267–288. [Google Scholar]
  • 30.Tibshirani R., Regression shrinkage and selection via the lasso, J. Roy. Statist. Soc: Ser. B 58 (1996), pp. 267–288. [Google Scholar]
  • 31.Wu D., Alternative tests of independence between stochastic regressors and disturbances, J. Econometrica 41 (1973), pp. 733. doi: 10.2307/1914093 [DOI] [Google Scholar]
  • 32.Xu-Qing L., Gao F., and Yu Z.-F., Improved ridge estimators in a linear regression model, J. Appl. Stat. 40 (2013), pp. 209–220. doi: 10.1080/02664763.2012.740623 [DOI] [Google Scholar]
  • 33.Yuzbasi B., Arashi M., and Akdeniz F., Penalized regression via the restricted bridge estimator, arXiv:1910.03660, 2019.

Articles from Journal of Applied Statistics are provided here courtesy of Taylor & Francis

RESOURCES