Abstract
In this paper, we introduce stochastic-restricted Liu predictors which will be defined by combining in a special way the two approaches followed in obtaining the mixed predictors and the Liu predictors in the linear mixed models. Superiorities of the linear combination of the new predictor to the Liu and mixed predictors are done in the sense of mean square error matrix criterion. Finally, numerical examples and a simulation study are done to illustrate the findings. In numerical examples, we took some arbitrary observations from the data as the prior information since we did not have historical data or additional information about the data sets. The results show that this case does the new estimator gain efficiency over the constituent estimators and provide accurate estimation and prediction of the data.
Keywords: Multicollinearity, mixed predictor, Liu predictor, stochastic-restricted Liu predictor, linear mixed model
1. Introduction
Many common statistical models can be expressed as linear models that incorporate both fixed effects, which are parameters associated with an entire population or with certain repeatable levels of experimental factors, and random effects, which are associated with individual experimental units drawn at random from a population. A model with both fixed effects and random effects is called a linear mixed model (LMM) [23]. LMMs provide a broad range of structures including longitudinal data, repeated measures data (including cross-over studies), growth and dose-response curve data, clustered (or nested) data, multivariate data and correlated data.
Let us consider the LMM
where is an vector of response variables measured on subject i, β is a parameter vector of fixed effects, and are and known design matrices of the fixed and random effects, respectively, is a random vector, the components of which are called random effects and is an random vector of errors.
Usually one assumes that
where and are independent, D and are and known positive definite (pd) matrices.
Let , , , where ⊕ represents the direct sum, and . Then, we can write the model more compactly as
| (1) |
which implies that
where , and , with ⊗ denoting the Kronecker product and , the identity matrix of order m. Then, under model (1), we get , where . For the convenience of the theoretical computations, we assume that the matrices G and W are known. If G and W are unknown, we replace G and W by their maximum likelihood (ML) or restricted maximum likelihood (REML) estimates and update the estimate of β by where is any estimator of β in LMM.
and are obtained as
| (2) |
| (3) |
by Henderson [6] and Henderson et al. [7]. The Henderson's estimator and predictor given by (2) and (3) are called, respectively, as the best linear unbiased estimator (BLUE) and the best linear unbiased predictor (BLUP).
In linear regression model, we usually assume that the variables of fixed effects design matrix are independent. However, in practice, there may be strong or near to strong linear relationships among the variables of fixed effects design matrix. In that case the independence assumptions are no longer valid, which causes the problem of multicollinearity. In the existence of multicollinearity, at least one main diagonal element of may be quite large, which in view of means that at least one element of may have a large variance, and may be far from its true value.
In statistical research, there have been many attempts to provide better estimators, for example, by the incorporation of prior information available in the form of exact or stochastic restrictions [24, p. 111]. If prior information comes from the theory and imposes restrictions among parameters that should hold in exact terms, it could be included in the model as a deterministic restriction. On the other hand, if the prior information comes from previous estimations of similar but different models or samples, it could be considered as a hint or a range of values that should contain the value of the parameter with some probability. If this information is not taken into account in the estimation despite being good, then the information will be wasted as well as the chance of improve the efficiency of the estimator. This is the idea behind stochastic restrictions approach, and the stochastic restrictions yield asymptotic efficiency gains under some specific assumptions about the asymptotics of prior information (shown in [27,31] for a linear model under normality of the errors). The collection and use of stochastic restriction also solve the multicollinearity problem (see [1,18]). The mixed estimation method suggested by Theil and Goldberger [31] in the linear regression model is extended to LMMs by Kuran and Özkale [12].
In addition to model (1), Kuran and Özkale [12] used the stochastic linear restriction as
| (4) |
where r is an random vector, R is a known prior information of rank and Φ is a random vector, independent of ϵ, with and , where V is a known pd matrix and proposed, respectively, the mixed estimator and the mixed predictor in LMMs as
| (5) |
where is given by Equation (2).
Another attempt to provide better estimation is introduced by Liu [13] called as the Liu estimator. And then, Kaçıranlar et al. [10] improved Liu's approach, and introduced restricted Liu estimator and discussed its stochastic properties. By using Swindel [29] and Özkale and Kaçıranlar's [20] approach in the linear regression models, Liu's approach in the linear regression model is enlarged to LMMs by Özkale and Kuran [21] in view of penalized log-likelihood. Thus, Özkale and Kuran [21] suggested, respectively, the Liu estimator and the Liu predictor in LMMs as
where , 0<d<1 is the Liu biasing parameter and is the BLUE given by Equation (2).
Hubert and Wijekoon [8] introduced another alternative Liu-type estimator which will be defined by combining in a special way the two approaches followed in obtaining the mixed estimator and the Liu estimator in the linear regression models. They called their new biased estimator as the stochastic restricted Liu estimator (SRLE). The mean squared error matrix of SPLR was compared with several other biased estimators, and the conditions needed for the superiority over these biased estimators were derived by Hubert and Wijekoon [8]. Liu estimator is also studied by authors in several models and estimators. The Logistic Liu Estimator (LLE) [14]; the Liu-Type Logistic Estimator (LTLE) [9]; the Almost Unbiased Liu Logistic Estimator (AULLE) [33]; the Restricted Logistic Liu Estimator (RLLE) [30] and the Stochastic Restricted Liu Maximum Likelihood Estimator (SRLMLE) [32] have been proposed in the linear regression literature.
Our primary aim in this article is to widen Hubert and Wijekoon's [8] idea under Özkale and Kaçıranlar's [20] approach in the linear regression models to LMMs and the article is organized as follows. In Section 2, we propose the estimator and predictor in linear form of mixed estimator, abbreviated respectively by LFME and LFMP, in LMMs by using Kuran and Özkale's [12] mixed estimation method and Özkale and Kuran's [21] Liu approach. Superiorities of the linear combinations of the predictors are done in the sense of mean square error (MSE) matrix criterion in Section 3. Numerical examples and a simulation study are done to illustrate the findings in Section 4 and Section 5, respectively. Finally, we give some conclusions in Section 6.
2. Stochastic restricted Liu predictors in linear mixed models
In this section, we estimate the parameter vectors of fixed and random effects via the penalized log-likelihood approach.
Under the assumptions of model (1), u and y are jointly Gaussian distributed as
| (6) |
Then, the conditional distribution of y given u is . Henderson et al. [7] developed a set of equations that simultaneously yield BLUE and BLUP. For this purpose, they maximize the joint density of y and u which will be for Equation (6) as
where denotes the determinant of a matrix.
After the log-joint distribution of is obtained as
we add a penalization term (that is; a penalization term is corresponding to stochastic linear restriction given by Equation (4)) with regularization parameter to ,
| (7) |
where is a diagonal matrix with elements as the Liu biasing parameters and is the mixed estimator given by Equation (5).
The objective function (7) looks for an estimator which maximizes in a class of estimators that is close to than the origin.
As a result of dropping the constant term and taking into consideration of the log function, Equation (7) can also be written as
| (8) |
where the first two terms carry out the Liu estimation idea in the linear regression models and the second term carries out the estimation procedure through the random effects.
Equating the partial derivatives of Equation (8) with respect to the elements of β and u to zero and using and to denote the LFME and LFMP gives
| (9) |
| (10) |
Equations (9) and (10) can compactly be written in matrix as
| (11) |
Using Gilmour et al.'s [5] approach, Equation (11) can be written as
| (12) |
where , , and Ģ is full rank with
where the superscript ‘+’ denotes the Moore-Penrose inverse.
Solving Equation (12), we get
| (13) |
where is found from the inverse formula of the partitioned matrix [26] as
where and .
After is replaced in Equation (13) and after algebraic simplifications, we get the LFME and the LFMP, respectively, as
For the special case of , we rewrite as
where 0<d<1 is the Liu biasing parameter. carries the idea of Hubert and Wijekoon [8].
3. The comparisons of the predictors in linear mixed models
Prediction of linear combinations of β and u can be expressed as for specific matrices L and . This type of prediction problem was investigated by Yang et al. [34], Pereira and Coelho [22] and Robinson [25].
Under the LFMP, the predictor of μ is expressible as
where . One of the criteria proposed for measuring the ‘betterness’ of is taken to be the MSE matrix criterion.
Following Štulajter [28], we can write the MSE matrix for as
| (14) |
where .
, , and can be derived, respectively, as
| (15) |
| (16) |
| (17) |
and
| (18) |
where
and . Then, Equations (14), (15), (16) and (17) are put in Equation (14) to obtain
| (19) |
where
| (20) |
In the same manner, we obtain
| (21) |
| (22) |
where
| (23) |
| (24) |
are given, respectively, by Özkale and Kuran [21] and Kuran and Özkale [12]. When we examine Equations (19), (21) and (22), we can say that the superiority of over and is equivalent to the superiority of over and , respectively.
3.1. The SRLE vs the Liu estimator
Theorem 3.1
The LFME is always superior to the Liu estimator in the MSE matrix criterion.
Proof.
By using Equations (20) and (23), we can write
Note that,
Following Theorem A.1 in Appendix 1, we say that since is nonnegative definite (nnd), is also nnd. Hence is nnd and that impiles the difference is a positive semidefinite (psd) matrix. Then, the proof is completed.
3.2. The SRLE vs the mixed estimator
Theorem 3.2
The LFME is superior to the mixed estimator for fixed 0<d<1 in the MSE matrix sense if and only if is satisfied where .
Proof.
By using Equations (20) and (24), we can write
where . By utilizing , we write the difference C as
(25) Equation (25) shows us that the matrix C is pd for any 0<d<1. Then, from the theorem of Farebrother [4], is nnd if and only if . Since , we obtain the necessary and sufficient condition as
which completes the proof. Here d can be selected anyway such as suggested by Özkale and Kuran [21].
4. Numerical examples
4.1. A hypothetical data analysis
In this analysis, we use the results of Los Alamos study of high efficiency particulate air (HEPA) filter cartridges presented by Kerschner et al. [11]. Such cartridges are used with commercial respirators to provide respiratory protection against dusts, toxic fumes, mists, radionuclides and other particulate matter.
The primary objective of the study was to determine whether the current standard aerosol used to test these filters could be replaced by any of alternate aerosols for quality-assurance testing of HEPA respirator cartridges. HEPA respirator filters fail a quality-assurance test if the challenge penetration is greater than some percent penetration breakpoint otherwise, it is considered to have passed. A secondary objective was to identify those factors that contribute most to the variability in the penetrations of the filters.
In a subset of the aerosol data set [2], two aerosols were crossed with two filter manufacturers. Within each manufacturer, three filters were used to evaluate the penetration of the two aerosols, so that there were six filters in total. By taking a filter nested within the two manufacturers, the following linear mixed model with independent random effects and independent random errors is given as
| (26) |
where is the percent penetration, μ is an intercept, is a fixed effect for the ith aerosol type , is a fixed effect for the jth filter manufacturer , is a random effect for the hth filter nested within the the jth manufacturer and is the error associated with the lth replication in the ijhlth cell , subject to the usual restrictions: .
Clearly, model (26) may be put into the form of model (1) with the vector of responses and approximate design matrices
| (27) |
| (28) |
Here, , and , , , , , , are vectors of ones and zeros where subscripts show the dimensions. y in (27) is a vector of the percent penetration values, is an arbitrary vector and with , .
It was seen that multicollinearity did not appear at the aerosol data examined by Beckman et al. [2] and Kerschner et al. [11]. But, to serve for our purposes, the data must have multicollinearity. Thus, by using this aerosol data and the model in (26), hypothetical data are produced as follows.
Since multicollinearity must be at the X matrix corresponding to fixed effects as explained in theory, the data corresponding to the fixed effects are generated using the following equation in McDonald and Galarneau [15]
where are taken from (27) and δ is specified so that correlation between any two fixed effects is given by . These effects are then standardized so that the fixed effects design matrix is in correlation form and for this hypothetical data, is considered.
The stochastic linear restrictions are taken as
where r = 5.000 is the 14th element of y vector in Equation (28), R is the 14th row of generated X matrix and is an arbitrary vector. Because of having the largest percent penetration value, the 14th element of y vector is chosen. But, the other elements of y vector and corresponding elements of the X matrix can also be taken as r and R, respectively.
Generally, since there is not too much difference between the results of ML and REML covariance estimates (see [19]), we choose REML approach in our hypothetical data analysis and we obtain , and . Then, from the formula , is obtained.
The eigenvalues of the matrix are obtained as , and . Then, the condition number, , where and indicate the minimum and maximum eigenvalues of is calculated as . Since the condition number is used to measure the extent of multicollinearity in the data and it is larger than 1000, it shows severe multicollinearity (see for example Montgomery et al. [16], p.298).
The Liu biasing parameter d is computed as which is defined by Özkale and Kuran [21] where and P is an orthogonal matrix such that with a diagonal matrix .
Table 1 presents the parameter estimates and scalar MSE values (trace of matrix MSE) when the variance of y is estimated by REML. From Table 1, we observe that LFME outperforms the other estimators which is followed by the mixed estimator.
Table 1. Parameter estimates and MSE scalar values for when the covariance parameters are estimated by REML.
| μ | 10.8672 | – | 20.0638 | – | 10.4350 | – | 19.2684 | – |
| – | – | – | – | |||||
| 6.9127 | – | 11.8321 | – | 6.6414 | – | 11.3702 | – | |
| – | – | – | 0.0198 | – | ||||
| – | 0.0736 | – | 0.0736 | – | 0.0731 | – | 0.0732 | |
| – | – | – | – | |||||
| – | – | – | – | |||||
| – | – | – | – | |||||
| – | 0.0516 | – | – | 0.0583 | – | |||
| – | 0.0023 | – | – | 0.0090 | – | |||
| scalarMSE | 16.8331 | – | 11.1831 | – | 15.5766 | – | 10.3698 | – |
To consider the superiorities of the estimators for other d values, plot of the scalar MSE values of , , , versus d are presented by Figure 1. Figure 1 demonstrates that although is better than , and are better than for d values respectively in intervals 0.342<d<1 and 0.362<d<1. And, we can also say that though is better than for d values in interval 0.524<d<1, is better than for d values in interval 0<d<1. For d values in interval 0<d<1, has smaller MSE scalar values than and it can be seen that as d values increase, the difference between MSE scalar values of and increases. is the best among the estimators for d values in the interval 0.524<d<1. Therefore, we say that there always exists a d value such that is better than , or .
Figure 1.
MSE scalar values of BLUE, mixed, Liu and LFME versus d.
The theorems in Section 3 are also examined. And so, under the matrix MSE criterion, is always better than . The condition given by Theorem 3.2 is computed as which is not smaller than , hence is not better than for under the matrix MSE criterion though it is better than under the scalar MSE criterion. However, there can be found a d value such that dominates in terms of the matrix MSE criterion; these d values for this data set are found to be greater than 0.9898.
The hypothetical data analysis summarizes that the SRLE dominates the BLUE, mixed and Liu estimators in the sense of the scalar MSE criterion for appropriate selected d value under multicollinearity.
4.2. Real data analysis: greenhouse gases data
Greenhouse gases trap heat and make the planet warmer. Human activities are responsible for almost all of the increase in greenhouse gases in the atmosphere over the past 150 years. The largest source of greenhouse gas emissions from human activities in the United States is from burning fossil fuels for electricity, heat, and transportation (see the official United States Environmental Protection Agency (EPA) website [3]). The transportation sector generates the largest share of greenhouse gas emissions in the United States (nearly 28.5 % of 2016 greenhouse gas emissions). Greenhouse gas emissions from transportation come from burning fossil fuel for cars, light/heavy trucks-buses, motorcycles and railways. The remaining greenhouse gas emissions from the transportation sector come from other modes of transportation, including commercial aircraft, ships and boats (see EPA website [3]).
In this analysis, we use data on 297 fuel combustion in transport (in million tonnes) randomly selected from 27 areas (Belgium, Bulgaria, European Union, Denmark, Germany, Estonia, Greece, Spain, France, Croatia, Italy, Ireland, Latvia, Lithuania, Luxembourg, Hungary, Netherlands, Austria, Poland, Portugal, Romania, Slovenia, Slovakia, Switzerland, Sweden, United Kingdom, Norway) which were regularly obtained during the period 2006–2016 inclusive.
To determine fuel combustion in transport (in million tonnes), we get repeated measurements from the list of fuel combustions, namely the cars, the light duty trucks, the heavy duty trucks-buses, the motorcycles and railways. The data are available from the official Eurostat website [3].
We define the response (y) as fuel combustion in transport (in million tonnes) and we get the explanatory variables as fuel combustions in cars (), light duty trucks (), heavy duty trucks-buses (), motorcycles () and railways (). That is, we can say that fuel combustions in cars, light duty trucks, heavy duty trucks-buses, motorcycles, railways are expressed as fixed effects and since the 27 areas are randomly selected from the areas, the areas factor effect on the response is expressed as random effect. Then, the random intercept and slope model (RISM) is given as
where denotes the ith observation of the jth area of the response, denotes the ith observation of the jth area of the explanatory variable , and shows time corresponding to .
The stochastic linear restriction is taken as
where r = 983.37034 is the th observation, R is the vector of th observations for and is an arbitrary vector. Because of having the largest cereal production value, the th observation is chosen. But, the other observations of vector and corresponding elements of the observations can also be taken as r and R, respectively.
Generally, since there is not too much difference between the results of ML and REML covariance estimates (see [19]), we prefer ML approach in our real data analysis and we get , and . Then, from the formula , is obtained.
The eigenvalues of the matrix are obtained as , 9.260924e + 02, , 2.571984 and 15.225476. Then, the condition number is calculated as and it shows severe multicollinearity (see for example Montgomery et al. [16], p.298).
The Liu biasing parameter is computed from Özkale and Kuran [21] as . Table 2 gives the parameter estimates and scalar MSE values when the variance of y is estimated by ML. From Table 2, we observe that LFME outperforms the mixed estimator, the Liu estimator and the BLUE in the sense of scalar MSE.
Table 2. Parameter estimates and MSE scalar values for when the covariance parameters are estimated by ML.
| 1.0204 | – | 1.0046 | – | 1.0212 | – | 1.0055 | – | |
| 1.0507 | – | 1.0406 | – | 1.0511 | – | 1.0410 | – | |
| 0.9380 | – | 0.9637 | – | 0.9374 | – | 0.9630 | – | |
| 3.3953 | – | 3.5860 | – | 3.3794 | – | 3.5688 | – | |
| 3.7790 | – | 4.1731 | – | 3.7579 | – | 4.1502 | – | |
| – | 0.1329 | – | 0.1014 | – | 0.1332 | – | 0.1017 | |
| – | – | – | – | |||||
| scalarMSE | 0.1511 | – | 0.1472 | – | 0.1497 | – | 0.1459 | – |
To consider the superiorities of the estimators for other d values, plot of the scalar MSE values of , , , versus d are presented by Figure 2. Figure 2 demonstrates that although is better than for d values in interval 0<d<1, and are better than for d values respectively in intervals 0.508<d<1 and 0.482<d<1. And, we can also say that though is better than for d values in interval 0.535<d<1, is better than for d values in interval 0.563<d<1. For d values in interval 0<d<1, has smaller scalar MSE values than and it can be seen that as d values increase, the difference between scalar MSE values of and increases. As compared to Figure 1, we see that the tendencies of the estimators towards each other are same.
Figure 2.
MSE scalar values of BLUE, mixed, Liu and LFME versus d.
In examining Theorem 3.1, we see that is always better than in the sense of matrix MSE. The condition given by Theorem 3.2 is computed as which is smaller than , hence is better than under the matrix MSE matrix criterion.
Under multicollinearity, this real data analysis summarizes that the LFME dominates the BLUE, mixed and Liu estimators in the sense of both the scalar MSE and matrix MSE criteria for appropriate selected d value.
5. A simulation study
In this section, we will discuss the simulation study to compare the performance of, , and and the performance of , , and in the sense of the estimated mean square error (EMSE) and the predicted mean square error (PMSE).
The fixed effects are computed from McDonald and Galarneau [15] as
where are independent standard normal pseudo-random numbers and γ is specified so that the correlation between any two fixed effects is given by . Three different sets of correlations are considered corresponding to , 0.95 and 0.99. The number of fixed effects are chosen as p = 2 and p = 4. To fix the generated explanatory variables, 2019 is used as the state number.
We consider m = 30, 60 subjects and , 10 observations per subject. The stochastic linear restrictions are taken as
where and are chosen arbitrary. The parameter vector is chosen as the normalized eigenvector corresponding to the largest eigenvalue of so that (see [17]). Then, the underlying model takes the following form with q = 2 random effects
where D = V are both taken as is the AR(1) process with and shows time which was taken as the same set of occasions, { for }. Three values of are investigated as 0.5, 1 and 10.
By using Özkale and Kuran [21], the Liu biasing parameter is selected as which is given by Theorem 4.1 where h is determined as multiplying the upper bound defined in Theorem 4.1 by 0.99 if and determined arbitrarily as if . This arbitrarily chosen is determined by inspiring the numerical examples presented in Section 4.
For each choice of m, , γ, p and , the experiment is replicated 500 times by generating response variable and the EMSE for any estimator of β and the PMSE for any predictor of u are calculated, respectively, as
where the subscript ŗ refers to the ŗth replication.
To see the degree of supremacy, after the EMSE and the PMSE results are obtained, the change percentage values are calculated as
where and are any two estimators and and are any two predictors.
The simulation results are summarized by Tables 3–10. Based on Tables 3–10, we have the following conclusions:
Table 4. Predicted MSE values with p = 2, m = 30 and .
| 0.90 | ||||||||
| 0.95 | ||||||||
| 0.99 |
Table 5. Estimated MSE values with p = 2, m = 60 and .
| 0.90 | ||||||||
| 0.95 | ||||||||
| 0.99 |
Table 6. Predicted MSE values with p = 2, m = 60 and .
| 0.90 | ||||||||
| 0.95 | ||||||||
| 0.99 |
Table 7. Estimated MSE values with p = 4, m = 30 and .
| 0.90 | ||||||||
| 0.95 | ||||||||
| 0.99 |
Table 8. Predicted MSE values with p = 4, m = 30 and .
| 0.90 | ||||||||
| 0.95 | ||||||||
| 0.99 |
Table 9. Estimated MSE values with p = 4, m = 60 and .
| 0.90 | ||||||||
| 0.95 | ||||||||
| 0.99 |
Table 3. Estimated MSE values with p = 2, m = 30 and .
| 0.90 | ||||||||
| 0.95 | ||||||||
| 0.99 |
Table 10. Predicted MSE values with p = 4, m = 60 and .
| 0.90 | ||||||||
| 0.95 | ||||||||
| 0.99 |
For fixed value of sample size, correlation and p, the EMSE values of , , and increase with the increase of .
When sample size, and p are fixed, the superiorities of over , and increase as increases. And, the superiority over and can be seen more clear than the superiority over .
The , , and values increase with the increase of p values under fixed value of sample size, correlation and .
As increases from 150 to 600 at fixed , and p, the EMSE values of , , and decreases.
As the degree of correlation increases under fixed values of sample size, and p, the , , and values increase.
The values are smaller than the , and values in all cases. This case is seen from the change percentages which are all positive. The change percentage indicates that as and increases at fixed sample size and p, the degrees of supremacies of over , and increase. Furthermore, we can say that the degree of supremacies of over and is too much than the degree of supremacy of over .
The , , and values increase with the increase of for fixed value of sample size, correlation and p. And, this increase can be also seen more clearly as p values increase.
With the increase of p values under fixed value of sample size, correlation and , the PMSE values of , , and increase.
As increases from 150 to 600 at fixed , and p, the , , and values decrease.
The PMSE values of , , and are very similar to each other in all cases. Although the values are smaller than the , and values, we see that the values are smaller than the , and values as p values increase. But, consequently, we say that the difference between the PMSE values of , , and is not too much.
The change percentage indicates that as and increases at fixed sample size and p, the degrees of supremacies of over , and increase. Moreover, we can say that the degree of supremacies of over and is too much than the degree of supremacy of over .
In Tables 3–10, employs better approximation (that is, values are closer) to the true values than , and .
6. Discussion
In this article, we propose the LFME and LFMP in LMMs with the help of Kuran and Özkale's [12] mixed estimation method and Özkale and Kuran's [21] Liu approach. Then, matrix MSE comparisons are done and lastly, the study is supported with both numerical examples and a simulation study.
This study has confirmed that, under multicollinearity, the use of the LFME outperforms the Liu estimator for any selected Liu biasing parameter and the mixed estimator for appropriate selected Liu biasing parameter in terms of mean square error. If there exists prior information as in the form of stochastic linear restrictions, one can use the LFME and LFMP in an LMM. The advantage of the LFME and LFMP over the Liu estimator and predictor is that LFME and LFMP consider the prior information and over the mixed estimator and predictor is that they gain efficiency.
Appendix 1. Matrix algebra.
Theorem A.1 Rao and Toutenburg [24], p. 302 —
Let A>0 and B>0. Then, B−A>0 if and only if .
Disclosure statement
No potential conflict of interest was reported by the authors.
References
- 1.Baksalary J.K. and Trenkler G., Nonnegative and positive definiteness of matrices modified by two matrices of rank one, Linear Algebra Appl. 151 (1991), pp. 169–184. doi: 10.1016/0024-3795(91)90362-Z [DOI] [Google Scholar]
- 2.Beckman R.J., Nachtsheim C.J., and Cook R.D., Diagnostics for mixed model analysis of variance, Technometrics 29 (1987), pp. 413–426. [Google Scholar]
- 3.Eurostat website . Available at http://appsso.eurostat.ec.europa.eu/nui/show.do?dataset, 2018.
- 4.Farebrother R.W., Further results on the mean square error of ridge regression, J. R. Stat. Soc. B 38 (1976), pp. 248–250. [Google Scholar]
- 5.Gilmour A.R., Thompson R., and Cullis B.R., Average information REML: an efficient algorithm for variance parameter estimation in linear mixed models, Biometrics 51 (1995), pp. 1440–1450. doi: 10.2307/2533274 [DOI] [Google Scholar]
- 6.Henderson C.R., Estimation of genetic parameters (abstract), Ann. Math. Stat. 21 (1950), pp. 309–310. [Google Scholar]
- 7.Henderson C.R., Kempthorne O., Searle S.R., and Von Krosig C.N., Estimation of environmental and genetic trends from records subject to culling, Biometrics 15 (1959), pp. 192–218. doi: 10.2307/2527669 [DOI] [Google Scholar]
- 8.Hubert M.H. and Wijekoon P., Improvement of the Liu estimator in linear regression model, Stat. Papers 47 (2006), pp. 471–479. doi: 10.1007/s00362-006-0300-4 [DOI] [Google Scholar]
- 9.Inan D. and Erdoğan B.E., Liu-Type logistic estimator, Commun. Stat. Simul. Comput. 42 (2013), pp. 1578–1586. doi: 10.1080/03610918.2012.667480 [DOI] [Google Scholar]
- 10.Kaçıranlar S., Sakallıoğlu S., Akdeniz F., Styan G.P.H., and Werner H.J., A new biased estimator in linear regression and a detailed analysis of the widely analyzed dataset on Portland Cement, Sankhya: The lndian J. Stat. 61B (1999), pp. 443–459. [Google Scholar]
- 11.Kerschner H.F., Ettinger H.J., DeField J.D., and Beckman R.J., A Comparative Study of HEPA Filter Efficiencies When Challenged with Thermal and Air-Jet Generated Di-2-Ethylhexyl Sebecate, Di-2-Ethylhexyl Phthalate and Sodium Chloride, Laboratory Report LA-9985-MS, Los Alamos National Laboratory, Los Alamos, 1984
- 12.Kuran Ö. and Özkale M.R., Gilmour's approach to mixed and stochastic restricted ridge predictions in linear mixed models, Linear Algebra Appl. 508 (2016), pp. 22–47. doi: 10.1016/j.laa.2016.06.040 [DOI] [Google Scholar]
- 13.Liu K., A new class of biased estimate in linear regression, Commun. Stat. Theor. Meth. 22 (1993), pp. 393–402. doi: 10.1080/03610929308831027 [DOI] [Google Scholar]
- 14.Mansson G., Kibria B.M.G., and Shukur G., On Liu Estimators for the Logit Regression Model, The Royal Institute of Technology, Centre of Excellence for Science and Innovation Studies (CESIS), Sweden, Paper No. 259, 2012
- 15.McDonald M.C. and Galarneau D.I., A Monte Carlo evaluation of some ridge-type estimators, J. Am. Stat. Assoc. 70 (1975), pp. 407–416. doi: 10.1080/01621459.1975.10479882 [DOI] [Google Scholar]
- 16.Montgomery D.C., Peck E.A., and Vining G.G., Introduction to Linear Regression Analysis, John Wiley and Sons, Inc., Hoboken, NJ, 2012. [Google Scholar]
- 17.Newhouse J.P. and Oman S.D., An evaluation of ridge estimators, Rand. Corporation R-716-PR, 1971, pp. 1–28.
- 18.Özkale M.R., A stochastic restricted ridge regression estimator, J. Multivariate Anal. 100 (2009), pp. 1706–1716. doi: 10.1016/j.jmva.2009.02.005 [DOI] [Google Scholar]
- 19.Özkale M.R. and Can F., An evaluation of ridge estimator in linear mixed models: an example from kidney failure data, J. Appl. Stat. 44 (2017), pp. 2251–2269. doi: 10.1080/02664763.2016.1252732 [DOI] [Google Scholar]
- 20.Özkale M.R. and Kaçıranlar S., The restricted and unrestricted two-parameter estimators, Commun. Stat. Theor. Meth. 36 (2007), pp. 2707–2725. doi: 10.1080/03610920701386877 [DOI] [Google Scholar]
- 21.Özkale M.R. and Kuran Ö., A further prediction method in linear mixed models: Liu prediction, Commun. Stat. Simul. Comput. (2018), doi: 10.1080/03610918.2018.1535071. [DOI] [Google Scholar]
- 22.Pereira L.N. and Coelho P.S., A small area predictor under area-level linear mixed models with restrictions, Commun. Stat. Theor. Meth. 41 (2012), pp. 2524–2544. doi: 10.1080/03610926.2011.648000 [DOI] [Google Scholar]
- 23.Pinherio J.C. and Bates D.M., Mixed-Effects in S and S-Plus, Springer-Verlag, New York, 2000. [Google Scholar]
- 24.Rao C.R. and Toutenburg H., Linear Models-Least Squares and Alternatives, Springer-Verlag, New York, 1995. [Google Scholar]
- 25.Robinson G.K., That BLUP is a good thing: the estimation of random effects (with discussion), Stat. Sci. 6 (1991), pp. 15–32. doi: 10.1214/ss/1177011926 [DOI] [Google Scholar]
- 26.Searle S.R., Matrix Algebra Useful for Statistics, John Wiley and Sons, New York, 1982. [Google Scholar]
- 27.Shiller R., A distributed lag estimator derived from smoothness priors, Econometrica 41 (1973), pp. 775–778. doi: 10.2307/1914096 [DOI] [Google Scholar]
- 28.Štulajter F., Predictions in nonlinear regression models, Acta Math. Univ. Comenianae LXVI (1997), pp. 71–81. [Google Scholar]
- 29.Swindel B.F., Good estimators based on prior information, Commun. Stat. Theor. Meth. 5 (1976), pp. 1065–1075. doi: 10.1080/03610927608827423 [DOI] [Google Scholar]
- 30.Şiray G.Ü., Toker S., and Kaçiranlar S., On the restricted Liu estimator in logistic regression model, Commun. Stat. Simul. Comput. 44 (2015), pp. 217–232. doi: 10.1080/03610918.2013.771742 [DOI] [Google Scholar]
- 31.Theil H. and Goldberger A.S., On pure and mixed statistical estimation in economics, Int. Econ. Rev. 2 (1961), pp. 65–78. doi: 10.2307/2525589 [DOI] [Google Scholar]
- 32.Varathan N. and Wijekoon P., Logistic Liu Estimator under stochastic linear restrictions, Stat. Papers 60 (2019), pp. 945–962. doi: 10.1007/s00362-016-0856-6 [DOI] [Google Scholar]
- 33.Xinfeng C., On the almost unbiased ridge and Liu estimator in the logistic regression model, International Conference on Social Science, Education Management and Sports Education, Atlantis Press, 2015, pp. 1663–1665
- 34.Yang H., Ye H., and Xue K., A further study of predictions in linear mixed models, Commun. Stat. Theor. Meth. 43 (2014), pp. 4241–4252. doi: 10.1080/03610926.2012.725497 [DOI] [Google Scholar]


