Skip to main content
Journal of Applied Statistics logoLink to Journal of Applied Statistics
. 2020 Oct 10;48(5):924–942. doi: 10.1080/02664763.2020.1833182

Improvement of mixed predictors in linear mixed models

Özge Kuran a,CONTACT, M Revan Özkale b
PMCID: PMC9041920  PMID: 35707446

Abstract

In this paper, we introduce stochastic-restricted Liu predictors which will be defined by combining in a special way the two approaches followed in obtaining the mixed predictors and the Liu predictors in the linear mixed models. Superiorities of the linear combination of the new predictor to the Liu and mixed predictors are done in the sense of mean square error matrix criterion. Finally, numerical examples and a simulation study are done to illustrate the findings. In numerical examples, we took some arbitrary observations from the data as the prior information since we did not have historical data or additional information about the data sets. The results show that this case does the new estimator gain efficiency over the constituent estimators and provide accurate estimation and prediction of the data.

Keywords: Multicollinearity, mixed predictor, Liu predictor, stochastic-restricted Liu predictor, linear mixed model

1. Introduction

Many common statistical models can be expressed as linear models that incorporate both fixed effects, which are parameters associated with an entire population or with certain repeatable levels of experimental factors, and random effects, which are associated with individual experimental units drawn at random from a population. A model with both fixed effects and random effects is called a linear mixed model (LMM) [23]. LMMs provide a broad range of structures including longitudinal data, repeated measures data (including cross-over studies), growth and dose-response curve data, clustered (or nested) data, multivariate data and correlated data.

Let us consider the LMM

yi=Xiβ+Ziui+εi,i=1,,m

where yi is an ni×1 vector of response variables measured on subject i, β is a p×1 parameter vector of fixed effects, Xi and Zi are ni×p and ni×q known design matrices of the fixed and random effects, respectively, ui is a q×1 random vector, the components of which are called random effects and εi is an ni×1 random vector of errors.

Usually one assumes that

uiiidNq0,σ2DandεiiidNni0,σ2Wi,i=1,,m

where ui and εi are independent, D and Wi are q×q and ni×ni known positive definite (pd) matrices.

Let y=(y1,,ym), X=(X1,,Xm), Z=i=1mZi, where ⊕ represents the direct sum, u=(u1,,um) and ε=(ε1,,εm). Then, we can write the model yi more compactly as

y=Xβ+Zu+ε (1)

which implies that

uεNqm+n0qm0n, σ2G00σ2W

where n=i=1mni, G=ImD and W=i=1mWi, with ⊗ denoting the Kronecker product and Im, the identity matrix of order m. Then, under model (1), we get yN(Xβ,σ2H), where H=ZGZ+W. For the convenience of the theoretical computations, we assume that the matrices G and W are known. If G and W are unknown, we replace G and W by their maximum likelihood (ML) or restricted maximum likelihood (REML) estimates and update the estimate of β by β~ where β~ is any estimator of β in LMM.

βˆ and uˆ are obtained as

βˆ=(XH1X)1XH1y (2)
uˆ=GZH1(yXβˆ) (3)

by Henderson [6] and Henderson et al. [7]. The Henderson's estimator and predictor given by (2) and (3) are called, respectively, as the best linear unbiased estimator (BLUE) and the best linear unbiased predictor (BLUP).

In linear regression model, we usually assume that the variables of fixed effects design matrix are independent. However, in practice, there may be strong or near to strong linear relationships among the variables of fixed effects design matrix. In that case the independence assumptions are no longer valid, which causes the problem of multicollinearity. In the existence of multicollinearity, at least one main diagonal element of (XH1X)1 may be quite large, which in view of Var(βˆ)=σ2(XH1X)1 means that at least one element of βˆ may have a large variance, and βˆ may be far from its true value.

In statistical research, there have been many attempts to provide better estimators, for example, by the incorporation of prior information available in the form of exact or stochastic restrictions [24, p. 111]. If prior information comes from the theory and imposes restrictions among parameters that should hold in exact terms, it could be included in the model as a deterministic restriction. On the other hand, if the prior information comes from previous estimations of similar but different models or samples, it could be considered as a hint or a range of values that should contain the value of the parameter with some probability. If this information is not taken into account in the estimation despite being good, then the information will be wasted as well as the chance of improve the efficiency of the estimator. This is the idea behind stochastic restrictions approach, and the stochastic restrictions yield asymptotic efficiency gains under some specific assumptions about the asymptotics of prior information (shown in [27,31] for a linear model under normality of the errors). The collection and use of stochastic restriction also solve the multicollinearity problem (see [1,18]). The mixed estimation method suggested by Theil and Goldberger [31] in the linear regression model is extended to LMMs by Kuran and Özkale [12].

In addition to model (1), Kuran and Özkale [12] used the stochastic linear restriction as

r=Rβ+Φ (4)

where r is an s×1 random vector, R is a known s×p prior information of rank sp and Φ is a random vector, independent of ϵ, with E(Φ)=0 and Var(Φ)=σ2V, where V is a known pd matrix and proposed, respectively, the mixed estimator and the mixed predictor in LMMs as

βˆme=(XH1X+RV1R)1(XH1y+RV1r)=βˆ+(XH1X)1R(V+R(XH1X)1R)1(rRβˆ)uˆmp=GZH1(yXβˆme). (5)

where βˆ is given by Equation (2).

Another attempt to provide better estimation is introduced by Liu [13] called as the Liu estimator. And then, Kaçıranlar et al. [10] improved Liu's approach, and introduced restricted Liu estimator and discussed its stochastic properties. By using Swindel [29] and Özkale and Kaçıranlar's [20] approach in the linear regression models, Liu's approach in the linear regression model is enlarged to LMMs by Özkale and Kuran [21] in view of penalized log-likelihood. Thus, Özkale and Kuran [21] suggested, respectively, the Liu estimator and the Liu predictor in LMMs as

βˆd=(XH1X+Ip)1(XH1y+dβˆ)=Fdβˆuˆd=GZH1(yXβˆd)

where Fd=(XH1X+Ip)1(XH1X+dIp), 0<d<1 is the Liu biasing parameter and βˆ is the BLUE given by Equation (2).

Hubert and Wijekoon [8] introduced another alternative Liu-type estimator which will be defined by combining in a special way the two approaches followed in obtaining the mixed estimator and the Liu estimator in the linear regression models. They called their new biased estimator as the stochastic restricted Liu estimator (SRLE). The mean squared error matrix of SPLR was compared with several other biased estimators, and the conditions needed for the superiority over these biased estimators were derived by Hubert and Wijekoon [8]. Liu estimator is also studied by authors in several models and estimators. The Logistic Liu Estimator (LLE) [14]; the Liu-Type Logistic Estimator (LTLE) [9]; the Almost Unbiased Liu Logistic Estimator (AULLE) [33]; the Restricted Logistic Liu Estimator (RLLE) [30] and the Stochastic Restricted Liu Maximum Likelihood Estimator (SRLMLE) [32] have been proposed in the linear regression literature.

Our primary aim in this article is to widen Hubert and Wijekoon's [8] idea under Özkale and Kaçıranlar's [20] approach in the linear regression models to LMMs and the article is organized as follows. In Section 2, we propose the estimator and predictor in linear form of mixed estimator, abbreviated respectively by LFME and LFMP, in LMMs by using Kuran and Özkale's [12] mixed estimation method and Özkale and Kuran's [21] Liu approach. Superiorities of the linear combinations of the predictors are done in the sense of mean square error (MSE) matrix criterion in Section 3. Numerical examples and a simulation study are done to illustrate the findings in Section 4 and Section 5, respectively. Finally, we give some conclusions in Section 6.

2. Stochastic restricted Liu predictors in linear mixed models

In this section, we estimate the parameter vectors of fixed and random effects via the penalized log-likelihood approach.

Under the assumptions of model (1), u and y are jointly Gaussian distributed as

uyN0Xβ,σ2GGZZGH. (6)

Then, the conditional distribution of y given u is y|uN(Xβ+Zu,σ2W). Henderson et al. [7] developed a set of equations that simultaneously yield BLUE and BLUP. For this purpose, they maximize the joint density of y and u which will be for Equation (6) as

fy,u=fy|ufu=(2πσ2)(n+qm)/2W1/2G1/2×exp12σ2[yXβZuW1yXβZu+uG1u]

where |.| denotes the determinant of a matrix.

After the log-joint distribution of f(y,u) is obtained as

logfy,u=logfy|u+logfu=12{(n+qm)log2π+(n+qm)logσ2+logW+logG+[yXβZu)W1yXβZu+uG1u]/σ2}

we add a penalization term (that is; a penalization term is corresponding to stochastic linear restriction given by Equation (4)) with regularization parameter δ=12σ20 to logf(y,u),

logfy,u12σ2(βDβˆme)(βDβˆme) (7)

where D=diag(d1,,dp) is a diagonal matrix with elements 0<di<1 i=1,,p as the Liu biasing parameters and βˆme is the mixed estimator given by Equation (5).

The objective function (7) looks for an estimator which maximizes logf(y,u) in a class of estimators that is close to Dβˆme than the origin.

As a result of dropping the constant term and taking into consideration of the log function, Equation (7) can also be written as

12σ2{yXβW1yXβ+(βDβˆme)(βDβˆme)}12σ2{u(ZW1Z+G1)u2yXβW1Zu} (8)

where the first two terms carry out the Liu estimation idea in the linear regression models and the second term carries out the estimation procedure through the random effects.

Equating the partial derivatives of Equation (8) with respect to the elements of β and u to zero and using βˆlfme and uˆlfmp to denote the LFME and LFMP gives

XW1(yXβˆlfme)+DβˆmeβˆlfmeXW1Zuˆlfmp=0 (9)
ZW1(yXβˆlfme)(ZW1Z+G1)uˆlfmp=0. (10)

Equations (9) and (10) can compactly be written in matrix as

XW1X+IpXW1ZZW1XZW1Z+G1βˆlfmeuˆlfmp=XW1y+DβˆmeZW1y. (11)

Using Gilmour et al.'s [5] approach, Equation (11) can be written as

CΨˆ=ωW1y+κ (12)

where Ψˆ=(βˆlfme,uˆlfmp), ω=(X,Z), κ=(Dβˆme,0)and C=ωW1ω+Ģ + is full rank with

G=Ip00GandG+=Ip00G1

where the superscript ‘+’ denotes the Moore-Penrose inverse.

Solving Equation (12), we get

Ψˆ=C1ωW1y+C1κ (13)

where C1 is found from the inverse formula of the partitioned matrix [26] as

C1=N´N´XH1ZGGZH1XN´Υ+GZH1XN´XH1ZG

where N´=(XH1X+Ip)1 and Υ=(ZW1Z+G1)1.

After C1 is replaced in Equation (13) and after algebraic simplifications, we get the LFME and the LFMP, respectively, as

βˆlfme=(XH1X+Ip)1(XH1y+Dβˆme)uˆlfmp=GZH1(yXβˆlfme).

For the special case of D=dIp, we rewrite βˆlfme as 

βˆlfme=(XH1X+Ip)1(XH1X+dIp)βˆme=Fdβˆme

where 0<d<1 is the Liu biasing parameter. βˆlfme carries the idea of Hubert and Wijekoon [8].

3. The comparisons of the predictors in linear mixed models

Prediction of linear combinations of β and u can be expressed as μ=Lβ+Mu for specific matrices LRp×1 and MRq×1. This type of prediction problem was investigated by Yang et al. [34], Pereira and Coelho [22] and Robinson [25].

Under the LFMP, the predictor of μ is expressible as

μˆlfmp=Lβˆlfme+Muˆlfmp=Qβˆlfme+MGZH1y

where Q=LMGZH1X. One of the criteria proposed for measuring the ‘betterness’ of μˆlfmp is taken to be the MSE matrix criterion.

Following Štulajter [28], we can write the MSE matrix for μˆlfmp as

MSE(μˆsrlp)=E[(μˆlfmpμ)(μˆlfmpμ)]=Var(μˆlfmp)+Var(μ)+bias(μˆlfmp)bias(μˆlfmp)Cov(μˆlfmp,μ)Cov(μ,μˆlfmp), (14)

where bias(μˆlfmp)=E(μˆlfmpμ).

Var(μˆlfmp), Var(μ), bias(μˆlfmp) and Cov(μˆlfmp,μ) can be derived, respectively, as

Var(μˆlfmp)=QVar(βˆlfme)Q+QCov(βˆlfme,y)H1ZGM+MGZH1Cov(y,βˆlfme)Q+σ2MGZH1ZGM, (15)
Var(μ)=Var(Lβ+Mu)=σ2MGM, (16)
bias(μˆlfmp)=E(μˆlfmpμ)=E(Qβˆlfme+MGZH1yLβMu)=Qbias(βˆlfme), (17)

and

Cov(μˆlfmp,μ)=Cov(Qβˆlfme+MGZH1y,Lβ+Mu)=QCov(βˆlfme,u)M+MGZ H1Cov(y,u)M (18)

where

Var(βˆlfme)=σ2Fd(XH1X+RV1R)1Fd,Cov(βˆlfme,y)=σ2Fd(XH1X+RV1R)1X,bias(βˆlfme)=(FdIp)β,Cov(βˆsrle,u)=σ2Fd(XH1X+RV1R)1XH1ZG

and Cov(y,u)=σ2ZG. Then, Equations (14), (15), (16) and (17) are put in Equation (14) to obtain

MSE(μˆlfmp)=QMSE(βˆlfme)Q+σ2M(GGZH1ZG)M (19)

where

MSE(βˆlfme)=Var(βˆlfme)+[bias(βˆlfme)][bias(βˆlfme)]=σ2Fd(XH1X+RV1R)1Fd+(FdIp)ββ(FdIp). (20)

In the same manner, we obtain

MSE(μˆd)=QMSE(βˆd)Q+σ2M(GGZH1ZG)M (21)
MSE(μˆmp)=QMSE(βˆme)Q+σ2M(GGZH1ZG)M (22)

where

MSE(βˆd)=σ2Fd(XH1X)1Fd+(FdIp)ββ(FdIp), (23)
MSE(βˆme)=σ2(XH1X+RV1R)1. (24)

are given, respectively, by Özkale and Kuran [21] and Kuran and Özkale [12]. When we examine Equations (19), (21) and (22), we can say that the superiority of MSE(μˆlfmp) over MSE(μˆd) and MSE(μˆmp) is equivalent to the superiority of MSE(βˆlfme) over MSE(βˆd) and MSE(βˆme), respectively.

3.1. The SRLE vs the Liu estimator

Theorem 3.1

The LFME is always superior to the Liu estimator in the MSE matrix criterion.

Proof.

By using Equations (20) and (23), we can write

MSE(βˆd)MSE(βˆlfme)=σ2Fd((XH1X)1(XH1X+RV1R)1)Fd.

Note that,

((XH1X+RV1R)1)1(XH1X)=RV1R.

Following Theorem A.1 in Appendix 1, we say that since RV1R is nonnegative definite (nnd), ((XH1X+RV1R)1)1(XH1X) is also nnd. Hence (XH1X)1(XH1X+RV1R)1 is nnd and that impiles the difference MSE(βˆd)MSE(βˆlfme) is a positive semidefinite (psd) matrix. Then, the proof is completed.

3.2. The SRLE vs the mixed estimator

Theorem 3.2

The LFME is superior to the mixed estimator for fixed 0<d<1 in the MSE matrix sense if and only if βE1βσ21d  is satisfied where E=XH1X(XH1X+RV1R)1+(XH1X+RV1R)1XH1X+(1+d)(XH1X+RV1R)1.

Proof.

By using Equations (20) and (24), we can write

MSE(βˆme)MSE(βˆlfme)=σ2C(FdIp)ββ(FdIp)

where C=Var(βˆme)Var(βˆlfme)=(XH1X+RV1R)1Fd(XH1X+RV1R)1Fd. By utilizing Fd, we write the difference C as

C=(XH1X+I)1[(XH1X+I)(XH1X+RV1R)1(XH1X+I)(XH1X+dI)(XH1X+RV1R)1(XH1X+dI)](XH1X+I)1=(XH1X+I)1[(1d)XH1X(XH1X+RV1R)1+(1d)(XH1X+RV1R)1XH1X+(1d2)(XH1X+RV1R)1](XH1X+I)1 (25)

Equation (25) shows us that the matrix C is pd for any 0<d<1. Then, from the theorem of Farebrother [4], MSE(βˆme)MSE(βˆlfme) is nnd if and only if β(FdIp)C1(FdIp)βσ2. Since FdIp=(d1)(XH1X+I)1, we obtain the necessary and sufficient condition as

(1d)β[XH1X(XH1X+RV1R)1+(XH1X+RV1R)1XH1X+(1+d)(XH1X+RV1R)1]1βσ2

which completes the proof. Here d can be selected anyway such as suggested by Özkale and Kuran [21].

4. Numerical examples

4.1. A hypothetical data analysis

In this analysis, we use the results of Los Alamos study of high efficiency particulate air (HEPA) filter cartridges presented by Kerschner et al. [11]. Such cartridges are used with commercial respirators to provide respiratory protection against dusts, toxic fumes, mists, radionuclides and other particulate matter.

The primary objective of the study was to determine whether the current standard aerosol used to test these filters could be replaced by any of alternate aerosols for quality-assurance testing of HEPA respirator cartridges. HEPA respirator filters fail a quality-assurance test if the challenge penetration is greater than some percent penetration breakpoint otherwise, it is considered to have passed. A secondary objective was to identify those factors that contribute most to the variability in the penetrations of the filters.

In a subset of the aerosol data set [2], two aerosols were crossed with two filter manufacturers. Within each manufacturer, three filters were used to evaluate the penetration of the two aerosols, so that there were six filters in total. By taking a filter nested within the two manufacturers, the following linear mixed model with independent random effects and independent random errors is given as

yijhl=μ+Ai+Mj+Fjh+εijhl (26)

where yijhl is the percent penetration, μ is an intercept, Ai is a fixed effect for the ith aerosol type (i=1,2), Mj is a fixed effect for the jth filter manufacturer (j=1,2), FjhN(0,γFjh) is a random effect for the hth filter nested within the the jth manufacturer (h=1,2,3) and εijhlN(0,ρεijhl) is the error associated with the lth replication in the ijhlth cell (l=1,2,3), subject to the usual restrictions: iAi=jMj=0.

Clearly, model (26) may be put into the form of model (1) with the vector of responses and approximate design matrices

y36×1=0.7500.7701.5201.580 (27)
X36×3=191919191919191919191919,Z36×6=13030609012070713131313080801501501501513131313131301501501209060313. (28)

Here, 19=(1,1,1,1,1,1,1,1,1), 03=(0,0,0) and 13, 06, 07, 08, 09, 012, 015 are vectors of ones and zeros where subscripts show the dimensions. y in (27) is a vector of the percent penetration values, β=(μ,A1,M1)=(2,1.5,9) is an arbitrary vector and u=(F11,F12,F13,F21,F22,F23) with uN(0,σ2G(γFjh)=σ2G), εN(0,σ2W(ρεijhl)=σ2W).

It was seen that multicollinearity did not appear at the aerosol data examined by Beckman et al. [2] and Kerschner et al. [11]. But, to serve for our purposes, the data must have multicollinearity. Thus, by using this aerosol data and the model in (26), hypothetical data are produced as follows.

Since multicollinearity must be at the X matrix corresponding to fixed effects as explained in theory, the data corresponding to the fixed effects are generated using the following equation in McDonald and Galarneau [15]

xij=1δ21/2xij+δxi4,i=1,,36, j=1,2,3

where xij are taken from (27) and δ is specified so that correlation between any two fixed effects is given by δ2. These effects are then standardized so that the fixed effects design matrix is in correlation form and for this hypothetical data, δ2=0.99 is considered.

The stochastic linear restrictions are taken as

r=Rβ+Φ,R=1.09501.09501.0950,Φ0,σ2Vυ=σ2V

where r = 5.000 is the 14th element of y vector in Equation (28), R is the 14th row of generated X matrix and β=21.59 is an arbitrary vector. Because of having the largest percent penetration value, the 14th element of y vector is chosen. But, the other elements of y vector and corresponding elements of the X matrix can also be taken as r and R, respectively.

Generally, since there is not too much difference between the results of ML and REML covariance estimates (see [19]), we choose REML approach in our hypothetical data analysis and we obtain GˆREML=0.2534×I7, WˆREML=9.9353×I37 and VˆREML=0.6547×I1. Then, from the formula H=ZGZ+W, HˆREML is obtained.

The eigenvalues of the matrix XHˆREML1X are obtained as λ1=11.8812, λ2=0.0084 and λ3=0.0258. Then, the condition number, λmax/λmin, where λmin and λmax indicate the minimum and maximum eigenvalues of XHˆREML1X is calculated as λmax/λmin=1.4154e+03. Since the condition number is used to measure the extent of multicollinearity in the data and it is larger than 1000, it shows severe multicollinearity (see for example Montgomery et al. [16], p.298).

The Liu biasing parameter d is computed as dˆ=1σˆ2/λi(λi+1)αˆi2/(λi+1)2=0.9596 which is defined by Özkale and Kuran [21] where αˆ=Pβˆ and P is an orthogonal matrix such that P(XH1X)P=Δ with a diagonal matrix Δ=diag(λ1,,λp).

Table 1 presents the parameter estimates and scalar MSE values (trace of matrix MSE) when the variance of y is estimated by REML. From Table 1, we observe that LFME outperforms the other estimators which is followed by the mixed estimator.

Table 1. Parameter estimates and MSE scalar values for dˆ=0.9596 when the covariance parameters are estimated by REML.

  βˆ uˆ βˆme uˆmp βˆdˆ uˆdˆ βˆlfme uˆlfmp
μ 10.8672 20.0638 10.4350 19.2684
A1 15.9017 28.3389 15.2609 27.1943
M1 6.9127 11.8321 6.6414 11.3702
Fintercept 0.0000 0.4277 0.0198 0.3912
F11 0.0736 0.0736 0.0731 0.0732
F12 0.0349 0.0349 0.0354 0.0354
F13 0.0387 0.0387 0.0392 0.0391
F21 0.1289 0.2595 0.1222 0.2477
F22 0.0516 0.0790 0.0583 0.0672
F23 0.0023 0.1283 0.0090 0.1165
scalarMSE 16.8331 11.1831 15.5766 10.3698

To consider the superiorities of the estimators for other d values, plot of the scalar MSE values of βˆ, βˆme, βˆd, βˆlfme versus d are presented by Figure 1. Figure 1 demonstrates that although βˆme is better than βˆ, βˆd and βˆlfme are better than βˆ for d values respectively in intervals 0.342<d<1 and 0.362<d<1. And, we can also say that though βˆlfme is better than βˆme for d values in interval 0.524<d<1, βˆme is better than βˆd for d values in interval 0<d<1. For d values in interval 0<d<1, βˆlfme has smaller MSE scalar values than βˆd and it can be seen that as d values increase, the difference between MSE scalar values of βˆlfme and βˆd increases. βˆlfme is the best among the estimators for d values in the interval 0.524<d<1. Therefore, we say that there always exists a d value such that βˆlfme is better than βˆ, βˆme or βˆd.

Figure 1.

Figure 1.

MSE scalar values of BLUE, mixed, Liu and LFME versus d.

The theorems in Section 3 are also examined. And so, under the matrix MSE criterion, βˆlfme is always better than βˆdˆ. The condition given by Theorem 3.2 is computed as βˆE1βˆ=10.8241 which is not smaller than σˆ2/(1dˆ)=2.6387, hence βˆlfme is not better than βˆme for dˆ=0.9596 under the matrix MSE criterion though it is better than βˆme under the scalar MSE criterion. However, there can be found a d value such that βˆlfme dominates βˆme in terms of the matrix MSE criterion; these d values for this data set are found to be greater than 0.9898.

The hypothetical data analysis summarizes that the SRLE dominates the BLUE, mixed and Liu estimators in the sense of the scalar MSE criterion for appropriate selected d value under multicollinearity.

4.2. Real data analysis: greenhouse gases data

Greenhouse gases trap heat and make the planet warmer. Human activities are responsible for almost all of the increase in greenhouse gases in the atmosphere over the past 150 years. The largest source of greenhouse gas emissions from human activities in the United States is from burning fossil fuels for electricity, heat, and transportation (see the official United States Environmental Protection Agency (EPA) website [3]). The transportation sector generates the largest share of greenhouse gas emissions in the United States (nearly 28.5 % of 2016 greenhouse gas emissions). Greenhouse gas emissions from transportation come from burning fossil fuel for cars, light/heavy trucks-buses, motorcycles and railways. The remaining greenhouse gas emissions from the transportation sector come from other modes of transportation, including commercial aircraft, ships and boats (see EPA website [3]).

In this analysis, we use data on 297 fuel combustion in transport (in million tonnes) randomly selected from 27 areas (Belgium, Bulgaria, European Union, Denmark, Germany, Estonia, Greece, Spain, France, Croatia, Italy, Ireland, Latvia, Lithuania, Luxembourg, Hungary, Netherlands, Austria, Poland, Portugal, Romania, Slovenia, Slovakia, Switzerland, Sweden, United Kingdom, Norway) which were regularly obtained during the period 2006–2016 inclusive.

To determine fuel combustion in transport (in million tonnes), we get repeated measurements from the list of fuel combustions, namely the cars, the light duty trucks, the heavy duty trucks-buses, the motorcycles and railways. The data are available from the official Eurostat website [3].

We define the response (y) as fuel combustion in transport (in million tonnes) and we get the explanatory variables as fuel combustions in cars (x1), light duty trucks (x2), heavy duty trucks-buses (x3), motorcycles (x4) and railways (x5). That is, we can say that fuel combustions in cars, light duty trucks, heavy duty trucks-buses, motorcycles, railways are expressed as fixed effects and since the 27 areas are randomly selected from the areas, the areas factor effect on the response is expressed as random effect. Then, the random intercept and slope model (RISM) is given as

yij=β1xij1+β2xij2+β3xij3+β4xij4+β5xij5+u1+u2tij+εij,i=1,,27,j=1,,11

where yij denotes the ith observation of the jth area of the response, xijs denotes the ith observation of the jth area of the explanatory variable xs, s=1,,5 and tij shows time corresponding to yij.

The stochastic linear restriction is taken as

r=Rβ+Φ,R=558.69609107.73599242.2381810.861148.14196Φ0,σ2Vυ=σ2V

where r = 983.37034 is the y31th observation, R is the vector of x31sth observations for s=1,,5 and β=11111 is an arbitrary vector. Because of having the largest cereal production value, the y31th observation is chosen. But, the other observations of yij vector and corresponding elements of the xijs observations can also be taken as r and R, respectively.

Generally, since there is not too much difference between the results of ML and REML covariance estimates (see [19]), we prefer ML approach in our real data analysis and we get GˆML=0.022027×I2, WˆML=4.146253×I298 and VˆML=1.155446×I1. Then, from the formula H=ZGZ+W, HˆML is obtained.

The eigenvalues of the matrix XHˆML1X are obtained as   λ1= 9.040303e+05, λ2= 9.260924e + 02, λ3= 2.903158e+02, λ4= 2.571984 and λ5= 15.225476. Then, the condition number is calculated as λmax/λmin=351490>1000 and it shows severe multicollinearity (see for example Montgomery et al. [16], p.298).

The Liu biasing parameter is computed from Özkale and Kuran [21] as dˆ=0.9812. Table 2 gives the parameter estimates and scalar MSE values when the variance of y is estimated by ML. From Table 2, we observe that LFME outperforms the mixed estimator, the Liu estimator and the BLUE in the sense of scalar MSE.

Table 2. Parameter estimates and MSE scalar values for dˆ=0.9812 when the covariance parameters are estimated by ML.

  βˆ uˆ βˆme uˆmp βˆdˆ uˆdˆ βˆlfme uˆlfmp
β1 1.0204 1.0046 1.0212 1.0055
β2 1.0507 1.0406 1.0511 1.0410
β3 0.9380 0.9637 0.9374 0.9630
β4 3.3953 3.5860 3.3794 3.5688
β5 3.7790 4.1731 3.7579 4.1502
u1 0.1329 0.1014 0.1332 0.1017
u2 0.0252 0.0323 0.0252 0.0324
scalarMSE 0.1511 0.1472 0.1497 0.1459

To consider the superiorities of the estimators for other d values, plot of the scalar MSE values of βˆ, βˆme, βˆd, βˆlfme versus d are presented by Figure 2. Figure 2 demonstrates that although βˆme is better than βˆ for d values in interval 0<d<1, βˆd and βˆlfme are better than βˆ for d values respectively in intervals 0.508<d<1 and 0.482<d<1. And, we can also say that though βˆlfme is better than βˆme for d values in interval 0.535<d<1, βˆme is better than βˆd for d values in interval 0.563<d<1. For d values in interval 0<d<1, βˆlfme has smaller scalar MSE values than βˆd and it can be seen that as d values increase, the difference between scalar MSE values of βˆlfme and βˆd increases. As compared to Figure 1, we see that the tendencies of the estimators towards each other are same.

Figure 2.

Figure 2.

MSE scalar values of BLUE, mixed, Liu and LFME versus d.

In examining Theorem 3.1, we see that βˆlfme is always better than βˆdˆ in the sense of matrix MSE. The condition given by Theorem 3.2 is computed as βˆE1βˆ=1.0240 which is smaller than σˆ2/(1dˆ)=17.5236, hence βˆlfme is better than βˆme under the matrix MSE matrix criterion.

Under multicollinearity, this real data analysis summarizes that the LFME dominates the BLUE, mixed and Liu estimators in the sense of both the scalar MSE and matrix MSE criteria for appropriate selected d value.

5. A simulation study

In this section, we will discuss the simulation study to compare the performance of βˆ, βˆme, βˆd and βˆlfme and the performance of uˆ, uˆme, uˆd and uˆlfmp in the sense of the estimated mean square error (EMSE) and the predicted mean square error (PMSE).

The fixed effects are computed from McDonald and Galarneau [15] as

xijk=(1γ2)1/2wijk+γwijp+1,i=1,,m,j=1,,ni,k=1,,p

where wijk are independent standard normal pseudo-random numbers and γ is specified so that the correlation between any two fixed effects is given by γ2. Three different sets of correlations are considered corresponding to γ2=0.90, 0.95 and 0.99. The number of fixed effects are chosen as p = 2 and p = 4. To fix the generated explanatory variables, 2019 is used as the state number.

We consider m = 30, 60 subjects and ni=5, 10 observations per subject. The stochastic linear restrictions are taken as

r=Rβ+Φ,Φ0,σ2Vυ=σ2V

where r=0.40.3 and R=10.50.51 are chosen arbitrary. The parameter vector β=(β1,,βp) is chosen as the normalized eigenvector corresponding to the largest eigenvalue of XH1X so that ββ=1 (see [17]). Then, the underlying model takes the following form with q = 2 random effects

yij=β1xij1+β2xij2+u1+u2tij+εij,uiiidN(0,σ2D),εijiidN(0,σ2Ini)

where D = V are both taken as 1ρρ1 is the AR(1) process with ρ=0.90 and tij shows time which was taken as the same set of occasions, {tij=j for i=1,,m,j=1,,ni}. Three values of σ2 are investigated as 0.5, 1 and 10.

By using Özkale and Kuran [21], the Liu biasing parameter is selected as dˆh which is given by Theorem 4.1 where h is determined as multiplying the upper bound defined in Theorem 4.1 by 0.99 if i=1p1/λi(λi+1)>2/(λp(λp+1)) and determined arbitrarily as dˆh=0.8650 if i=1p1/λi(λi+1)<2/(λp(λp+1)). This arbitrarily chosen dˆh=0.8650 is determined by inspiring the numerical examples presented in Section 4.

For each choice of m, ni, γ, p and σ2, the experiment is replicated 500 times by generating response variable and the EMSE for any estimator β~ of β and the PMSE for any predictor u~ of u are calculated, respectively, as

EMSE(β~)=1500r=1500(β~(r)β)(β~(r)β)PMSE(u~)=1500r=1500(u~(r)u)(u~(r)u)

where the subscript (ŗ) refers to the ŗth replication.

To see the degree of supremacy, after the EMSE and the PMSE results are obtained, the change percentage values are calculated as

1EMSE(β~1)EMSE(β~2).1001PMSE(u~1)PMSE(u~2).100

where β~1 and β~2 are any two estimators and u~1 and u~2 are any two predictors.

The simulation results are summarized by Tables 310. Based on Tables 310, we have the following conclusions:

Table 4. Predicted MSE values with p = 2, m = 30 and ni=5.

γ2 σ2 uˆ uˆme uˆdˆh uˆlfmp uˆ/uˆlfmp uˆme/uˆlfmp uˆdˆh/uˆlfmp
0.90 0.5110 0.0129649890.0218123300.269327347 0.0129640290.0218051620.269362811 0.0129654410.0218142160.269370583 0.0129645050.0218072680.269428009 0.00370.02320.0373 0.00360.00960.0242 0.00720.03180.0213
0.95 0.5110 0.0129605870.0218070390.269340232 0.0129584170.0217919760.269194164 0.0129610780.0218088280.269363964 0.0129589690.0217942620.269246984 0.01240.05850.0346 0.00420.01040.0196 0.01620.06670.0434
0.99 0.5110 0.0129581490.0218047640.269366971 0.0129653610.0218099790.269113379 0.0129595670.0218090420.269366531 0.0129668720.0218150920.269136232 0.067320.04730.0856 0.01160.02340.0084 0.05630.02770.0854

Table 5. Estimated MSE values with p = 2, m = 60 and ni=10.

γ2 σ2 βˆ βˆme βˆdˆh βˆlfme βˆ/βˆlfme βˆme/βˆlfme βˆdˆh/βˆlfme
0.90 0.5110 0.0093852960.0187691450.187545841 0.0090309510.0173801460.097858940 0.0093646040.0186870260.180471159 0.0090111610.0173047250.094349996 3.98637.802249.6923 0.21910.43393.5857 3.77427.397147.7201
0.95 0.5110 0.0183341970.0366658300.366395276 0.0169445130.0313918030.116000194 0.0182520750.0363431100.341960717 0.0168688420.0311169050.108666493 7.992415.133770.3417 0.44650.87566.3221 7.578414.380168.2225
0.99 0.5110 0.0900147790.1800180181.798970566 0.0618670670.0904642680.085052218 0.0880996680.1729553461.505693191 0.0605536150.0869313000.072424028 32.729251.709695.9741 2.12303.905314.8475 31.266949.737795.1899

Table 6. Predicted MSE values with p = 2, m = 60 and ni=10.

γ2 σ2 uˆ uˆme uˆdˆh uˆlfmp uˆ/uˆlfmp uˆme/uˆlfmp uˆdˆh/uˆlfmp
0.90 0.5110 0.0027615270.0052653380.042150581 0.0027609350.0052615100.041556416 0.0027614850.0052650720.042107279 0.0027608980.0052612630.041516515 0.02270.07731.5042 0.00130.00460.0960 0.02120.07231.4029
0.95 0.5110 0.0027616020.0052654180.042152520 0.0027608610.0052609390.041539116 0.0027615530.0052651260.042108122 0.0027608180.0052606740.041500225 0.02830.09001.5474 0.00150.00500.0936 0.02660.08451.4436
0.99 0.5110 0.0027614170.0052650250.042153788 0.0027598480.0052578300.042309344 0.0027613220.0052645730.042194581 0.0027597760.0052575120.042359572 0.05940.14260.4881 0.00260.00600.1187 0.05590.13410.3910

Table 7. Estimated MSE values with p = 4, m = 30 and ni=5.

γ2 σ2 βˆ βˆme βˆdˆh βˆlfme βˆ/βˆlfme βˆme/βˆlfme βˆdˆh/βˆlfme
0.90 0.5110 0.1032745760.2062752202.053570779 0.0895413750.1557102240.481597125 0.0965476140.1812075411.072799065 0.0837349280.1368892690.258139478 18.920033.637587.4297 6.484612.087146.3992 13.270824.457175.9377
0.95 0.5110 0.2056209210.4106984494.088803667 0.1577048360.2572663980.564946843 0.1805701680.3244605551.852567988 0.1385753740.2036066230.250745900 32.606350.424393.8674 12.129920.857655.6160 23.256737.247686.4649
0.99 0.5110 1.0246486302.04660314120.376157839 0.4156903580.5142955060.420561330 0.6680054791.1520879477.927083175 0.2700214400.2835005060.141381774 73.647486.147799.3061 35.042644.875966.3826 59.577975.392498.2164

Table 8. Predicted MSE values with p = 4, m = 30 and ni=5.

γ2 σ2 uˆ uˆme uˆdˆh uˆlfmp uˆ/uˆlfmp uˆme/uˆlfmp uˆdˆh/uˆlfmp
0.90 0.5110 0.1639311210.3323412453.883180462 0.1633801310.3327870433.890053063 0.1638025060.3325971483.886868865 0.1632616490.3330385103.893748122 0.40830.20980.2721 0.07250.07550.0949 0.33010.13270.1769
0.95 0.5110 0.1639374460.3323515793.883202839 0.1632808240.3306468113.862361232 0.1637958560.3319564343.876337677 0.1631654370.3303733173.858445026 0.47090.59520.6375 0.07060.08270.1013 0.38480.47690.4615
0.99 0.5110 0.1639440950.3323620723.883208286 0.1630862590.3305393233.863445116 0.1637703230.3319581613.876519827 0.1630568550.3305082203.859643872 0.54110.55770.6068 0.01800.00940.0983 0.43560.43670.4353

Table 9. Estimated MSE values with p = 4, m = 60 and ni=10.

γ2 σ2 βˆ βˆme βˆdˆh βˆlfme βˆ/βˆlfme βˆme/βˆlfme βˆdˆh/βˆlfme
0.90 0.5110 0.0256302280.0512587210.512390711 0.0246796030.0475900940.290100771 0.0252112530.0496034330.382396641 0.0242765680.0460555910.216901368 5.281410.150757.6687 1.63303.224425.2324 3.70747.152443.2784
0.95 0.5110 0.0509676110.1019317631.018926171 0.0476261290.0893054990.426523475 0.0493117390.0954707330.632346202 0.0460807520.0836563490.263886154 9.588117.929074.1015 3.24486.325638.1309 6.552112.374858.2687
0.99 0.5110 0.2537202490.5074232115.072276675 0.1874250510.2982922500.548580815 0.2162308560.3779393172.190539699 0.1597998050.2221533800.213480744 37.017356.219395.7912 14.739325.524961.0849 26.097541.219890.2544

Table 3. Estimated MSE values with p = 2, m = 30 and ni=5.

γ2 σ2 βˆ βˆme βˆdˆh βˆlfme βˆ/βˆlfme βˆme/βˆlfme βˆdˆh/βˆlfme
0.90 0.5110 0.0341722600.0683553040.684626643 0.0295468130.0516356080.130726927 0.0338898370.0672634240.616032929 0.0293054540.0508272020.120487228 14.241925.642682.4010 0.81681.56557.8328 13.527324.435680.4414
0.95 0.5110 0.0667746590.1335712701.337913932 0.0501092970.0781978340.111981552 0.0656855690.1294847331.144462731 0.0492992960.0758484040.100152233 26.170643.215092.5142 1.61643.004410.5636 24.946541.422891.2489
0.99 0.5110 0.3277407890.6555893286.567047387 0.1040035900.1049196600.069167955 0.3062164380.5873363095.123817483 0.0972436590.0942719330.060878465 70.329085.620299.0729 6.499710.148411.9845 68.243483.949298.8118

Table 10. Predicted MSE values with p = 4, m = 60 and ni=10.

γ2 σ2 uˆ uˆme uˆdˆh uˆlfmp uˆ/uˆlfmp uˆme/uˆlfmp uˆdˆh/uˆlfmp
0.90 0.5110 0.0763013010.1526916521.542243891 0.0762046940.1524274231.537362490 0.0762974660.1526806441.541908579 0.0762016190.1524205391.537675547 0.13060.17750.2962 0.00400.00450.0203 0.12560.17030.2745
0.95 0.5110 0.0763019760.1526929901.542255152 0.0761390630.1522579151.536095808 0.0763048040.1527003691.542372187 0.0761445390.1522793181.537763255 0.20630.27090.2912 0.00710.01400.1085 0.21000.27570.2988
0.99 0.5110 0.0763027900.1526946031.542268935 0.0759875870.1519617781.537643517 0.0763034430.1526948061.542476207 0.0760121020.1520649551.540231580 0.38090.41230.1321 0.03220.06780.1683 0.38180.41240.1455
  1. For fixed value of sample size, correlation and p, the EMSE values of βˆ, βˆme, βˆd and βˆlfme increase with the increase of σ2.

  2. When sample size, σ2 and p are fixed, the superiorities of βˆlfme over βˆ, βˆme and βˆd increase as γ2 increases. And, the superiority βˆlfme over βˆ and βˆd can be seen more clear than the superiority βˆlfme over βˆme.

  3. The EMSE(βˆ), EMSE(βˆme), EMSE(βˆd) and EMSE(βˆlfme) values increase with the increase of p values under fixed value of sample size, correlation and σ2.

  4. As n=i=1mni increases from 150 to 600 at fixed γ2, σ2 and p, the EMSE values of βˆ, βˆme, βˆd and βˆlfme decreases.

  5. As the degree of correlation increases under fixed values of sample size, σ2 and p, the EMSE(βˆ), EMSE(βˆme), EMSE(βˆd) and EMSE(βˆlfme) values increase.

  6. The EMSE(βˆlfme) values are smaller than the EMSE(βˆ), EMSE(βˆme) and EMSE(βˆlfme) values in all cases. This case is seen from the change percentages which are all positive. The change percentage indicates that as σ2 and γ2 increases at fixed sample size and p, the degrees of supremacies of βˆlfme over βˆ, βˆme and βˆd increase. Furthermore, we can say that the degree of supremacies of βˆlfme over βˆ and βˆd is too much than the degree of supremacy of βˆlfme over βˆme.

  7. The PMSE(uˆ), PMSE(uˆme), PMSE(uˆd) and PMSE(uˆlfme) values increase with the increase of σ2 for fixed value of sample size, correlation and p. And, this increase can be also seen more clearly as p values increase.

  8. With the increase of p values under fixed value of sample size, correlation and σ2, the PMSE values of uˆ, uˆme, uˆd and uˆlfme increase.

  9. As n=i=1mni increases from 150 to 600 at fixed γ2, σ2 and p, the PMSE(uˆ), PMSE(uˆme), PMSE(uˆd) and PMSE(uˆlfme) values decrease.

  10. The PMSE values of uˆ, uˆme, uˆd and uˆlfme are very similar to each other in all cases. Although the PMSE(uˆme) values are smaller than the PMSE(uˆ), PMSE(uˆd) and PMSE(uˆlfme) values, we see that the PMSE(uˆlfme) values are smaller than the PMSE(uˆ), PMSE(uˆme) and PMSE(uˆd) values as p values increase. But, consequently, we say that the difference between the PMSE values of uˆ, uˆme, uˆd and uˆlfme is not too much.

  11. The change percentage indicates that as σ2 and γ2 increases at fixed sample size and p, the degrees of supremacies of uˆlfme over uˆ, uˆme and uˆd increase. Moreover, we can say that the degree of supremacies of uˆlfme over uˆ and uˆd is too much than the degree of supremacy of uˆlfme over uˆme.

  12. In Tables 3–10, βˆlfme employs better approximation (that is, βˆlfme values are closer) to the true values than βˆ, βˆme and βˆd.

6. Discussion

In this article, we propose the LFME and LFMP in LMMs with the help of Kuran and Özkale's [12] mixed estimation method and Özkale and Kuran's [21] Liu approach. Then, matrix MSE comparisons are done and lastly, the study is supported with both numerical examples and a simulation study.

This study has confirmed that, under multicollinearity, the use of the LFME outperforms the Liu estimator for any selected Liu biasing parameter and the mixed estimator for appropriate selected Liu biasing parameter in terms of mean square error. If there exists prior information as in the form of stochastic linear restrictions, one can use the LFME and LFMP in an LMM. The advantage of the LFME and LFMP over the Liu estimator and predictor is that LFME and LFMP consider the prior information and over the mixed estimator and predictor is that they gain efficiency.

Appendix 1. Matrix algebra.

Theorem A.1 Rao and Toutenburg [24], p. 302 —

Let A>0 and B>0. Then, BA>0 if and only if A1B1>0.

Disclosure statement

No potential conflict of interest was reported by the authors.

References

  • 1.Baksalary J.K. and Trenkler G., Nonnegative and positive definiteness of matrices modified by two matrices of rank one, Linear Algebra Appl. 151 (1991), pp. 169–184. doi: 10.1016/0024-3795(91)90362-Z [DOI] [Google Scholar]
  • 2.Beckman R.J., Nachtsheim C.J., and Cook R.D., Diagnostics for mixed model analysis of variance, Technometrics 29 (1987), pp. 413–426. [Google Scholar]
  • 3.Eurostat website . Available at http://appsso.eurostat.ec.europa.eu/nui/show.do?dataset, 2018.
  • 4.Farebrother R.W., Further results on the mean square error of ridge regression, J. R. Stat. Soc. B 38 (1976), pp. 248–250. [Google Scholar]
  • 5.Gilmour A.R., Thompson R., and Cullis B.R., Average information REML: an efficient algorithm for variance parameter estimation in linear mixed models, Biometrics 51 (1995), pp. 1440–1450. doi: 10.2307/2533274 [DOI] [Google Scholar]
  • 6.Henderson C.R., Estimation of genetic parameters (abstract), Ann. Math. Stat. 21 (1950), pp. 309–310. [Google Scholar]
  • 7.Henderson C.R., Kempthorne O., Searle S.R., and Von Krosig C.N., Estimation of environmental and genetic trends from records subject to culling, Biometrics 15 (1959), pp. 192–218. doi: 10.2307/2527669 [DOI] [Google Scholar]
  • 8.Hubert M.H. and Wijekoon P., Improvement of the Liu estimator in linear regression model, Stat. Papers 47 (2006), pp. 471–479. doi: 10.1007/s00362-006-0300-4 [DOI] [Google Scholar]
  • 9.Inan D. and Erdoğan B.E., Liu-Type logistic estimator, Commun. Stat. Simul. Comput. 42 (2013), pp. 1578–1586. doi: 10.1080/03610918.2012.667480 [DOI] [Google Scholar]
  • 10.Kaçıranlar S., Sakallıoğlu S., Akdeniz F., Styan G.P.H., and Werner H.J., A new biased estimator in linear regression and a detailed analysis of the widely analyzed dataset on Portland Cement, Sankhya: The lndian J. Stat. 61B (1999), pp. 443–459. [Google Scholar]
  • 11.Kerschner H.F., Ettinger H.J., DeField J.D., and Beckman R.J., A Comparative Study of HEPA Filter Efficiencies When Challenged with Thermal and Air-Jet Generated Di-2-Ethylhexyl Sebecate, Di-2-Ethylhexyl Phthalate and Sodium Chloride, Laboratory Report LA-9985-MS, Los Alamos National Laboratory, Los Alamos, 1984
  • 12.Kuran Ö. and Özkale M.R., Gilmour's approach to mixed and stochastic restricted ridge predictions in linear mixed models, Linear Algebra Appl. 508 (2016), pp. 22–47. doi: 10.1016/j.laa.2016.06.040 [DOI] [Google Scholar]
  • 13.Liu K., A new class of biased estimate in linear regression, Commun. Stat. Theor. Meth. 22 (1993), pp. 393–402. doi: 10.1080/03610929308831027 [DOI] [Google Scholar]
  • 14.Mansson G., Kibria B.M.G., and Shukur G., On Liu Estimators for the Logit Regression Model, The Royal Institute of Technology, Centre of Excellence for Science and Innovation Studies (CESIS), Sweden, Paper No. 259, 2012
  • 15.McDonald M.C. and Galarneau D.I., A Monte Carlo evaluation of some ridge-type estimators, J. Am. Stat. Assoc. 70 (1975), pp. 407–416. doi: 10.1080/01621459.1975.10479882 [DOI] [Google Scholar]
  • 16.Montgomery D.C., Peck E.A., and Vining G.G., Introduction to Linear Regression Analysis, John Wiley and Sons, Inc., Hoboken, NJ, 2012. [Google Scholar]
  • 17.Newhouse J.P. and Oman S.D., An evaluation of ridge estimators, Rand. Corporation R-716-PR, 1971, pp. 1–28.
  • 18.Özkale M.R., A stochastic restricted ridge regression estimator, J. Multivariate Anal. 100 (2009), pp. 1706–1716. doi: 10.1016/j.jmva.2009.02.005 [DOI] [Google Scholar]
  • 19.Özkale M.R. and Can F., An evaluation of ridge estimator in linear mixed models: an example from kidney failure data, J. Appl. Stat. 44 (2017), pp. 2251–2269. doi: 10.1080/02664763.2016.1252732 [DOI] [Google Scholar]
  • 20.Özkale M.R. and Kaçıranlar S., The restricted and unrestricted two-parameter estimators, Commun. Stat. Theor. Meth. 36 (2007), pp. 2707–2725. doi: 10.1080/03610920701386877 [DOI] [Google Scholar]
  • 21.Özkale M.R. and Kuran Ö., A further prediction method in linear mixed models: Liu prediction, Commun. Stat. Simul. Comput. (2018), doi: 10.1080/03610918.2018.1535071. [DOI] [Google Scholar]
  • 22.Pereira L.N. and Coelho P.S., A small area predictor under area-level linear mixed models with restrictions, Commun. Stat. Theor. Meth. 41 (2012), pp. 2524–2544. doi: 10.1080/03610926.2011.648000 [DOI] [Google Scholar]
  • 23.Pinherio J.C. and Bates D.M., Mixed-Effects in S and S-Plus, Springer-Verlag, New York, 2000. [Google Scholar]
  • 24.Rao C.R. and Toutenburg H., Linear Models-Least Squares and Alternatives, Springer-Verlag, New York, 1995. [Google Scholar]
  • 25.Robinson G.K., That BLUP is a good thing: the estimation of random effects (with discussion), Stat. Sci. 6 (1991), pp. 15–32. doi: 10.1214/ss/1177011926 [DOI] [Google Scholar]
  • 26.Searle S.R., Matrix Algebra Useful for Statistics, John Wiley and Sons, New York, 1982. [Google Scholar]
  • 27.Shiller R., A distributed lag estimator derived from smoothness priors, Econometrica 41 (1973), pp. 775–778. doi: 10.2307/1914096 [DOI] [Google Scholar]
  • 28.Štulajter F., Predictions in nonlinear regression models, Acta Math. Univ. Comenianae LXVI (1997), pp. 71–81. [Google Scholar]
  • 29.Swindel B.F., Good estimators based on prior information, Commun. Stat. Theor. Meth. 5 (1976), pp. 1065–1075. doi: 10.1080/03610927608827423 [DOI] [Google Scholar]
  • 30.Şiray G.Ü., Toker S., and Kaçiranlar S., On the restricted Liu estimator in logistic regression model, Commun. Stat. Simul. Comput. 44 (2015), pp. 217–232. doi: 10.1080/03610918.2013.771742 [DOI] [Google Scholar]
  • 31.Theil H. and Goldberger A.S., On pure and mixed statistical estimation in economics, Int. Econ. Rev. 2 (1961), pp. 65–78. doi: 10.2307/2525589 [DOI] [Google Scholar]
  • 32.Varathan N. and Wijekoon P., Logistic Liu Estimator under stochastic linear restrictions, Stat. Papers 60 (2019), pp. 945–962. doi: 10.1007/s00362-016-0856-6 [DOI] [Google Scholar]
  • 33.Xinfeng C., On the almost unbiased ridge and Liu estimator in the logistic regression model, International Conference on Social Science, Education Management and Sports Education, Atlantis Press, 2015, pp. 1663–1665
  • 34.Yang H., Ye H., and Xue K., A further study of predictions in linear mixed models, Commun. Stat. Theor. Meth. 43 (2014), pp. 4241–4252. doi: 10.1080/03610926.2012.725497 [DOI] [Google Scholar]

Articles from Journal of Applied Statistics are provided here courtesy of Taylor & Francis

RESOURCES