Abstract
Time dynamic varying coefficient models play an important role in applications of biology, medicine, environment, finance, etc. Traditional methods, such as kernel smoothing and spline smoothing, are popular. But explicit expressions are unavailable using these methods, and the convergence rate of coefficient function estimators is slow. To address these problems, we expand the varying component with appropriate basis functions. And then we solve a sparse regression problem via a sequential thresholded least-squares estimator. The “parameterization” leads to explicit expressions and fast computation speed. Convergence of the sequential thresholded least squares algorithm is guaranteed. The asymptotic distribution of the coefficient function estimator is derived under certain assumptions. Our simulation shows the proposed method has higher precision and computing speed. Finally, our proposed method is applied to the study of PM concentration in Beijing. We analyze the relationship between PM and other impact factors.
Keywords: Sparse regression, varying coefficient models, basis expansion
1. Introduction
There has been much research on varying coefficient models since they were proposed by [9]. Varying coefficient model is an extension of the classical linear model. Their objective is to examine how regression coefficients change over time or other covariates. Park et al. [16] gave an overview of methodological and theoretical developments of earlier studies. To construct a varying coefficient model, one strategy is to let all regression coefficients depend on a mutual covariate. Zhang and Fan [25] proposed a two-step local polynomial least squares estimation procedure. Chiang et al. [5] proposed a componentwise smoothing spline method. The other is to let different regression coefficients depend on different covariates. Marginal integration technique [13] and smooth backfitting technique [14] are used frequently.
The problem, how a covariate affects response Y over time in the dynamical longitudinal analysis, is of interest. Varying coefficient models are useful and play an important role to solve the problem. The simplest functional regression model is
| (1) |
where are unknown coefficient functions and ε is a mean zero stochastic process. Here the coefficient functions depend only on time t. Lee et al. [11] studied a varying coefficient model with the response variable dependent on the covariates as follows
| (2) |
where , , , and Zhang et al. [26] proposed the following model
| (3) |
and a new backfitting algorithm was developed to estimate the unknown regression functions with the convergence rate . Lee et al. [10] studied a new varying coefficient model
| (4) |
where ε is a stochastic process such that . Different from model (1), the coefficient functions change with time and a covariate. Kernel smoothing technique was developed to estimate this model. And the convergence rate of the coefficient function estimators was also . Although similar to model (3), model (4) takes nonlinear interaction effects into account.
The estimation procedure of kernel smoothing depends on the selection of bandwidths. Spline smoothing methods depend on knots and smoothing parameters. And the backfitting algorithms are computationally cumbersome. Furthermore, varying coefficient models are semiparametric and have bad interpretability. We are likely to estimate such models by deep learning methods [1] due to their ability to mine nonlinear structures hidden in the data. However, we will not develop this point here for lack of interpretability.
Brunton et al. [3] combined sparsity-promoting and machine learning techniques to discover governing equations from noisy measurement data. Their proposed data-driven modeling framework, called sparse identification of nonlinear dynamics (short for SINDy), could identify nonlinear low-order models. Motivated by this, we propose our data-driven estimator for time-dynamic varying coefficient models. We expand functional regression coefficients globally by constructing a library of candidate basis functions. This is a basis expansion method. The basis functions are chosen more flexibly based on domain knowledge. Then, the time-dynamic varying coefficient model becomes a sparse linear regression model. One advantage is that an explicit functional expression may be found with strong interpretability. Even though the constructed functions are not in the correct functional space in some cases, the proposed procedure can still obtain an approximate estimator.
We can use regularization techniques, such as LASSO [21], to fit a sparse linear regression model. There is ample literature on high dimensional linear regression, see [17] for an overview. Besides, [3] proposed a sequentially least squares algorithm to compute the sparse regression problem. Zhang and Schaeffer [22] provided some theoretical results on the behavior and convergence of SINDy. In further study, [18] found that LASSO performed poorly with highly correlated data. And they approximated the problem to a ridge regression problem with hard thresholding, called sequential threshold ridge regression (STRidge). We use sequential thresholded least squares estimator to estimate the varying functional regression coefficients in this research.
This paper is organized as follows. Section 2 introduces time dynamic varying coefficient models and our proposed estimation procedure. consistency and some other theoretical properties are shown in Section 3. Simulation studies are conducted to verify the efficiency of this approach in Section 4. We analyze the air quality data set using our method in Section 5. Finally, we draw some conclusions and directions for future research in Section 6.
2. Model
In this section, we take the model (4) for an example to introduce our estimation procedure, then extend it to model (1) and (3). The observed response and covariates in the longitudinal setting satisfy the model
| (5) |
where T denotes the random variable that represents the independent time points allowed to be different for different subjects, the processes , , and Y are observed, the coefficients are bivariate functions and vary with time T. Some terms, such as polynomials, often appear in in practice. If we can construct possible basis functions, the bivariate function will expand in a linear combination of the basis functions.
2.1. Sparse identification of nonlinear dynamics
Sparse identification of nonlinear dynamics (SINDy) is developed for discovering ordinary differential equations and partial differential equations in engineering. Let's have a brief review of sparse identification of nonlinear dynamics. Consider the dynamical system of some time series data as follows
SINDy combines sparsity regression and nonlinear function libraries to identify a nonlinear and dynamical system from time-series data. The main idea is to convert the nonlinear model identification problem to a linear system
| (6) |
where the column full rank matrix is a data-driven dictionary whose columns are candidate functions of the given data u, the unknown vector represents the coefficients of the selected terms in the governing equation f, and the vector is an approximation to the first-order time derivative . A sparse vector ω which approximately solves (6) is generated by the following iterative scheme:
| (7) |
| (8) |
where is a thresholding parameter, and is the support set of ω. The calculating process is illustrated in Algorithm 1 [3].
2.2. Data-Driven estimator for dynamic varying coefficient models
We rewrite the model (5) as follows
| (9) |
where , . Let y be the matrix
We vectorize the response y to be , let , a library of candidate nonlinear functions can be constructed as follows
| (10) |
where are univariate functions or bivariate functions from to . These functions can be chosen with prior knowledge. Here, prior information can be viewed as domain knowledge. We can cooperate with domain experts or search the literature to know prior knowledge. For example, the right-hand side of ordinary differential equations often appears polynomials, then we can choose polynomials as candidate functions. If we want to discover the dynamic transmission model of COVID-19 [12], SIR (susceptible, infective, and recovered), SEIR (susceptible, exposed, infective, and recovered), SEIJR (susceptible, exposed, infective, diagnosed, and recovered), and SIRD (susceptible, infective, recovered and dead) models, which are designed for infectious diseases, can be regarded as our prior knowledge. We select polynomials in this case. Even though the constructing functions in some cases are not in the correct functional space, the proposed procedure can still gain an approximate estimator.
Remark 2.1
Any continuous function can be well approximated by either polynomial base or trigonometric base. It is also available to include both in the library. We aim to get a global functional expression of the model of interest. Basis functions, such as B-spline functions, cannot be used in our research. Because B-spline methods recover data interval by interval, leading to a local solution.
We take Example 4.1 in Section 4 [10] for an example, three varying coefficients are specified as follows
| (11) |
An equivalent formulation of (11) is
it is interesting to find each is polynomial. Then
If each varying coefficient is expanded in three-order polynomial functions, we get
Thus the original varying coefficient model now becomes a 30-dimensional linear sparse regression model. The true coefficients are illustrated in Table 1.
Table 1.
Basis functions and corresponding coefficients in Example 4.1.
| 1 | 0 | 3 | 2.275 |
| t | 0.25 | 1 | 3.5 |
| z | 0 | 7 | 3.5 |
| tt | 0.5 | 0 | 3.5 |
| tz | 3 | 2 | 0 |
| zz | 0 | 2 | 0 |
| ttt | 0 | 0 | 0 |
| ttz | 0 | 0 | 0 |
| tzz | 1 | 0 | 0 |
| zzz | 0 | 0 | 0 |
Remark 2.2
We assume the constructed library of candidate functions contains true function . It will be feasible as long as the set of basis functions can have more terms than . We tend to pick a moderately large set. The computation cost will be expensive if the set is too large. Basis functions should be three-order or greater than three-order in (11). The higher the order is, the harder it is for computation.
We assume true functions are a subset of the functions in (10), then model (5) is
where , and ° denotes a new product
| (12) |
Let , , , and , we rewrite (9) as
| (13) |
where y is a vector of observed outcomes, Θ is an matrix of predictors, is the unknown parameter of interest, and ε is a vector of additive noise with mean 0 and variance .
As estimates for , we use sequential thresholded least squares estimator
| (14) |
where λ is a tuning parameter, and the calculating process is illustrated in Algorithm 2. Refer to [3] for more details.
For the tuning parameter λ, we can use k-fold cross-validation or AIC criteria [2] to determine it. Sequential thresholded least squares estimator in (14) is equivalent to the penalized least squares as follows
where is much smaller when λ is a small value, which has little effect on the results. We fix λ to a small value, such as 0.05, in our simulation. The results do not change much when λ is a small value.
Remark 2.3
We use sequential thresholded least squares estimator for the reason it is equivalent to penalized regression. penalized least squares estimator penalizes the number of variables in a regression model. It has an advantage over other penalties for sparse regressions in nature. More sparse solutions can be obtained using penalty. There is an enormous amount of literature studying sparse regression in linear models. STRidge in [18] is one of them developed for sparse linear regression models. The calculation procedure is similar between sequential thresholded least squares and STRidge. But STRidge needs to specify one more hyperparameter.
2.3. Estimator of non-structured varying coefficient models
The estimation procedure of (4) is not difficult when we expand (4) in an appropriate basis. But it is hard to decide which covariate interacts with both time and another covariate arises. This problem won't occur in (1) and (3). To deal with it, we consider the following model
| (15) |
where is an unknown function and ε is a stochastic process such that . We expand the model (15) with appropriate basis functions. An explicit expression can still be obtained though there is severe sparsity of the basis functions. Furthermore, it would be helpful for us to decide which covariate interacts with time and another covariate if we just use three-order polynomials as basis functions. This is a pre-selection process. Then we update the estimation in model (4) using Algorithm 2.
3. Theoretical properties
In the work of [22], they showed the SINDy algorithm could find local minimizers of (16). Thus, it has theoretical guarantees summarized below in the setting of our dynamic varying coefficient models.
Lemma 3.1
Let with and . Assume that is the sequence generated by Algorithm 2. Define the objective function F by
(16) We have
converges to a fixed point of the iterative scheme defined by (16) at most mp steps;
a fixed point of the scheme is a local minimizer of F;
a global minimizer of F is a fixed point of the scheme;
the inequality holds unless the iterations are stationary.
Proof.
For the proof, we refer the reader to [22].
In (13), let be the true model with size , , where denotes the cardinality of . For , let be the dimensional sub-vector of Ψ indexed by S, and let be the matrix obtained from Θ by extracting columns corresponding to S. If the estimate of satisfies as , selection consistency is obtained. Then we can say SINDy consistently selects the correct model .
By Lemma 3.1, the solution of SINDy algorithm equals to penalized least squares solution. The penalty directly penalizes the number of nonzero components in (16), which is appealing in sparse regression. Though penalized least squares has a good property, it is discontinuous and the computation complexity is NP-hard. To overcome such difficulty, many researchers [4,7,8] suggested to relax the regularization and considered the regularization instead, for example, the LASSO [21]. Dicker et al. [6] proposed the seamless- (SELO) penalty, a smooth function on , very close to the penalty. They also showed the model selection consistency and asymptotical normality. Applying to our time dynamic varying coefficient models, Lemma 3.2 holds.
Lemma 3.2
Assume that true basis functions of each coefficient are a subset of the functions in (10) and the size of library m is fixed, then under
condition 1: and ;
condition 2: where ;
condition 3: There exist positive constants such that where and are the smallest and largest eigenvalues of respectively;
condition 4:
we have: .
Proof.
For the proof we refer the reader to [6].
According to Lemma 3.2, the SINDy algorithm in (16) approximately selects the true model with probability 1. Shen et al. [19] gave an error bound for a global minimizer of (16) using penalty. Lemma 3.3 can be obtained for our time dynamic varying coefficient models.
Lemma 3.3
Let and under (16) with then
where
Moreover, if for some constant then
where with and constant which is called an -band with upper and lower radii u and l .
Proof.
For the proof we refer the reader to [19].
According to Lemma 3.3, model selection consistency of (16) is clear. Now, we analyze the asymptotic normality of (16) in Theorem 3.4.
Theorem 3.4
If the model selection consistency of (16) mentioned in Lemmas 3.2 and 3.3 is guaranteed, then
in distribution, where .
Proof.
Let , the derivative of L to is
and solve , it is the ordinary least squares solution
By basic calculus,
Thus , ,
in distribution, where
Denote , , size of is , . Let be a submatrix of G,
Note , and is a map from to . Let
By Theorem 3.4 and the transformation , it is evident that in distribution. So
where is a vector, and is a diagonal matrix with elements , .
Thus the convergence rate of our proposed method is , while the convergence rate of backfitting algorithms is . Our proposed method has a faster convergence rate. According to asymptotic distributions, backfitting algorithms are biased, while our proposed method is unbiased.
4. Simulation
Simulation studies are conducted in this section to check the performance of the proposed estimator. We use matlab 2019a on a server with Intel(R) core(TM) i5-10210U CPU @1.60 GHz. Our proposed method aims at time dynamic varying coefficient regression models. So we compare with methods developed for time dynamic varying coefficient regression models. We focus on two types of varying coefficient models with Case 1 in [10] and Case 2 in [26].
In Examples 4.1 and 4.2, we generate 50 pseudo samples of sizes n = 200 and 400 from the model (4) with d = 3. The numbers of time points , , are having a discrete uniform distribution supported on the set . The time points , , , are sampled from , the uniform distribution on the unit interval , independently across i and l. The values of the linear covariates at are obtained according to
where V's and ζ's are random variables to be specified below. The values of the nonlinear covariates at are
where W's and η's are random variables to be specified below. The random vectors are with
and are with
The random variables are , and are independent of . They are sampled also independent of V's and W's. The errors in are generated by with being , and are independent of .
We consider two regression functions specified below.
Example 4.1
Example 4.2
We use basis functions as illustrated in Table 1 in Example 4.1. And we add sine, cosine basis functions for each in Example 4.2 as follows It can be seen the set of constructing candidate basis functions consists of the set of true functions in Example 4.1. But in Example 4.2, it is not the situation, so we can only give approximate results.
Figures 1 and 2 depict the true and the estimated coefficient functions for sample size n = 400 averaged over the 50 pseudo samples using our proposed method. The estimated surfaces in the figures suggest that our method estimates the true coefficients well. We compare our proposed method with that of [10], denoted by backfitting algorithm 1. We compute the mean integrated squared errors (MISE), the integrated squared biases (ISB), the integrated variances (IV), the number of iterations, and the computation time of the estimators. The formulas of , , and , j = 1, 2, 3, are given by
and . The BFA depends on the selected bandwidths. We select the optimal bandwidths , , that minimize the asymptotic mean integrated squared error (AMISE), where . For a given , we choose , dependent on , which minimizes . Denote such minima by . Then, we choose the optimal which minimizes . The optimal , are correspondingly obtained. We use Epanechnikov kernel, the convergence criteria for is
where is the tth iteration result of true , the threshold value ϵ is 0.00001. We summarize values of bandwidths in Figure 3.
Figure 1.
The lower panels are the true coefficient functions, and the upper panels are their estimates for n = 400 averaged over 50 pseudo samples in Example 4.1.
Figure 2.
The lower panels are the true coefficient functions, and the upper panels are their estimates for n = 400 averaged over 50 pseudo samples in Example 4.2.
Figure 3.
Values of bandwidth in Examples 4.1 and 4.2 in simulations.
Remark 4.1
The convergence threshold is a number close to zero in the numerical calculation, such as and . It depends on researcher's tolerence. Sometimes the convergence threshold is also called the tolerence. Researchers give threshold values without any justification in most cases. For example, the convergence threshold in [10] is ; the convergence threshold in [26] is . We set the threshold to considering the calculation time and accuracy. Higher accuracy requires more computation time. As we mentioned before, the convergence rate of backfitting algorithms is . Even though we set the threshold to , it is not going to make a qualitative difference.
Tables 2 and 3 summarize the results. The values of MISE, ISB, and IV decrease as the sample size n increases. Our method is higher in accuracy and more rapid in speed.
Table 2.
Comparison of the mean integrated squared errors (MISE), the integrated squared biases (ISB), the integrated variances (IV), the number of iterations, and the computing time in Example 4.1.
| n = 200 | n = 400 | |||
|---|---|---|---|---|
| Our proposed method | Backfitting algorithm 1 | Our proposed method | Backfitting algorithm 1 | |
| MISE | 0.0105 | 0.0331 | 0.0033 | 0.0259 |
| MISE | 0.0209 | 0.0668 | 0.0096 | 0.0186 |
| MISE | 0.0032 | 0.0270 | 0.0024 | 0.0123 |
| MISE | 0.0346 | 0.1269 | 0.0153 | 0.0568 |
| ISB | 0.0019 | 0.0217 | 0.0010 | 0.0176 |
| ISB | 0.0074 | 0.0460 | 0.0038 | 0.0091 |
| ISB | 0.0013 | 0.0234 | 0.0010 | 0.0083 |
| IV | 0.0086 | 0.0114 | 0.0023 | 0.0083 |
| IV | 0.0135 | 0.0208 | 0.0058 | 0.0095 |
| IV | 0.0019 | 0.0036 | 0.0014 | 0.0041 |
| Time | 0.0327s | 110.6247s | 0.0422s | 257.85s |
| Iterations | − | 7.5200 | − | 7.7800 |
Table 3.
Comparison of the mean integrated squared errors (MISE), the integrated squared biases (ISB), the integrated variances (IV), the number of iterations, and the computing time in Example 4.2.
| n = 200 | n = 400 | |||
|---|---|---|---|---|
| Our proposed method | Backfitting algorithm 1 | Our proposed method | Backfitting algorithm 1 | |
| MISE | 0.0281 | 0.0655 | 0.0102 | 0.0357 |
| MISE | 0.0781 | 0.1059 | 0.0234 | 0.0358 |
| MISE | 0.0041 | 0.0183 | 0.0009 | 0.0113 |
| MISE | 0.1103 | 0.1897 | 0.0345 | 0.0828 |
| ISB | 0.0106 | 0.0215 | 0.0016 | 0.0173 |
| ISB | 0.0502 | 0.0598 | 0.0105 | 0.0151 |
| ISB | 0.0007 | 0.0085 | 0.0007 | 0.0064 |
| IV | 0.0175 | 0.0439 | 0.0086 | 0.0184 |
| IV | 0.0279 | 0.0461 | 0.0129 | 0.0207 |
| IV | 0.0035 | 0.0098 | 0.0008 | 0.0049 |
| Time | 0.0766s | 140.665s | 0.1335s | 330.8357s |
| Iterations | − | 8.3800 | − | 8.5600 |
In Examples 4.3 and 4.4, two longitudinal covariates are generated by
where are having a multivariate normal distribution with mean and covariance
and and are and independent of . We take d = 2 and consider the following two regression functions.
Example 4.3
.
Example 4.4
The stochastic components are generated by , where are independent random variables having distribution with , , and
Then, the response data are obtained, where are i.i.d random noises having distribution.
If we use polynomials as basis functions, true coefficients in Example 4.3 are shown in Table 4. In Example 4.4, time T is expanded in third-order polynomials, sine and cosine functions. Sine and cosine functions include , and are expanded in third-order polynomials. So
where consists of third-order polynomials, sine and cosine functions, and are third-order polynomials.
Table 4.
Basis functions and corresponding coefficients in Example 4.3.
| 1 | 1.25 | 0 | 0 |
| t | 1 | 0 | 0 |
| 0 | 0 | 0 | |
| 0 | 0 | 0 | |
| 1 | 0 | 0 | |
| 0 | 0 | 0 | |
| 0 | 0 | 0 | |
| 0 | 0 | 0 | |
| 0 | 0 | 0 | |
| 0 | 0 | 1 | |
| 0 | 0 | 0 | |
| 0 | 0 | 0 | |
| 0 | 0 | 0 | |
| 0 | 1 | 0 | |
| 0 | 0 | 0 | |
| 0 | 0 | ||
| 0 | 0 | 0 | |
| 0 | 0 | 0 | |
| 0 | 0 | 0 | |
| 0 | 0 | 0 |
Figures 4 and 5 depict the true and the estimated coefficient functions for sample size n = 400 averaged over the 50 pseudo samples using our proposed method. The estimated surfaces in the figures suggest that our method estimates the true coefficients well. We compare our proposed method with that of [26], denoted by backfitting algorithm 2. The bandwidths selection method and convergence criteria are the same as Examples 4.3 and 4.4. Values of bandwidths are illustrated in Figure 6.
Figure 4.
The lower panels are the true coefficient functions, and the upper panels are their estimates for n = 400 averaged over 50 pseudo samples in Example 4.3.
Figure 5.
The lower panels are the true coefficient functions, and the upper panels are their estimates for n = 400 averaged over 50 pseudo samples in Example 4.4.
Figure 6.
Values of bandwidth in Examples 4.3 and 4.4 in simulations.
Tables 5 and 6 summarize the results. The values of MISE, ISB, and IV decrease as the sample size n increases. Our method is higher in accuracy and more rapid in speed.
Table 5.
Comparison of the mean integrated squared errors (MISE), the integrated squared biases (ISB), the integrated variances (IV), the number of iterations, and the computing time in Example 4.3.
| n = 200 | n = 400 | |||
|---|---|---|---|---|
| Our proposed method | Backfitting algorithm 2 | Our proposed method | Backfitting algorithm 2 | |
| MISE | 0.0296 | − | 0.0043 | − |
| MISE | 0.0088 | 0.0498 | 0.0029 | 0.0347 |
| MISE | 0.0076 | 0.0340 | 0.0008 | 0.0272 |
| MISE | 0.0164 | 0.0838 | 0.0037 | 0.0619 |
| ISB | 0.0026 | − | 0.0007 | − |
| ISB | 0.0032 | 0.0174 | 0.0005 | 0.0111 |
| ISB | 0.0009 | 0.0250 | 0.0005 | 0.0213 |
| IV | 0.0270 | − | 0.0036 | − |
| IV | 0.0056 | 0.0324 | 0.0024 | 0.0235 |
| IV | 0.0075 | 0.0090 | 0.0003 | 0.0059 |
| Time | 0.0048s | 25.7134s | 0.0062s | 43.9962s |
| Iterations | − | 4.5600 | − | 4.8800 |
Table 6.
Comparison of the mean integrated squared errors (MISE), the integrated squared biases (ISB), the integrated variances (IV), the number of iterations, and the computing time in Example 4.4.
| n = 200 | n = 400 | |||
|---|---|---|---|---|
| Our proposed method | Backfitting algorithm 2 | Our proposed method | Backfitting algorithm 2 | |
| MISE | 0.0621 | − | 0.0077 | − |
| MISE | 0.0391 | 0.0895 | 0.0066 | 0.0692 |
| MISE | 0.0039 | 0.0222 | 0.0012 | 0.0160 |
| MISE | 0.0430 | 0.1117 | 0.0078 | 0.0852 |
| ISB | 0.0561 | − | 0.0023 | − |
| ISB | 0.0299 | 0.0518 | 0.0015 | 0.0496 |
| ISB | 0.0006 | 0.0162 | 0.0004 | 0.0148 |
| IV | 0.0059 | − | 0.0054 | − |
| IV | 0.0092 | 0.0377 | 0.0051 | 0.0196 |
| IV | 0.0033 | 0.0060 | 0.0008 | 0.0012 |
| Time | 0.0589s | 48.2093s | 0.0924s | 127.2423s |
| Iterations | − | 6.5600 | − | 7.2600 |
5. Real data analysis
The data set comes from the UCI machine learning repository, called Beijing Multi-Site Air-Quality Data Data Set [24]. It includes hourly air pollutants data from 12 nationally controlled air-quality monitoring sites. The air-quality data are collected by the Beijing Municipal Environmental Monitoring Center. Variables consist of concentration , concentration , concentration , concentration , CO concentration and concentration . The meteorological data in each air-quality site are matched with the nearest weather station from the China Meteorological Administration, including temperature (degree Celsius), pressure (hPa), dew point temperature (degree Celsius), precipitation, wind direction, and wind speed (m/s). The period is from March 1st, 2013 to February 28th, 2017. Missing data are denoted as NA. One of the interesting purposes is to determine the relationship between and other impact factors.
According to domain knowledge and previous research, meteorology factors and gaseous pollutants are the main contributors to the formation process of . For gaseous pollutants, , , and CO are selected in our models. The reason we do not consider is that [15] concluded had little effect on . Besides, the correlation coefficient between and is , weakly correlated. For meteorology factors, temperature and wind speed are chosen. Dew point temperature and pressure are also weakly correlated to the concentration of , with correlation coefficients 0.02 and 0.01, respectively.
The original data set is from 12 nationally-controlled air-quality monitoring sites, we only analyze data from Wanshouxigong station in this study. There are 30666 hourly samples without missing values. Viewed as longitudinal data, there are n = 1461 daily observations with each day's time point during the period. We scale the interval of time point from to , and the concentration of CO is divided by 100. We split the data into a training set with 25000 randomly sampled data and a test set with the remaining data. We consider different types of varying coefficient models to analyze the relationship between the concentration of and the selected covariates. The measure of assessment is the relative mean squared error (RMSE) given by
It can been seen the smaller the RMSE is, the better the model fits. We do 100 simulations and calculate the mean RMSE as the final RMSE in test data.
Model 1: additive model, refer to [20].
we use ‘gam’ in R package ‘mgcv’ to fit Model 1, the RMSE is 0.2998.
Model 2: varying coefficient model in (1)
the RMSE is 0.2930 when basis functions of are three-order polynomials and sin functions.
Model 3: varying coefficient model in (3)
the RMSE is 0.2836 when basis functions of are three-order polynomials and sin functions. Model 2 is a special case of Model 3. By comparing RMSE, Model 3 is better than Model 2. Figure 7 depicts the estimated surfaces of different components. The concentration of increases as , CO, and temperature increase. The result is consistent with the fact and CO are pollutants that compose . The temperature has a higher impact on the concentration of at noon than in the morning and evening when the temperature is high. In contrast, it has a higher impact in the morning and evening than at noon when the temperature is low.
Figure 7.
Estimates of coefficient functions in Model 3.
Model 4: varying coefficient model in (4)
different from the above-mentioned models, the problems which covariate should be the linear component and which covariate should be the varying coefficient component, are unclear for Model 4. We use cubic polynomials as the library of functions for full data, and find out the coefficients of terms and are not zero. According to the result in reference [24], air quality will be better once the wind blows away . Cooling is often accompanied by strong winds. This result is consistent with our life experience in Beijing. So Model 4 makes sense according to our data-driven pre-estimation result and research in [24].
If we only use cubic polynomials, the RMSE is 0.2935; if we use cubic polynomials and sine functions simultaneously, the RMSE is 0.2916; if we use second-order polynomials, the RMSE is 0.3020; if we use both second-order polynomial and sine functions, the RMSE is 0.3020. By comparing RMSE, we choose cubic polynomials and sine functions in this case. Figure 8 depicts the estimated surface of each component.
Figure 8.
Estimates of coefficient functions in Model 4.
By comparing RMSE of four models, Model 3 performs best.
6. Discussion
Designing efficient and accurate algorithms to estimate varying coefficient models is of significance as more and more data is available. We focus on different types of dynamic varying coefficient models and propose a new procedure to estimate the varying component efficiently in our research. The main idea is to expand the varying component with appropriate basis functions and then solve a sparse regression problem. Finally, we can get an explicit expression of the model under certain conditions. Furthermore, the proposed procedure can also find out an expression without model assumptions as shown in the real data analysis.
Constructing a library of functions, especially when the number of covariates is large is a challenge to us. To deal with this, we can incorporate domain knowledge and other related methods. Still taking the concentration of as an example, the existing research results show meteorology factors and gaseous pollutants are the main contributors to the formation process of , so the feature extraction process will depend on these two aspects. Curse of dimensionality will occur when the number of covariates is large. To handle this, before implementing the above-mentioned procedure, we can delete some unimportant variables as in the variable selection and feature screening procedure. This is based on the sparsity assumption. Sensitivity to noise is also troublesome. One technique is to use absolute loss function or check function in quantile regression to get robust estimation. And the other is to take advantage of Bayesian inference [23]. We will study robust estimation under the frame of SINDy in our future research.
Acknowledgements
We would like to thank three anonymous referees for their constructive comments and suggestions.
Funding Statement
The work was partially supported by the National Natural Science Foundation of China [grant number 11861042], and the China Statistical Research Project [grant number 2020LZ25].
Disclosure statement
No potential conflict of interest was reported by the author(s).
References
- 1.Bauer B. and Kohler M., On deep learning as a remedy for the curse of dimensionality in nonparametric regression, Ann. Statist. 47 (2019), pp. 2261–2285. [Google Scholar]
- 2.Bozdogan H., Model selection and Akaike's information criterion (AIC): The general theory and its analytical extensions, Psychometrika 52 (1987), pp. 345–370. [Google Scholar]
- 3.Brunton S.L., Proctor J.L., and Kutz J.N., Discovering governing equations from data by sparse identification of nonlinear dynamical systems, Proc. Natl. Acad. Sci. U. S. A 113 (2016), pp. 3932–3937. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Cands E.J., Romberg J.K., and Tao T., Stable signal recovery from incomplete and inaccurate measurements, Commun. Pure Appl. Math 59 (2006), pp. 1207–1223. [Google Scholar]
- 5.Chiang C.T., Rice J.A., and Wu C.O., Smoothing spline estimation for varying coefficient models with repeatedly measured dependent variables, J. Am. Stat. Assoc 96 (2001), pp. 605–619. [Google Scholar]
- 6.Dicker L., Huang B., and Lin X., Variable selection and estimation with the seamless penalty, Stat. Sin. 23 (2013), pp. 929–962. [Google Scholar]
- 7.Donoho D.L., Compressed sensing, IEEE Trans. Inf. Theor 52 (2006), pp. 1289–1306. [Google Scholar]
- 8.Donoho D.L., High-dimensional centrally symmetric polytopes with neighborliness proportional to dimension, Discret. Comput. Geom 35 (2005), pp. 617–652. [Google Scholar]
- 9.Hastie T. and Tibshirani R., Varying-coefficient models, J. R. Stat. Soc. Ser. B-Stat. Methodol 55 (1996), pp. 757–796. [Google Scholar]
- 10.Lee K., Lee Y.K., Park B.U., and Yang S.J., Time-dynamic varying coefficient models for longitudinal data, Comput. Stat. Data Anal. 123 (2018), pp. 50–65. [Google Scholar]
- 11.Lee Y.K., Mammen E., and Park B.U., Projection-type estimation for varying coefficient regression models, Bernoulli 18 (2012), pp. 177–205. [Google Scholar]
- 12.Liang J., Zhang X., Wang K., Tang M., and Tian M., Discovering dynamic models of covid-19 transmission, Transbound. Emerg. Dis. (2021). 10.1111/tbed.14263. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Linton O. and Nielsen J.P., A kernel method of estimating structured nonparametric regression based on marginal integration, Biometrika 82 (1995), pp. 93–100. [Google Scholar]
- 14.Mammen E., Linton O., and Nielsen J., The existence and asymptotic properties of a backfitting projection algorithm under weak conditions, Ann. Stat. 27 (2000), pp. 1443–1490. [Google Scholar]
- 15.Mei B. and Tian M., Analysis of influencing factors on pm2.5 in beijing based on spatio-temporal model, J. Appl. Statist. Manage. 37 (2018), pp. 571–586. [Google Scholar]
- 16.Park B.U., Mammen E., Lee Y.K., and Lee E.R., Varying coefficient regression models: A review and new developments, Int. Stat. Rev. 83 (2015), pp. 36–64. [Google Scholar]
- 17.Peter Bühlmann S.v.d.G., Statistics for High-Dimensional Data: Methods, Theory and Applications, Berlin, Heidelberg, Springer, 2011. [Google Scholar]
- 18.Rudy S.H., Brunton S.L., Proctor J.L., and Kutz J.N., Data-driven discovery of partial differential equations, Sci. Adv. 3 (2017), pp. e1602614. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Shen X., Pan W., Zhu Y., and Zhou H., On constrained and regularized high-dimensional regression, Ann. Inst. Stat. Math 65 (2013), pp. 807–832. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Su J.G., Hopke P.K., Tian Y., Baldwin N., and Rich D.Q., Modeling particulate matter concentrations measured through mobile monitoring in a deletion/substitution/addition approach, Atmos. Environ. 122 (2015), pp. 477–483. [Google Scholar]
- 21.Tibshirani R., Regression shrinkage and selection via the lasso, J. R. Stat. Soc. Ser. B-Stat. Methodol. 58 (1996), pp. 267–288. [Google Scholar]
- 22.Zhang L. and Schaeffer H., On the convergence of the Sindy algorithm, Multiscale Model. Simul. 17 (2019), pp. 948–972. [Google Scholar]
- 23.Zhang S. and Lin G., Robust data-driven discovery of governing physical laws with error bars, Proc. R. Soc. A-Math. Phys. Eng. Sci. 474 (2018), pp. 20180305. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Zhang S., Guo B., Dong A., He J., Xu Z., and Chen S.X., Cautionary tales on air-quality improvement in Beijing, Proc. R Soc. Math. Phys. Eng. Sci. 473 (2017), pp. 20170457. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Zhang W. and Fan J., Statistical estimation in varying coefficient models, Ann. Stat. 27 (1999), pp. 1491–1518. [Google Scholar]
- 26.Zhang X., Park B.U., and Wang J.L., Time-varying additive models for longitudinal data, J. Am. Stat. Assoc. 108 (2013), pp. 983–998. [Google Scholar]








