Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2021 Feb 16;16(2):e0246159. doi: 10.1371/journal.pone.0246159

HDSI: High dimensional selection with interactions algorithm on feature selection and testing

Rahi Jain 1, Wei Xu 1,2,*
Editor: Fabio Rapallo3
PMCID: PMC7886179  PMID: 33592034

Abstract

Feature selection on high dimensional data along with the interaction effects is a critical challenge for classical statistical learning techniques. Existing feature selection algorithms such as random LASSO leverages LASSO capability to handle high dimensional data. However, the technique has two main limitations, namely the inability to consider interaction terms and the lack of a statistical test for determining the significance of selected features. This study proposes a High Dimensional Selection with Interactions (HDSI) algorithm, a new feature selection method, which can handle high-dimensional data, incorporate interaction terms, provide the statistical inferences of selected features and leverage the capability of existing classical statistical techniques. The method allows the application of any statistical technique like LASSO and subset selection on multiple bootstrapped samples; each contains randomly selected features. Each bootstrap data incorporates interaction terms for the randomly sampled features. The selected features from each model are pooled and their statistical significance is determined. The selected statistically significant features are used as the final output of the approach, whose final coefficients are estimated using appropriate statistical techniques. The performance of HDSI is evaluated using both simulated data and real studies. In general, HDSI outperforms the commonly used algorithms such as LASSO, subset selection, adaptive LASSO, random LASSO and group LASSO.

Introduction

Classical statistical models have been the mainstay for data analysis. However, the growth in dataset sizes both in terms of sample size (n) and feature dimension (p) had triggered some challenges for traditional approaches of statistical data analysis. In the case of p>n, classical approaches cannot control for false discovery rate of identified features. In the case of n>p sample size, an increase in the size of a feature set would exponentially increase the feature combination set that needs evaluation which leads to an exponential increase in requirement for computation time and resources [1]. Furthermore, the original input feature set might not contain complete information. Hence, the incorporation of interaction terms in the feature set might be necessary [2], which exacerbates the challenge of processing the feature combination set.

Feature selection has been used as a common approach to address the high dimensionality challenge by identifying the significant features from the input feature space for performing final statistical analysis. Different strategies are in use for performing feature selection which can be categorized into three primary types and two sub-types, as shown in Fig 1. The first primary approach is expert-based feature selection which relies upon the use of experience of experts to shortlist variables for final model development [3, 4]. It is subjective and becomes more challenging to perform with an increase in feature set and interaction terms.

Fig 1. Feature selection approaches suggested in literature with examples.

Fig 1

Statistics-based feature selection is another primary approach which uses various statistical techniques to perform the feature selection. One strategy is to select features based on the intrinsic properties of the features like multicollinearity [5] and distribution [6]. Another strategy is to select features based on their statistical significance (i.e. p-value) during univariate analysis [4, 5]. Linear regression-based screening is a common approach in univariate analysis [7]. The third strategy is to select features based on their importance in the model during multivariate analysis [4, 5]. Some of the conventional approaches used in multivariate analysis are subset selection and penalized regression [2, 8]. However, they have certain limitations like subset selection approach can work only if p<<n and LASSO regression cannot select features more than n [9]. Besides classic LASSO, many extensions have been proposed. Elastic LASSO can be used for p>n cases, however, it may not be immune from selecting the noise variables [10]. Group LASSO is another approach which has been used to improve the performance of LASSO in selecting variables which need to be selected in a group such as marginal and interaction terms. However, they follow a selection hierarchy and may not select interaction terms if marginal features are not selected. Further, the groups need to be predefined [9].

One of the sub-type approaches is expert-statistics hybrid based feature selection. This approach incorporates the domain knowledge in the feature selection process. The domain knowledge and statistical analysis sequence can occur in two ways. One strategy is to first assign importance to the features based on the domain knowledge followed by implementation of statistical analysis on the feature set for feature selection. This approach is inherent in Bayesian regression-based approaches [11]. The second strategy is to perform statistical analysis for preliminary feature selection. The final features are selected from the preliminary selected feature set based on their importance estimated from the domain knowledge. Differential gene expression analysis commonly employs this strategy [12]. However, the issues of expert-based feature selection approach prevail.

The third primary approach is machine learning (ML) based feature selection which can deal with high dimensional data but focus less on interpretability of the selection process. Random Forest is a common ML technique which can perform feature selection [13]. It provides the importance of a feature in model prediction, which can be a metric for feature selection. The technique allows the use of different ensemble methods like bagging [14], random subspace sampling [15] or both [16] for model building. Bagging and random subspace sampling methods enable creating robust models in noisy settings [17].

Consequently, the ML ensemble methods have been integrated with classical statistical techniques to create ML-statistics hybrid based feature selection as the other sub-type approach. BoLASSO is one such technique which performs LASSO regression on different bootstrap samples of n [18]. The features are selected based on the number of models in which feature was selected. regRSM is another technique in which subsamples of feature set are created with/without weighted selection probability followed with linear regression on each of the feature subsets [19]. The feature selection is based on their performance in t-statistic metric across different models. Random LASSO technique is a two-step procedure which integrates both bagging and random subspace sampling with LASSO for feature selection [20]. Bootstrapping and random subspace sampling creates subsamples of a dataset. In the first step, application of LASSO on each subsample provides the importance of each feature. In the second step, application of LASSO or adaptive LASSO on each subsample enables in obtaining the final coefficient estimate of the features.

All these techniques do not directly consider the interaction effects in the feature selection process. In the case of LASSO-based techniques, they may select only the interaction feature while ignoring their marginal features. Further, random LASSO is comprehensive in feature selection and could outperform elastic net [20], but it is also computationally intensive due to its two-step approach. Random LASSO performs both feature selection and estimation but does not give the user flexibility to stop the process at feature selection or use techniques other than LASSO or adaptive LASSO. It does not provide a guideline to address the interaction terms in the model as well as the significance of estimated coefficients. Further, the estimates may suffer from systematic bias as during feature selection and estimation it assigns zero value to features not sampled. Additionally, its selection step may have interpretability issues as it uses a user assigned threshold value to determine the feature performance.

In this paper, we focus on addressing the challenge of incorporating interaction effects in the feature selection approach with lesser computational intensiveness as compared to random LASSO. We propose a novel strategy, named High Dimensional Selection with Interactions (HDSI) algorithm for improved feature selection. Our approach combines bootstrapping and random subspace sampling with classical statistical model selection techniques with the in-built capability to handle interaction terms. Further, the approach addresses the shortcomings of random LASSO by allowing use of multiple statistical techniques in feature selection, reduce the systematic bias in feature selection and improve interpretability of feature selection process. The paper is organized as follows. The proposed strategy is explained in the Methodology Section with the evaluation of the method performance in the Simulation Study Section. The strategy is tested using real data studies in the Real Data Studies Section, followed by the Conclusion and Discussion Section.

Methodology

In this section, first, we will describe some of the existing feature selection algorithms such as LASSO, random forest and random LASSO and is followed by the proposed HDSI algorithm.

LASSO

LASSO is a penalized regression method proposed by (Tibshirani, 1996) [21] which enables the features selection. Its criterion performs L1-penalization of the regression coefficients, minβi=1n(yij=1pβjxij)2+λj=1p|βj|, where i represents the ith subject of totally n subjects, βj represents the regression coefficient of jth feature in p feature set. yi represents the response feature and xi = (xi1,….,xip) represents a p-dimensional vector of features. λ is a non-negative tuning parameter.

The L1-norm penalty has the singularity of the derivative at |βj| = 0, so LASSO tries to shrink the βj towards zero and some estimated βj will be precisely zero at sufficiently large λ. However, this method has a few limitations. Firstly, in cases of p>n, it can only select maximum n features. Secondly, among the highly correlated variables, it may choose arbitrarily only one variable and drop the other variables [10]. Thirdly, in case of interactions, these methods could select the interaction terms while dropping the main features, since LASSO is neutral to the pattern in which features exist [22]. In cases where interpretive models are desired, methods that allow retention of main features is preferable.

Random forest

Random Forest is the extension of the decision tree method (a non-parametric approach) [16]. In the decision-tree technique, variables selection is made parsimoniously based on a series of logical criteria to separate the data into subsets and estimate the outcome in each subset. The random forest creates multiple decision-trees. These multiple decision-trees could be prepared by either bootstrapping the samples, randomly selecting the features or both. The random forest estimates the overall importance of features based on the influence of feature on the performance of multiple decision-trees using some metric [16, 23, 24]. The features can be selected using the feature importance scores. However, the approach may show bias due to the scale of measurement of features [24]. Further, while, the random forest consider interaction terms in building the model, it does not directly provide the importance score of the interaction effects [25].

Random LASSO

Random LASSO is a dual procedure method as shown in Algorithm 1 [20]. In the first procedure, the importance of features is estimated using bootstrapping and random subspace sampling. LASSO is used for feature estimation in each bootstrap. Then, in the second procedure, bootstrapping and weighted random subspace sampling is performed to get the estimates of the coefficients of features. LASSO (or, Adaptive LASSO) is used for feature estimation in each bootstrap. The final estimates of the features are computed by averaging multiple estimates obtained from bootstrapping over the total number of bootstraps.

Algorithm 1: Random LASSO
Procedure I Generate the Importance Scores for the Features
I(a) Bootstrap B samples with size n from the original dataset.
I(b) For each Bootstrap sample, randomly select q(1) features (q(1)n) from original p features.
I(c) Apply LASSO to estimate coefficient, β^ij(1)|i={1,,B},j={1,,p}. The coefficients of unselected features for each bootstrap sample are considered zero.
I(d) Compute the importance score, Ij=i=1B|β^ij(1)|/B.
Procedure II Generate the final coefficient estimates of the Features
II(a) Bootstrap another set of B samples with size n from the original dataset
II(b) For each Bootstrap sample, randomly select q(2) features (q(2)n) from original p features with feature selection probability proportional to importance scores, Ij.
II(c) Apply LASSO (or, Adaptive LASSO with weight, wj=Ij1) to estimate coefficient, β^ij(2)|i={1,,B},j={1,,p}. The coefficients of unselected features for each bootstrap sample are considered zero.
II(d) Compute the final coefficient estimate, β^j=i=1B|β^ij(2)|/B. The features with final coefficients above predefined threshold are selected.

High dimensional selection with interactions (HDSI)

Random LASSO is a comprehensive approach for feature selection in high dimensional settings, but it has certain limitations. One of the limitations is that it is computationally intensive and provides little flexibility in the coefficient estimation of selected features as it allows use of only LASSO (or, Adaptive LASSO) for coefficient estimation. Secondly, random LASSO does not consider any interaction terms in the feature selection process. Thirdly, random LASSO does not provide any statistical guideline for various activities like the number of bootstrap samples, feature selection and its significance estimation. Fourthly, random LASSO assigns zero value to coefficients of the features unselected during random selection. These values may create systematic bias as they are used to calculate the importance scores of features.

The HDSI methodology (Algorithm 2) is developed to address its limitations. Fig 2 provides a graphical representation of HDSI. The method generates random samples by bootstrapping the original dataset and random subspace sampling of the features. Interaction terms of sampled features are generated for each sample. The statistical modeling is performed on each sample using appropriate feature selection techniques like penalized regression or subset selection. The results of samples are pooled to determine the statistical significance of the estimated coefficients of features and select the significant features. The final coefficient estimation of the selected features could be done with another set of appropriate statistical modeling technique like simple linear regression. The proposed method is discussed below for more details.

Fig 2. Graphical representation of HDSI methodology.

Fig 2

p: number of features in the dataset, n: sample size of the dataset, p*: number of features (including sampled features and its interaction terms) used for modeling and n*: each bootstrapped dataset has the same sample size n used for modeling. The green colour steps represent the proposed method and blue colour steps represent the process to estimate the coefficient of selected features.

Procedure flexibility

Random LASSO uses a two-stage procedure to perform feature selection and coefficient estimation. However, double bootstrapping makes the method computationally expensive. Besides that, while adaptive LASSO has global oracle property, its performance is dependent on the estimator used for assigning weights to the coefficients of predictors [26]. Further, random LASSO may bias the coefficients [20]. Hence, the user might prefer using other techniques on selected features for coefficient estimation, but the whole process needs to complete to get the feature selection results.

HDSI modifies random LASSO to increase procedure flexibility and reduce computation intensiveness. The first modification is using the importance scores from Step I(d) as a metric for the feature selection. This modification allows the user to obtain preliminary results on feature selection and enable them to decide on the coefficient estimation step. The second modification is to allow the use of techniques other than LASSO for coefficient estimation in Procedure I. This will enable the user to use different techniques in the same procedure and address the LASSO limitations.

Algorithm 2: HDSI
Procedure I Feature Selection
I(a) Bootstrap B samples with size n from the original dataset.
I(b) For each Bootstrap sample, randomly select q features (qn) from original p features.
I(c) Prepare the χ interaction set from q features.
I(d) Create a final sample feature set, p* = χq.
I(e) Apply LASSO (or, any other feature selection technique) to estimate coefficient, β^ij|i={1,,B},j={1,..,p,..,k=2ω(pk)}. The coefficients of unselected features for each bootstrap sample are considered missing.
I(f) Compute the mean coefficient estimate, β^j=i=1Bβ^ij/bj, where bj is the number of bootstrap samples containing the jth feature.
Compute jth feature minimum coefficient of determination (R2), minRj2=Min(Rij2)|i{1,,B}.
I(g) Select significant features based on the quantile of coefficient estimate and minimum R2 value.
I(h) Add missing marginal features of significant interaction terms in the final selected feature set.

Feature selection

Random LASSO performs the feature selection in its first stage of the procedure. Similar to random LASSO (Step I(a) and Step I(b), Algorithm 1), HDSI (Step I(a) and Step I(b), Algorithm 2) generates random samples from the dataset of n sample size and p feature space. Multiple samples of size n from the original dataset is created with a replacement through bootstrapping. For each of these samples, a feature sample of size q from original feature space is created without replacement through random subspace sampling.

While, random LASSO performs LASSO based statistical modeling for coefficient estimation of each feature q in each sample (Step I(c), Algorithm 1), HDSI incorporates interaction terms before statistical modeling (Step I(c) and Step I(d), Algorithm 2). HDSI generates all possible interaction terms between q features. A final sample feature set of size p* is created for each bootstrap sample which is the combination of q features and its interaction terms. The statistical modeling of p* features is performed (Step I(e), Algorithm 2). Different modeling techniques like LASSO, adaptive LASSO, regression and subset selection could be used for getting coefficient estimates of features in each of the bootstrap samples. HDSI uses three different techniques for feature selection, namely LASSO based (HDSI_L), adaptive LASSO based (HDSI_AL) and regression based (HDSI_R).

Random LASSO pools the coefficient estimates of all features from all the bootstraps to estimate the feature performance. It calculates the importance score of each feature by averaging its estimates from bootstrapped datasets over the total number of bootstraps (Step I(d), Algorithm 1). Random LASSO considers the coefficients of unselected features in a bootstrap sample as zero rather than missing, which causes systematic bias in the estimation of regression coefficients. HDSI estimates the feature performance using two metrics, namely coefficient estimates and model coefficient of determination (R2). HDSI treats the coefficients and model R2 of unselected features in a bootstrap sample as missing (Step I(f), Algorithm 2). Consequently, during the computation step of the jth feature, bootstrap samples with the missing feature were dropped from the computation, as shown below:

β^j=i=1Bβ^ij/bj (1)
minRj2=Min(Rij2)|i{1,,B} (2)

where bj is the number of bootstrap samples containing the jth feature, β^j represents an averaged coefficient estimation of the jth feature over bj bootstrap samples and minRj2 represents a minimum model R2 of the jth feature over bj bootstrap samples.

Random LASSO does not provide any statistical guideline to determine the significance of estimated coefficients. Since, coefficient estimate distribution is unknown, in HDSI the upper and lower quantile values of coefficients are estimated using β^ij|i=1,,bj values as sample data. If zero value is not present between the lower and upper quantile values, β^j is considered to have a significant non-zero value.

Finally, the features with performance above a predefined threshold are selected. Random LASSO uses heuristic rather than the statistical threshold to select the features based on the estimated coefficients (Step II(d), Algorithm 1). HDSI selects the features which fulfil two criteria (Step I(g), Algorithm 2). Firstly, the coefficient estimates between lower and upper quantile values should not contain zero. The quantile (Qi) is a hyperparameter which needs to be optimized where lower quantile is Qi/2, and upper quantile is 100−Qi/2. Secondly, the features should have occurred in models with a high coefficient of determination, i.e. R2. A heuristic cut-off value (Rf) for considering high R2 can be used. However, the heuristic cut-off value may not be easy to interpret. Accordingly, the heuristic cut-off value is transformed into μminR2+RfσminR2, where μminR2(=j=1p*minRj2/p*) is the mean value of minRj2 values of p* features. σMinR2 is the standard deviation of minRj2 values of p* features as given below.

σminR2=j=1p*(minRj2μminR2)2p*1 (3)

minRj2 of the feature should be higher than μminR2+RfσminR2 for selection. The hyperparameter Rf can take value from [−∞,∞]. Among the selected features, some interaction terms could get selected while their corresponding marginal features are not. In such cases, the final set of selected features incorporates the unselected marginal features of selected interaction terms (Step I(h), Algorithm 2).

Interaction effects

Random LASSO does not provide any guideline in dealing with interactions among the features. Further, non-group LASSO based LASSO methods have issues in dealing with interactions since LASSO is neutral to feature pattern; hence any pattern of features can exist [22]. Thus, a model based on LASSO and adaptive LASSO cannot deal with feature interactions. In HDSI, after the selection of q features for a bootstrap sample, all the possible k = {2,…,ω} level interaction terms, χ, among the q features are created. A new sample feature set, p* (= χq) is used for coefficient estimation.

χ=k=2ω(qk) (4)

Number of feature samples

The pooling criteria is dependent on the feature performance in different models. Accordingly, it is vital to ensure that every feature is sampled multiple times. The current random LASSO algorithm provides no statistical guideline to determine the number of bootstrap times a feature should be selected and modeled.

The HDSI method considers the hypothetical population mean value of a coefficient as zero, then uses it as a reference value against which the estimated mean value of coefficient could be compared. Since it is a one-sample case, the sample size of the coefficient values for a feature could be estimated from following Lehr’s equation, L = 8/Δ2 [27]. L is the minimum number of times a feature should be selected (i.e., the sample size of coefficient values) and Δ is the effect size. As Cohen’s rule of thumb, Δ could be equal to 0.2, 0.5 or 0.8 for ‘small’, ‘medium’ and ‘large’ effect sizes [28]. The probability of a feature to be included in a sample of q features for any bootstrap is ρ = q/p. In cases when interaction terms are considered ρ can be calculated as follows:

ρ=k=2ω(qk)/k=2ω(pk) (5)

q and p are not added into the calculation because the marginal features would be selected when the interaction terms are selected. Each B can be considered an independent trial for selecting a feature. Then, the probability of a feature to get selected L times, (Pr(X = L)), in B trials is equivalent to the probability mass function of a binomial distribution, (BL)ρL(1ρ)BL. Since, L is the minimum number of desired selections of a feature, the cumulative distribution function should be used as follows for calculating B:

Pr(XL)=1m=0L1Pr(X=m) (6)
Bf(Pr(XL),L) (7)

Feature estimation

Random LASSO estimates the coefficient values of selected features by performing the second stage of the procedure (Algorithm 1). The second stage repeats the steps of the first stage of the procedure with some modifications. Firstly, weighted random subspace sampling rather than random subspace sampling performs the sampling of features. Secondly, adaptive LASSO (recommended) performs statistical modeling. In both modifications, the feature weight is proportional to its importance score obtained from step 1(d). The importance score obtained in Step 2(d) is the estimated coefficient value of the features. The features with the coefficient value above the heuristically determined threshold value are selected.

HDSI does not perform stage two to reduce computation time. Further, the statistical technique will also depend on the user problem statement. The method gives the user flexibility in choosing the appropriate statistical modeling technique for feature estimation. The current study uses ordinary least squares based regression.

Hyperparameters

HDSI requires three hyperparameters, namely, number of features in a sample (q), coefficient estimate quantile threshold (Qi) and minimum R2 threshold (Rf). The values of these hyperparameters depend on the dataset. Hence, hyperparameter optimization is needed for optimal performance. Hyperparameter optimization is done in three steps. In first step, Qi and Rf are kept constant at value five and value zero respectively and q is optimized. Randomly, multiple values are generated for q and value with the best predictive performance is selected. In second step, q and Rf are kept constant at the best value obtained from first step and value zero respectively, while Qi is optimized. Randomly, multiple values are generated for Qi and value with the best predictive performance is selected. In third step, q and Qi are kept constant at the best value obtained from step one and two respectively, while Rf is optimized. Randomly, multiple values are generated for Rf and value with the best predictive performance is selected.

Simulation studies

Simulated data are used to demonstrate the performance of the proposed method and compare it with other methods. The current study considers only two-way interactions, but the approach can be demonstrated for higher-order interactions too. The simulation data are generated from the regression model, y = β0+β1x1+⋯+βpxp+β12x12+⋯+ϵ. ε~N(0, σ2), x1,…,xp~N(0, 1) and {x12,x13,…,x(p−1)p} represents the two-way interactions between features {(x1,x2), (x1,x3),…, (xp−1, xp)}. Coefficient values are zero for features unless mentioned (Table 1). Covariance matrix is defined to create multicollinearity in the model with non-zero covariance among {x1,…,x5} and zero covariance among all other cases as shown below:

[x1x1x1x2..x1x5x2x1x2x2x2x5x5x1x5x1x5x5x6x1x6x1x6x5xpx1]=[10.30.30.60.6000.310.30.20.10.00.30.310.20.10.00.60.20.210.10.00.60.10.10.110.0000001.0.......000000001]
Table 1. Description of the simulation data.
Scenario Features Effect β (Non-Zero coefficients) p Train set Test set σ2
Marginal Terms
1 Yes {β1, β2, β3, β12} = {0.2, 0.3, 0.4, 0.3} 25 500 500 0.25
2 Yes {β1, β2, β3, β12} = {0.2, 0.3, 0.4, 0.3} 50 500 500 0.25
3 Yes {β1, β2, β3, β12} = {0.2, 0.3, 0.4, 0.3} 100 500 500 0.25
4 No {β3, β12} = {0.4, 0.3} 50 500 500 0.25

The data for each feature is generated from the multivariate normal distribution. Table 1 shows the different settings considered for evaluating the different models. The number of features, p, considered across different settings varied from 25–100. The training and test dataset used in each scenario is 500. Scenario 4 does not consider the marginal effect of the interaction terms. The number of target variables is put less than n to enable comparison with standard methods.

Hyperparameter optimal range

Scenario 1 dataset is used to perform the hyperparameter optimization for identifying the optimal ranges of q, Qi and Rf. Different hyperparameter combinations are tried with HDSI_AL technique on the training dataset. The hyperparameter combination with best mean predictive performance on the five-fold cross-validated dataset is selected. The best hyperparameters predictive performance is tested on test data. Root mean square error (RMSE) is used to measure predictive performance. Table 2 shows the results obtained from 10 trials.

Table 2. Optimal region for hyperparameters.

HDSI Technique Hyperparameter (trials = 10) RMSE
q Qi Rf
Mean (±2SD) [Min, Max] Mean (±2SD) [Min, Max] Mean (±2SD) [Min, Max] Mean (95% CI)
HDSI_AL 12 (±14) [5,20] 7.22 (±11.01) [0.73,13.96] 0.98 (±2.16) [-0.90,2.13] 0.23(0.19–0.27)
HDSI_R 15 (±4) [9,18] 6.13 (±6.52) [3.08,15.25] 0.98 (±1.42) [-0.36,1.74] 0.25(0.18–0.31)

Further, the hyperparameter optimization process is repeated with HDSI_R technique for the same dataset. The hyperparameter range identified for each parameter is within two standard deviation range identified with HDSI_AL technique. Additionally, the RMSE performance obtained from the two techniques is similar. The search region for optimal values of hyperparameters q, Qi and Rf is [2,26], (0, 18.24] and [-1.18,3.14] respectively. The optimal region of q is truncated and depends upon feature space as q cannot take values less than 2 or more than p. Similarly, the minimal value of Qi is truncated at zero as quantile intervals are not negative.

Bootstraps analysis

Scenario 1 dataset is used to showcase the importance of bootstraps in the model performance. The performance of HDSI_AL and HDSI_R is compared for different effect size and consequently, different number of bootstraps as shown in Table 3. The mean values of hyperparameters q, Qi and Rf optimal region are used for analysis. RMSE is used to measure predictive performance. The results from 10 trials suggest that increase in the number of bootstraps reduces the selection of noise variables and improves predictive performance. Effect size, Δ = 0.5, which corresponds to 186 bootstraps, can eliminate all noise features and only select target features using HDSI_AL technique. In the case of HDSI_R, Δ = 0.5 can eliminate almost all noise features and only select target features. Nevertheless, an increase in bootstraps helps in reducing the noise features selection. Additionally, results suggest that HDSI could provide good model performance even when globally optimal hyperparameters are not used.

Table 3. HDSI performance in different bootstraps.

L Δ Bootstraps Performance (Trials = 10)
HDSI_AL HDSI_R
Selected Features RMSE (95% CI) Selected Features RMSE (95% CI)
Marginal (Range) Interactions (Range) Marginal (Range) Interactions (Range)
1 2.8 14 6 (3–11) 4 (1–8) 0.23 (0.19–0.27) 18 (10–24) 22 (8–42) 0.27 (0.25–0.30)
5 1.3 40 3 (3–4) 1 (1–2) 0.23 (0.19–0.27) 10 (5–13) 8 (3–12) 0.23 (0.19–0.27)
13 0.8 80 3 (3–4) 1 (1–2) 0.23 (0.19–0.27) 6 (3–7) 4 (1–5) 0.23 (0.19–0.27)
32 0.5 186 3 (3–3) 1 (1–1) 0.23 (0.19–0.27) 4 (3–5) 2 (1–3) 0.23 (0.19–0.27)
200 0.2 1006 3 (3–3) 1 (1–1) 0.23 (0.19–0.27) 3 (3–4) 1 (1–2) 0.23 (0.19–0.27)

HDSI comparison with standard methods

The performance of HDSI_L, HDSI_AL and HDSI_R are compared with various standard methods, namely random LASSO, LASSO, adaptive LASSO, group LASSO and regression. The large effect size is considered for the study. The simulation studies are performed in R. Random LASSO is performed by modifying the existing algorithm available at GitHub (https://github.com/samskhan/KSULasso/tree/master/R). The code was modified to allow the algorithm to take different q values and provide the intercept term. Further, the original code was unable to run if the importance score of features after the procedure I is zero, so negligible value (1/(px106)) is assigned to it. The number of bootstrap samples used for the random LASSO is 200, as suggested by [20] in their paper. The R package glmnet is used to perform LASSO and adaptive LASSO [29]. Ridge regression precedes adaptive LASSO (except, when adaptive LASSO is used in random LASSO) for obtaining weights for adaptive LASSO as suggested by [26]. The R package glinternet [30] is used to perform group LASSO as it considers interaction terms [31]. The R package MASS [32] is used to perform forward subset selection.

The performance of different methods is evaluated based on multiple criteria. The first criterion is the ability of a method to select true features and reject noise features. Accordingly, the number of target and noise features selected by a method is calculated. The second criterion is the prediction performance of a method. Root mean square error (RMSE) and R2 between the estimated outcome and actual outcome are used as evaluation parameters.

Table 4 shows that among the standard methods, LASSO, Adaptive LASSO and random LASSO showed similar performance in terms of feature selection. They had successfully identified the marginal variables with non-zero coefficients. Similar results were reported by [20] in their paper. Further, LASSO, adaptive LASSO and random LASSO did not select any non-zero marginal coefficients in all the scenarios. These methods outperformed the regression and group LASSO, which consistently selected the noise variables. However, other than Group LASSO, no other standard method was able to identify the interaction variables. HDSI with different selection techniques, i.e. HDSI_L, HDSI_AL and HDSI_R, outperformed the standard methods as they consistently selected the target variables while rejecting almost all noise variables. Therefore, it seems that HDSI might be able to leverage the performance of existing statistical feature selection methods.

Table 4. Feature selection performance of different approaches in simulated scenarios.

Scenario Performance Parameter (Number of Features Selected) Standard HDSI
LASSO Adaptive LASSO Group LASSO Random LASSO Regression HDSI_L HDSI_AL HDSI_R
1 Marginal (p = 25) 3 3 25 4 3 4 3 7
Target (s = 3) 3 3 3 3 3 3 3 3
Noise (s = 22) 0 0 22 1 0 1 0 4
Interaction (s = 300) 0 0 78 0 0 2 1 5
Target (s = 1) 0 0 1 0 0 1 1 1
Noise (s = 299) 0 0 77 0 0 1 0 4
Total Feature Selection (s = 4) 3 3 103 4 3 6 4 12
2 Marginal (p = 50) 3 3 50 3 5 3 3 7
Target (s = 3) 3 3 3 3 3 3 3 3
Noise (s = 47) 0 0 47 0 2 0 0 4
Interaction (s = 1225) 0 0 223 0 0 1 1 6
Target (s = 1) 0 0 1 0 0 1 1 1
Noise (s = 1224) 0 0 222 0 0 0 0 5
Total Feature Selection (s = 4) 3 3 273 3 5 4 4 13
3 Marginal (p = 100) 3 3 98 3 3 3 3 6
Target (s = 3) 3 3 3 3 3 3 3 3
Noise (s = 97) 0 0 95 0 0 0 0 3
Interaction (s = 4950) 0 0 263 0 0 1 1 6
Target (s = 1) 0 0 1 0 0 1 1 1
Noise (s = 4949) 0 0 262 0 0 0 0 5
Total Feature Selection (s = 4) 3 3 361 3 3 4 4 12
4 Marginal (p = 50) 1 1 50 1 1 3 3 24
Target (s = 3) 1 1 3 1 1 3 3 3
Noise (s = 47) 0 0 47 0 0 0 0 21
Interaction (s = 1225) 0 0 281 0 0 1 1 19
Target (s = 1) 0 0 1 0 0 1 1 1
Noise (s = 1224) 0 0 280 0 0 0 0 18
Total Feature Selection (s = 4) 1 1 331 1 1 4 4 43

Table 5 shows that outcome prediction performance of HDSI is better than the standard methods in all the scenarios. Thus, HDSI seems to be a better option as compared to standard methods in models with interaction terms having non-zero coefficients. Among the standard methods, only group LASSO performance can detect interaction terms and outperformed other standard methods. The performance of HDSI_L, HDSI_AL and HDSI_R in test dataset is coincidentally is identical. Further, HDSI could enable in expanding the modeling functionality of basic statistical approaches like simple linear regression to high dimensional settings if the number of target features is less than n. However, the main limitation is that HDSI process is computationally intensive as compared to standard methods.

Table 5. Outcome prediction performance of different approaches in simulated scenarios.

Scenario Performance Parameter (Outcome prediction) Standard HDSI
LASSO Adaptive LASSO Group LASSO Random LASSO Regression HDSI_L HDSI_AL HDSI_R
1 Test Dataset
RMSE 0.47 0.47 0.27 0.51 0.43 0.26 0.26 0.26
R2 0.72 0.71 0.90 0.71 0.72 0.90 0.90 0.90
2 Test Dataset
RMSE 0.41 0.40 0.28 0.50 0.42 0.25 0.25 0.25
R2 0.76 0.76 0.87 0.75 0.73 0.89 0.89 0.89
3 Test Dataset
RMSE 0.42 0.42 0.27 0.61 0.40 0.26 0.26 0.26
R2 0.75 0.75 0.88 0.72 0.74 0.89 0.89 0.89
4 Test Dataset
RMSE 0.41 0.40 0.29 0.48 0.42 0.25 0.25 0.26
R2 0.57 0.57 0.74 0.57 0.52 0.80 0.80 0.77

Real data studies

We implement the HDSI methods and compare them with other methods on five real-world datasets. The features in these studies can be textual, continuous or categorical with many missing values. For simplicity, the study only uses continuous features and remove features with a large number of missing values.

Dataset I is Community Health Status Indicators (CHSI) dataset (available at https://healthdata.gov/dataset/community-health-status-indicators-chsi-combat-obesity-heart-disease-and-cancer) which contains USA county-level data on various demographics and health parameters to help in making informed decisions in combating obesity, heart disease and cancer. The dataset contains data on 578 features for 3141 US counties. The final dataset has a sample size and feature size of 1156 and 55, respectively.

Dataset II and IV are National Social Life, health and Aging Project (NSHAP) datasets for Wave 3 (2015–2016) (available at https://www.icpsr.umich.edu/icpsrweb/NACDA/studies/ 36873) and Wave 1 (2005–2006) (available at https://www.icpsr.umich.edu/icpsrweb/NACDA/studies/20541), respectively. The datasets contain data of USA population related to health, social life and well being of older Americans. The Dataset II contains data on 1470 features for 4377 residents. The final dataset of Dataset II has a sample size and feature size of 1292 and 19, respectively. The Dataset IV contains data on 820 features for 3005 residents. The final dataset of Dataset IV has a sample size and feature size of 1511 and 27, respectively.

Dataset III is Study of Women’s Health Across the Nation (SWAN), 2006–2008 dataset (available at https://www.icpsr.umich.edu/icpsrweb/NACDA/studies/32961) which contains multi-site data for middle-aged women in USA on various physical, biological, psychological and social parameters. The dataset contains data on 887 features for 2245 respondents. The final dataset has a sample size and feature size of 1571 and 32, respectively.

Dataset V is Hawaii Aging with HIV Cardiovascular Study dataset (available at https://www.icpsr.umich.edu/icpsrweb/NAHDAP/studies/36389) which focus on determining the atherosclerosis development in HIV positive adults with age 40 and over residing in the Hawaii state, USA. The dataset has 248 features related to demographics and health indicators for a sample size of 110. The final dataset has a sample size and feature size of 104 and 21, respectively.

Table 6 provides a detailed summary of the five datasets. Each of the datasets is split into a training dataset (ntr) and a test dataset (nte). A large effect size is considered for estimating B for reduced computation time. Different methods are compared based on their prediction performance in the test dataset repeated over 30 trials.

Table 6. Summary of the real datasets.

Dataset Marginal Features (p) Outcome feature Sample size
Total (n) Train (ntr) Test (nte)
Dataset I 55 Percentage of unhealthy days 1156 925 231
Dataset II 19 Height 1292 1034 258
Dataset III 32 Body Mass Index 1571 1257 314
Dataset IV 26 Height 1511 1209 302
Dataset V 21 Framingham Risk Score 104 84 20

Tables 7 and 8 summarise the results of feature selection methods. Firstly, the HDSI methods have performed similar or better than the standard methods, which indicates that the HDSI methods have the potential to compete with existing methods for feature selection task. The RMSE of HDSI is less than or at par with the standard methods for all datasets. In terms of R-squared metric, the performance of proposed and standard methods is similar. Secondly, the HDSI methods have consistently identified the interaction terms. Among the standard methods, group LASSO has also identified the interactions terms, but a high RMSE and many interaction terms indicate overfitting. Thirdly, the performance of different techniques in HDSI varied with the datasets. The similar variation in performance of standard methods based on datasets is observed. HDSI gives a robust performance with different datasets while accommodating the variable performance of specific techniques.

Table 7. Feature selection performance of different methods on the real datasets.

Methods Dataset
I II III IV V
Marginal Features (μ (Range))
Standard LASSO 11 (5–30) 9 (8–11) 8 (7–10) 10 (7–18) 6 (3–10)
Adaptive LASSO 14 (6–31) 11 (10–13) 10 (8–12) 12 (8–17) 7 (4–11)
Group LASSO 53 (50–55) 19 (19–19) 7 (7–8) 26 (25–26) 20 (19–20)
Random LASSO 25 (19–31) 17 (14–19) 12 (8–14) 17 (14–21) 1 (0–2)
Regression 8 (6–13) 7 (6–10) 8 (7–9) 8 (6–11) 5 (3–6)
HDSI HDSI_L 9 (6–15) 2 (0–2) 7 (5–9) 11 (8–14) 1 (1–2)
HDSI_AL 8 (0–13) 1 (0–2) 6 (0–9) 10 (8–14) 1 (1–1)
HDSI_R 32 (24–38) 12 (9–14) 18 (12–24) 24 (19–26) 6 (2–12)
Interaction Features (μ (Range))
Standard LASSO 0 (0–0) 0 (0–0) 0 (0–0) 0 (0–0) 0 (0–0)
Adaptive LASSO 0 (0–0) 0 (0–0) 0 (0–0) 0 (0–0) 0 (0–0)
Group LASSO 255 (232–270) 89 (76–99) 7 (6–8) 156 (147–166) 58 (53–64)
Random LASSO 0 (0–0) 0 (0–0) 0 (0–0) 0 (0–0) 0 (0–0)
Regression 0 (0–0) 0 (0–0) 0 (0–0) 0 (0–0) 0 (0–0)
HDSI HDSI_L 3 (1–7) 1 (0–1) 3 (1–6) 3 (1–4) 0 (0–0)
HDSI_AL 3 (0–6) 1 (0–1) 2 (0–5) 3 (2–5) 0 (0–0)
HDSI_R 31 (18–44) 14 (9–17) 17 (11–22) 48 (29–61) 4 (1–8)

Table 8. RMSE performance of different methods on the real datasets for test data.

Methods Dataset
I II III IV V
RMSE (μ (95% CI))
Standard LASSO 0.92 (0.9–0.93) 3.85 (3.75–3.94) 0.51 (0.5–0.53) 3.57 (3.5–3.64) 0.06 (0.06–0.06)
Adaptive LASSO 0.91 (0.9–0.93) 3.83 (3.74–3.92) 0.51 (0.49–0.53) 3.57 (3.51–3.64) 0.06 (0.06–0.07)
Group LASSO 0.95 (0.93–0.96) 3.65 (3.56–3.74) 0.21 (0.2–0.22) 5.08 (3.85–6.31) 0.12 (0.09–0.14)
Random LASSO 0.98 (0.96–1.00) 3.93 (3.84–4.02) 1.08 (1.03–1.12) 3.89 (3.83–3.96) 0.07 (0.07–0.07)
Regression 0.91 (0.89–0.92) 3.74(3.65–3.83) 0.50 (0.48–0.52) 3.55(3.49–3.61) 0.06 (0.06–0.06)
HDSI HDSI_L 0.91 (0.89–0.92) 3.77 (3.67–3.87) 0.19 (0.15–0.24) 3.45 (3.38–3.51) 0.06 (0.06–0.07)
HDSI_AL 0.91 (0.90–0.93) 3.83 (3.73–3.92) 0.31 (0.23–0.40) 3.46 (3.4–3.53) 0.06 (0.06–0.07)
HDSI_R 0.90 (0.89–0.92) 3.6 (3.5–3.7) 0.12 (0.11–0.13) 6.77 (4.03–9.52) 0.07 (0.07–0.07)
R2 (μ (95% CI))
Standard LASSO 0.45 (0.44–0.47) 0.27 (0.25–0.29) 1 (0.99–1) 0.35 (0.34–0.37) 0.35 (0.31–0.4)
Adaptive LASSO 0.45 (0.44–0.47) 0.28 (0.26–0.29) 1 (0.99–1) 0.36 (0.34–0.37) 0.35 (0.31–0.4)
Group LASSO 0.44 (0.42–0.45) 0.34 (0.32–0.36) 1 (1–1) 0.3 (0.25–0.35) 0.18 (0.13–0.24)
Random LASSO 0.43 (0.42–0.45) 0.26 (0.25–0.28) 0.98 (0.98–0.98) 0.32 (0.31–0.34) 0.19 (0.15–0.24)
Regression 0.46 (0.44–0.48) 0.31 (0.29–0.32) 1 (1–1) 0.36 (0.34–0.37) 0.38 (0.33–0.43)
HDSI HDSI_L 0.46 (0.44–0.48) 0.29 (0.27–0.31) 1 (1–1) 0.39 (0.38–0.41) 0.13 (0.1–0.17)
HDSI_AL 0.46 (0.44–0.47) 0.27 (0.25–0.29) 1 (1–1) 0.39 (0.37–0.4) 0.16 (0.12–0.2)
HDSI_R 0.47 (0.45–0.49) 0.35 (0.34–0.37) 1 (1–1) 0.23 (0.18–0.28) 0.2 (0.14–0.25)

Conclusion and discussion

An innovative method, HDSI, is proposed to perform variable selection, including the interaction terms in high dimensional settings. The method is inspired by the random forest method provided by [16]. HDSI randomly samples both data and features, along with the incorporation of interaction terms. The method offers the flexibility of generating predictive models using existing modeling techniques available in the literature. The pooling of predictive models developed from different samples addresses the many limitations of shrinkage methods like LASSO and subset selection methods like forward selection. Firstly, it reduces the sample size restriction in feature selection. The current methodology segments the high dimensional feature space to low dimensional feature space to enable the application of classical statistical approaches on the high dimension feature set. Hence, the individual model is restricted by the number of main effects and interaction effects it can accommodate. However, HDSI as a whole is not restricted, since it pools results from multiple restricted models. Secondly, it enables more efficient selection and estimation of interaction terms from existing statistical modeling techniques like LASSO and ordinary least square regression. HDSI does not change the existing methods. Instead, it changes the ecosystem in which the standard methods operate, which enables them to operate on high dimensional data. Thirdly, it is less susceptible to multicollinearity issue. The sampling of feature space separates multicollinear features into different samples. Hence, multicollinear features could be more efficiently selected or removed from the model.

The simulation studies and real-world studies show that HDSI can outperform existing methods in the feature selection and consequently, the prediction performance. Further, the success of HDSI in real data settings demonstrates its practical relevance. Future research could focus on addressing some of the limitations of the current study. The main objective of the study is to explain and propose a method to handle interaction terms during feature selection in high dimensional settings. Hence, the current study has not extensively tested the HDSI with different types of datasets like temporal datasets, categorical outcomes and time to event outcomes and features like categorical features. Such evaluation of current method could determine the robustness of HDSI in real-world scenarios.

Another limitation is that the study has not tried integrating other types of statistical learning methods like glasso, decision trees, support vector machines, artificial neural network in HDSI framework. So, it could be an area of exploration to comprehensively determine the capability of HDSI in enhancing the capability of other techniques.

The HDSI model allows the use of different statistical techniques to build models. However, it may not be able to address all the limitations of any given technique. For instance, LASSO results may not be consistent across the bootstraps [33]. Future research could try to develop HDSI approaches which can deal with consistency limitations of techniques.

Data Availability

All real study dataset files are available from the ICPSR database (accession number(s) Dataset I: https://healthdata.gov/dataset/community-health-status-indicators-chsi-combat-obesity-heart-disease-and-cancer, Dataset II: https://www.icpsr.umich.edu/web/NACDA/studies/36873, Dataset III: https://www.icpsr.umich.edu/icpsrweb/NACDA/studies/32961, Dataset IV: https://www.icpsr.umich.edu/icpsrweb/NACDA/studies/20541, Dataset V: https://www.icpsr.umich.edu/icpsrweb/NAHDAP/studies/36389).

Funding Statement

The authors received the following funding support: Canadian Network for Research and Innovation in Machining Technology, Natural Sciences and Engineering Research Council of Canada, RGPIN-2017-06672, Dr Wei Xu; Prostate Cancer Canada, Dr Wei Xu.

References

  • 1.Fan J, Li R. Statistical challenges with high dimensionality: feature selection in knowledge discovery. Proceedings of the International Congress of Mathematicians Madrid, August 22–30, 2006. Madrid; 2007. pp. 595–622. [Google Scholar]
  • 2.Tavallali P, Razavi M, Brady S. A non-linear data mining parameter selection algorithm for continuous variables. PLoS One. 2017; e0187676 10.1371/journal.pone.0187676 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Walter S, Tiemeier H. Variable selection: Current practice in epidemiological studies. Eur J Epidemiol. 2009;24: 733–736. 10.1007/s10654-009-9411-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Heinze G, Wallisch C, Dunkler D. Variable selection–A review and recommendations for the practicing statistician. Biometrical J. 2018;60: 431–449. 10.1002/bimj.201700067 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Heinze G, Dunkler D. Five myths about variable selection. Transpl Int. 2017;30: 6–10. 10.1111/tri.12895 [DOI] [PubMed] [Google Scholar]
  • 6.Donoghue JR. Univariate Screening Measures for Cluster Analysis. Multivariate Behav Res. 1995;30: 385–427. 10.1207/s15327906mbr3003_5 [DOI] [PubMed] [Google Scholar]
  • 7.Desboulets LDD. A review on variable selection in regression analysis. Econometrics. 2018;6: 1–23. [Google Scholar]
  • 8.Morozova O, Levina O, Uusküla A, Heimer R. Comparison of subset selection methods in linear regression in the context of health-related quality of life and substance abuse in Russia. BMC Med Res Methodol. 2015;15: 1–17. 10.1186/1471-2288-15-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Emmert-streib F, Dehmer M. High-Dimensional LASSO-Based Computational Regression Models: Regularisation, Shrinkage, and Selection. Mach Learn Knowl Extr. 2019;1: 359–383. [Google Scholar]
  • 10.Zou H, Hastie T. Regularization and variable selection via the elastic net. J R Stat Soc Ser B Stat Methodol. 2005;67: 301–320. [Google Scholar]
  • 11.Mitchell TJ, Beauchamp JJ. Bayesian variable selection in linear regression. J Am Stat Assoc. 1988;83: 1023–1032. [Google Scholar]
  • 12.Zycinski G, Barla A, Squillario M, Sanavia T, Di Camillo B, Verri A. Knowledge Driven Variable Selection (KDVS)—a new approach to enrichment analysis of gene signatures obtained from high-throughput data. Source Code Biol Med. 2013;8: 1–14. 10.1186/1751-0473-8-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Liaw A, Wiener M. Classification and Regression by randomForest. R news. 2002;2: 18–22. [Google Scholar]
  • 14.Breiman L. Bagging Predictors. Mach Learn. 1996;140: 123–140. [Google Scholar]
  • 15.Ho TK. The Random Subspace Method for Constructing Decision Forests. IEEE Trans Pattern Anal Mach Learn. 1998;20: 832–844. [Google Scholar]
  • 16.Breiman L. Random Forests. Mach Learn. 2001;45: 5–32. [Google Scholar]
  • 17.Kotsiantis S, Kanellopoulos D. Combining bagging, boosting and random subspace ensembles for regression problems. Int J Innov Comput Inf Control. 2012;8: 3953–3961. [Google Scholar]
  • 18.Bach FR. Bolasso: Model Consistent Lasso Estimation through the Bootstrap. Proceedings of the 25th International Conference on Machine Learning. Helsinki; 2008. pp. 33–40.
  • 19.Teisseyre P, Kłopotek RA, Mielniczuk J. Random Subspace Method for high-dimensional regression with the R package regRSM. Comput Stat. 2016;31: 943–972. [Google Scholar]
  • 20.Wang BS, Nan B, Rosset S, Zhu J. Random lasso. Ann Appl Stat. 2011;5: 468–485. 10.1214/10-AOAS377 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Tibshirani R. Regression shrinkage and selection via the lasso. J R Stat Soc Ser B. 1996;58: 267–288. [Google Scholar]
  • 22.Bien J, Taylor J, Tibshirani R. A lasso for hierarchical interactions. Ann Stat. 2013;41: 1111–1141. 10.1214/13-AOS1096 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Ishwaran H. Variable importance in binary regression trees and forests. Electron J Stat. 2007;1: 519–537. [Google Scholar]
  • 24.Strobl C, Boulesteix AL, Zeileis A, Hothorn T. Bias in random forest variable importance measures: Illustrations, sources and a solution. BMC Bioinformatics. 2007;8 10.1186/1471-2105-8-25 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Basu S, Kumbier K, Brown JB, Yu B. Iterative random forests to discover predictive and stable high-order interactions. Proc Natl Acad Sci U S A. 2018;115: 1943–1948. 10.1073/pnas.1711236115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Zou H. The adaptive lasso and its oracle properties. J Am Stat Assoc. 2006;101: 1418–1429. [Google Scholar]
  • 27.Lehr R. Sixteen S‐squared over D‐squared: A relation for crude sample size estimates. Statistics in Medicine. 1992. pp. 1099–1102. 10.1002/sim.4780110811 [DOI] [PubMed] [Google Scholar]
  • 28.Cohen J. Statistical power analysis for the behavioral sciences 2nd ed. Hillsdale, New Jersey: Lawrence Earlbaum Associates; 1988. [Google Scholar]
  • 29.Friedman JH, Hastie T, Tibshirani R. Regularization Paths for Generalized Linear Models via Coordinate Descent. J Stat Softw. 2010;33: 1–22. [PMC free article] [PubMed] [Google Scholar]
  • 30.Lim M, Hastie T. glinternet: Learning Interactions via Hierarchical Group-Lasso Regularization. R Packag version 109. 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Lim M, Hastie TJ. Learning interactions through hierarchical group-lasso regularisation. J Comput Graph Stat. 2015;24: 627–654. 10.1080/10618600.2014.938812 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Venables WN, Ripley BD. Modern Applied Statistics with S. Fourth. New York: Springer; 2002. [Google Scholar]
  • 33.Knight K, Fu W. Asymptotics for Lasso-type estimators. Ann Stat. 2000;28: 1356–1378. [Google Scholar]

Decision Letter 0

Fabio Rapallo

14 Aug 2020

PONE-D-20-13743

HDSI: High Dimensional Selection with Interactions Algorithm on Feature Selection and Testing

PLOS ONE

Dear Dr. Xu,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the revision process. You have two detailed reports, with several suggestions on both the technical descriprion of your method and the presentation of the paper. Also please consider to make available the source code.Please submit your revised manuscript by Sep 28 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Fabio Rapallo, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: N/A

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors proposed a resampling-based method to facilitate feature selection and prediction for high-dimensional data. Technically, the proposed method is an ensemble method and more like a framework, which can be flexible to incorporate any machine learning methods with feature selection including lasso, adaptive lasso and elastic net. Application efficiency should be the key point for this method. However, the paper lacks extensive evaluation and the method itself needs some parameters which are hard or heuristic to determine in real application. More efforts are needed to demonstrate the effectiveness in simulations and real applications. Below are my itemized comments.

Major:

1: In table 3 and table 6, the result of random lasso is far worse than other methods including regular lasso. Random lasso uses two-step bootstrapping to get better estimation with the sacrifice of computing time. It is hard to believe that the performance of random lasso is so poor. Many issues may be reconsidered. a): In the HDSI algorithm, it generates the interaction effect after subspace sampling. That is, after the selection of q features for a bootstrapped sample, all the possible interaction terms among the q features are created. I wonder whether you conduct this procedure similarly to random lasso. In random lasso, you should also generate interaction terms after subspace sampling q features. b): The algorithm in random lasso is quite straightforward and you may directly write your own code to implement it if there is no available R package. The author of the github page you mention is not the original author of random lasso. I doubt the correctness of this github code especially the author mentioned that the code is unable to run unless modification.

2: The number of features selected in each bootstrapped sample, q, is a crucial parameter. In the simulation and real applications, the author directly fixed it at 5 or 15, which is not an appropriate way. The q must be evaluated by cross validation for each simulated data or real application.

3: I believe the number of bootstrapped datasets is not a crucial parameter and there is no value added to estimate the best number of bootstrapped datasets. In random forest, there will be no overfitting along with the increase of number of trees. If computing power allows, the user should choose an adequate large number of trees for the random forest to be stable. Here for HDSI, the concept is similar, and I suggest there is no need to estimate this parameter. If the authors insist to design an algorithm to estimate the appropriate number of bootstrapped datasets, at least simulation studies are needed to show that this parameter indeed has real impact to the performance.

4: For the feature selection part, authors proposed to use three criteria, quantile, confidence interval and R^2 simultaneously, which induce too many parameters which needs to be specified in real applications. The guidance and extensive evaluation should be given.

Minor:

1: Line 160: boostrapped->bootstrapped

2: Please put i=1….B behind min or max in all the notations like min(Rsq_ij) or max(Rsq_ij).

3: The form between line 147 and 148 looks not very nice. Please reformat and polish this form. Also, the form between line 184 and 185.

Reviewer #2: (Since PLoS One does not allow reviewers to use a compiled PDF file as the only 'comments for the authors', I am pasting the exported text below and also attaching the compiled PDF file for the authors' convenience, because some symbols will be surely rendered incorrectly in text-only format.)

The authors have proposed a new method (HDSI) for handling interaction effects in high- dimensional regression by building on the idea of random lasso but making substantial changes to the implementation. Overall, I think it is a new method that attempts to solve an important problem and illustrates its performance through multiple real data analysis and one simulation study.

In what follows, I will restrict my review to technical details and correctness and presentation. Firstly, I think the presentation is a bit pedestrian and could have been better. In particular,

when writing the algorithmic version of HDSI. Also, the motivation can be improved, the benefits of using the proposed method and the issues with random Lasso should be emphasized more in the introduction.

My comments on the paper are below, not necessarily in the order of importance.

1.Since the main contribution of this paper is proposing a new method for handling high- dimensional regression that relies on many tuning parameters, it would be useful to supply the codes for implementing the proposed method in a simulated or real data-set, either through a public repository such as github or an R package. This will also ensure reproducibility and greater use of the method by the community at large.

2.Lasso was proposed in 1996 and since then there have been many, many variants of the Lasso, including extensions for more complicated structures or non-convex penalties or more recent debiased Lasso or LAVA. It seems some discussion of the literature on sparsity is warranted. See Tibshirani [2014].

3.The main motivation for HDSI (or random lasso) is that Lasso can not select more than n fea- tures in large p, small n problems, which is resolved by methods like Elastic net. This brings a natural question, why not compare with elastic net as well, or more advanced regularization methods like de-biased Lasso?

4.Page 13: q2 is a bad notation. Maybe make this q(2)?

5.page 13, line 152: what is meant by ‘little flexibility in coefficient estimation’? Can you elaborate or add a reference?

6.page 14, line 174: same as before? Either explain why random Lasso under-estimates or provide a reference. As a matter of fact, Lasso has a non-vanishing bias in the tails [Carvalho et al., 2009] but it’s not always ‘under’-estimation.

7.Algorithm 2: How do you make sure that the 2q is less than n? It seems this poses a re- striction on how many main effects and interactions can be considered in the model and the simulation study conducted has the problem set to avoid situations where the size of model with interaction terms is larger than the sample size.

8.Algorithm 2: Too many terms that are explained / defined much later in the text, e.g. L, bj,

Rsqij. Please define them before using them in an algorithm.

9.Also comment on Algo 2: Lasso-based CI’s are not reliable [Chatterjee & Lahiri, 2011].

10.page 17, line 221: if you are calculating bootstrap-based CI’s, why care about Normal distri- bution of βˆj ? In fact, bootstrap works better when the statistics under consideration is either non-Gaussian or non-linear.

11.page 18, line 242: R2 does not follow a Normal distribution, might not even be symmetric.

Why use the µ ± f × σ approach? Also, how do you choose f ? (Please do not use ∗ for denoting a product, if necessary use ×.)

12.As explained before, the simulation study is designed in a way where the number of true non- zero β’s and interaction terms are very small (2-4), and it is not clear how that is against the main motivation (limitation of Lasso to select ≤ n coefficients).

13.Among the methods compared (e.g. Table 2), did the regression or lasso/random lasso have any interaction terms in the model at all? It seems only the group lasso had a provision for including interaction terms. If that’s the case, most of the comparisons are meaningless.

14.On Table 2, n is used to denote the number of non-zero coefficients. Statisticians use n for sample size and I think even this paper has n before to denote sample size. Please consider using a different notation.

15.Finally, a minor point: the separation/dichotomy between Statistics-based and ML-based method (Lasso vs Random Forest) is artificial (at least to this reviewer). One can club them both as supervised learning methods or more generally, statistical learning.

References

CARVALHO, C. M., POLSON, N. G. & SCOTT, J. G. (2009). Handling sparsity via the horseshoe.

Journal of Machine Learning Research W&CP 5, 73–80.

CHATTERJEE, A. & LAHIRI, S. N. (2011). Bootstrapping lasso estimators. Journal of the American Statistical Association 106, 608–625.

TIBSHIRANI, R. J. (2014). In praise of sparsity and convexity. Past, Present, and Future of Statis- tical Science , 497–505.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Yujia Li

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: PLoS_one_HDSI.pdf

PLoS One. 2021 Feb 16;16(2):e0246159. doi: 10.1371/journal.pone.0246159.r002

Author response to Decision Letter 0


27 Sep 2020

We would like to thank the reviewers for the valuable suggestions and comments which have significantly improved the content and presentation of the paper. We have addressed the specific comments in the point by point response letter.

Attachment

Submitted filename: HDSI_Response_1.0_WX_RJ.docx

Decision Letter 1

Fabio Rapallo

2 Nov 2020

PONE-D-20-13743R1

HDSI: High Dimensional Selection with Interactions Algorithm on Feature Selection and Testing

PLOS ONE

Dear Dr. Xu,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Both reviewers acknowledge major improvements with respect to the previous version, but some comments still needs the authors' attention. In particular, you should address the comment(s) about the hyperparameters. 

Please submit your revised manuscript by Dec 17 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Fabio Rapallo, Ph.D.

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I am appreciative of the expanded comments, but dismayed that the investigators did not take into account the rather substantial and insightful comments from both reviewers.

There are some main examples to this:

1: In section “Hyperparameters”, authors only mentioned the need of optimizing the hyperparameters without a practical guidance. There are three parameters needed to be determined: $q$, $Q_i$ and $R_f$, Line 336-340 mentions the hyperparameters are obtained by five-fold CV and heuristic optimization. Could authors state clearly what heuristic procedure they did? In real data (Line 416), authors mention the three parameters are selected by grid search. I am thinking an exhaustive cross validation of the combination of the three parameters seems overwhelming. I suggest authors can clearly state how they select three parameters in “Hyperparameters” Section.

2: Following 1, I suggest authors do some sensitivity analysis to the hyperparameters, giving a practical region of each hyperparameter (i.e. $Q_i$ between 0.05-0.2 works well). With this, users can quickly pick a suitable parameter sets to explore. Also, it helps to reduce the computational load if users want to do cross validation over combinations of three hyperparameters (the number of combinations can be largely reduced).

3: Authors should spend more time to proofread. For example, line 39, p and n should use mathematical format: $p$ and $n$. Line 228: Min_RSq_j is not professional. Line 395, the space is too large and looks strange. There are many flaws like this and authors should proofread carefully and make the writing more professional.

4: Following 3, I feel that many equations can fit into the text. There are 16 equations numbered and most of them are not the key part of this manuscript and it is better to fit them into the text.

5: For number of bootstrapped datasets ($B$), my suggestion is to do a sensitivity analysis using simulation (i.e. B=20, 50, 100, 200, 500, 1000, 2000). At the beginning (B=20), the result will be unstable, but at certain point, the result will be stable (for example, B=500). This sensitivity analysis will be in complementary with the rule of thumb authors proposed.

Reviewer #2: I think almost all the comments that I made on the earlier version were addressed in the current draft. I am still a little bothered by the fact that the current method is essentially bounded by limitation of Lasso of selecting p <= n variables inside the HDSI algorithm. It seems using other variable selection methods might bypass this limitation. I have just a couple of minor suggestions.

a) Include the answer to Q. 7 in the main manuscript, that is, the segment: "The current methodology segments the high dimensional feature space to low dimensional feature space to enable the application of classical statistical approaches on the high dimension feature set. Hence, the individual model is restricted by the number of main effects and interaction effects that it can accommodate. However, HDSI as a whole is not restricted, since it pools results from multiple restricted models."

b) On HDSI algorithm, change "feature technique" to "feature selection technique".

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Feb 16;16(2):e0246159. doi: 10.1371/journal.pone.0246159.r004

Author response to Decision Letter 1


11 Dec 2020

We are very grateful to you for your effort in handling our paper entitled “HDSI: High Dimensional Selection with Interactions Algorithm on Feature Selection and Testing” [PONE-D-20-13743R1]. With the kind help from you, the Academic Editor Dr. Fabio Rapallo, and two reviewers, the paper has been carefully revised and all the comments are fully dealt with. We have submitted the revised manuscript in the system, along with this cover letter, the response letter to the two reviewers and a document to track changes in the manuscript. After revising the paper by taking all comments from you, the Academic Editor and two reviewers into account, we think that it might be suitable for consideration for publication in PLOS One.

Attachment

Submitted filename: HDSI_Response_2.0_WX_RJ.docx

Decision Letter 2

Fabio Rapallo

15 Jan 2021

HDSI: High Dimensional Selection with Interactions Algorithm on Feature Selection and Testing

PONE-D-20-13743R2

Dear Dr. Xu,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Fabio Rapallo, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors have now addressed all the questions/concerns. These additions/modifications have improved the manuscript.

Reviewer #2: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Acceptance letter

Fabio Rapallo

3 Feb 2021

PONE-D-20-13743R2

HDSI: High Dimensional Selection with Interactions Algorithm on Feature Selection and Testing

Dear Dr. Xu:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Fabio Rapallo

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: PLoS_one_HDSI.pdf

    Attachment

    Submitted filename: HDSI_Response_1.0_WX_RJ.docx

    Attachment

    Submitted filename: HDSI_Response_2.0_WX_RJ.docx

    Data Availability Statement

    All real study dataset files are available from the ICPSR database (accession number(s) Dataset I: https://healthdata.gov/dataset/community-health-status-indicators-chsi-combat-obesity-heart-disease-and-cancer, Dataset II: https://www.icpsr.umich.edu/web/NACDA/studies/36873, Dataset III: https://www.icpsr.umich.edu/icpsrweb/NACDA/studies/32961, Dataset IV: https://www.icpsr.umich.edu/icpsrweb/NACDA/studies/20541, Dataset V: https://www.icpsr.umich.edu/icpsrweb/NAHDAP/studies/36389).


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES