Next Article in Journal
Robust Higher-Order Nonsingular Terminal Sliding Mode Control of Unknown Nonlinear Dynamic Systems
Previous Article in Journal
Enhanced ResNet50 for Diabetic Retinopathy Classification: External Attention and Modified Residual Branch
Previous Article in Special Issue
Applying the Laplace Transform Procedure, Testing Exponentiality against the NBRUmgf Class
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sure Independence Screening for Ultrahigh-Dimensional Additive Model with Multivariate Response

1
School of Statistics, Capital University of Economics and Business, Beijing 100070, China
2
Department of Biostatistics, School of Public Health, Peking University, Beijing 100191, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(10), 1558; https://doi.org/10.3390/math13101558
Submission received: 30 March 2025 / Revised: 2 May 2025 / Accepted: 6 May 2025 / Published: 9 May 2025
(This article belongs to the Special Issue Parametric and Nonparametric Statistics: From Theory to Applications)

Abstract

:
This paper investigated an ultrahigh-dimensional feature screening approach for additive models with multivariate responses. We proposed a nonparametric screening procedure based on random vector correlations between each predictor and multivariate response, and we established the theoretical results of sure screening and ranking consistency properties under regularity conditions. We also developed an iterative sure independence screening algorithm for convenient and efficient implementation. Extensive finite-sample simulations and a real data example demonstrate the superiority of the proposed procedure over 58–100% of existing candidates. On average, the proposed method outperforms 79% of existing methods across all scenarios considered.

1. Introduction

Many scientific fields work with ultrahigh-dimensional data, where the data dimensionality far exceeds the sample size. Examples include microarray analysis, tumor classification, and biomedical imaging. Analyzing such data presents challenges in terms of computational expediency, statistical accuracy, and algorithmic stability, as pointed out in Fan and Lv [1], Fan et al. [2], Li et al. [3] and Liu et al. [4].
Several methods have been proposed to tackle the challenges associated with handling ultrahigh-dimensional data. Fan and Lv [1] introduced the concept of sure screening, which screens for a property that all the important variables survive after applying a variable screening procedure with probability tending to 1, and they proposed the sure independence screening (SIS) method based on the Pearson correlation for Gaussian predictors and responses in linear models. They also developed an iterative sure independence screening (ISIS) method to improve its effectiveness. Fan and Song [5] extended SIS and ISIS to generalized linear models by ranking the maximum marginal likelihood estimates. Zhu et al. [6] proposed a sure-independent ranking and screening (SIRS) procedure for selecting active predictors in multi-index models. Fan et al. [7] extended correlation learning to marginal nonparametric learning and introduced a nonparametric independence screening (NIS) procedure for sparse additive models, utilizing B-spline expansion for approximating the nonparametric projections. They also developed an iterative nonparametric independence screening (INIS) procedure to improve the performance of NIS in terms of false positive rates. Yuan and Billor [8] proposed a functional version of SIS (functional SIS) by applying the SIS method to functional regression models with a scalar response and functional predictors. Cui et al. [9] introduced a modified SIS (EVSIS) to select important predictors in ultrahigh-dimensional linear models with measurement errors.
The methods mentioned above are model-based. On the other hand, there are model-free feature screening procedures that use the measure of independence. For example, Li et al. [10] developed a sure independence screening procedure based on distance correlation (DC-SIS). Shao and Zhang [11] employed martingale difference correlation to construct a screening procedure (MDC-SIS). Cui et al. [12] proposed a screening method (MV-SIS) using a mean–variance index for ultrahigh-dimensional discriminant analysis. Chen and Lu [13] proposed a quantile-composited feature screening (QCS) procedure for ultrahigh-dimensional grouped data by transforming the continuous predictor to a Bernoulli variable. Liu et al. [14] introduced a screening method based on the proposed generalized Jaccard coefficient (GJAC-SIS) in ultrahigh-dimensional survival data. Sang and Dang [15] utilized the Gini distance correlation between a group of covariates with responses to establish a grouped screening method for discriminant analysis in ultrahigh-dimensional data. Wu and Cui [16] developed feature screening procedures on the basis of the Hellinger distance (HD-SIS and AHD-SIS), which are applicable to both discrete and continuous response variables. Zhong et al. [17] offered a groupwise feature screening procedure via the semi-distance correlation for ultrahigh-dimensional classification problems. Tian et al. [18] proposed a conditional mean independence screening procedure via variations in the conditional mean.
Out of the screening methods mentioned earlier, only DC-SIS can handle multivariate responses. However, there is a growing need for ultrahigh-dimensional feature selection for multivariate response in applications such as pathway analysis. If we overlook the correlation among the multivariate response and only apply the screening procedure to each component of the response, we may miss some important predictors. As a result, Li et al. [3] developed a projection screening procedure to select variables for linear models with multivariate response and ultrahigh-dimensional covariates. Using the generalized correlation between each predictor and multivariate response, Liu et al. [4] proposed a screening method form ultrahigh-dimensional additive models (GCPS and IGCPS). Li et al. [19] generalized the martingale difference correlation (GMDC) to measure the conditional mean independence and conditional quantile independence, and they utilized it as a marginal utility to develop high-dimensional variable screening (GMDC-SIS and quantile GMDC-SIS) procedures.
Due to the lack of a priori information about the model structure, a more flexible class of nonparametric models, such as the additive model, can be used to promote the flexibility of models. Some studies on additive model with multivariate response are listed as follows. Liu et al. [20] proposed a novel approach for multivariate additive models by treating the coefficients in the spline expansions as a third-order tensor and combined dimension reduction with penalized variable selection to deal with ultrahigh dimensionality. They applied the proposed method to a breast cancer study and selected several important genes that have significant associations between the measured genome copy numbers and the level of gene transcripts. Desai et al. [21] introduced a new method for the simultaneous selection and estimation of multivariate sparse additive models with correlated errors. The method was utilized to investigate the protein–mRNA expression levels, and it helped to explore and establish nonlinear mRNA–protein associations.
Therefore, in this study, we focus on an additive model for sparse ultrahigh-dimensional data with multivariate response. A nonparametric screening procedure is proposed to measure the relationship between the multivariate response and each predictor simultaneously in the additive model. In particular, we obtain a normalized B-spline basis for each predictor and then compute the random vector correlation between the multivariate response and this basis. The predictors are then ranked based on the random vector correlation. The proposed method has sure screening and ranking consistency properties under some regularity conditions. Furthermore, to address the issue of the marginal method selecting unimportant predictors, which is due to their high correlations with important ones, an iterative sure independence screening method is also developed to improve the finite sample’s performance.
The remainder of this paper is organized as follows. In Section 2, we proposed sure independence screening based on the random vector correlation (RVC-SIS) procedure for sparse additive models with multivariate response and study its theoretical properties. Section 3 presents an iterative sure independence screening (RVC-ISIS) procedure to enhance the performance of RVC-SIS. In Section 4, we conduct numerical studies and real data analysis to demonstrate the new procedure’s finite sample performance in comparison to some existing methods. Section 5 provides an application example. Finally, we discuss the performance and limitations of the proposed method in Section 6 and conclude in Section 7. The technical proofs are presented in Appendix A.

2. Sure Independence Screening Using Random Vector Correlation

Suppose Y = ( Y 1 , , Y q ) R q is from the following additive model:
Y = μ + j = 1 p f j ( X j ) + ε ,
where X = ( X 1 , , X p ) R p is a predictor vector, f j ( · ) = ( f 1 j ( · ) , , f q j ( · ) ) denotes the unknown vector functions, ε = ( ε 1 , , ε q ) is the random error with the zero mean, and μ is a q-vector intercept term. Note that f l j is a nonparametric smoothing functions; we use a cubic B-spline parameterization method to approximate them. Specifically, let { B j k , k = 1 , , d n } be a normalized B-spline basis with B j k 1 , where d n is the number of the basis function, and · is the sup norm. Then, f l j ( X j ) can be well approximated by
f n l j ( X j ) = k = 1 d n β l j k B j k ( X j ) = B j β l j ,
with β l j = ( β l j 1 , , β l j d n ) and B j = ( B j 1 ( X j ) , , B j d n ( X j ) ) , which motivates us to utilize the random vector correlation (RVC hereafter) defined by Escoufier [22] as a marginal utility screening index:
ω j = tr ( Σ Y B j Σ B j Y ) tr ( Σ Y Y 2 ) tr ( Σ B j B j 2 ) = : V ( Y , B j ) V ( Y , Y ) V ( B j , B j ) ,
where Σ Y B j is the population covariance matrix between Y and B j , Σ Y Y 2 is the square covariance matrix of Y and Σ B j B j 2 and the squared covariance matrix of B j , and tr denotes the trace operator. Intuitively, if Y and X j are independent, then ω j = 0 . Otherwise, ω j would be positive. This desired property allows us to use ω j to propose a feature screening method for additive models with multivariate responses.
Suppose that X i = ( X i 1 , , X i p ) and Y i = ( Y i 1 , , Y i q ) , 1 i n are the samples from X and Y , respectively. Then, the sample version of B j is B j i = ( B j 1 ( X i j ) , , B j d n ( X i j ) ) . An unbiased estimator of V ( Y , B j ) is given by
V ^ n ( Y , B j ) = 1 n ( n 3 ) i k A i k B i k ,
where A i k and B i k are
A i k = a i k 1 n 2 l = 1 n a i l 1 n 2 s = 1 n a s k + 1 ( n 1 ) ( n 2 ) s , l = 1 n a s l , B i k = b i k 1 n 2 l = 1 n b i l 1 n 2 s = 1 n b s k + 1 ( n 1 ) ( n 2 ) s , l = 1 n b s l ,
with a i k = B j i B j k 2 / 2 and b i k = Y i Y k 2 / 2 . The sample RVC between Y and B j is defined by
ω ^ j = V ^ n ( Y , B j ) V ^ n ( Y , Y ) V ^ n ( B j , B j ) , j = 1 , , p .
Thus, a sure independence screening procedure based on a random vector can be carried out according to { ω ^ j } j = 1 p .
To study the sure screening property of the RVC-SIS procedure, we define the subset of the truly important predictors as
A = { j : E ( f j ( X j ) ) > 0 , 1 j p } .
The following conditions are imposed to facilitate the proof of the properties:
(C1)
The distribution of X is absolutely continuous, and its density is bounded away from zero and infinity on [ a , b ] p , where a and b are finite real numbers.
(C2)
The nonparametric components { f l j , 1 j p , 1 l q } belong to F , where the rth derivative, f ( r ) , exists and satisfies the Lipschitz Condition with exponent α : | f ( r ) ( t ) f ( r ) ( t ) | K | t t | α for t , t [ a , b ] for some positive constant K. Here, r is a non-negative integer and α ( 0 , 1 ] such that d = r + α > 0.5 .
(C3)
j A f l j is bound: l = 1 , , q .
(C4)
The random error vector ε satisfies E { exp ( t ε 2 ) } < for all 0 < t 2 t 0 .
(C5)
min j A max 1 l q E [ { E ( Y l | X j ) } 2 ] c 1 d n n 2 κ for some 0 < κ < d / ( 2 d + 1 ) and c 1 > 0 .
(C6)
There exist ξ ( 0 , 1 ) and c > 0 such that d n 2 d 1 c 1 ( 1 ξ ) n 2 κ / C 1 and 2 c < c 1 D 1 ξ / D 2 V ( Y , Y ) . Here, C 1 , D 1 , and D 2 are given in Fact 1 and Fact 3 of [7].
Based on Condition (C1), we can infer that the marginal density of X j is bounded on the interval [ a , b ] , away from zero and infinity. The Lipschitz Condition (C2) requires the function to be sufficiently smooth to determine the convergence rate for its spline estimates. Condition (C4) controls the tail distribution of the random error vector to ensure the sure screening property. Condition ( C 5 ) requires that the signal of the important predictors is not too weak. Condition (C6) is imposed to show the good approximation of f n l j ( X j ) relative to f l j ( X j ) . Together with Conditions (C4) and (C6), Condition (C3) ensures uniform control in the probability inequality. These conditions are given in light of Li et al. [3], Liu et al. [4], Fan et al. [7] and the reference therein.
Note that it will be shown that
min j A ω j 2 c d n 1 / 2 n 2 κ
under the conditions aforementioned. We thus select a set of predictors
A ^ = { j : ω ^ j c d n 1 / 2 n 2 κ , j = 1 , , p }
from a theoretical perspective. In addition, there exists an efficient way to choose a subset of variables A ^ at a practical application level. To be specific, we sort { ω ^ j } j = 1 p in descending order and pick out the first and largest [ n / log ( n ) ] to form A ^ , where [ a ] denotes the integer part of a.
Theorem 1
(Sure Screening Property). Under Conditions (C1)–(C4), for 0 < γ < 1 / 2 2 κ and c 2 > 0 , there exist some positive constants c 3 and c 4 such that
Pr ( max 1 j p | ω ^ j ω j | c 2 d n 1 / 2 n 2 κ ) O ( p [ exp { c 3 n 1 4 κ 2 γ d n 2 } + n exp ( c 4 n γ ) ] ) .
Along with Conditions (C5)–(C6), we have that
Pr ( A A ^ ) 1 O ( s n [ exp { c 3 n 1 4 κ 2 γ d n 2 } + n exp ( c 4 n γ ) ] ) ,
where s n = | A | is the cardinality of A .
Theorem 1 states that the proposed procedure is capable of handling the NP-dimensionality log p = o ( n 1 4 κ 2 γ d n 2 + n γ ) . To balance the two terms on the right side of (1), we have n 1 4 κ 2 γ d n 2 = n γ . As a result, (1) can be expressed as
Pr ( max 1 j p | ω ^ j ω j | c 2 d n 1 / 2 n 2 κ ) O ( p [ exp { c 3 n ( 1 4 κ ) / 3 d n 2 / 3 } ] ) .
This implies that we can handle the NP-dimensionality log p = o ( n ( 1 4 κ ) / 3 d n 2 / 3 ) , indicating that dimensionality depends on the the minimum true signal strength and the number of basis functions.
The proposed procedure enjoys the ranking consistency property.
Theorem 2
(Ranking Consistency Property). If Conditions (C1)–(C6) hold, then we have
lim inf n { min j A ω ^ j max j A c ω ^ j } > 0 in probability .
Compared to the sure screening property, Theorem 2 demonstrates a stronger theoretical result. It indicates, with overwhelming probability, that the truly active predictors tend to have larger magnitudes of ω ^ j than the inactive ones. In other words, the active predictors can be ranked at the top by our screening method.

3. Iterative Sure Independence Screening Using Random Vector Correlation

Three main challenges identified by Fan and Lv [1] may also arise with RVC-SIS. Firstly, RVC-SIS may prioritize unimportant predictors that are highly correlated with important predictors over other important predictors that are relatively weakly related to the response. Secondly, RVC-SIS may fail to select an important predictor that is marginally uncorrelated but jointly correlated with the response, leading to its exclusion from the estimated model. Finally, all the important predictors may not be screened out by RVC-SIS due to the collinearity among predictors. To address these three challenges, we use the joint information of the covariates more comprehensively in variable selection rather than solely relying on marginal information, to develop the RVC-ISIS method that is summarized in Algorithm 1.
Algorithm 1 RVC-ISIS Algorithm
0:
Start with an empty important set A ^ = .
1:
First, use RVC-SIS to select d 1 predictors with the first d 1 largest ω ^ j values.
   Add the indices of these selected predictors to A ^ .
2:
Calculate X n e w = { I n X A ^ ( X A ^ X A ^ ) 1 X A ^ } X A ^ c .
3:
Conduct RVC-SIS procedure for Y and X n e w . Choose d 2 predictors in the same way as Step 1 and update A ^ by adding the indices of these selected d 2 predictors.
4:
Repeat Step 2 and Step 3 until the total number of selected predictors reaches a predetermined value, such as | A ^ | = [ n / log ( n ) ] .
5:
Output the final indices of the selected predictors A ^ .
RVC-ISIS focuses on Step 3, which operates in two ways: Firstly, when an active variable is correlated with other active variables but marginally independent of the response, Step 3 helps make it relevant to the response, and then, it becomes marginally detectable. Secondly, when many irrelevant variables are highly correlated with the active variables, Step 3 utilizes the correlated active variables selected in Step 1 to make the rest of the active variables easier to detect, rather than detecting the irrelevant variables.

4. Numerical Studies

In this section, we evaluate the finite-sample performance of our proposed screening procedure through a series of simulation studies and a real data example. We compare its performance with Naive-SIS1, Naive-ISIS1, Naive-SIS2, Naive-ISIS2, PS, and IPS in Li et al. [3]; GCPS and IGCPS in [4]; DC-SIS in Li et al. [10]; and DC-ISIS in Zhong and Zhu [23]. Our simulation studies are conducted using R version 4.1.2. We set the total number of predictors as p = 2000 and the sample size as n = 400 . For our proposed method, we take [ n 1 / 5 ] as the number of internal knots. We adopt the criteria from Liu et al. [24] to assess the performance of the proposed method:
(i)
R j : The average rank of X j in terms of the sorted list via the screening procedure based on 1000 replications. The higher the ranking of X j , the greater the probability of being selected.
(ii)
M: The minimum model size to include all active predictors. In other words, M represents the largest rank of the true predictors: M = max j A R j , where A is the true model. We report the 5 % , 25 % , 50 % , 75 % , and 95 % quantiles of M from 1000 repetitions.
(iii)
P j : The proportion of important predictor X j , j A being selected for a given model size [ n / log ( n ) ] in the 1000 replications.
(iv)
P a : The proportion of all active predictors being selected into the submodel with size [ n / log ( n ) ] over 1000 simulations.
Example 1.
Following Li et al. [3], we consider a linear model with multivariate responses as follows.
Y 1 = β 11 X 1 + β 12 X 2 + + β 15 X 5 + ε 1 , Y q n = β q n , 1 X 1 + β q n , 2 X 2 + + β q n , 5 X 5 + ε q n .
The dimension of the response is q n = [ n α ] , α U ( 0.2 , 0.9 ) . The truly active predictors set is M = { X 1 , , X 5 } . X = ( X 1 , , X p ) is drawn from a multivariate normal distribution N ( 0 p × 1 , Σ ) , where Σ is a p × p covariance matrix with elements σ i j = 0 . 8 | i j | . The nonzero coefficient is given by β k j = ( 1 ) W k j ( log ( n ) + | Z k j | ) , k = 1 , , q n , j = 1 , , 5 , where W k j is a binary random variable with Pr ( W k j = 1 ) = 0.6 and Z k j N ( 0 , 1 ) . The random error term ε k follows the standard normal distribution N ( 0 , 1 ) .
Example 2.
We apply the screening procedures to the additive model when the components of the multivariate response are strongly correlated:
Y 1 = 5 f 1 ( X 1 ) + 8 f 2 ( X 2 ) + 4 f 3 ( X 3 ) + 6 f 4 ( X 4 ) + ε 1 , Y 2 = 4 f 1 ( X 1 ) + 6 f 2 ( X 2 ) + 5 f 3 ( X 3 ) + 8 f 4 ( X 4 ) + ε 2 , Y 3 = 6 f 1 ( X 1 ) + 4 f 2 ( X 2 ) + 3 f 3 ( X 3 ) + 5 f 4 ( X 4 ) + ε 3 ,
with
f 1 ( x ) = exp ( sin ( π x ) ) , f 2 ( x ) = exp ( x ) ( 2 x 1 ) 2 , f 3 ( x ) = sin ( 2 π x ) 2 sin ( 2 π x )
and
f 4 ( x ) = 0.1 sin ( 2 π x ) + 0.2 cos ( 2 π x ) + 0.3 sin 2 ( 2 π x ) + 0.4 cos 3 ( 2 π x ) + 0.5 sin 3 ( 2 π x ) .
The predictor X = ( X 1 , , X p ) is generated by
X j = W j + U 2 , j = 1 , , p ,
where W 1 , , W p and U are i.i.d. U(0,1). The truly active predictors set is M = { X 1 , , X 4 } . The random error term ε k is from the standard normal distribution N ( 0 , 1 ) .
Example 3.
The following model is taken from Fan and Lv [1] and Li et al. [3].
Y 1 = 5 X 1 + 5 X 2 + 5 X 3 15 X 4 0.5 + ε 1 , Y 2 = 4 X 1 + 6 X 2 + 8 X 3 18 X 4 0.5 + ε 2 , Y 3 = 5 X 1 + 4 X 2 + 6 X 3 15 X 4 0.5 + ε 3 .
X = ( X 1 , , X p ) follows a multivariate normal distribution N ( 0 p × 1 , Σ ) . The covariance matrix Σ satisfies σ 11 = = σ p p = 1 and the off-diagonal elements σ i j = 0.5 , i j , except for σ 4 j = σ j 4 = 0.5 , j 4 . The random error term ε k follows a standard normal distribution N ( 0 , 1 ) . The set of truly active predictors is M = { X 1 , , X 4 } .
Example 4.
We adopt the simulation model in [1,3] as follows.
Y 1 = 5 X 1 + 5 X 2 + 5 X 3 15 X 4 0.5 + 1.25 X 5 + ε 1 , Y 2 = 4 X 1 + 6 X 2 + 8 X 3 18 X 4 0.5 + 1.5 X 5 + ε 2 , Y 3 = 5 X 1 + 4 X 2 + 6 X 3 15 X 4 0.5 + 1.25 X 5 + ε 3 .
The covariance matrix Σ is the same as in Example 3, except that σ 5 j = σ j 5 = 0 , j 5 . X 5 is uncorrelated with any X j , j 5 . The random error term ε k is from the standard normal distribution N ( 0 , 1 ) .
In Example 1, the model is linear, and the response is multivariate. All methods perform well under this scenario, as shown in Table 1, Table 2 and Table 3.
In Example 2, the model is nonlinear. Each component of the multivariate response has the same nonlinear function of the active predictors. The only difference is the corresponding coefficients. It is easy to check, in this example, that the components of the response are strongly correlated. Note that the smaller the rank of a predictor, the easier it is to be selected. Thus, we can conclude from Table 1 that our proposed method RVC-SIS and RVC-ISIS outperforms the remaining ones in terms of the rank of the important predictors. Naive-SIS1, Naive-SIS2, and PS perform badly in selecting the important predictors X 1 X 4 . Naive-ISIS1 and Naive-ISIS2 still fail to detect X 1 X 2 , while IPS still exhibits bad performances in choosing X 1 due to the large rank. The rank of active variables X 2 X 4 in GCPS is too high to be chosen, while the rank of X 2 in IGCPS is low enough to be selected. Although DC-SIS behaves badly in selecting X 2 in this context, DC-ISIS dramatically improves its performance. From Table 2, we can see that DC-ISIS, RVC-SIS, and RVC-ISIS can pick up X 1 X 4 with an overwhelming probability. Naive-SIS1, Naive-SIS2, and PS select these predictors with little chance. Naive-ISIS1 and Naive-ISIS2 still fail to detect X 1 X 2 , while IPS cannot identify X 1 . GCPS selects X 1 with overwhelming probability, while IGCPS chooses X 2 X 4 with a large probability. Table 3 demonstrates that only DC-ISIS, RVC-SIS, and RVC-ISIS include all important predictors X 1 X 4 with a low M, indicating that the four predictors have a higher priority of being selected by these screening methods and thus are ranked at the top.
In Example 3, it is evident that predictor X 4 is marginally uncorrelated with Y at the population level, but it is important to Y . Table 1 shows that Naive-SIS1, Naive-SIS2, PS, DC-SIS, GCPS, and RVC-SIS consistently identify X 4 with a larger rank, while the proposed iterative screening procedures of these methods screen out X 4 with a smaller rank. This indicates that the iterative screening procedure can effectively tackle the second challenge that is mentioned at the beginning of Section 3. Table 2 demonstrates that Naive-SIS1, Naive-SIS2, PS, DC-SIS, GCPS, and RVC-SIS select X 4 with little probability, while their respective iterative screening procedures can pick it up with an overwhelming probability. It can be seen from Table 3 that the model size of Naive-SIS1, Naive-SIS2, PS, DC-SIS, GCPS, and RVC-SIS is consistently large due to the incapability of identifying X 4 .
In Example 4, it is worth noting that important predictor X 4 is again uncorrelated with the response Y , and X 5 has a very small correlation with response Y . In fact, variable X 5 has the same proportion of contributions to the response as noise ε i does. Similar conclusions can be drawn from these tables due to the existence of X 4 . Furthermore, these screening methods have a lower priority of selecting X 5 than X 1 X 3 due to the small contribution of X 5 to the response.

5. Application

In order to demonstrate the application of our proposed method, we used yeast cell-cycle gene expression data in Spellman et al. [25] and chromatin immunoprecipitation on chip (CHIP-chip) data from Lee et al. [26]. The yeast cell-cycle gene expression data consist of 542 genes from a factor-based experiment in which 18 mRNA levels are measured every 7 min for 119 min.
The chromatin immunoprecipitation on chip (ChIP-chip) data contain the binding information for 106 transcription factors (TFs), which are related to cell regulation. These two datasets are combined as the yeast cell-cycle dataset available in the R package spls. For further details about this dataset, please refer to Chun and Keles [27].
First, the data are randomly split into a training set of 300 and a test set of 242. We apply all screening methods on the training set to select the active predictors. To obtain the threshold in the screening procedures, we introduce auxiliary variables proposed by Zhu et al. [6]. That is, we independently generate d = 2000 auxiliary variables Z N d ( 0 , I d ) such that Z is independent of both Y and X . Consider the ( p + d ) -dimensional vector ( X , Z ) as the predictors and Y as the response. Then, we calculate ω j ^ for j = 1 , , p + d . Note that Z is truly inactive via artificial construction, and we can define a threshold
C d = max j = p + l , p + d ω j ^
to select the important predictors. We repeated the random data segmentation 100 times to avoid the impact of random segmentation on the results and comparisons.
Note that there are 21 known and experimentally verified TFs related to the cell-cycle process, as stated in Wang et al. [28]. We then compare the number of confirmed TFs selected in the screening procedure mentioned earlier. The results are reported in Figure 1. From this figure, it can be concluded that RVC-ISIS performs best among all screening procedures, and RVC-SIS performs best among the screening procedures except the iterative ones.
Furthermore, we applied the test set to assess the performance of the screening procedures in terms of the mean square error of the predictions, which can be found in Figure 2 and Table 4. It can be seen that these iterative screening procedures uniformly outperform the original non-iterative screening ones.

6. Discussion

Through extensive simulations and a real data example, we see that the proposed procedure outperforms 58–100% of the existing candidates such as the Naive-SIS1, Naive-ISIS1, Naive-SIS2, Naive-ISIS2, PS and IPS by Li et al. [3], GCPS and IGCPS by Liu et al. (2020) [4], DC-SIS by Li et al. [10], and DC-ISIS by Zhong and Zhu [23]. On average, the proposed method outperforms 79% of existing methods across both linear and nonlinear scenarios. In particular, the proposed method has an enormous advantage in the nonlinear scenarios, with multivariate responses being strongly correlated, like in Example 2.
There exists some limitations in the proposed screening method as well. First, as Muthyala et al. (2024) [29] pointed out, although sure independence screening methods have been proven to be particularly effective in the natural sciences and clinical studies, its widespread adoption has been limited by performance inefficiencies and the challenges posed by its R-based or FORTRAN-based implementation, especially in modern computing environments with massive data. More fast and efficient implementation using new techniques, such as PyTorch 3.12 —based implementations, are necessary for the proposed RVC-SIS and RVC-ISIS algorithms, which is critical for projects like chronic thromboembolic pulmonary hypertension studies [30] with ultrahigh-dimensional multimodal medical data, including radiomics features, genomics data, laboratory omics data, and so on.
Second, this paper only considered the sure independence screening for additive models of continuous multivariates response. However, in clinical practice, such as the pulmonary thromboembolism study, the majority of clinical outcomes is of the binary or survival type. The proposed method and algorithm cannot be directly applied to either scenarios with multivariate, binary, or multilevel categorical outcomes or the scenarios with survival outcomes. However, it is relatively easy to adaptive the proposed RVC-SIS and RVC-ISIS to other types of outcomes using routine methods of related statistical fields.
Last but not the least, we do not investigate the performance of our proposed method for the skewed and heavy-tailed distributed data [31]. Further exploring the interactions between (among) predictors and detecting the corresponding ones is beyond the scope of this paper. These could be interesting topics for future research.

7. Conclusions

This paper investigated a nonparametric screening method for sparse, ultrahigh-dimensional additive models with multivariate responses. We proved that the RVC-SIS exhibits sure screening and ranking consistency properties under certain regularity conditions. In addition, we show, through extensive simulation study, that the proposed RVC-ISIS procedure effectively detects the truly important predictors even in the presence of complex correlations among predictors. The proposed method in this article has great potential for applications in developing nonlinear models with both ultrahigh-dimensional predictors and multivariate responses, especially in the chronic thromboembolic pulmonary hypertension study, where there exist multi-organ linkage responses, such as the right-sided heart failure in hearts, chronic thromboembolic pulmonary hypertension in lungs, and cerebral thrombosis in the brain.

Author Contributions

Methodology, Y.C. and B.L.; Software, Y.C.; Validation, B.L.; Formal analysis, Y.C. and B.L.; Investigation, B.L.; Resources, B.L.; Writing—original draft, Y.C. and B.L.; Writing—review & editing, B.L.; Funding acquisition, B.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Noncommunicable Chronic Diseases-National Science and Technology Major Project (No. 2023ZD0506600) and the Capital’s Funds for Health Improvement and Research (grant number: 2024-1G-4251).

Data Availability Statement

The original data presented in the study are publicly available in R package ‘spls’ (https://cran.r-project.org/web/packages/spls/index.html, accessed on 5 May 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Proof of Theorem 1. 
We first demonstrate the consistency of the numerator of ω ^ j . It is easy to show that V ^ n ( Y , B j ) is a U statistic of order four: that is,
V ^ n ( Y , B j ) = 1 n 4 s < t < u < v h ( Z s , Z t , Z u , Z v ) ,
where
h ( Z 1 , Z 2 , Z 3 , Z 4 ) = 1 12 { ( 2 Y 1 Y 2 + 2 Y 3 Y 4 Y 1 Y 3 Y 1 Y 4 Y 2 Y 3 Y 2 Y 4 ) ( B j 1 B j 2 + B j 3 B j 4 ) + ( 2 Y 1 Y 3 + 2 Y 2 Y 4 Y 1 Y 2 Y 1 Y 4 Y 2 Y 3 Y 3 Y 4 ) ( B j 1 B j 3 + B j 2 B j 4 ) + ( 2 Y 1 Y 4 + 2 Y 2 Y 3 Y 1 Y 2 Y 1 Y 3 Y 2 Y 4 Y 3 Y 4 ) ( B j 1 B j 4 + B j 2 B j 3 ) }
with Z i = ( Y i , B j i ) .
Rewrite the U-statistic V ^ n ( Y , B j ) as follows:
V ^ n ( Y , B j ) = 1 n 4 s < t < u < v h ( Z s , Z t , Z u , Z v ) 1 { | h ( Z s , Z t , Z u , Z v ) |   M } + 1 n 4 s < t < u < v h ( Z s , Z t , Z u , Z v ) 1 { | h ( Z s , Z t , Z u , Z v ) | >   M } = : V ^ n 1 ( Y , B j ) + V ^ n 2 ( Y , B j ) ,
where M will be specified later. Accordingly, V ( Y , B j ) can be decomposed into two parts:
V ( Y , B j ) = E [ h ( Z s , Z t , Z u , Z v ) 1 { | h ( Z s , Z t , Z u , Z v ) |   M } ] + E [ h ( Z s , Z t , Z u , Z v ) 1 { | h ( Z s , Z t , Z u , Z v ) | >   M } ] = : V 1 ( Y , B j ) + V 2 ( Y , B j ) .
Evidently, V ^ n 1 ( Y , B j ) and V ^ n 2 ( Y , B j ) are unbiased estimators of V 1 ( Y , B j ) and V 2 ( Y , B j ) , respectively.
First, we demonstrate the consistency of V ^ n 1 ( Y , B j ) . Note that V ^ n 1 ( Y , B j ) is a U-statistic of order four. We can therefore conclude, based on theorem 5.6.1.A of Serfling [32], that Pr ( V ^ n 1 ( Y , B j ) V 1 ( Y , B j ) ε ) exp ( n ε 2 / ( 8 M 2 ) ) for any given ε > 0 . In light of the symmetry of U statistics, it is straightforward to obtain
Pr ( | V ^ n 1 ( Y , B j ) V 1 ( Y , B j ) | ε ) 2 exp ( n ε 2 / ( 8 M 2 ) ) .
Next, we consider the consistency of V ^ n 2 ( Y , B j ) . Notice that B j k 1 . Using the Cauchy–Schwarz inequality, we have | h ( Z 1 , Z 2 , Z 3 , Z 4 ) | d n ( Y 1 2 + Y 2 2 + Y 3 2 + Y 4 2 ) . Together with Markov’s inequality, it is not dfficult to show that { V 2 ( Y , B j ) } 2 E { h 2 ( Z 1 , Z 2 , Z 3 , Z 4 ) } [ E ( exp ( t Y 1 2 ) ) ] 4 / exp ( t M / d n ) for any t > 0 . By choosing M = C n γ d n for 0 < γ < 1 / 2 2 κ , | V 2 ( Y , B j ) | ε / 2 provided that n is sufficiently large and Conditions (C3)–(C4) hold. Meanwhile, we can confirm that the events satisfy { | V ^ n 2 ( Y , B j ) |   ε / 2 } { Y i 2   >   M / ( 4 d n ) , for some 1 i n } . Otherwise, if Y i 2   M / ( 4 d n ) holds for all 1 i n , then we can conclude that | h ( Z s , Z t , Z u , Z v ) |   M . Therefore | V ^ n 2 ( Y , B j ) | =   0 , which contradicts event | V ^ n 2 ( Y , B j ) |     ε / 2 . Under Conditions (C3)–(C4), for any t > 0 , there exists a positive constant C such that Pr ( Y i 2     M / ( 4 d n ) ) C exp ( t M / ( 4 d n ) ) . Consequently, Pr ( | V ^ n 2 ( Y , B j ) |   ε / 2 ) C n exp ( t M / ( 4 d n ) ) . which implies that
Pr ( | V ^ n 2 ( Y , B j ) V 2 ( Y , B j ) |   ε ) C n exp ( t M / ( 4 d n ) )
Recall that M = C n γ d n . By combining (A1) and (A2), we have
Pr ( | V ^ n ( Y , B j ) V ( Y , B j ) |   ε ) = O { exp ( c 3 ε 2 n 1 2 γ d n 2 ) + n exp ( c 4 n γ ) }
for some positive constants c 3 and c 4 .
Similarly, we can determine the convergence rate of the denominator. Furthermore, recalling Fact 3 in Fan et al. [7], we can show that V ( B j , B j ) = O ( d n 1 ) . Hence, we obtain the convergence rate of ω ^ j , which is similar to the form of (A3). To be precise, Pr ( | ω ^ j ω j |   d n 1 / 2 ε ) = O { exp ( c 3 ε 2 n 1 2 γ d n 2 ) + n exp ( c 4 n γ ) } . Let ε = c 2 n 2 κ , where any c 2 > 0 and κ satisfies 0 < 2 κ + γ < 1 / 2 ; we have
Pr ( max 1 j p | ω ^ j ω j | c 2 d n 1 / 2 n 2 κ ) p max 1 j p Pr ( | ω ^ j ω j |   c 2 d n 1 / 2 n 2 κ ) O ( p [ exp { c 3 n 1 4 κ 2 γ d n 2 } + n exp ( c 4 n γ ) ] ) .
To deal with the second part, we first show that min j A ω j 2 c d n 1 / 2 n 2 κ . By invoking Condition (C5), for any j A , there exists l { 1 , , q } such that E [ { E ( Y l | X j ) } 2 ] c 1 d n n 2 κ . Under Conditions (C1), (C2), and (C6), we can conclude from Lemma 1 and Fact 3 in Fan et al. [7] that V ( Y , B j ) c 1 D 1 ξ n 2 κ , which implies that ω j 2 c d n 1 / 2 n 2 κ provided that 0 < c < c 1 D 1 ξ / D 2 V ( Y , Y ) / 2 . Let us introduce the event ζ n = { max j A | ω ^ j ω j | c d n 1 / 2 n 2 κ } . It is worth noting that ζ n { A A ^ } . Therefore,
Pr ( A A ^ ) Pr ( ζ n ) = 1 Pr ( ζ n c ) 1 O ( s n [ exp { c 3 n 1 4 κ 2 γ d n 2 } + n exp ( c 4 n γ ) ] ) .
where the last inequality follows from (A3). □
Proof of Theorem 2. 
Recall the definition of A ; it is clear that ω j > 0 for j A and ω j = 0 for j A c . Under Conditions (C1), (C2), (C5), and (C6), we can show that min j A ω j 2 c d n 1 / 2 n 2 κ . Then, it is straightforward to verify
Pr ( min j A ω ^ j max j A c ω ^ j ) Pr { ( max j A c ω ^ j max j A c ω j ) ( min j A ω ^ j min j A ω j ) 2 c d n 1 / 2 n 2 κ } Pr { max j A c | ω ^ j ω j | + max j A | ω ^ j ω j | 2 c d n 1 / 2 n 2 κ } Pr { max 1 j p | ω ^ j ω j | c d n 1 / 2 n 2 κ } .
Using the result of Theorem 1 and Fatou’s Lemma, we can derive
Pr { lim inf n ( min j A ω ^ j max j A c ω ^ j ) 0 } lim n Pr ( min j A ω ^ j max j A c ω ^ j ) = 0 .
The proof is now completed. □

References

  1. Fan, J.; Lv, J. Sure independence screening for ultrahigh-dimensional feature space. J. R. Stat. Soc. Ser. (Stat. Methodol.) 2008, 70, 849–911. [Google Scholar] [CrossRef]
  2. Fan, J.; Samworth, R.; Wu, Y. Ultrahigh-dimensional feature selection: Beyond the linear model. J. Mach. Learn. Res. 2009, 10, 2013–2038. [Google Scholar] [PubMed]
  3. Li, X.; Cheng, G.; Wang, L.; Lai, P.; Song, F. Ultrahigh-dimensional feature screening via projection. Comput. Stats Data Anal. 2017, 114, 88–104. [Google Scholar] [CrossRef]
  4. Liu, S.; Li, X.; Zhang, J. Ultrahigh-dimensional feature screening for additive model with multivariate response. J. Stat. Comput. Simul. 2020, 90, 775–799. [Google Scholar] [CrossRef]
  5. Fan, J.; Song, R. Sure independence screening in generalized linear models with NP-dimensionality. Ann. Stat. 2010, 38, 3567–3604. [Google Scholar] [CrossRef]
  6. Zhu, L.P.; Li, L.; Li, R.; Zhu, L.X. Model-free feature screening for ultrahigh-dimensional data. J. Am. Stat. Assoc. 2011, 106, 1464–1475. [Google Scholar] [CrossRef]
  7. Fan, J.; Feng, Y.; Song, R. Nonparametric independence screening in sparse ultra-high-dimensional additive models. J. Am. Stat. Assoc. 2011, 106, 544–557. [Google Scholar] [CrossRef]
  8. Yuan, Y.; Billor, N. Sure independent screening for functional regression model. Commun. Stat.-Simul. Comput. 2024, 1–20. [Google Scholar] [CrossRef]
  9. Cui, H.; Zou, F.; Ling, L. Feature screening and error variance estimation for ultrahigh-dimensional linear model with measurement errors. Commun. Math. Stat. 2025, 13, 139–171. [Google Scholar] [CrossRef]
  10. Li, R.; Zhong, W.; Zhu, L. Feature screening via distance correlation learning. J. Am. Stat. Assoc. 2012, 107, 1129–1139. [Google Scholar] [CrossRef]
  11. Shao, X.; Zhang, J. Martingale difference correlation and its use in high-dimensional variable screening. J. Am. Stat. Assoc. 2014, 109, 1302–1318. [Google Scholar] [CrossRef]
  12. Cui, H.; Li, R.; Zhong, W. Model-free feature screening for ultrahigh-dimensional discriminant analysis. J. Am. Stat. Assoc. 2015, 110, 630–641. [Google Scholar] [CrossRef]
  13. Chen, S.; Lu, J. Quantile-Composited feature screening for ultrahigh-dimensional data. Mathematics 2023, 11, 2398. [Google Scholar] [CrossRef]
  14. Liu, R.; Deng, G.; He, H. Generalized Jaccard feature screening for ultra-high dimensional survival data. AIMS Math. 2024, 9, 27607–27626. [Google Scholar] [CrossRef]
  15. Sang, Y.; Dang, X. Grouped feature screening for ultrahigh-dimensional classification via Gini distance correlation. J. Multivar. Anal. 2024, 204, 105360. [Google Scholar] [CrossRef]
  16. Wu, J.; Cui, H. Model-free feature screening based on Hellinger distance for ultrahigh-dimensional data. Stat. Pap. 2024, 65, 5903–5930. [Google Scholar] [CrossRef]
  17. Zhong, W.; Li, Z.; Guo, W.; Cui, H. Semi-distance correlation and its applications. J. Am. Stat. Assoc. 2024, 119, 2919–2933. [Google Scholar] [CrossRef]
  18. Tian, Z.; Lai, T.; Zhang, Z. Variation of conditional mean and its application in ultrahigh-dimensional feature screening. Commun. Stat.-Theory Methods 2025, 54, 352–382. [Google Scholar] [CrossRef]
  19. Li, L.; Ke, C.; Yin, X.; Yu, Z. Generalized martingale difference divergence: Detecting conditional mean independence with applications in variable screening. Comput. Stat. Data Anal. 2023, 180, 107618. [Google Scholar] [CrossRef]
  20. Liu, L.; Lian, H.; Huang, J. More efficient estimation of multivariate additive models based on tensor decomposition and penalization. computational statistics and data analysis. J. Mach. Learn. Res. 2024, 25, 1–27. [Google Scholar]
  21. Desai, N.; Baladandayuthapani, V.; Shinohara, R.; Morris, J. Covariance Assisted Multivariate Penalized Additive Regression (CoMPAdRe). J. Comput. Graph. Stat. 2024, 1–10. [Google Scholar] [CrossRef]
  22. Escoufier, Y. Le traitement des variables vectorielles. Biometrics 1973, 29, 751–760. [Google Scholar] [CrossRef]
  23. Zhong, W.; Zhu, L. An iterative approach to distance correlation-based sure independence screening. J. Stat. Comput. Simul. 2015, 85, 2331–2345. [Google Scholar] [CrossRef]
  24. Liu, J.; Li, R.; Wu, R. Feature selection for varying coefficient models with ultrahigh-dimensional covariates. J. Am. Stat. Assoc. 2014, 109, 266–274. [Google Scholar] [CrossRef]
  25. Spellman, P.T.; Sherlock, G.; Zhang, M.Q.; Iyer, V.R.; Anders, K.; Eisen, M.B.; Brown, P.O.; Botstein, D.; Futcher, B. Comprehensive identification of cell cycle–regulated genes of the yeast Saccharomyces cerevisiae by microarray hybridization. Mol. Biol. Cell 1998, 9, 3273–3297. [Google Scholar] [CrossRef]
  26. Lee, T.I.; Rinaldi, N.J.; Robert, F.; Odom, D.T.; Bar-Joseph, Z.; Gerber, G.K.; Hannett, N.M.; Harbison, C.T.; Thompson, C.M.; Young, R.A.; et al. Transcriptional regulatory networks in Saccharomyces cerevisiae. Science 2002, 298, 799–804. [Google Scholar] [CrossRef] [PubMed]
  27. Chun, H.; Keles, S. Sparse partial least squares regression for simultaneous dimension reduction and variable selection. J. R. Stat. Soc. Ser. (Stat. Methodol.) 2010, 72, 3–25. [Google Scholar] [CrossRef]
  28. Wang, L.; Chen, G.; Li, H. Group SCAD regression analysis for microarray time course gene expression data. Bioinformatics 2007, 23, 1486–1494. [Google Scholar] [CrossRef]
  29. Muthyala, M.; Sorourifar, F.; Paulson, J.A. TorchSISSO: A PyTorch-based implementation of the sure independence screening and sparsifying operator for efficient and interpretable model discovery. Digit. Chem. Eng. 2024, 13, 100198. [Google Scholar] [CrossRef]
  30. Xie, W.; Yu, Y.; Huang, Q.; Yan, X.; Yang, Y.; Xiong, C.; Liu, Z.; Wan, J.; Gong, S.; Wang, L.; et al. Epidemiology and management patterns of chronic thromboembolic pulmonary hypertension in China: A systematic literature review and meta-analysis. Chin. Med J. 2025, 138, 1000–1002. [Google Scholar] [CrossRef]
  31. Song, Y.; Zhou, W.; Zhou, W.X. Large-scale inference of multivariate regression for heavy-tailed and asymmetric data. Stat. Sin. 2023, 33, 1831–1852. [Google Scholar] [CrossRef]
  32. Serfling, R.J. Approximation Theorems of Mathematical Statistics; John Wiley & Sons: Hoboken, NJ, USA, 1980. [Google Scholar]
Figure 1. The number of confirmed TFs being selected by different screening methods, obtained from 100 random data partitions.
Figure 1. The number of confirmed TFs being selected by different screening methods, obtained from 100 random data partitions.
Mathematics 13 01558 g001
Figure 2. The mean squared errors of prediction based on 100 random partitions of the data.
Figure 2. The mean squared errors of prediction based on 100 random partitions of the data.
Mathematics 13 01558 g002
Table 1. The average rank R j of the true predictor X j for Examples 1–4.
Table 1. The average rank R j of the true predictor X j for Examples 1–4.
Method R 1 R 2 R 3 R 4 R 5
Example 1Naive-SIS14.1082.482.0022.4884.056
Naive-ISIS13.8562.7422.2022.3423.878
Naive-SIS23.712.7262.4082.7223.634
Naive-ISIS24.2363.0562.4082.5462.886
PS2.942.9982.992.9883.084
IPS3.6561.0161.9863.2583.798
DC-SIS4.532.4421.8622.4644.154
DC-ISIS4.382.3472.0362.8983.956
GCPS4.6532.4621.3622.4744.468
IGCPS4.0382.8471.3622.9984.854
RVC-SIS4.1562.451.8982.4944.114
RVC-ISIS3.8322.552.1342.2564.234
Example 2Naive-SIS11006.022617.5341709.5441774.618-
Naive-ISIS1927.04465.648107.87248.554-
Naive-SIS21078.528654.6781737.1421800.202-
Naive-ISIS21032.746509.788149.61641.562-
PS1076.362865.076734.618570.108-
IPS1000.356156.8249.4825.864-
DC-SIS125.704385.022.0121-
DC-ISIS19.25414.5962.3161-
GCPS1.035677.842345.654887.356-
IGCPS1.03528.64777.33696.256-
RVC-SIS8.6188.122.2321-
RVC-ISIS7.7564.612.5361-
Example 3Naive-SIS12.7462.2481.0281002.92-
Naive-ISIS14.2322.7561.0285.376-
Naive-SIS22.6462.3581.0521003.12-
Naive-ISIS23.8363.1121.0525.332-
PS1.5562.861.5841014.08-
IPS1.5563.9481.5965.386-
DC-SIS2.7062.2841.028982.866-
DC-ISIS4.1122.8781.0285-
GCPS2.7322.3421.126856.562-
IGCPS4.1262.9641.1265.051-
RVC-SIS2.7342.2361.032965.084-
RVC-ISIS4.2162.7581.0325-
Example 4Naive-SIS12.9082.3061.0261010.83812.572
Naive-ISIS14.2082.8061.02641.2966.024
Naive-SIS22.6682.4521.0361011.00612.306
Naive-ISIS23.7963.2241.03656.8086.044
PS1.5522.8621.586977.96233.456
IPS1.5523.9041.58672.2385.638
DC-SIS2.8942.3161.032968.66222.386
DC-ISIS4.0682.9261.0323.0006.016
GCPS2.8252.2121.2621126.12436.746
IGCPS4.5342.6471.2623.8988.956
RVC-SIS3.0542.2641.044989.4422.552
RVC-ISIS4.1342.8381.0443.0006.106
NOTE: “-” denotes that there is no corresponding content because of the difference in M .
Table 2. The selecting rate P j of the true predictors and P a for Examples 1–4.
Table 2. The selecting rate P j of the true predictors and P a for Examples 1–4.
Method P 1 P 2 P 3 P 4 P 5 P a
Example 1Naive-SIS1111111
Naive-ISIS1111111
Naive-SIS2111111
Naive-ISIS2111111
PS111111
IPS111111
DC-SIS111111
DC-ISIS111111
GCPS111111
IGCPS111111
RVC-SIS111111
RVC-ISIS111111
Example 2Naive-SIS10.0260.16800-0
Naive-ISIS10.0540.2340.7980.916-0.015
Naive-SIS20.0280.17600-0
Naive-ISIS20.0550.2200.7020.916-0.02
PS00.0700.1100.130-0
IPS0.022 0.752 11-0.016
DC-SIS0.6280.26811-0.158
DC-ISIS0.9860.99811-0.984
GCPS10.1560.2780.002-0
IGCPS10.9960.8830.824-0.648
RVC-SIS0.9640.98211-0.966
RVC-ISIS0.9880.99611-0.986
Example 3Naive-SIS11110-0
Naive-ISIS11111-1
Naive-SIS21110-0
Naive-ISIS21111-1
PS1110-0
IPS1111-1
DC-SIS1110-0
DC-ISIS1111-1
GCPS1110-0
IGCPS1111-1
RVC-SIS1110-0
RVC-ISIS1111-1
Example 4Naive-SIS1111010
Naive-ISIS1111111
Naive-SIS2111010
Naive-ISIS2111111
PS111010
IPS111111
DC-SIS111010
DC-ISIS111111
GCPS111010
IGCPS111111
RVC-SIS111010
RVC-ISIS111111
Table 3. The quantiles of M = max j A R j for Examples 1–4.
Table 3. The quantiles of M = max j A R j for Examples 1–4.
Method 5 % 25 % 50 % 75 % 95 %
Example 1Naive-SIS155556
Naive-ISIS155556
Naive-SIS255556
Naive-ISIS255556
PS55555
IPS55555
DC-SIS55555
DC-ISIS55555
GCPS55555
IGCPS55555
RVC-SIS55556
RVC-ISIS55556
Example 2Naive-SIS11668.451990.75196919942000
Naive-ISIS1277.966910151407.251887.05
Naive-SIS21713.41913197519952000
Naive-ISIS2261.25724.9510691466.51846.1
PS811.951215.251488.51739.251957.15
IPS122512.759821476.51940.4
DC-SIS22.95122.75315.5653.51317.2
DC-ISIS5591240.1
GCPS280.56634.85948.321203.461568.12
IGCPS20.8638.9557.78286.96500.85
RVC-SIS444746
RVC-ISIS444738
Example 3Naive-SIS1430.65628.5887.51326.251879.55
Naive-ISIS155555
Naive-SIS2444.25639903.51343.51878.15
Naive-ISIS255555
PS458.85682.759391325.751829.3
IPS44455
DC-SIS337.755908711389.251828.3
DC-ISIS55555
GCPS268.65457.54671.88999.581509.98
IGCPS55555
RVC-SIS284.65596.758801299.251811.35
RVC-ISIS55555
Example 4Naive-SIS1426.8639.590313941840.2
Naive-ISIS166666
Naive-SIS2421.9641.25890.51368.251835.35
Naive-ISIS266666
PS373.85676.759351253.251809.25
IPS555.567
DC-SIS304.9609.759221324.51809.05
DC-ISIS66666
GCPS298.25807.751130.421628.751811.50
IGCPS66666
RVC-SIS258.95572.5934.51424.251876.15
RVC-ISIS66666
Table 4. Averaged mean squared errors (AMSEs) of the prediction over 100 random data partitions and the corresponding standard deviations of the screening procedures.
Table 4. Averaged mean squared errors (AMSEs) of the prediction over 100 random data partitions and the corresponding standard deviations of the screening procedures.
MethodAMSEStandard Deviation
Naive-SIS10.19970.0281
Naive-ISIS10.19590.0286
Naive-SIS20.19910.0278
Naive-ISIS20.18780.0288
PS0.18980.0652
IPS0.17490.0604
DC-SIS0.18220.0124
DC-ISIS0.17720.0126
GCPS0.16680.0223
IGCPS0.16290.0221
RVC-SIS0.16010.0136
RVC-ISIS0.14220.0134
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.; Liang, B. Sure Independence Screening for Ultrahigh-Dimensional Additive Model with Multivariate Response. Mathematics 2025, 13, 1558. https://doi.org/10.3390/math13101558

AMA Style

Chen Y, Liang B. Sure Independence Screening for Ultrahigh-Dimensional Additive Model with Multivariate Response. Mathematics. 2025; 13(10):1558. https://doi.org/10.3390/math13101558

Chicago/Turabian Style

Chen, Yongshuai, and Baosheng Liang. 2025. "Sure Independence Screening for Ultrahigh-Dimensional Additive Model with Multivariate Response" Mathematics 13, no. 10: 1558. https://doi.org/10.3390/math13101558

APA Style

Chen, Y., & Liang, B. (2025). Sure Independence Screening for Ultrahigh-Dimensional Additive Model with Multivariate Response. Mathematics, 13(10), 1558. https://doi.org/10.3390/math13101558

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop