Next Article in Journal
Digital h-Fibrations and Some New Results on Digital Fibrations
Previous Article in Journal
Modulation Transfer between Microwave Beams: Asymptotic Evaluation of Integrals with Pole Singularities near a First-Order Saddle Point
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Construction and Analysis for the Optimal Supersaturated Designs with a Large Number of Inert Factors

1
School of Mathematics and Statistics, Qingdao University, Qingdao 266071, China
2
School of Mathematics and Computing Science, Guilin University of Electronic Technology, Guilin 541004, China
*
Authors to whom correspondence should be addressed.
Axioms 2024, 13(3), 179; https://doi.org/10.3390/axioms13030179
Submission received: 9 January 2024 / Revised: 20 February 2024 / Accepted: 6 March 2024 / Published: 8 March 2024

Abstract

:
Supersaturated designs (SSDs) refer to those designs in which the run size is much smaller than the main effects to be estimated. They are commonly used to identify a few, but critical, active factors from a large set of potentially active ones, keeping the cost as low as possible. In this regard, the development of new construction and analysis methods has recently seen a rapid increase. In this paper, we provide some methods to construct equi- and mixed-level E ( f   N O D ) optimal SSDs with a large number of inert factors using the substitution method. The proposed methods are easy to implement, and many new SSDs can then be constructed from them. We also study a variable selection method based on the screening-selection network (SSnet) method for regression problems. A real example is analyzed to illustrate that it is able to effectively identify active factors. Eight different analysis methods are used to analyze the data generated from the proposed designs. Three scenarios with different setups of parameters are designed, and the performances of each method are illustrated by extensive simulation studies. Among all these methods, SSnet produced the most satisfactory results according to power.

1. Introduction

In industry, it is often necessary to consider a large number of potentially relevant factors, of which only a small number are significant. However, due to practical restrictions, such as financial costs and time constraints, the actual number of test runs is often far from sufficient to estimate all potentially relevant factors, so the question of how to screen a small number of significant factors with a smaller number of tests becomes a critically important one. In this regard, supersaturated design (SSD), as a powerful tool for scenarios when the run size is much smaller than the main effects to be estimated, is used to reduce the cost of conducting experiments and to identify the active factors as correctly as possible. Therefore, the development of good supersaturated designs and variable screening methods are highly needed.
An SSD is fundamentally a fractional factorial design in which the number of potential effects is larger than the number of runs. Satterthwaite [1] was a pioneer who introduced the concept of SSDs in random balanced designs. Booth and Cox [2] conducted a comprehensive analysis of SSDs and introduced the E ( s 2 ) criterion as a measure to evaluate 2-level SSDs. However, further research on such designs remained limited until Lin [3] and Wu [4]. They innovatively constructed SSDs by Hadamard matrices. With their potential in factor screening experiments, SSDs have gained growing popularity in recent years. Research on mixed-level SSDs has been extensive, with early contributions by Fang, Lin, and Liu [5,6], who introduced the E ( f   N O D ) criterion and the fractions of saturated orthogonal arrays (FSOA) method for efficiently constructing such designs. The evaluation of mixed-level SSDs has been approached by Yamada and Lin [7], who utilized the χ 2 criterion. Fang et al. [6] further derived lower bounds and illustrated sufficient conditions to achieve the lower bounds. They demonstrated that both the E ( s 2 ) and the χ 2 criteria are special cases of E ( f   N O D ) criterion. Zou and Qin [8] provided new lower bounds of the E ( s 2 ) criterion and the E ( f   N O D ) criterion. Consequently, this paper exclusively employs the E ( f   N O D ) criterion as the measure of choice. Georgiou, Koukouvinos, and Mantas [9] introduced a novel approach to constructing multi-level SSDs that are optimal under the E ( f   N O D ) criterion, utilizing resolvable balanced incomplete block designs. Thereafter, researchers have paid great attention to constructing mixed-level SSDs. Liu and Cai [10] presented a special substitution method, dividing the rows of a design into blocks and substituting the levels of another design with these blocks, yielding E ( f   N O D ) -optimal mixed-level SSDs. Sun, Lin, and Liu [11] achieved numerous optimal SSDs based on balanced designs and the construction of differential transposition, utilizing the Kronecker sum. For more details on the construction of SSDs, we refer to the works of Georgiou [12], Drosou and Koukouvinos [13], Jones et al. [14], Chatterjee and Koukouvinos [15], Khurana and Banerjee [16], Daraz et al. [17], Li et al. [18]. Further research is needed on the construction methods for E ( f   N O D ) -optimal SSDs with more flexible row and column numbers, especially in situations with a large number of inert factors.
The identification of active factors in SSDs poses a challenging problem, mainly attributed to the non-full rank and the potential for complex covariance between columns of the design matrix. Despite the existence of various methods for variable selection in factor screening experiments, a definitive optimal solution remains elusive. A prominent and widely used variable selection method is the least absolute shrinkage and selection operation (LASSO) [19], which effectively penalizes the squared loss of the data by incorporating an l 1 -norm penalty on the parameter vector. After LASSO was introduced, Fan and Li [20] introduced a class of variable selection procedures through nonconcave penalized likelihood, utilizing the smoothly clipped absolute deviation (SCAD) penalty. However, Li and Lin [21] demonstrated that direct application of the SCAD penalty for analyzing SSDs is not feasible due to the regularity conditions outlined in their paper, which necessitate a full rank design matrix. To address this limitation, Li and Lin [21] adopted the shrinkage estimation and the selection method proposed by Fan and Li [20], incorporating the SCAD penalty. They further extended the nonconcave penalized likelihood approaches to least squares, specifically focusing on scenarios where the design matrix is SSD. Zou [22] proposed a novel approach known as adaptive LASSO (Ada LASSO) in the context of finite parameters. It involves solving a convex optimization problem with an l 1 constraint, as described by Fan and Lv [23]. In another study, Candes and Tao [24] introduced the Dantzig selector (DS) method, which identifies the most relevant subset of variables or active factors by solving a straightforward convex program. Based on DS, Phoa, Pan, and Xu [25] conducted simulations to demonstrate the effectiveness of the DS method in analyzing SSDs, showcasing its superior efficiency compared to other approaches, such as the SCAD method elaborated by Das et al. [26]. The minimax concave penalty (MCP) method, introduced by Zhang [27], is highly advantageous as it ensures convexity in regions with low variable density while maintaining the desired thresholds for unbiasedness and variable selection. Drosou and Koukouvinos [28] proposed a novel method called support vector regression based on recursive feature elimination (SVR-RFE). This method is specifically designed to analyze SSDs for regression purposes; the method was shown to be efficient and accurate for two-level and mixed-level (two-level and three-level) SSDs. Errore et al. [29] assessed different model selection strategies and concluded that both the LASSO and the DS approach can successfully identify active factors. Ueda and Yamada [30] discussed the combination of SSDs and data analysis methods. In a pioneering contribution, Pokarowski et al. [31] introduced the screening-selection network (SSnet), an innovative approach that enhances the predictive power of LASSO by accommodating general convex loss functions. This novel method extends the applicability of LASSO beyond normal linear models, including logistic regression, quantile regression, and support vector machines. To effectively identify the most influential coefficients, the authors propose a novel technique where they meticulously order the absolute values of nonzero coefficients obtained through LASSO. Subsequently, the final model is selected from a small nested family using the generalized information criterion (GIC), ensuring optimal model performance. How the SSnet method performs in the analysis of equi- and mixed-level SSDs needs to be explored.
In order to evaluate the performance of the factor screening, we use power as the measure, which is defined by the proportion of active main effects successfully detected. This is extremely important because the use of SSDs is mainly to screen out the factors that should be considered for further investigation. The higher the power, the lower the probability of missing a truly active factor, which is more important in the case of SSDs. [13]
In this paper, we employ the substitution method, along with the generalized Hadamard matrix, to construct E ( f   N O D ) -optimal designs. Notably, the construction method can obtain more new equi- and mixed-level SSDs with a large number of inert factors. Meanwhile, the constructed designs possess the desirable characteristic of having no fully aliased columns. We also employ the SSnet method for analyzing SSDs for the first time. The results of the numerical study and real data analysis reveal that the SSnet method has a higher power than existing commonly used methods under different sparsity levels of effects. The SSnet method can be considered as an advantageous tool due to its extremely good performance in identifying active factors in SSDs.
The rest of the article is organized as follows. In Section 22, notation and definitions are introduced. In Section 33, supersaturated designs are extended to supersaturated designs in higher dimensions by the substitution method. In Section 44, we present the principles of the SSnet method and test whether it can select salient factors in variable selection. We also compare the power consumption of each variable selection method in SSD to validate the effectiveness of SSnet in this section. A conclusion is made in Section 5, where we discuss the most interesting findings of this paper.

2. Symbols and Definitions

A difference matrix is an λ s × r matrix with entries from a finite Abelian group ( G , + ) with s elements. It has the property that the vector difference between any two distinct columns, which is calculated by the addition operation on Abelian groups G, contains every element of G equally often (exactly λ times) [32]. For any difference matrix, if we subtract the first column from any column then we can obtain a normalized difference matrix. Its first column consists entirely of the zero element of G [33]. A matrix M is called a generalized Hadamard matrix if and only if both M and its transposition M   are difference matrices. We denote a normalized generalized Hadamard matrix used in this paper as M ( N , s ) , which is a N × N array with s elements, and its first column consists entirely of 0 s . For any prime power, s, and integers m n 1 , theorem 6.6 of Hedayat et al. [33] shows the generalized Hadamard matrix M ( s m , s n ) exists. Lemma 3 of Li et al. [18] provides the construction method of the generalized Hadamard matrix M ( ( 2 s ) n , s ) . A generalized Hadamard matrix M ( 3 2 , 3 ) based on Galois field G F ( 3 , + ) is shown below, where the vector difference between any two columns contains 0 , 1 , and 2 equally often.
M ( 3 2 , 3 ) = 0 0 0 0 0 0 0 0 0 0 1 2 0 2 0 0 1 2 0 2 1 0 2 1 0 2 1 0 0 0 2 2 2 1 1 1 0 1 2 2 0 1 1 2 0 0 2 1 2 1 0 1 0 2 0 0 0 1 1 1 2 2 2 0 1 2 1 2 0 2 0 1 0 2 1 1 0 2 2 1 0
A mixed-level design, D ( n ; q 1 , , q m ) , has n runs and m factors with q 1 , , q m levels, respectively, where the q j -level is taken from a set of q j symbols, { 0 , , q j 1 } . When j = 1 m ( q j 1 ) = n 1 , D is called a saturated design, and when j = 1 m ( q j 1 ) > n 1 , it is called a supersaturated design. When some q j ’s are equal, we use the notation D ( n ; q 1 m 1 q t m t ) , indicating m i factors having q i levels, i = 1 , , t . If all the q j ’s are equal, the design is said to be equi-level and denoted by D ( n ; q m ) . A design is called balanced if each column has the equal occurrence property of the q j symbols. Throughout this paper, we only consider balanced designs.
For a design consisting of D ( n ; q 1 m 1 q t m t ) , we assume that the ith column, d i , and the jth column, d j , have a q i level and a q j level, respectively. The E ( f   N O D ) criterion is defined as
E ( f   N O D ) = 1 i < j m f   N O D ( d i , d j ) m 2 ,
where m 2 = m ( m 1 ) / 2 , f   N O D ( d i , d j ) = a = 1 q i b = 1 q j ( n a b ( d i , d j ) n / q i q j ) 2 , n a b ( d i , d j ) is the number of ( a , b ) -pairs in ( d i , d j ) , and n q i q j stands for the average frequency of level-combinations in ( d i , d j ) . Here, the subscript “NOD” stands for nonorthogonality of the design. The f   N O D ( d i , d j ) value provides a nonorthogonality measure for ( d i , d j ) . Thus, E ( f   N O D ) measures the average nonorthogonality among the columns of D. If and only if f   N O D ( d i , d j ) = 0 , the d i and d j columns are orthogonal. If and only if E ( f   N O D ) = 0 , the design is orthogonal, that is, f   N O D ( d i , d j ) = 0 for any i j .
Let λ h l be the coincidence number between rows h and l, which is defined to be the number of k’s, such that d h k = d l k . Fang et al. [34] expressed E ( f   N O D ) as
E ( f   N O D ) = h , l = 1 , h l n λ h l 2 m ( m 1 ) + C f ,
where C f = n m m 1 n 2 m ( m 1 ) ( i = 1 t m i q i + i = 1 t m i ( m i 1 ) q i 2 + i = 1 t j = 1 , j i t m i m j q i q j ) , the lower bound of which is
E ( f   N O D ) n ( n 1 ) m ( m 1 ) [ ( λ + 1 λ ) ( λ λ ) + λ 2 ] + C f ,
where λ denotes the integer part of λ ,
λ = n i = 1 t m i / q i m n 1 .
Note that when q 1 = · · · = q t = q , λ = m ( n / q 1 ) / ( n 1 ) . They present a sufficient condition for a design being E ( f   N O D ) -optimal.
Lemma 1. 
If the difference among all coincidence numbers between distinct rows of design D does not exceed one, then D is E ( f   N O D ) -optimal.
A design with equal coincidence numbers between different rows is called an equidistant design. We know that equidistant designs are E ( f   N O D ) -optimal designs. In the whole paper, we use the E ( f   N O D ) criterion.

3. Generation of SSDs by Substitution Method

In this section, the generalized Hadamard matrix is extended to SSDs with a large number of inert factors using a special substitution method.
Let A be a s × ( s 1 ) matrix with a cyclic structure denoted as D ( s ; s s 1 ) , where s 2 . The elements in the first row are all 0 and the second row is ( 1 , 2 , , s 1 ) . Each remaining row of this matrix is obtained from each previous row by moving the first element to the last position and cyclicly permutating the other elements to the left. It is obvious that the coincidence number of A is zero.
Construction 1. 
Step 1. Let A be the design D ( s ;   s s 1 ) .
Step 2. Let B be the design obtained by ignoring the first column (all-zero column) of the generalized Hadamard matrix M ( s k , s ) , which is an s k × s k 1 equidistant design also denoted as D ( s k ;   s s k 1 ) .
Step 3. Substitute the ith level of B by the ith row of A, i = 1 , , s , respectively.
Step 4. We obtain a new design P, denoted by D ( s k ;   s ( s 1 ) ( s k 1 ) ) .
Theorem 1. 
The new design, D ( s k ;   s ( s 1 ) ( s k 1 ) ) , obtained by Construction 1 is an E ( f   N O D ) -optimal design.
Proof of Theorem 1. 
Since the transposition of the generalized Hadamard matrix is a difference matrix, the number 0 occurs s k 1 times in the vector difference between any two different rows of M ( s k , s ) . It is natural that the coincidence numbers of B and A in Construction 1 take the constant λ B = s k 1 1 , λ A = 0 , respectively. P is obtained via substituting the ith level of B by the ith row of A. For any two rows of h and l of B, the same level is substituted by the same row of A, so the coincidence number of these is ( s 1 ) ( s k 1 1 ) . The different levels are substituted by the different rows of A, so the coincidence number of these is 0 because of λ A = 0 . Therefore, the coincidence number of the generated design P is λ P = ( s 1 ) ( s k 1 1 ) , which is the product of the factors of A and the coincidence number of B. We know that the design is E ( f   N O D ) -optimal when the coincidence number is a constant. Thus, P is an E ( f   N O D ) -optimal design. □
Specifically, when B is also D ( s ; s s 1 ) in Construction 1, the E ( f   N O D ) -optimal design D ( s ;   s ( s 1 ) 2 ) can be obtained by the substitution method with the coincidence number of 0.
Example 1. 
Take design D ( 3 ;   3 2 ) (refer to Table 1) as both A ( 1 ) and B ( 1 ) . Then the generated design P ( 1 ) is a D ( 3 ;   3 4 ) (refer to Table 2). E f N O D P ( 1 ) = 2 reaches the lower bound.
Construction 2. 
Step 1. Let A be the design D ( s ;   s s 1 ) . Let P 1 be the design D ( s k ;   s ( s 1 ) ( s k 1 ) ) obtained by Construction 1.
Step 2. For i 1 , let B i = P i .
Step 3. Substitute the ith level of B i by the ith row of A.
Step 4. We obtain a new design, P i + 1 , denoted by D ( s k ;   s ( s 1 ) i + 1 ( s k 1 ) ) .
Step 5. Repeat Steps 2–4 until the required number of columns is reached.
Corollary 1. 
For i 1 , the new design D ( s k ;   s ( s 1 ) i ( s k 1 ) ) obtained by Construction 2 is an E ( f   N O D ) -optimal design.
Proof of Corollary 1. 
The coincidence numbers of B i and A in Construction 2 take the constant λ B i = ( s 1 ) i ( s k 1 1 ) , λ A = 0 , respectively. The coincidence number of the generated design P i + 1 is λ P i + 1 = ( s 1 ) i + 1 ( s k 1 1 ) , which is the product of the factors of A and the coincidence number of P i . Thus, P i + 1 is an E ( f   N O D ) -optimal design. □
Lemma 2 
(Li et al. [18], Corollary 1). If s is a prime power, the generalized Hadamard matrix M ( 2 s , s ) can be constructed by Lemma 3(ii) or Remark 2, then M s M s s M l (where s is the Kronecker sum defined on the Abelian group G) is a generalized Hadamard matrix, M ( ( 2 s ) l , s ) .
By deleting the first column of M ( ( 2 s ) l , s ) , we can obtain the E ( f   N O D ) -optimal design D ( ( 2 s ) l ; s ( 2 s ) l 1 ) , and it is an equidistant design.
Construction 3. 
Step 1. Let A be the design D ( s ;   s s 1 ) .
Step 2. Let B be D ( ( 2 s ) l ;   s ( 2 s ) l 1 ) obtained by Lemma 2.
Step 3. Substitute the ith level of B by the ith row of A, respectively.
Step 4. We obtain a new design P, denoted by D ( ( 2 s ) l ;   s ( ( 2 s ) l 1 ) ( s 1 ) ) .
Theorem 2. 
The new design D ( ( 2 s ) l ;   s ( ( 2 s ) l 1 ) ( s 1 ) ) obtained by Construction 3 is an E ( f   N O D ) -optimal design.
Proof of Theorem 2. 
The coincidence number of A is λ A = 0 and the coincidence number of B is λ B = 2 l s l 1 1 . Now consider the coincidence number between any two distinct rows of h and l of the generated design P is a constant ( 2 l s l 1 1 ) ( s 1 ) , so P is E ( f   N O D ) -optimal. □
Example 2. 
Let v = 3 , α = 2 , and l = 2 in Lemma 3(ii) of Li et al. [18]; we can obtain the generalized Hadamard matrix M ( 6 , 3 ) in Lemma 2. Then, A ( 2 ) is a D ( 36 ;   3 35 ) (refer to Table A6) and B ( 2 ) is a D ( 3 ;   3 2 ) (refer to Table 1), then the generated design P ( 2 ) is a D ( 36 ;   3 70 ) (refer to Table A8). E f N O D P ( 2 ) = 12.52174 reaches the lower bound.
In the above theorems, there is a strict limitation that design B and the E ( f   N O D ) -optimal designs that can be constructed are all equi-level SSDs. We now introduce two methods for constructing E ( f   N O D ) -optimal mixed-level SSDs.
Lemma 3 
(Fang et al. [6], Theorem 4). Let X be a saturated orthogonal array, L n ( q m ) , and q > 2 be a prime power, and let n = q t , m = ( n 1 ) / ( q 1 ) , t 2 ; the E ( f   N O D ) -optimal mixed-level supersaturated design D ( p q t 1 ;   p 1 q m 1 ) is obtained from F S O A and the p ( 2 p < q ) level factor is orthogonal to the q level factor, and it is also an equidistant design.
Construction 4. 
Step 1. Let A be the design D ( s ;   s s 1 ) .
Step 2. Let B be D ( p q t 1 ;   p 1 q m 1 ) obtained by Lemma 3, where s = k [ p , q ] and [ p , q ] is the lowest common multiple of p and q, k 1 . Divide B into s sub-designs by rows; these sub-designs are denoted by b 1 , , b s , each with p q t 1 / s rows and m columns.
Step 3. Substitute the ith level of A by the ith sub-design of B, respectively.
Step 4. We obtain a new design P denoted by D ( p q t 1 ;   p s 1 q ( m 1 ) ( s 1 ) ) .
Theorem 3. 
The new design D ( p q t 1 ;   p s 1 q ( m 1 ) ( s 1 ) ) obtained by Constructon 4 is an E ( f   N O D ) -optimal SSD.
Proof of Theorem 3. 
The coincidence number between any two rows is a constant, λ = ( s 1 ) [ q t 1 + ( m 1 ) p q t 2 m ] / ( p q t 1 1 ) ; thus, D is E ( f   N O D ) -optimal. □
Example 3. 
Let s = 6 , p = 2 , q = 3 , t = 2 , m = 4 , then B ( 3 ) is the matrix D ( 6 ;   2 1 3 3 ) (refer to Table A1). Take design D ( 6 ;   6 5 ) (refer to Table A2) as A ( 3 ) ; thus, the generated design P ( 3 ) is D ( 6 ;   2 5 3 15 ) (refer to Table A3). E f N O D P ( 3 ) = 2.526316 reaches the lower bound.
Example 4. 
Let s = 12 , p = 3 , q = 4 , t = 2 , m = 5 . B ( 4 ) is the design D ( 12 ;   3 1 4 4 ) (refer Table A4), A ( 4 ) is the design D ( 12 ;   12 11 ) (refer to Table A5); thus, the generated design P ( 4 ) is the design D ( 12 ;   3 11 4 44 ) (refer to Table A7). E f N O D P ( 4 ) = 6.651852 reaches the lower bound.
Construction 5. 
Step 1. Let A be the design D ( s ;   s s 1 ) .
Step 2. Let B be the ( 0 2 q ( c , F ) , D 2 , , D 2 q ) in Corollary 3 of Li et al. [18], which can be denoted by D ( ( 2 q ) 2 ;   ( 2 q ) 1 q ( 2 q ) ( 2 q 1 ) p 1 m 1 p r m r ) , where q is a prime power, c = ( 0 , 1 , , 2 q 1 ) . D is the design obtained by deleting the first 2 q columns of M ( ( 2 q ) 2 , q ) in Lemma 2. Divide D into 2 q 1 sub-designs by columns, and D i is the ( i 1 ) -th sub-design, i = 2 , , 2 q . F is an equidistant design D ( 2 q ; p 1 m 1 p r m r ) with a constant coincidence number λ. s = k [ 2 q , p 1 , , p r ] . Divide B into s sub-designs by rows.
Step 3. Substitute the ith level of A by the ith sub-design of B, respectively.
Step 4. We obtain a new design P denoted by
D ( ( 2 q ) 2 ;   ( 2 q ) s 1 q ( 2 q ) ( 2 q 1 ) ( s 1 ) p 1 m 1 ( s 1 ) p r m r ( s 1 ) ) .
Theorem 4. 
If i = 1 r m i = λ + 2 q 3 . The new design P obtained by Construction 5 is an E ( f   N O D ) -optimal SSD.
Proof of Theorem 4. 
When i = 1 r m i = λ + 2 q 3 , design B is an equidistant design with a constant coincidence number, λ + 4 q 2 . The coincidence number of design P is a constant ( s 1 ) ( λ + 2 q 3 ) , so it is optimal. □
Example 5. 
Let A ( 5 ) be the design D ( 6 ;   6 5 ) (Table 3), F is the equidistant design D ( 6 ;   2 1 3 3 ) (refer Table A2). Let v = 3 , α = 2 , l = 2 , c = ( 0 , 1 , 2 , 3 , 4 , 5 ) , B ( 5 ) is the design D ( 36 ;   2 1 3 33 6 1 ) . Thus, the generated design P ( 5 ) is D ( 36 ;   2 5 3 165 6 5 ) . E f N O D P ( 5 ) = 14.90837 reaches the lower bound.
In fact, using the similar method of Construction 2, we can further expand Constructions 3–5 to obtain more new SSDs with flexible columns.
Corollary 2. 
Since the B designs in Constructions 3–5 have no completely aliased columns, the generated designs constructed by the above construction methods have no completely aliased columns.
Actually, most B designs in Construction 1 have no completely aliased columns, and the generated design constructed by Constructions 1 and 2 have no completely aliased columns.
Proof of Corollary 2. 
Let a i , b i , p i be the i-th column of design A , B , P and b l = ( b l 1 , , b l ( s 1 ) ) be the l-th sub-design of design B. Then the f N O D value between p i and p j is f N O D ( p i , p j ) = 2 2 n s 2 ( n 1 ) f N O D ( b i , b j ) . Where i = ( u 1 ) ( s 1 ) + i , j = ( u 1 ) ( s 1 ) + j , u = 1 , , ( 2 s ) n 1 , i , j = 1 , , s 1 and i j . It is easy to find that the f N O D for generating design P is determined by the f N O D of B. Because B performs well in its maximum value f N O D , there is no fully aliased column in the generated design P. □
In this section, we provided five construction methods to obtain many new equi- and mixed-level SSDs with a large number of inert factors by the substitution method. The proposed methods are easy to implement and most obtained SSDs have no fully aliased columns.

4. Variable Selection Method for SSD

For the problem of variable selection in SSDs, there remains a significant challenge. In this section, to facilitate a comprehensive comparison of variable selection methods for active factor identification on SSDs, a novel variable selection method, SSnet, is first used to perform factor screening on the Williams Rubber dataset, comparing it to seven other widely recognized and applicable methods for SSDs. These methods include Lq-penalty, Ada LASSO [22], SCAD [21], MCP [27], Ds [25], SVR-RFE [28], and ISIS [13]. This was done in order to validate the effectiveness and rationality of SSnet in factor identification. Then, through simulations, the constructed SSDs are used as simulated data to compare these seven different variable selection methods and provide power values for different scenarios of the active factor quantities. We conclude that SSnet has a strong ability to identify active factors on SSDs.

4.1. SSnet

Pokarowski et al. [31] introduced an innovative approach aimed at enhancing the performance of the LASSO method in predictive models with convex loss functions. This method involves systematically ordering the nonzero coefficients of the LASSO based on their absolute values under a specific penalty. Subsequently, the final model is selected from a compact nested family using GIC.
The SS algorithm employs two selection procedures for variable selection. Firstly, the TL method extracts only the absolute values of the LASSO estimator for variables surpassing a specific threshold. The second step, termed the SS procedure, identifies the minimum GIC value among nested families constructed using sequences comprising nonzero LASSO coefficients. Fundamentally, the SS algorithm can be viewed as a GIC-based LASSO approach with adaptive thresholding. To practically implement the SS algorithm, the SSnet algorithm utilizes the g l m n e t R package to calculate LASSO values across a range of penalty values. Subsequently, the optimal model is selected from a small family by minimizing the GIC.
In this paper, we demonstrate the applicability of the SSnet-based variable selection method to SSDs. We assess its performance through real-world examples and simulations and compare it with other popular methods without making general claims.

4.2. Practical Example

This example utilizes an SSD constructed by Lin [3] and the rubber data obtained from Williams [35]. The SSD presented in Table 4 is employed in this study. The design consists of 24 factors and 14 runs. However, as it contains two identical columns, namely 13 and 16, one of them needs to be eliminated. Consequently, the final analysis will retain the remaining 23 factors. In this case, column 13, corresponding to the 13-factor, was selected for deletion. Additionally, the labels of factors 14–24 were amended to 13–23.
The original analysis was based on Williams’ study of a 28-trial versus 24-factor Plackett–Burma design. The rubber data have been previously scrutinized by numerous researchers, including Abraham, Chipman, and Vijayan [36], who employed full subset regression to narrow down the search for the optimal subset within a model with five or fewer factors; Bettie, Fong, and Lin [37], who compared various model selection methods; Li and Lin [38], who adopted a stepwise SCAD approach; and Phoa et al. [25], who introduced the DS method. In this study, the SSnet algorithm is utilized to investigate this two-level supersaturated design. A concise summary of the analysis of the rubber data in each model is presented in Table 4, displaying the respective sizes and variables for each method, as well as the selected model.
The presence of factor 14 in all selected models highlights its significance, making it a noteworthy factor in the SSD. Comparing the various methods used, it becomes evident that SSnet efficiently identifies a significant set of factors while maintaining a smaller model size. This serves as solid evidence for the soundness and accuracy of the SSnet method to variable selection.

4.3. Simulation

To compare the performance of the proposed SSnet method with the other five methods, we generate data from a unique main effects linear model, y = X β + ϵ , with X being an n × m SSD matrix, and a random error, ϵ N ( 0 , 1 ) , i = 1 , 2 , , n ; y is the observed data. We consider two factor models: β of the active factor is generated by N ( μ , 1 ) , with an absolute value of 2 or 5 for μ , and β of the other inactive factors equal to 0. To demonstrate the performance of various methods in screening active factors under different sparsity levels, the three scenarios are used based on the number of active factors. Experience has shown that the number of runs used in the screening design should be at least twice the number of potentially significant effects. Therefore, the maximum number of active factors is set to be n / 2 . The analysis of SSDs is based on the principle of sparse effects [39]. This principle advocates that relatively few of the effects studied in the screening design are active, so the number of active factors in the effect sparse scenario is set to be less than the 1 / 3 of n / 2 ( n / 6 ). The number of active factors in the intermediate complexity scenario is between the 1 / 3 of n / 2 ( n / 6 ) and the 1 / 2 of n / 2 ( n / 4 ). Meanwhile, this selection and construction of three scenarios is similar to the setting of Marley and Woods [40] and Drosou and Koukouvinos [13,28], which makes it convenient to compare with methods in existing papers. The active factors are randomly selected from X and the number of that is set according to the following scenarios:
  • Scenario I (Effect sparse scenario): p = 1 , , n / 6 .
  • Scenario II (Moderately complex scenarios): p = n / 6 + 1 , , n / 4 .
  • Scenario III (Massive effect scenario): p = n / 4 + 1 , , n / 2 .
The number of active factor for Scenario I is taken as 1 n / 6 , with each quantity of active factors simulated 1000 times and averaged. We use Mean power to represent the overall average of the corresponding power under each quantity of active factors. Similarly, for Scenario II, take n / 6 + 1 n / 4 , and for Scenario III, take n / 4 + 1 n / 2 . Without loss of generality, we select three different SSDs as the input matrix in this section, shown respectively as Examples 2, 4, and 5.
Example 6. 
In this example, we choose the generated design D ( 36 ; 3 70 ) of Example 2 as the input SSD matrix X. We choose the number of active factors to be 1∼6, 7∼9, and 10∼18 in three different scenarios, respectively.
Figure 1 shows the mean power of SSD D ( 36 ; 3 70 ) in various scenarios. Model 1 is a model with random coefficients of small magnitude equal to 2 and variance equal to 1. This leads to a more complicated model with coefficients as small as the error term. Model 2 is a model with random coefficients but with relatively large magnitude equal to 5 and variance equal to 1.
Scenario I.
In Model 1, SSnet shows the highest mean power of approximately 97%; SCAD, Lq-penalty, MCP, Ds, ISIS, and Ada LASSO have almost the same mean power of approximately 90%; while SVR-RFE performs poorly with a mean power of 85%. Similarly, SSnet in Model 2 shows excellent performance, with mean power almost reaching 1; other methods have also reached over 95%. These figures indicate that the reliability of these methods is consistently high, even when the activity factors are sparse.
Scenario II.
Similar to Scenario I, SSnet exhibits the highest mean power in this scenario. In Model 1, only the mean power of SSnet and SCAD reached 90%, followed by MCP, DS, and ISIS, with a mean power above 85%, and Lq-penalty, Ada LASSO, and SVR-RFE, with a mean power around 80%. In Model 2, SSnet shows the highest mean power of approximately 98%; the other methods, except for ISIS, had a mean power above 90%. It is evident that SSnet dominates among these eight methods in moderately complex situations.
Scenario III.
In massive effect scenarios, the differences between each method are more significant, exhibiting different mean power levels. SSnet stands out, with a significantly higher mean power compared to the other methods.
In Example 6, SSnet demonstrated excellent factor screening ability on equi-level SSDs. Next, we will continue to use mixed-level SSDs to enhance the persuasiveness of this conclusion.
Example 7. 
In this example, the generation design D ( 12 ; 3 11 4 44 ) of Example 3 is considered as the input SSD matrix. We choose the number of active factors to be 1∼2, 3, and 4∼6 in three different scenarios for the two models, respectively. The results, focusing on mean power, are shown in Figure 2.
In Scenario I, SSnet has the highest mean power, with only SSnet achieving a mean power value of 90% in Model 1, and SSnet approaching 1 in Model 2. In both Scenario II and Scenario III, SSnet performed the best in terms of mean power. In this example, the mean power of SSnet is optimal, significantly surpassing other methods.
From Example 7, it can be seen that SSnet has a significant ability to identify active factors on mixed-level SSDs. Let us further verify this conclusion.
Example 8. 
In this example, consider the generated design D ( 36 ; 2 5 3 165 6 5 ) of Example 5 as the input SSD matrix. We choose the number of active factors to be 1∼6, 7∼9, and 10∼18 in three different scenarios for both models; the results, focusing on mean power, are shown in Figure 3.
It is clear that Model 1 always exhibits lower mean power than Model 2, indicating that the model with larger random coefficients is more capable of recognizing in scenarios with sparse active factors. SSnet always has the highest average power in scenes 2 and 3. Although Lq-penalty has a slightly higher average power than SSnet in Scenario I, its mean power in Scenarios II and III decreases significantly, especially in Scenario III, which is only 30%, which is very unstable. On the other hand, SSnet has a high mean power in all scenarios and is robust.
In conclusion, whether in equi-level design or mixed-level SSDs, the ability of SSnet to identify active factors has been proven to be robust, maintaining consistently high mean power across all scenarios. It is evident that SSnet is particularly sensitive to random coefficients. It performs better in scenarios where active factors are sparse in SSDs and where there are larger random coefficients.

5. Conclusions

SSDs are commonly used to select a small number of active factors from a large number of candidates. In this paper, many new equi- and mixed-level SSDs with a large number of inert factors are constructed by the substitution method of the generalized Hadamard matrix, and most SSDs have no fully aliased columns. The proposed methods are easy to implement and the number of columns is more flexible. These SSDs can serve as a good check when identifying active factors using variable selection methods. Note that the substitution method can generate large designs using small equidistant designs. Changing the source small design to other types of design can result in more SSDs or other space-filling designs, which is worthy of further study. This paper investigates the application of SSnet as a model selection method in SSDs, marking its novelty as it is employed for the first time in this context. The main emphasis during screening tests lies in accurately identifying valid effects rather than excluding invalid ones, focusing primarily on power. The major advantage of SSnet lies in its capacity to deliver robust performance against power in SSDs despite its simplicity. Simulation studies have consistently demonstrated that SSnet produces excellent outcomes and can be meaningfully compared with the existing analysis methods proposed in the literature. Additionally, it demonstrates exceptional performance, not only in scenarios with extremely sparse real models but also in cases with moderately active effects. Furthermore, this approach exhibits the capability to effectively handle large-scale designs comprising numerous factors, including both equi- and mixed-level designs.

Author Contributions

Conceptualization, M.L. and X.H.; methodology, M.L. and X.W.; software, X.W. and X.H.; validation, W.Z.; formal analysis, M.L and X.W.; investigation, W.Z.; resources, M.L.; data curation, X.H.; writing—original draft preparation, X.H. and M.L.; writing—review and editing, M.L. and X.W.; visualization, X.W.; supervision, W.Z.; project administration, M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Shandong Provincial Natural Foundation of China (ZR2021QA053, ZR2021QA044) and the National Natural Science Foundation of China (12101346).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. D ( 6 ; 2 1 3 3 ) .
Table A1. D ( 6 ; 2 1 3 3 ) .
1234
10012
20120
30201
41021
51102
61210
Table A2. D ( 6 ; 6 5 ) .
Table A2. D ( 6 ; 6 5 ) .
12345
100000
212345
323451
434512
545123
651234
Table A3. D ( 6 ; 2 5 3 15 ) .
Table A3. D ( 6 ; 2 5 3 15 ) .
1234567891011121314151617181920
100120012001200120012
201200201102111021210
302011021110212100120
410211102121001200201
511021210012002011021
612100120020110211102
Table A4. D ( 12 ; 3 1 4 4 ) .
Table A4. D ( 12 ; 3 1 4 4 ) .
12345
100000
201111
302222
403333
513021
611203
712130
810312
920123
1023210
1121032
1222301
Table A5. D ( 12 ; 12 11 ) .
Table A5. D ( 12 ; 12 11 ) .
1234567891011
100000000000
21234567891011
32345678910111
43456789101112
54567891011123
65678910111234
76789101112345
87891011123456
98910111234567
109101112345678
111011123456789
121112345678910
Table A6. D ( 36 ; 3 35 ) .
Table A6. D ( 36 ; 3 35 ) .
1234567891011121314151617181920212223242526272829303132333435
100000000000000000000000000000000000
212120012120012120012120012120012120
321102021102021102021102021102021102
422011022011022011022011022011022011
501221001221001221001221001221001221
610212010212010212010212010212010212
700000111111222222111111222222000000
812120120201201012120201201012012120
921102102210210021102210210021021102
1022011100122211200100122211200022011
1101221112002220110112002220110001221
1210212121020202101121020202101010212
1300000222222111111111111000000222222
1412120201012120201120201012120201012
1521102210021102210102210021102210021
1622011211200100122100122022011211200
1701221220110112002112002001221220110
1810212202101121020121020010212202101
1900000222222222222000000111111111111
2012120201012201012012120120201120201
2121102210021210021021102102210102210
2222011211200211200022011100122100122
2301221220110220110001221112002112002
2410212202101202101010212121020121020
2500000000000111111222222222222111111
2612120012120120201201012201012120201
2721102021102102210210021210021102210
2822011022011100122211200211200100122
2901221001221112002220110220110112002
3010212010212121020202101202101121020
3100000111111000000222222111111222222
3212120120201012120201012120201201012
3321102102210021102210021102210210021
3422011100122022011211200100122211200
3501221112002001221220110112002220110
3610212121020010212202101121020202101
Table A7. D ( 12 ; 3 11 4 44 ) .
Table A7. D ( 12 ; 3 11 4 44 ) .
12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455
10000000000000000000000000000000000000000000000000000000
20111102222033331302111203121301031220123232102103222301
30222203333130211120312130103122012323210210322230101111
40333313021112031213010312201232321021032223010111102222
51302111203121301031220123232102103222301011110222203333
61120312130103122012323210210322230101111022220333313021
71213010312201232321021032223010111102222033331302111203
81031220123232102103222301011110222203333130211120312130
92012323210210322230101111022220333313021112031213010312
102321021032223010111102222033331302111203121301031220123
112103222301011110222203333130211120312130103122012323210
122230101111022220333313021112031213010312201232321021032
Table A8. D ( 36 ; 3 70 ) .
Table A8. D ( 36 ; 3 70 ) .
12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970
10000000000000000000000000000000000000000000000000000000000000000000000
21221122100001221122100001221122100001221122100001221122100001221122100
32112120021002112120021002112120021002112120021002112120021002112120021
42121001212002121001212002121001212002121001212002121001212002121001212
50012212112000012212112000012212112000012212112000012212112000012212112
61200211221001200211221001200211221001200211221001200211221001200211221
70000000000121212121212212121212121121212121212212121212121000000000000
81221122100122100210012210012001221122100210012210012001221001221122100
92112120021120021211200211200002112120021211200211200002112002112120021
102121001212120000122121211212210000120000122121211212210000002121001212
110012212112121221000021212100121200121221000021212100121200000012212112
121200211221122112002100210021120012122112002100210021120012001200211221
130000000000212121212121121212121212121212121212000000000000212121212121
141221122100210012001221122100210012122100210012001221122100210012001221
152112120021211200002112120021211200120021211200002112120021211200002112
162121001212211212210000120000122121120000122121002121001212211212210000
170012212112212100121200121221000021121221000021000012212112212100121200
181200211221210021120012122112002100122112002100001200211221210021120012
190000000000212121212121212121212121000000000000121212121212121212121212
201221122100210012001221210012001221001221122100122100210012122100210012
212112120021211200002112211200002112002112120021120021211200120021211200
222121001212211212210000211212210000002121001212120000122121120000122121
230012212112212100121200212100121200000012212112121221000021121221000021
241200211221210021120012210021120012001200211221122112002100122112002100
250000000000000000000000121212121212212121212121212121212121121212121212
261221122100001221122100122100210012210012001221210012001221122100210012
272112120021002112120021120021211200211200002112211200002112120021211200
282121001212002121001212120000122121211212210000211212210000120000122121
290012212112000012212112121221000021212100121200212100121200121221000021
301200211221001200211221122112002100210021120012210021120012122112002100
310000000000121212121212000000000000212121212121121212121212212121212121
321221122100122100210012001221122100210012001221122100210012210012001221
332112120021120021211200002112120021211200002112120021211200211200002112
342121001212120000122121002121001212211212210000120000122121211212210000
350012212112121221000021000012212112212100121200121221000021212100121200
361200211221122112002100001200211221210021120012122112002100210021120012

References

  1. Satterthwaite, F. Random balance experimentation. Technometrics 1959, 1, 111–137. [Google Scholar] [CrossRef]
  2. Booth, K.H.V.; Cox, D.R. Some systematic supersaturated designs. Technometrics 1962, 4, 489–495. [Google Scholar] [CrossRef]
  3. Lin, D.K.J. A new class of supersaturated designs. Technometrics 1993, 35, 28–31. [Google Scholar] [CrossRef]
  4. Wu, C.F.J. Construction of supersaturated designs through partially aliased interactions. Biometrika 1993, 80, 661–669. [Google Scholar] [CrossRef]
  5. Fang, K.T.; Lin, D.K.J.; Liu, M.Q. Optimal Mixed-Level Supersaturated Design and Computer Experiment; Technical Report MATH-286; Hong Kong Baptist University: Hong Kong, China, 2000. [Google Scholar]
  6. Fang, K.T.; Lin, D.K.J.; Liu, M.Q. Optimal mixed-level supersaturated design. Metrika 2003, 58, 279–291. [Google Scholar] [CrossRef]
  7. Yamada, S.; Lin, D.K.J. Construction of mixed-level supersaturated design. Metrika 2002, 56, 205–214. [Google Scholar] [CrossRef]
  8. Zou, N.; Qin, H. A Note on Supersaturated Design. Acta Math. Appl. Sin. Engl. Ser. 2022, 38, 729–739. [Google Scholar] [CrossRef]
  9. Georgiou, S.; Koukouvinos, C.; Mantas, P. On multi-level supersaturated designs. J. Stat. Plan. Inference 2006, 136, 2805–2819. [Google Scholar] [CrossRef]
  10. Liu, M.Q.; Cai, Z.Y. Construction of mixed-level supersaturated designs by the substitution method. Stat. Sin. 2009, 19, 1705–1719. [Google Scholar]
  11. Sun, F.S.; Lin, D.K.J.; Liu, M.Q. On construction of optimal mixed-level supersaturated designs. Ann. Stat. 2011, 39, 1310–1333. [Google Scholar] [CrossRef]
  12. Georgiou, S.D. Supersaturated designs: A review of their construction and analysis. J. Stat. Plan. Inference 2014, 144, 92–109. [Google Scholar] [CrossRef]
  13. Drosou, K.; Koukouvinos, C. Sure independence screening for analyzing supersaturated designs. Commun. Stat.-Simul. Comput. 2018, 48, 1979–1995. [Google Scholar] [CrossRef]
  14. Jones, B.; Lekivetz, R.; Majumdar, D.; Nachtsheim, C.J.; Stallrich, J.W. Construction, properties, and analysis of group-orthogonal supersaturated designs. Technometrics 2020, 62, 403–414. [Google Scholar] [CrossRef]
  15. Chatterjee, K.; Koukouvinos, C. Construction of mixed-level supersaturated split-plot designs. Metrika 2021, 84, 949–967. [Google Scholar] [CrossRef]
  16. Khurana, S.; Banerjee, S. The development of a method for constructing supersaturated design and nearly orthogonal design with mixed-level orthogonal design. Curr. Top. Math. Comput. Sci. 2021, 10, 46–52. [Google Scholar]
  17. Daraz, U.; Chen, E.; Tang, Y. Supersaturated designs with less β-aberration. Commun. Stat.-Theory Methods 2022. [Google Scholar] [CrossRef]
  18. Li, M.; Liu, M.Q.; Sun, F.S.; Zhang, D. Construction of optimal supersaturated designs via generalized Hadamard matrices. Commun. Stat.-Theory Methods Taylor Fr. J. 2022, 51, 2565–2579. [Google Scholar] [CrossRef]
  19. Tibshirani, R. Regression Shrinkage and Selection via the LASSO. J. R. Stat. Soc. Ser. B-Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  20. Fan, J.; Li, R. Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties. J. Am. Stat. Assoc. 2001, 96, 1348–1360. [Google Scholar] [CrossRef]
  21. Li, R.; Lin, D.K.J. Data analysis in supersaturated designs. Stat. Probab. Lett. 2002, 59, 135–144. [Google Scholar] [CrossRef]
  22. Zou, H. The Adaptive Lasso and Its Oracle Properties. J. Am. Stat. Assoc. 2006, 101, 1418–1429. [Google Scholar] [CrossRef]
  23. Fan, J.; Lv, J. Sure Independence Screening for Ultrahigh Dimensional Feature Space. J. R. Stat. Soc. Ser. B Stat. Methodol. 2008, 70, 849–911. [Google Scholar] [CrossRef] [PubMed]
  24. Candes, E.; Tao, T. The Dantzig selector: Statistical estimation when p is much larger than n. Ann. Stat. 2007, 35, 2313–2351. [Google Scholar]
  25. Phoa, F.K.H.; Pan, Y.H.; Xu, H. Analysis of supersaturated designs via the Dantzig selector. J. Stat. Plan. Inference 2009, 139, 2362–2372. [Google Scholar] [CrossRef]
  26. Das, U.; Gupta, S.; Gupta, S. Screening active factors in supersaturated designs. Comput. Stat. Data Anal. 2014, 77, 223–232. [Google Scholar] [CrossRef]
  27. Zhang, C.H. Nearly unbiased variable selection under minimax concave penalty. Ann. Stat. 2010, 38, 894–942. [Google Scholar] [CrossRef] [PubMed]
  28. Drosou, K.; Koukouvinos, C. A new variable selection method based on SVM for analyzing supersaturated designs. J. Qual. Technol. 2019, 51, 21–36. [Google Scholar] [CrossRef]
  29. Errore, A.; Jones, B.; Li, W.; Nachtsheim, C.J. Using definitive screening designs to identify active firstand second-order factor effects. J. Qual. Technol. 2017, 49, 244–264. [Google Scholar] [CrossRef]
  30. Ueda, A.; Yamada, S. Discussion on supersaturated design and its data analysis method in computer experiments. Total Qual. Sci. 2022, 7, 60–68. [Google Scholar] [CrossRef]
  31. Pokarowski, P.; Rejchel, W.; Soltys, A.; Frej, M.; Mielniczuk, J. Improving Lasso for model selection and prediction. Scand. J. Stat. 2022, 49, 831–863. [Google Scholar] [CrossRef]
  32. Bose, R.C.; Bush, K.A. Orthogonal arrays of strength two and three. Ann. Math. Stat. 1952, 23, 508–524. [Google Scholar] [CrossRef]
  33. Hedayat, A.S.; Sloane, N.J.A.; Stufken, J. Orthogonal Arrays: Theory and Applications. Springer Ser. Stat. Springer-Verlag. J. Stat. Plan. Inference 1999, 139, 2362–2372. [Google Scholar]
  34. Fang, K.T.; Ge, G.N.; Liu, M.Q.; Qin, H. Combinatorial constructions for optimal supersaturated designs. Discret. Math. 2004, 279, 191–202. [Google Scholar] [CrossRef]
  35. Williams, K.R. Designed experiments. Rubber Age 1968, 100, 65–71. [Google Scholar]
  36. Abraham, B.; Chipman, H.; Vijayan, K. Some Risks in the Construction and Analysis of Supersaturated Designs. Ann. Stat. 1999, 41, 135–141. [Google Scholar] [CrossRef]
  37. Beattie, S.D.; Fong, D.K.H.; Lin, D.K.J. A Two-Stage Bayesian Model Selection Strategy for Supersaturated Designs. Technometrics 2002, 44, 55–63. [Google Scholar] [CrossRef]
  38. Li, R.; Lin, D.K.J. Analysis Methods for Supersaturated Design: Some Comparisons. J. Data Sci. 2003, 1, 249–260. [Google Scholar] [CrossRef]
  39. Box, G.; Meyer, R. Dispersion effects from fractional designs. Technometrics 1986, 28, 19–27. [Google Scholar] [CrossRef]
  40. Marley, C.J.; Woods, D.C. A comparison of design and model selection methods for supersaturated experiments. Comput. Stat. Data Anal. 2010, 54, 3158–3167. [Google Scholar] [CrossRef]
Figure 1. Mean power in each scenario for SSD D ( 36 ;   3 70 ) .
Figure 1. Mean power in each scenario for SSD D ( 36 ;   3 70 ) .
Axioms 13 00179 g001
Figure 2. Mean power in each scenario for SSD D ( 12 ;   3 11 4 44 ) .
Figure 2. Mean power in each scenario for SSD D ( 12 ;   3 11 4 44 ) .
Axioms 13 00179 g002
Figure 3. Mean power in each scenario for SSD D ( 36 ;   2 5 3 165 6 5 ) .
Figure 3. Mean power in each scenario for SSD D ( 36 ;   2 5 3 165 6 5 ) .
Axioms 13 00179 g003
Table 1. D ( 3 ;   3 2 ) .
Table 1. D ( 3 ;   3 2 ) .
12
100
212
321
Table 2. D ( 3 ;   3 4 ) .
Table 2. D ( 3 ;   3 4 ) .
1234
10000
21221
32112
Table 3. Williams rubber data.
Table 3. Williams rubber data.
1234567891011121314151617181920212223Response
1111−1−1−111111−1−1−111−1−11−1−1−11133
21−1−1−1−1−1111−1−1−1111−11−1−111−1−162
311−111−1−1−1−11−111111−1−1−1−111−145
411−11−11−1−1−111−1−111−1111−1−1−1−152
5−1−11111−111−1−1−1−1111−1−11−111156
6−1−111111−1111−111−1111111−1−147
7−1−1−1−11−1−11−11−111−1111111−1−1188
8−111−1−11−11−11−1−1−1−1−1−1−11−1111−1193
9−1−1−1−1−111−1−1−111−11−111−1−1−1−11132
101111−1111−1−1−1111−11−11−11−1−1153
11−11−111−1−111−11−11−1−1−111−1−1−111276
121−1−1−1111−11111−1−11−1−11−11111145
1311111−11−11−1−11−1−1−1−11−111−11−1130
14−1−11−1−1−1−1−1−1−1111−1−1−1−1−11−11−1−1127
Table 4. Results of significant factor screening for each model.
Table 4. Results of significant factor screening for each model.
Variable Selection MethodSizeSignificant Factors
Williams44 10 14 19
SCAD81 4 7 10 11 12 14 19
Ds114
MCP312 14 19
SVR-RFE77 8 9 11 14 15 19
Ada LASSO62 5 8 14 16 22
Lq-penalty214 16
ISIS69 14 15 18 19 20
SSnet74 7 9 12 14 15 19
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hong, X.; Zhou, W.; Wang, X.; Li, M. Construction and Analysis for the Optimal Supersaturated Designs with a Large Number of Inert Factors. Axioms 2024, 13, 179. https://doi.org/10.3390/axioms13030179

AMA Style

Hong X, Zhou W, Wang X, Li M. Construction and Analysis for the Optimal Supersaturated Designs with a Large Number of Inert Factors. Axioms. 2024; 13(3):179. https://doi.org/10.3390/axioms13030179

Chicago/Turabian Style

Hong, Xiaoqing, Weiping Zhou, Xiao Wang, and Min Li. 2024. "Construction and Analysis for the Optimal Supersaturated Designs with a Large Number of Inert Factors" Axioms 13, no. 3: 179. https://doi.org/10.3390/axioms13030179

APA Style

Hong, X., Zhou, W., Wang, X., & Li, M. (2024). Construction and Analysis for the Optimal Supersaturated Designs with a Large Number of Inert Factors. Axioms, 13(3), 179. https://doi.org/10.3390/axioms13030179

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop