Next Article in Journal
Chaotic Behaviour, Sensitivity Assessment, and New Analytical Investigation to Find Novel Optical Soliton Solutions of M-Fractional Kuralay-II Equation
Previous Article in Journal
Variational Bayesian Variable Selection in Logistic Regression Based on Spike-and-Slab Lasso
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Sparse Twin Parametric Insensitive Support Vector Regression Model

1
College of Mathematics and Information Science, Zhengzhou University of Light Industry, Zhengzhou 450002, China
2
School of Science and Technology, University of Camerino, 62032 Camerino, Italy
3
College of Software Engineering, Zhengzhou University of Light Industry, Zhengzhou 450001, China
*
Authors to whom correspondence should be addressed.
Mathematics 2025, 13(13), 2206; https://doi.org/10.3390/math13132206
Submission received: 2 June 2025 / Revised: 1 July 2025 / Accepted: 3 July 2025 / Published: 6 July 2025

Abstract

This paper proposes a novel sparse twin parametric insensitive support vector regression (STPISVR) model, designed to enhance sparsity and improve generalization performance. Similar to twin parametric insensitive support vector regression (TPISVR), STPISVR constructs a pair of nonparallel parametric insensitive bound functions to indirectly determine the regression function. The optimization problems are reformulated as two sparse linear programming problems (LPPs), rather than traditional quadratic programming problems (QPPs). The two LPPs are originally derived from initial L1-norm regularization terms imposed on their respective dual variables, which are simplified to constants via the Karush–Kuhn–Tucker (KKT) conditions and consequently disappear. This simplification reduces model complexity, while the constraints constructed through the KKT conditions— particularly their geometric properties—effectively ensure sparsity. Moreover, a two-stage hybrid tuning strategy—combining grid search for coarse parameter space exploration and Bayesian optimization for fine-grained convergence—is proposed to precisely select the optimal parameters, reducing tuning time and improving accuracy compared to a singlemethod strategy. Experimental results on synthetic and benchmark datasets demonstrate that STPISVR significantly reduces the number of support vectors (SVs), thereby improving prediction speed and achieving a favorable trade-off among prediction accuracy, sparsity, and computational efficiency. Overall, STPISVR enhances generalization ability, promotes sparsity, and improves prediction efficiency, making it a competitive tool for regression tasks, especially in handling complex data structures.

1. Introduction

Since support vector machine (SVM) was first proposed by Vapnik et al. (1995) [1], it has become one of the most widely used machine learning algorithms, renowned for its excellent statistical properties and superior generalization performance. Support vector regression (SVR) [2,3,4,5,6,7,8], an extension of SVM for regression, applies the concept of support vectors (SVs) to handle regression tasks. By mapping samples from the input space into a high-dimensional feature space, SVR constructs a regression function with strong generalization capability. The SVs are then utilized to predict outcomes, effectively modeling nonlinear relationships within the data.
Although SVR has been successfully applied across various fields, it requires solving a large single quadratic programming problem (QPP) with three sets of constraints, each as many as the data points. As the dataset size grows and the feature dimensions increase, both the computational complexity and storage requirements of SVR escalate significantly, posing substantial challenges for large-scale data applications. To overcome these limitations, researchers have proposed numerous SVR variations from different perspectives [9,10,11,12,13], aiming to enhance its performance.
Among these variations, Peng [9] proposed twin support vector regression (TSVR), inspired by the concept of the twin support vector machine (TWSVM) [14]. TSVR indirectly optimizes the regression function through a pair of ϵ -insensitive down-bound and up-bound functions. TSVR fundamentally differs from classical SVR in that it solves two smaller-sized QPPs instead of a single large QPP, as required by traditional SVR. This strategy theoretically enables TSVR to operate faster than classical SVR [9]. However, in practice, the prediction speed of TSVR is noticeably slower than that of classical SVR due to the formulation of its augmented vectors [11]. Additionally, TSVR assumes a uniform noise level across the entire domain of the training data, which limits its applicability in more complex scenarios. In the same year, Hao [12] introduced a novel parametric-insensitive v-support vector regression (par-v-SVR) by incorporating a parametric-insensitive loss function into the framework of v-support vector regression (v-SVR) [13]. This approach is better suited for datasets with heteroscedastic noise compared to classical SVR or v-SVR. Nevertheless, despite this improvement, par-v-SVR still involves solving a large QPP, resulting in high computational complexity and a learning speed comparable to that of classical SVR.
Based on the twin parameter margin support vector machine (TPMSVM) [15], Peng [11] proposed the twin parametric-insensitive support vector regression (TPISVR), which successfully combines the advantages of both TSVR [9] and par-v-SVR [12]. The core strategy of TPISVR involves solving two smaller-sized QPPs, similar to TSVR, which significantly enhances its learning speed compared to par-v-SVR and classical SVR. At the same time, TPISVR effectively handles heteroscedastic noise structures, similar to par-v-SVR. An additional advantage of TPISVR is its parameters can control the fractions of SVs and errors of two bounds. Moreover, TPISVR achieves comparable generalization performance to par-v-SVR, TSVR, and SVR. However, similar to TSVR, TPISVR loses its sparsity due to the formulation of their augmented vectors [11], which results in slower prediction speed and suggests that its generalization ability still requires further improvement.
In recent years, sparse learning has emerged as a significant research direction in the fields of machine learning. Its primary objective is to optimize the computational efficiency of models while maintaining, or even enhancing, their predictive performance. This is particularly crucial for an algorithm where prediction speed is a decisive factor, especially in applications requiring real-time processing. As powerful tools for supervised learning, SVM, SVR, and their variations have attracted significant attention. The core idea behind these models is to reduce the number of SVs or features, enhance the sparsity of model solutions, and improve prediction speed. Researchers have explored various sparse methods from different perspectives, resulting in a range of innovative approaches aimed at improving operational efficiency, interpretability, and generalization ability. Among these, sparse regularization techniques have attracted particular attention due to their ability to induce sparsity [16,17,18,19,20]. Significant progress has been made by incorporating different sparse norms and integrating them with additional algorithmic strategies, thereby advancing the field of sparse learning [19,20,21,22,23,24,25]. For instance, Huang et al. [19] introduced a sparse learning algorithm for support vector machines (SVM) by incorporating a 0-pseudonorm regularization term. Although this method effectively promotes sparsity, it leads to significant computational challenges, requiring iterative reweighted optimization techniques to find approximate solutions. To improve sparsity and model interpretability, Bi et al. [20] proposed a two-stage sparse SVM framework that applies L1-norm regularization to both the primal weights and the dual variables. By reformulating the optimization as a linear programming problem (LPP), their method effectively reduces dimensionality and the number of SVs while improving generalization. However, representing the L1-norm requires introducing auxiliary variables, which increases the number of optimization variables and thus the overall algorithmic complexity. Nevertheless, the transformation of the original optimization problem into an LPP represents a powerful idea, as it enables efficient sparse model construction and effective variable selection within a well-understood optimization framework [26,27]. Building upon this foundation, Qu et al. (2024) [24] further investigated solution sparsity in the linear case. They proposed the Lin-STPMSVM model, which directly applies the L1-norm to the dual variables, offering a more efficient framework for sparse classification.
Despite significant advancements in sparsity research, real-world applications still face several key challenges, such as enhancing sparsity while maintaining model stability, balancing complexity and performance, and effectively handling complex data structures, including heteroscedastic noise, nonuniform distributions, and diverse sparsity requirements. Motivated by these challenges and inspired by prior work, we propose a sparse twin parametric insensitivity support vector regression model (STPISVR). Leveraging the strengths of the TPISVR framework, STPISVR introduces structural innovations to enhance solution sparsity and improve prediction efficiency.
The main contributions of this work are summarized as follows:
  • An effective STPISVR model is introduced which encourages sparse solutions while retaining TPISVR’s ability for complex data structures, including heteroscedasticity.
  • The optimization problems are reformulated as two sparse LPPs with constraints derived from the KKT conditions. One of these constraints corresponds to an L1-norm expression, which is shown to be a constant through further analysis of the KKT conditions. As a result, the explicit L1-norm regularization term is removed, and the resulting LPPs inherently produce sparse solutions due to their underlying geometric structure.
  • This reformulation reduces computational complexity via two mechanisms: enhanced solution sparsity induced by the LPP structure and improved efficiency in solving LPPs instead of traditional QPPs.
  • A two-stage hybrid parameter tuning strategy is adopted to improve both tuning accuracy and computational efficiency.
  • Extensive experiments demonstrate that STPISVR significantly reduces SVs, enabling faster predictions without sacrificing accuracy, and achieves a superior trade-off among accuracy, sparsity, and computational efficiency, especially for complex data structures.
This paper is organized as follows: Section 2 provides a brief overview of related works; Section 3 introduces the STPISVR model and presents its relevant properties; Section 4 offers a discussion and comparison with several related models; in Section 5, we conduct numerical experiments on both synthetic and benchmark datasets to evaluate the performance of our sparse model; and finally, Section 6 presents conclusions and outlines directions for future research.

2. Related Works

In this section, we briefly review several related models that form the foundation for our proposed sparse method. For clarity and ease of reference, Table 1 lists some important notations used throughout this paper. In addition, all vectors in this article are assumed to be column vectors unless otherwise stated.

2.1. SVR Model

The classical SVR aims to determine a linear regression function in the feature space:
f ( x ) = w T φ ( x ) + θ ,
by introducing the kernel technique, the primal problem for classical SVR in nonlinear cases (for linear cases, the kernel function φ ( · ) reduces to the identity function; the other models discussed later follow a similar approach) [2,3,4,5,6,7] is given by:
min w , θ , ξ + , ξ 1 2 w 2 2 + ρ e T ( ξ + + ξ ) subject   to Y ( φ ( X ) w + θ e ϵ e + ξ + ) , ξ + 0 , ( φ ( X ) w + θ e ) Y ϵ e + ξ , ξ 0 .
By applying KKT conditions [5], the Wolfe dual problems [28] of SVR are
min α + , α 1 2 ( α + α ) T K ( α + α ) Y T ( α + α ) + ϵ e T ( α + + α ) subject   to e T ( α + α ) = 0 , 0 α + , α ρ e .
After solving the above optimization problem, we can compute
w = φ ( X ) T ( α + α )
and
θ = 1 2 1 | N + | p N + y p w T φ ( x p ) + 1 | N | q N y q w T φ ( x q ) ,
where N ± are the index sets satisfying α p + ( 0 , ρ ) , p I , or α q ( 0 , ρ ) , q I . Finally, the regression functions are given by:
f ( x ) = w T φ ( x ) + θ = ( α + α ) T K ( X , x ) + θ .

2.2. TPISVR Model

The following two nonparallel functions in the feature space H are considered for the TPISVR model [11]:
f d ( x ) = ( w d ) T φ ( x ) + θ d , f u ( x ) = ( w u ) T φ ( x ) + θ u .
The function f d ( x ) and f u ( x ) are called the parametric insensitive down-bound function and up-bound function regression, respectively. The primal problems of TPISVR are as follows:
min w d , θ d , ξ d 1 2 w d 2 2 v d | I | e T ( φ ( X ) w d + θ d e ) + ρ d | I | e T ξ d subject   to Y φ ( X ) w d + θ d e ξ d , ξ d 0
and
min w u , θ u , ξ u 1 2 w u 2 2 + v u | I | e T ( φ ( X ) w u + θ u e ) + ρ u | I | e T ξ u subject   to Y φ ( X ) w u + θ u e + ξ u , ξ u 0 .
By introducing the Lagrangian function for (8) and (9), respectively, and incorporating the KKT conditions [11], we obtain the dual problems of TPISVR:
min α d 1 2 ( α d ) T K α d v d | I | e T K α d + Y T α d subject   to 0 α d ρ d | I | e , e T α d = v d
and
min α u 1 2 ( α u ) T K α u v u | I | e T K α u Y T α u subject   to 0 α u ρ u | I | e , e T α u = v u .
After optimizing the dual problems (10) and (11), we can compute
w d = φ ( X ) T v d | I | e α d , w u = φ ( X ) T v u | I | e + α u ,
and
θ d = 1 | N d | p N d ( y p ( w d ) T φ ( x p ) ) , θ u = 1 | N u | q N u ( y q ( w u ) T φ ( x q ) ) ,
where N d and N u are the index sets satisfying α p d ( 0 , ρ d | I | ) , p I , and α q u ( 0 , ρ u | I | ) , q I , respectively. Then
f d ( x ) = ( w d ) T φ ( x ) + θ d = v d | I | e T K ( X , x ) ( α d ) T K ( X , x ) + θ d
and
f u ( x ) = ( w u ) T φ ( x ) + θ u = v u | I | e T K ( X , x ) + ( α u ) T K ( X , x ) + θ u .
Finally, the estimated regression function of TPISVR is as follows:
f ( x ) = 1 2 f d ( x ) + f u ( x ) = 1 2 α u α d + v d v u | I | e T K ( X , x ) + 1 2 ( θ d + θ u ) .

3. STPISVR

In this section, we introduce our STPISVR model and present some properties of it. For ease of comparison, we first propose the sparse SVR (SSVR) model, which will be formally shown in Equation (19). By incorporating the KKT conditions, we simplify the model formulation, making its interpretation more intuitive.

3.1. SSVR Model

We substitute (4) into the constraints of (2) first, and then, replace the first term 1 2 w 2 2 of the objective function in (2) with 1 2 ( α + 1 + α 1 ) . Finally, we combine with the KKT conditions (see the constraints in (3)) to obtain
min α + , α , ξ + , ξ , θ 1 2 ( α + 1 + α 1 ) + ρ e T ( ξ + + ξ ) subject   to Y K ( α + α ) θ e ϵ e + ξ + , ξ + 0 , K ( α + α ) + θ e Y ϵ e + ξ , ξ 0 , e T ( α α + ) = 0 , 0 α + , α ρ e .
Problem (17) is the embryonic form of our SSVR model. From the constraints e T ( α α + ) = 0 and α + 0 , α 0 , we have
α + 1 + α 1 = 2 e T α + .
Thus, (17) can be rewritten as
min α + , α , ξ + , ξ , θ e T α + + ρ e T ( ξ + + ξ ) subject   to Y K ( α + α ) θ e ϵ e + ξ + , ξ + 0 , K ( α + α ) + θ e Y ϵ e + ξ , ξ 0 , e T ( α α + ) = 0 , 0 α + , α ρ e .
This optimization problem (19) defines the SSVR model used for comparison later. After solving this optimization problem, we can calculate w using Equation (4) without the need for additional computation of θ , as θ has already been determined during the optimization process. Consequently, the final regression function of the SSVR is the same as that of the classical SVR.
Remark 1. 
In the primal problem (2) of SVR, the term 1 2 w 2 2 controls structural complexity. Similarly, in our SSVR model (17), the term 1 2 ( α + 1 + α 1 ) aims to minimize the number of nonzero elements in α + and α , thereby reducing structural complexity. Thus, both models share the same objective.
Remark 2. 
Unlike the classical SVR, which requires solving a QPP, our SSVR model is directly simplified into an LPP. Further analysis on this point will be provided in the subsequent sections.

3.2. STPISVR Model

As previously mentioned, the STPISVR model inherits the core of TPISVR by constructing a pair of nonparallel parametric insensitive bound functions, specifically the lower-bound and upper-bound regression functions. The final regression function remains consistent with Equation (16).
We now proceed to construct the STPISVR model. By substituting the first formula in (12) into both the second term of the objective function and the first constraint of problem (8), respectively, we obtain the following:
v d | I | e T φ ( X ) w d + θ d e = v d v d | I | 2 e T K e + 1 | I | e T K α d θ d
and
Y φ ( X ) w d + θ d e ξ d = v d | I | K e K α d + θ d e ξ d .
Combining (20), (21) with (8) and (10), and replacing the regularization term 1 2 w d 2 2 with α d 1 , we obtain the following optimization problem corresponding to (8):
min α d , θ d , ξ d α d 1 + v d 1 | I | e T K α d θ d v d | I | 2 e T K e + ρ d | I | e T ξ d subject   to Y v d | I | K e K α d + θ d e ξ d , ξ d 0 , e T α d = v d , 0 α d ρ d | I | e .
A similar procedure yields the following optimization problem corresponding to (9):
min α u , θ u , ξ u α u 1 + v u 1 | I | e T K α u + θ u v u | I | 2 e T K e + ρ u | I | e T ξ u subject   to Y v u | I | K e + K α u + θ u e + ξ u , ξ u 0 , e T α u = v u , 0 α u ρ u | I | e .
Problems (22) and (23) represent the initial version of the STPISVR model.
Notably, in the optimization problem (22), the constraints α d 0 and e T α d = v d allow us to easily obtain the following results:
α d 1 = e T α d = v d .
This term, thus, becomes a constant, independent of the optimization variable. By removing the constant term irrelevant to the optimization variables, problem (22) simplifies to the linear form:
min α d , θ d , ξ d v d | I | e T K α d v d θ d + ρ d | I | e T ξ d subject   to Y v d | I | K e K α d + θ d e ξ d , ξ d 0 , e T α d = v d , 0 α d ρ d | I | e .
Similarly, we have
α u 1 = e T α u = v u ,
and problem (23) simplifies to (27) using the same reasoning:
min α u , θ u , ξ u v u | I | e T K α u + v u θ u + ρ u | I | e T ξ u subject   to Y v u | I | K e + K α u + θ u e + ξ u , ξ u 0 , e T α u = v u , 0 α u ρ u | I | e .
Problems (25) and (27) are our final STPISVR model. Once solved, we can calculate the values of w d and w u according to Equation (12), where θ d and θ u do not need to be recalculated, as they have already been obtained during the optimization process. After these values are obtained, we can then derive the down-bound and up-bound functions, as shown in (15), and final regression function, as given in (16).
Let us now revisit the initial version of the model (22), noting that (25) is just a simplified form derived from it. The first term α d 1 in the objective function is originally introduced to promote sparsity in the model, which in turn simplifies its structural complexity. As discussed in Remark 1, this simplification of structural complexity aligns with the role of the 1 2 w d 2 2 term in the primal problem (8) of TPISVR, despite their different mathematical forms. However, during the derivation, this L1-norm regularization term in (22) is simplified to a constant via the KKT conditions, and therefore, disappears from the final optimization problem (25), which clearly leads to a more efficient reduction in optimization complexity. Nonetheless, it is important to note that the disappearance of this term does not affect the sparsity of the final model; the sparsity is effectively guaranteed by the linear programming structure and the constraints derived from the KKT conditions, which will be further analyzed in Section 4.1. The second term of the objective function in (22), derived from (20), aims to maximize the projection of the mean value of the training sample points on f d ( x ) , causing f d ( x ) to be as large as possible. The third term of the objective function in (22) minimizes the sum of error variables, mitigating overfitting to the training samples. Based on (21) and (14), the first constraint in (22) can be rewritten as
Y f d ( X ) ξ d ,
which means that the estimated value of the sample point x on f d ( x ) cannot exceed the response value of the sample point; otherwise, a nonnegative value ξ x d is introduced to measure this error. The final two constraints on α d are derived from KKT conditions.
Remark 3. 
The objective functions in (25) and (27) are linear, in contrasts to the dual QPPs in TPISVR.
Remark 4. 
The terms α d 1 and α u 1 in (22) and (23) may be replaced by other sparse norms.
Remark 5. 
In the linear case, where φ ( X ) = X , the optimization problems (25) and (27) reduce to the following form:
min α d , θ d , ξ d v d | I | e T X X T α d v d θ d + ρ d | I | e T ξ d subject   to Y v d | I | X X T e X X T α d + θ d e ξ d , ξ d 0 , e T α d = v d , 0 α d ρ d | I | e ,
and
min α u , θ u , ξ u v u | I | e T X X T α u + v u θ u + ρ u | I | e T ξ u subject   to Y v u | I | X X T e + X X T α u + θ u e + ξ u , ξ u 0 , e T α u = v u , 0 α u ρ u | I | e .
Next, we propose two modular algorithms for STPISVR: a training algorithm that learns the prediction function and records sparsity and training time (Algorithm 1) and a testing algorithm that evaluates performance and inference time (Algorithm 2). The detailed metric definitions are provided in Section 5.
Algorithm 1 STPISVR training algorithm.
Input: Training set D = { X , Y } ; parameter combination θ = ( v d , v u , ρ d , ρ u , γ (Gaussian kernel parameter))
Output: Prediction function f ( x ) ; training time (Tr-time); number of SVs (Num-SVs)
1.   Start timer.
2.   Solve optimization problem (25) to obtain α d and θ d .
3.   Solve optimization problem (27) to obtain α u and θ u .
4.   Construct bound functions f d ( x ) and f u ( x ) using Equations (14) and (15), respectively.
5.   Compute final regression function f ( x ) = 1 2 f d ( x ) + f u ( x ) via Equation (16).
6.   Stop timer, record Tr-time.
7.   Count Num-SVs from α d and α u .
8.   Return f ( x ) , Tr-time and Num-SVs.
Algorithm 2 STPISVR Testing Algorithm
Input: Test set T = { X test , Y test } ; trained prediction function f ( x )
Output: Root mean squared error (RMSE); coefficient of determination ( R 2 ); test time (Te-time); Zone Width between insensitive boundaes (ZoneWidth, if needed)
1.   Start timer.
2.   For each test sample x i in X test , Compute prediction y ^ i = f ( x i ) .
3.   Calculate RMSE and R 2 .
4.   Stop timer, record Te-time.
5.   Return RMSE, R 2 , Te-time and ZoneWidth (if computed).

3.3. Properties of the STPISVR Model

To theoretically analyze the proposed STPISVR model given in (25) and (27), we present two results to help us further understand its behavior.
As is well known, setting the parameters of the model in machine learning algorithms is crucial for achieving optimal performance. The following important theorem will illustrate how the parameters control the fractions of SVs and errors. Before presenting the theorems, let us first define the following two key terms.
Definition 1. 
The fractions of down-SVs and up-SVs are defined as:
1 | I | | { x p : α p d > 0 , p I } | and 1 | I | | { x q : α q u > 0 , q I } | ,
respectively.
Definition 2. 
The fractions of down-errors and up-errors are:
1 | I | | { x p : f d ( x p ) < 0 , p I } | and 1 | I | | { x q : f u ( x q ) > 0 , q I } | ,
respectively.
Theorem 1. 
Let f d ( x ) and f u ( x ) be what the STPISVR obtains as the nondegenerate parametric insensitive down-bound and up-bound regression functions, respectively; then, the following statements hold:
(i)
v d ρ d and v u ρ u are the lower bounds on the fractions of down-SVs and up-SVs, respectively.
(ii)
v d ρ d and v u ρ u are the upper bounds on the fractions of down-errors and up-errors, respectively.
Proof. 
From problem (29), we have
0 α p d ρ d | I | , p I .
(i) Suppose the number of down-SVs is denoted as s d . From problem (29), we have v d = e T α d = p I α p d s d ρ d | I | . Therefore, we can deduce that v d ρ d s d | I | . This shows that v d ρ d is a lower bound on the fraction of down-SVs.
Similarly, by the same reasoning, we can conclude that v u ρ u is a lower bound on the fraction of up-SVs. Thus, conclusion (i) is proved.
(ii) Suppose the number of all the points x p with f d ( x p ) < 0 is t d , where p I . Again, from problem (29), we have v d = e T α d = p I α p d t d ρ d | I | , and v d ρ d t d | I | can be obtained immediately. Therefore, v d ρ d is an upper bound on the fraction of down-errors.
By the same token, we can obtain that v u ρ u is an upper bound on the fraction of up-errors. Thus, conclusion (ii) is also proved.    □
Remark 6. 
Theorem 1 states that the values of v d ρ d v u ρ u can control the fractions of down-SVs (or up-SVs) and errors.
Remark 7. 
Obviously, v d ρ d and v u ρ u for STPISVR. For an overlarge value of v d ρ d v u ρ u , the sparsity and errors of the model can get worse, even leading f d ( x ) f u ( x ) to shift up (move down), we will see this point in the next theorem. However, if the value of v d ρ d v u ρ u is too small, it may cause overfitting, which will affect the generalization ability of the model. Thus, for the STPISVR model, a very important work is still to set and determine the parameters as other machine learning algorithms.
To further explain the theoretical properties of our STPISVR model, we set v d = v u = v and ρ d = ρ u = ρ , which is a common practice in real applications. Now, we consider the following new optimization problem first:
min α d , α u , θ d , θ u , ξ d , ξ u 2 v + v | I | e T K ( α d + α u ) + v s . ( θ u θ d ) 2 v 2 | I | 2 e T K e + ρ | I | e T ( ξ d + ξ u ) , subject   to Y f d ( X ) ξ d , ξ d 0 , e T α d = v , 0 α d ρ | I | e , Y f u ( X ) + ξ u , ξ u 0 , e T α u = v , 0 α u ρ | I | e ,
In this case, the objective function is the sum of the objective functions in (22) and (23), reflecting the results after combining (24) and (26), and the constraints are the unions of those in (22) and (23) when v d = v u = v and ρ d = ρ u = ρ . Clearly, problem (34) is optimized if both (25) and (27) (or (22) and (23)) are optimized.
For a data point ( x , y ) , we introduce
g ( x ) = f u ( x ) f d ( x ) 2 ,
which represents half of the longitudinal distance between the down- and up-insensitive regression bounds at the sample point x, when f u ( x ) f d ( x ) (we will discuss other cases in Theorem 2), reflecting the size of the insensitive interval. The larger the value of g ( x ) , the wider the regression interval, and the stronger the robustness of the model to noise and outliers. Conversely, the smaller the value of g ( x ) , the narrower the regression interval, the stringent the model fits the data, and the more sensitive the model.
Substituting (14) and (15) into (35), we have
g ( x ) = v | I | e T K ( X , x ) + 1 2 K ( X , x ) T ( α u + α d ) + 1 2 ( θ u + θ d ) .
Thus, problem (34) can be rewritten as:
min α d , α u , θ d , θ u , ξ d , ξ u 2 v + 2 v | I | e T g ( X ) + ρ | I | e T ( ξ d + ξ u ) subject   to f ( X ) Y g ( X ) + ξ d , ξ d 0 , e T α d = v , 0 α d ρ | I | e , Y f ( X ) g ( X ) + ξ u , ξ u 0 , e T α u = v , 0 α u ρ | I | e .
The following theorem will help us understand how to set the parameter range for STPISVR more precisely when v d = v u = v and ρ d = ρ u = ρ .
Theorem 2. 
Let f d ( x ) and f u ( x ) be the nondegenerate parametric insensitive down-bound and up-bound regression functions for STPISVR. Assuming v d = v u = v and ρ d = ρ u = ρ , then we have
(i)
f u ( x ) > f d ( x ) for any data point ( x , y ) if v ρ < 1 2 .
(ii)
f u ( x ) < f d ( x ) for any data point ( x , y ) if v ρ > 1 2 .
(iii)
f u ( x ) = f d ( x ) for any data point ( x , y ) if v ρ = 1 2 .
Proof. 
For a data point ( x , y ) , when v d = v u = v and ρ d = ρ u = ρ , the empirical loss of the optimization problem (37) is
2 v g ( x ) + ρ ( ξ d ) x + ( ξ u ) x ,
and the constraints are
f ( x ) y g ( x ) + ( ξ d ) x , ( ξ d ) x 0 , e T α d = v , 0 α d ρ | I | e , y f ( x ) g ( x ) + ( ξ u ) x , ( ξ u ) x 0 , e T α u = v , 0 α u ρ | I | e .
From the KKT conditions, at least one of ( ξ d ) x and ( ξ u ) x must be equal to zero. For example, assume ( ξ u ) x = 0 . Then, from (39), we have
f ( x ) y = g ( x ) + ( ξ d ) x , ( ξ d ) x 0 ,
y f ( x ) g ( x ) .
We have the following
(i) if v ρ < 1 2 , we have 2 v < ρ , which causes g ( x ) to be larger than ( ξ d ) x in order to minimize the empirical loss at point ( x , y ) ; then, we have g ( x ) > ( ξ d ) x , and f d ( x ) < f u ( x ) by considering ( ξ d ) x 0 .
(ii) if v ρ > 1 2 , then 2 v > ρ , which causes g ( x ) be less than ( ξ d ) x in order to minimize the empirical loss at point ( x , y ) . In this time, if g ( x ) = 0 , we have that g ( x ) ( 2 v g ( x ) + ρ ( ξ d ) x ) = 2 v ρ = 0 , and this contradicts v ρ > 1 2 . Thus, g ( x ) < 0 , and f d ( x ) > f u ( x ) .
(iii) if v ρ = 1 2 , we have f d ( x ) = f u ( x ) . In this case, the STPISVR model degenerates into a general regression model with no insensitive parametric boundary regions.    □
Remark 8. 
Theorem 1 states that both the values of v d ρ d and v u ρ u should not exceed 1. From Theorem 2, we see more clearly that when v d = v u = v and ρ d = ρ u = ρ , v ρ should be less than 1 2 in order to obtain reasonable insensitive down-bound function f d ( x ) and up-bound function f u ( x ) ; otherwise, f d ( x ) will shift up and f u ( x ) will move down.
Remark 9. 
According to (38), under the conditions of Theorem 2, when v ρ 1 2 , STPISVR minimizes both g ( x ) and ( ξ d ) x + ( ξ u ) x . On the one hand, it imposes that the parameter insensitive down-bound function and up-bound regression function are as close as possible; on the other hand, the sum of errors at each point x as small as possible. The optimization process aims to determine an optimal balance between these two goals.

4. Discussion and Comparison

To better understand our proposed sparse model, this section presents a discussion and comparison with several related algorithms regarding sparsity, prediction speed, and computational complexity.

4.1. Sparsity

As mentioned earlier, TPISVR loses sparsity due to its structure. In contrast, both SSVR and STPISVR are sparse models, achieving this by introducing the L1-norm of the dual variables as a regularization term. This results in fewer SVs compared to the traditional models, yielding sparser solutions. Although the L1-norm regularization term in STPISVR becomes a constant and does not explicitly appear in the final objective function, sparsity is preserved and even enhanced by the following mechanisms: (1) The optimization problems are ultimately formulated as LPPs with L1-norm equality constraints and box constraints; (2) According to linear programming theory, an optimal solution lies at a vertex of the feasible region, which, in high-dimensional spaces, is typically sparse—that is, many components of the optimization variables (i.e., the dual variables) are zero. Therefore, even without an explicit regularization term, the formulation and geometry of the problem naturally promote sparse solutions. Additionally, SSVR is based on SVR and simultaneously optimizes α + and α , whereas STPISVR is derived from TPISVR and optimizes α d and α u separately. Theoretically, SSVR is expected to yield sparser models than STPISVR. which is further supported by experimental results later.

4.2. Prediction Speed

The prediction speed of a machine learning algorithm is closely linked to its sparsity. Both SSVR and STPISVR benefit from sparsity due to having fewer SVs compared to the traditional models, SVR and TPISVR. This directly reduces the number of SVs, resulting in faster prediction speeds during the test phase.
In typical scenarios, we simplify the models by setting v d = v u and ρ d = ρ u in both TPISVR and STPISVR, which allows the regressor (16) to be rewritten as:
f ( x ) = 1 2 ( α u α d ) K ( X , x ) + 1 2 ( θ d + θ u ) .
From (6) and (42), it is clear that prediction speed depends primarily on the number of nonzero dual vectors, i.e., the number of SVs.
Although SSVR may be slightly faster than STPISVR due to its higher degree of sparsity, nevertheless, both models still outperform SVR and TPISVR in terms of prediction speed. Faster prediction speeds are particularly advantageous for real-time prediction tasks once the model is trained, as they enable more efficient and timely predictions. Thus, the sparse nature of SSVR and STPISVR results in faster prediction speeds, making these models well-suited for real-time applications after training.

4.3. Computational Complexity

The computational complexity of SVR and TPISVR is generally O ( | I | 3 ) , with TPISVR typically being faster than SVR [11]. The complexities of STPISVR and SSVR are also O ( | I | 3 ) .
In STPISVR, as shown in (25) and (27), all optimization variables are solved simultaneously without separately computing bias terms. Although STPISVR involves more constraints ( 4 | I | + 1 ) compared to TPISVR, the linearity of the objective function and the constant regularization term help control computational costs. In fact, solving QPPs is generally more computationally demanding than solving LPPs under similar conditions. For linear optimization, complexity depends on both the number of variables and constraints [29]. STPISVR’s each optimization problem involves 2 | I | + 1 variables and 4 | I | + 1 constraints—more constraints than in TPISVR. This increase in the number of constraints may leads to higher computational load as the training sample size grows. However, the linear objective in STPISVR provides an advantage over the quadratic form in TPISVR, keeping overall complexity manageable at O ( | I | 3 ) and ensuring practical feasibility for most real-world applications. This reflects a trade-off between the growth in linear constraints and the simplification from the linear objective. While more constraints increase computational burden, the simpler objective reduces complexity relative to quadratic formulations, maintaining scalability for large-scale problems.
By contrast, SSVR involves 4 | I | + 1 variables and 8 | I | + 1 constraints, significantly exceeding those of STPISVR, leading to higher computational costs —particularly for large sample sizes. Thus, SSVR is the most computationally expensive among the four algorithms during the training phase. Despite STPISVR’s slightly higher training cost compared to TPISVR, its reduced number of SVs results in faster and more efficient prediction during the testing phase. Therefore, STPISVR is well-suited for real-time prediction tasks. Ultimately, the improved sparsity of STPISVR offsets the additional training complexity, offering better generalization and prediction efficiency—making the slight increase in training cost a worthwhile trade-off.

5. Numerical Experiments

To validate the sparsity and generalization performance of the proposed STPISVR model, we compare the results of STPISVR, SVR, TPISVR, and SSVR across several synthetic datasets and 10 benchmark datasets. All experiments are conducted using MATLAB 2024A on a PC equipped with an Intel Core i7 processor and 4 GB of RAM. Since the training speed is not only related to the computational complexity of the model itself, but also closely related to the solver selected, selecting the appropriate solver for different models can speed up the training process. In this paper, we use Mosek 10.2 solver for solving QPPs, and Gurobi 12.0.1 solver for solving LPPs to ensure solver–model alignment and improve overall training efficiency.
Given the critical importance of parameter tuning, which directly influences model training and prediction performance, we employ a two-stage combinatorial tuning strategy with five-fold cross-validation at each stage. The tuning set is composed of 50% of the data randomly selected from the training set. In the first stage, we perform a grid search to select the top five parameter combinations based on performance. Then, leveraging the grid search results, we significantly narrow the parameter range and apply Bayesian optimization [30] for 200 iterations to identify the optimal parameter combination. This hybrid approach combines the broad filtering power of grid search with the precision of Bayesian optimization, yielding more reliable and accurate identification of optimal parameters. Moreover, this strategy helps address the challenge of parameter instability in sparse models, especially those involving L1-regularization. The effectiveness of this approach is further evaluated through ablation studies presented later. Once the optimal parameters are identified, the tuning data is reintegrated into the full training set to retrain the model and learn the final regressor.
Algorithm 3 outlines the two-stage parameter tuning procedure in detail.
For the synthetic datasets, unless otherwise specified, we always generate 10 distinct training sets and 10 corresponding test sets for each example. The models are trained and tested across these datasets, and the results are reported as the mean ± standard deviation over the 10 test sets. For the benchmark datasets, we apply 10-fold cross-validation on each dataset to train the models and assess generalization performance, with results also presented as the mean ± standard deviation across the 10 folds.
For consistency and simplicity, we set ρ d = ρ u = ρ for all models, and v d = v u = v for both TPISVR and STPISVR. Across all models, at the grid search stage, ρ is selected from the set { 10 i i = 2 ,   1 ,   ,   4 } , while v / ρ for TPISVR and STPISVR is selected from the set { 0.05 ,   0.1 ,   0.15 ,   ,   0.5 } . For nonlinear cases, the Gaussian kernel parameter is chosen from the set { 2 8 ,   2 7 ,   ,   2 8 } . The insensitive tube parameter ϵ for SVR and SSVR is selected from the set { 0.01 ,   0.02 ,   ,   0.2 } .
Algorithm 3 Two-stage parameter tuning strategy for STPISVR.
Input: Full training set D; grid parameter space Θ grid = { ( v d , v u , ρ d , ρ u , γ ) } ; number of cross-validation folds K
Output: Optimal parameter combination θ * = ( v d * , v u * , ρ d * , ρ u * , γ * )
1.   Randomly select 50% of D as tuning subset D tuning .
2.   Stage 1: Grid Search
3.   For each θ in Θ grid :
4.       Perform K-fold cross-validation on D tuning :
5.          For each fold:
6.            Train STPISVR on training fold using Algorithm 1 with parameter θ .
7.            Evaluate on validation fold using Algorithm 2 to obtain RMSE.
8.       Compute average RMSE over all folds.
9.   Select top 5 candidates with the lowest average RMSE.
10.   Stage 2: Bayesian Optimization
11.   Randomly split D tuning into training set D tuning _ tr (40%) and validation set D tuning _ val (10%).
12.   Define Bayesian parameter search space Θ bayes based on the top 5 candidates from grid search.
13.   Initialize b e s t _ s c o r e using the lowest RMSE of a top candidate from grid search, and set θ * accordingly.
14.   Set n o _ i m p r o v e _ c o u n t = 0 .
15.   Repeat for up to 200 iterations:
16.       Propose candidate parameter θ candidate via Bayesian optimization.
17.       Train STPISVR on D tuning _ tr using Algorithm 1 with θ candidate .
18.       Evaluate on D tuning _ val using Algorithm 2 to obtain RMSE.
19.       If RMSE <  b e s t _ s c o r e :
20.            Update b e s t _ s c o r e RMSE, θ * θ candidate , and reset n o _ i m p r o v e _ c o u n t = 0 .
21.       Else:
22.            Increment n o _ i m p r o v e _ c o u n t by 1.
23.       Update Bayesian optimizer with ( θ candidate , RMSE ) .
24.       If n o _ i m p r o v e _ c o u n t 20 , break loop.
25.   Return  θ * = ( v d * , v u * , ρ d * , ρ u * , γ * ) .
In the following experiments, we evaluate the performance of the four algorithms using several key metrics: root mean squared error (RMSE), coefficient of determination ( R 2 ), number of SVs (Num-SVs), training time (Tr-time), and test time (Te-time). Additionally, Zone Width between insensitive boundaries (ZoneWidth) is considered for TPISVR and STPISVR. The definitions of the RMSE, R 2 , and ZoneWidth metrics are as follows:
RMSE = 1 | I | p = 1 | I | ( y p y ^ p ) , R 2 = 1 p = 1 | I | ( y p y ^ p ) 2 p = 1 | I | ( y p y ¯ p ) 2 , ZoneWidth = 1 | I | p = 1 | I | | f u ( x p ) f d ( x p ) | = 1 | I | p = 1 | I | g ( x p ) ,
where y p ^ = f ( x p ) , p I , represents the regression value of the model for the pth sample x p , p I , and y ¯ = 1 | I | p = 1 | I | y p represents the mean of y p , p I . RMSE reflects the square error of the regression model, and generally, R 2 [ 0 ,   1 ] is a measure of how well the model fits the dataset. Clearly, for a given method, the smaller the RMSE value and the larger the R 2 value, the better the prediction performance.

5.1. Synthesis Datasets

Example 1. 
The first example is on the regression estimation of the Sinc function, which is defined as:
y = h 1 ( x ) = S i n c ( x ) = s i n x x , x [ 4 π , 4 π ] .
To effectively reflect the performance of our method, for a sample x p , p I , two types of noises are introduced, and
y p = h 1 ( x p ) + e p , x p U [ 4 π , 4 π ] , p I ,
where e p , p I , represents noise. We consider two kinds of noises:
Type A: e p N ( 0 , 0 . 1 2 ) , p I ;
Type B: e p U ( 0.2 , 0.2 ) , p I .
For each type of noise, we graduate 10 independent groups of samples, each comprising 300 training samples in this example and 300 test samples according to (45) and the above two kind of noise distribution.
Figure 1 and Figure 2 (unless otherwise specified, all graphical results presented in this paper are obtained with the optimal parameters) illustrate the regression performance of SVR, SSVR, TPISVR, and STPISVR on one of the ten test sets in Example 1, corresponding to Type A and Type B, respectively. It is evident from Figure 1 and Figure 2 that SSVR and STPISVR utilize significantly fewer SVs than SVR and TPISVR. Among these methods, STPISVR achieves the best regressors under both cases. From Figure 1 and Figure 2, it is interesting to note that the SVs identified by SSVR and STPISVR are neither strictly beyond the upper or lower bounds, nor constrained to the boundaries as in the traditional methods. Instead, they represent prototype points of the dataset. This phenomenon arises because, during the optimization process, α and α + in SSVR, as well as α d and α u in STPISVR, no longer act as simple dual variables, but take on more structural roles in optimization. This fundamental shift marks the key difference between our sparse models and conventional approaches.
Table 2 reports the average test results from 10 groups for each method on Example 1. To better measure the model performance, experiments were also conducted using noiseless test sets for Type A and Type B. The samples of the noiseless test set are the same as those of the noise test set, but the response values are obtained by (44). It can be observed from Table 2 that STPISVR demonstrates the best overall performance across all test scenarios, achieving the lowest RMSE and the highest R 2 values among the four algorithms. While SSVR features the fewest SVs, which aligns with the analysis of its superior sparsity and the shortest prediction time, its predictive performance is inferior to that of STPISVR. The very small number of SVs may cause underfitting, resulting in poorer prediction performance. In terms of training time (in our experiments, the SVR model used MATLAB’s optimized ‘fitrsvm’ function; although SVR results are reported, its training time is excluded from comparisons, which applies to subsequent examples as well), in this example, STPISVR benefits from the computational efficiency of a linear model, resulting in significantly faster training time compared to TPISVR. Among all the methods, SSVR achieves the shortest training time; however, its prediction performance is inferior to STPISVR, although it still outperforms SVR. Overall, STPISVR demonstrates the best comprehensive performance. Furthermore, when compared specifically with TPISVR, STPISVR surpasses TPISVR across all evaluation metrics on this example. STPISVR not only utilizes fewer SVs but also achieves shorter training time and prediction time while consistently delivering superior predictive performance.
Table 2 also presents the results for Type A of Example 1 using only grid search, evaluated on both noisy and noiseless test sets. Compared to the hybrid two-stage tuning strategy, the hybrid approach consistently achieves better prediction accuracy and stability, highlighting the advantage of combining broad parameter search with local refinement.
To further evaluate the effectiveness of our proposed STPISVR model, we tested the four models on a dataset with a heteroscedastic structure. We consider the following function as our Example 2:
Example 2. 
y = h 2 ( x ) = 0.2 s i n ( 2 π x ) + 0.2 x 2 + 0.3 , x [ 0 , 1 ] .
For a sample point x p , p I = { 1 , 2 , , 200 } , we set
y p = h 2 ( x p ) + e p , e p = 0.1 ( x p ) 2 + 0.05 ϵ p , p I ,
where e p , p I represents noise with the specified heteroscedastic structure. We consider the following two kind of noises:
Type C: x p = 0.005 p , ϵ p U ( 1 , 1 ) , p I .
Type D: x p U ( 0 , 1 ) , ϵ p N ( 0 , 0.25 ) , p I .
For each type of noise, we again generate 10 groups of samples and each consisting of 200 training samples and 200 test samples in this example, and the training set and test set are generated in a manner similar to Example 1.
In Example 2, the performance of SVR, SSVR, TPISVR, and STPISVR on one of the ten test sets is depicted in Figure 3 and Figure 4, corresponding to Type C and Type D, respectively. It can be observed that SSVR and STPISVR exhibit greater sparsity, while TPISVR and STPISVR achieve better fitting results compared to SVR and SSVR. This outcome aligns with previous observations, as TPISVR and STPISVR are better suited for datasets with heteroscedastic structures.
Additional results are presented in Table 3, including those on the noiseless test sets, as in Example 1. It can be seen that, similar to Example 1, STPISVR demonstrates the best overall prediction performance. Compared to TPISVR, STPISVR is sparser in several scenarios, exhibiting faster prediction speeds. Moreover, the training time of STPISVR is still lower than that of TPISVR on this example. SSVR performs well in terms of obtaining fewer SVs, having a shorter training time, and having a faster prediction time; however, in terms of generalization ability, it is inferior to both TPISVR and STPISVR. Notably, in this example, the prediction performance of TPISVR ranks second only to STPISVR. The reason is that both models are better suited for heteroscedastic datasets than SVR and SSVR, consistent with the results shown in Figure 3 and Figure 4.
To further demonstrate the overall performance of our STPISVR method, we continue the analysis using Example 2 (Type C), examining it from different perspectives. The results are reported in Figure 5, Figure 6 and Figure 7.
Figure 5 shows the relationships between the parameters ( ρ , v ) and the fraction of SVs and errors of STPISVR on Type C. Here, we present results for down-SVs and down-errors only, as up-SVs and up-errors exhibit similar patterns. In this figure, the kernel parameter is fixed at its optimal value. As can be seen from the three-dimensional graph of Figure 5, both the fraction of SVs and errors increase with the increase of v / ρ , but as ρ increases, the fraction of SVs rises, while the fraction of errors decreases. To highlight the results effectively, we set ρ = 0.1 and examine how v / ρ affects the fraction of SVs and errors (see the two-dimensional graph in Figure 5). As suggested by Theorem 1, v / ρ controls both the fraction of SVs and the errors well.
The relation between ( ρ , v ) and the ZoneWidth for STPISVR are shown in Figure 6. When v / ρ increases and is less than 0.5, the width values and its standard deviation decrease. Instead, when v / ρ is larger than 0.5, the width increases. The reason is that if v / ρ > 0.5 , f u ( x ) moves down and f d ( x ) shift up, f u ( x ) < f d ( x ) . This confirms the results of Theorem 2. In addition, this example also shows that for smaller or larger ρ , the width of the parametric insensitive zone is relatively stable, stabilizing at larger or smaller values, respectively.
Figure 7 presents the relationship between ( ρ , v ) and RMSE and R 2 , respectively, which further demonstrates that, over a large range of ρ , the RMSE stabilizes at a smaller value, while R 2 stabilizes at a larger value. These findings provide valuable guidance for setting appropriate parameter ranges to optimize the performance of our sparse regressor.
To investigate how the parameters of the STPISVR model influence the evaluation indicators, and more specifically, how different evaluation metrics change with increasing noise variance, we conduct an experiment using another type of noise (denoted as Type E) again for Example 2:
Type E: x p U ( 0 , 1 ) , ϵ p U ( a , a ) , p I , I = { 1 ,   2 ,   ,   200 } . Here, we set the value of a as 0.2 ,   0.4 ,   0.6 ,   ,   2.0 , respectively, then the corresponding noise standard deviations of ϵ p are given by: 2 300 , 3 300 , 4 300 , , 20 300 , respectively.
Figure 8 illustrates the impact of different noise variances on the evaluation indicators of the STPISVR model for heteroscedastic structured data. It is evident that the width of the parametric insensitive zone increases as the variance of ϵ p increases, whereas the proportion of SVs remains largely unaffected by changes in ϵ p . This observation indicates that the STPISVR model can automatically adjust the width of the parametric insensitive zone in response to variations in noise variance. Additionally, when we use the noise test set, as noise intensity increases, the RMSE increases and R 2 decreases. However, the magnitude of these changes is relatively minor when using a noiseless test set, indicating that the model is robust against increasing noise intensity.
To study how evaluation metrics change with increasing training sample size, we concentrate our attention to the case where a = 2.0 , corresponding to the largest noise variance in Figure 8, and we increase our dataset size from 200 up to 2000 samples by 200 in each step. The test dataset with the size of 200 samples is fixed. The results are presented in Figure 9 and Figure 10.
Figure 9 illustrates the trends in RMSE, R 2 , the proportion of SVs to training samples, training time, and test time of the four algorithms as the training sample size increases. From Figure 9, we observe that the RMSE ( R 2 ) values of SVR, TPISVR, and STPISVR steadily decrease (increase) as the training sample size grows. Among these, STPISVR exhibits the best performance, followed by TPISVR. In contrast, the RMSE ( R 2 ) values of SSVR fluctuate significantly, and even with larger sample sizes, its performance is worse than any other methods. Regarding sparsity, STPISVR exhibits a higher level of sparsity than TPISVR. As for training time, when the sample size remains below 1000, STPISVR continues to benefit from the computational efficiency due to the linearity of the model. Consistent with the results from previous examples, its training time remains lower than that of TPISVR and even outperforms the other algorithms. However, as the training sample size increases, its training time shows an upward trend; nonetheless, the overall computational cost remains acceptable. In contrast, TPISVR maintains a relatively stable increase in training time. For SSVR, previous examples have shown that it achieves the shortest training time due to the smaller sample size. However, in this case, when the sample size exceeds 400, its training cost rises significantly and continues to increase rapidly as the sample size grows. This result is largely consistent with our previous analysis of computational complexity. In terms of testing speed, STPISVR is significantly faster than TPISVR due to its higher sparsity. Moreover, as the sample size increases, the gap in test time between STPISVR and TPISVR widens further, highlighting the advantages brought by sparsity of STPISVR. Overall, STPISVR demonstrates the best performance on this example, striking a balance between accuracy, sparsity, and efficiency. It achieves a lower proportion of SVs and a faster prediction speed than TPISVR, though at the cost of higher computational complexity for larger-scale datasets. Nevertheless, this trade-off proves worthwhile, especially for regression tasks that require real-time predictions after training.
Figure 10 compares the trends of STPISVR and TPISVR on the parametric insensitive zone; the fraction of down-SVs and down-errors (results for up-SVs and up-errors are similar) decreases as the sample size increases. Moreover, to compare the trends of sparsity between TPISVR and STPISVR more clearly as the sample size increases, we also present the number of SVs of the two methods here. From Figure 10, we can highlight the following key aspects. First, the number of SVs for both methods increases with the training sample size. However, the difference between the two methods becomes more pronounced as the sample size grows, highlighting the sparsity advantage of STPISVR over TPISVR with larger datasets. Second, for STPISVR, the fractions of down-SVs and down-errors converge toward y = 0.0588 (the optimal value of v / ρ in STPISVR) from above and below, respectively, as the training sample size increases. A similar pattern is observed for TPISVR (converging to y = 0.0960 ). This outcome further confirms the conclusion of Theorem 1. Third, the widths of the parametric insensitive zones remain relatively stable for both methods. However, the width for STPISVR is consistently a little larger than that of TPISVR, which contributes to the superior performance of STPISVR.

5.2. Benchmark Datasets

To further evaluate the performance of our sparse models, we conduct experiments using four algorithms on 10 publicly available benchmark datasets: Auto-mpg (D1), AutoPrice (D2), Boston (D3), Concrete (D4), Daily (D5), Machine (D6), Mcs (D7), Motorcycle (B8), Yacht (D9), and Stock1 (D10). For a fair comparison, all dataset features were rescaled to the range [0,1] before the experiment. The results on these benchmark datasets are presented in Table 4.
As shown in Table 4, consistent with the results from synthetic datasets, the proposed STPISVR model achieves the best overall performance among the four methods. It records the lowest RMSE and highest R 2 values on 6 of the 10 benchmark datasets. If compared with TPISVR alone, from the values of RMSE and R 2 , the predictive performance of STPISVR exceeds TPISVR in 8 out of 10 datasets, and is only slightly lower than TPISVR on the other 2 datasets (D5 and D6), i.e., it is almost the same as TPISVR.These results suggest that STPISVR is better suited for a wider range of scenarios.
Regarding the number of SVs, Table 4 reveals that STPISVR consistently requires fewer SVs than TPISVR across all datasets, even on dataset D2, which has the fewest SVs among the four algorithms. Similarly, SSVR achieves greater sparsity than SVR and proves to be the sparsest model on 9 out of the 10 datasets. This finding aligns with earlier theoretical analyses and results from synthetic datasets.
In terms of training time, STPISVR outperforms TPISVR in all 10 datasets, once again demonstrating its advantage in training speed for small and medium sized datasets. A closer examination reveals that on datasets D3, D4, and D10, the training time of SSVR exceeds that of STPISVR. Notably, the sample size of these three datasets surpasses 500, and in particular, the training time of SSVR on D4 is nearly three times that of STPISVR. These findings indicate that while SSVR is faster only for very small-scale datasets, it becomes computationally expensive and less efficient as the dataset size increases.
Among the 4 algorithms, SSVR achieves the shortest prediction time on 8 out of 10 datasets, which is naturally attributed to its high sparsity. Although STPISVR exhibits slightly lower sparsity than SSVR and has a more complex regression function, it still achieves the shortest prediction time on 5 of the 10 datasets. Moreover, in 3 cases, its mean training time matches that of SSVR while exhibiting lower variance, highlighting its greater stability. Compared to TPISVR, STPISVR consistently achieves a faster prediction speed, further emphasizing its computational efficiency. Additionally, for the dataset D8, we conducted an ablation comparing grid search alone with the hybrid two-stage tuning strategy, confirming the latter’s superiority in prediction accuracy and stability.
In summary, STPISVR consistently outperforms the other methods in terms of prediction accuracy, sparsity, training time, prediction speed, and overall stability across these datasets.
To provide a more intuitive understanding of these results, we graphically visualized the outcomes for the dataset D8 as Figure 11. The visualization highlights that both STPISVR and SSVR achieve greater sparsity. Additionally, STPISVR and TPISVR display a better data-fitting effect compared to SVR and SSVR on this heteroscedastic dataset. Overall, STPISVR demonstrates superior performance, combining the advantages of sparsity and predictive accuracy effectively.

6. Conclusions

In this paper, the STPISVR model is presented, an efficient sparse regression model that refines TPISVR by reformulating its primal problem into an equivalent optimization in the dual space. In contrast to TPISVR, STPISVR promotes sparsity explicitly by incorporating a 1-norm regularization term on the dual variable, which is subsequently simplified to a constant through the application of the KKT conditions. This reformulation enables direct optimization over the dual variables, reducing the complexity of the original QPPs to LPPs, thereby improving computational efficiency.
Theoretical analysis and experimental evaluations confirm that STPISVR outperforms traditional regression models by achieving higher sparsity, faster prediction speeds, and superior prediction performance. The model’s ability to balance sparsity and computational efficiency makes it particularly advantageous for a wider range of datasets, including those with heteroscedastic noise structures. Notably, the results highlight that while STPISVR retains the efficiency of a linear model on small- and medium-sized datasets, while maintaining manageable computational cost as data scale increases. Moreover, it consistently achieves a favorable trade-off between prediction accuracy and model simplicity.
Future work includes exploring alternative sparse regularization strategies, such as different norms, pseudo-norms, and their continuous nonlinear approximations, as well as adaptive penalties, to further enhance performance across diverse regression tasks. Additionally, the model evaluation will be extended to a broader range of large-scale, high-dimensional real-world datasets to gain deeper insights into its robustness and adaptability. Further theoretical investigations will be conducted to refine the model’s computational complexity and establish formal guarantees, such as statistical consistency and generalization bounds, to enhance its efficiency, robustness, and reliability. Finally, direct comparisons with state-of-the-art sparse regression models will be carried out to better demonstrate STPISVR’s practical advantages in various application scenarios.

Author Contributions

Conceptualization, S.Q. and Y.G.; Formal analysis, S.Q. Methodology, S.Q. and R.D.L.; Validation, S.Q., Y.G. and M.H.; Visualization, S.Q. and P.L.; Investigation, Y.G.; Software, The simulations were conducted using MATLAB R2022a. Y.G., M.H. and P.L.; Writing—original draft, S.Q. and Y.G.; Writing—review and editing, S.Q., R.D.L., P.L. and M.H.; Supervision, R.D.L. and M.H.; Funding acquisition, R.D.L. and P.L. All authors have read and agreed to the published version of the manuscript.

Funding

The European Union—TextGenerationEU under the Italian Ministry of University and Research (MUR) National Innovation Ecosystem grant ECS00000041—VITALITY—CUP J13C22000430001; National Natural Science Foundation Youth Fund project (China, No.61802352).

Data Availability Statement

The original contributions presented in this study are included in the article; further inquiries can be directed to the corresponding author.

Acknowledgments

The authors would like to express gratitude to the reviewers for their valuable opinions and suggestions, as well as to the editorial staff for their hard work, which has been conducive to improving the quality of this article.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar]
  2. Drucker, H.; Burges, C.J.; Kaufman, L.; Smola, A.; Vapnik, V. Support vector regression machines. Adv. Neural Inf. Process. Syst. 1996, 9, 155–161. [Google Scholar]
  3. Smola, A.J.; Schölkopf, B. A tutorial on support vector regression. Stat. Comput. 2004, 14, 199–222. [Google Scholar] [CrossRef]
  4. Pontil, M.; Rifkin, R.; Evgeniou, T. From Regression to Classification in Support Vector Machines. 1998. Available online: https://dspace.mit.edu/bitstream/handle/1721.1/7258/AIM-1649.pdf?sequence=2&isAllowed=y (accessed on 30 May 2025).
  5. Basak, D.; Pal, S.; Patranabis, D.C. Support vector regression. Neural Inf. Process.-Lett. Rev. 2007, 11, 203–224. [Google Scholar]
  6. Awad, M.; Khanna, R.; Awad, M.; Khanna, R. Support vector regression. In Efficient Learning Machines: Theories, Concepts, and Applications for Engineers and System Designers; Apress: Berkeley, CA, USA, 2015; pp. 67–80. [Google Scholar]
  7. Zhang, F.; O’Donnell, L.J. Support vector regression. In Machine Learning; Academic Press: Cambridge, MA, USA, 2020; pp. 123–140. [Google Scholar]
  8. Huang, H.; Wei, X.; Zhou, Y. An overview on twin support vector regression. Neurocomputing 2022, 490, 80–92. [Google Scholar] [CrossRef]
  9. Peng, X. TSVR: An efficient twin support vector machine for regression. Neural Netw. 2010, 23, 365–372. [Google Scholar] [CrossRef]
  10. Hao, P.Y. Pair-v-svr: A novel and efficient pairing nu-support vector regression algorithm. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2503–2515. [Google Scholar] [CrossRef]
  11. Peng, X. Efficient twin parametric insensitive support vector regression model. Neurocomputing 2012, 79, 26–38. [Google Scholar] [CrossRef]
  12. Hao, P.Y. New support vector algorithms with parametric insensitive/margin model. Neural Netw. 2010, 23, 60–73. [Google Scholar] [CrossRef]
  13. Schölkopf, B.; Smola, A.J.; Williamson, R.C.; Bartlett, P.L. New support vector algorithms. Neural Comput. 2000, 12, 1207–1245. [Google Scholar] [CrossRef]
  14. Khemchandani, R.; Chandra, S. Twin support vector machines for pattern classification. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 905–910. [Google Scholar]
  15. Peng, X. TPMSVM: A novel twin parametric-margin support vector machine for pattern recognition. Pattern Recognit. 2011, 44, 2678–2692. [Google Scholar] [CrossRef]
  16. Zhu, J.; Rosset, S.; Tibshirani, R.; Hastie, T. 1-norm support vector machines. Adv. Neural Inf. Process. Syst. 2003, 16, 49–56. [Google Scholar]
  17. Tang, Q.; Li, G. Sparse L0-norm least squares support vector machine with feature selection. Inf. Sci. 2024, 670, 120591. [Google Scholar] [CrossRef]
  18. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B Stat. Methodol. 2005, 67, 301–320. [Google Scholar] [CrossRef]
  19. Huang, K.; Zheng, D.; Sun, J.; Hotta, Y.; Fujimoto, K.; Naoi, S. Sparse learning for support vector classification. Pattern Recognit. Lett. 2010, 31, 1944–1951. [Google Scholar] [CrossRef]
  20. Bi, J.; Bennett, K.; Embrechts, M.; Breneman, C.; Song, M. Dimensionality reduction via sparse support vector machines. J. Mach. Learn. Res. 2003, 3, 1229–1243. [Google Scholar]
  21. Santos, J.D.A.; Barreto, G.A. Novel sparse LSSVR models in primal weight space for robust system identification with outliers. J. Process Control 2018, 67, 129–140. [Google Scholar] [CrossRef]
  22. Cui, L.; Shen, J.; Yao, S. The Sparse Learning of The Support Vector Machine. J. Phys. Conf. Ser. 2021, 2078, 012006. [Google Scholar] [CrossRef]
  23. Moosaei, H.; Mousavi, A.; Hladík, M.; Gao, Z. Sparse l1-norm quadratic surface support vector machine with universum data. Soft Comput. 2023, 27, 5567–5586. [Google Scholar] [CrossRef]
  24. Qu, S.; De Leone, R.; Huang, M. Sparse Learning for Linear Twin Parameter-margin Support Vector Machine. In Proceedings of the 2024 3rd Asia Conference on Algorithms, Computing and Machine Learning, Shanghai, China, 22–24 March 2024; pp. 50–55. [Google Scholar]
  25. Zhang, Z.; Zhen, L.; Deng, N.; Tan, J. Sparse least square twin support vector machine with adaptive norm. Appl. Intell. 2014, 41, 1097–1107. [Google Scholar] [CrossRef]
  26. Smola, A.; Scholkopf, B.; Ratsch, G. Linear programs for automatic accuracy control in regression. In Proceedings of the 1999 Ninth International Conference on Artificial Neural Networks ICANN 99. (Conf. Publ. No. 470), Edinburgh, UK, 7–10 September 1999; Volume 2, pp. 575–580. [Google Scholar]
  27. Breiman, L. Prediction games and arcing algorithms. Neural Comput. 1999, 11, 1493–1517. [Google Scholar] [CrossRef] [PubMed]
  28. Wolfe, P. A duality theorem for non-linear programming. Q. Appl. Math. 1961, 19, 239–244. [Google Scholar] [CrossRef]
  29. Nesterov, Y.; Nemirovskii, A. Interior-Point Polynomial Algorithms in Convex Programming; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 1994. [Google Scholar]
  30. Shahriari, B.; Swersky, K.; Wang, Z.; Adams, R.P.; de Freitas, N. Taking the human out of the loop: A review of Bayesian optimization. Proc. IEEE 2015, 104, 148–175. [Google Scholar] [CrossRef]
Figure 1. Regression results for Example 1 (Type A).
Figure 1. Regression results for Example 1 (Type A).
Mathematics 13 02206 g001
Figure 2. Regression results for Example 1 (Type B).
Figure 2. Regression results for Example 1 (Type B).
Mathematics 13 02206 g002
Figure 3. Regression results for Example 2 (Type C).
Figure 3. Regression results for Example 2 (Type C).
Mathematics 13 02206 g003
Figure 4. Regression results for Example 2 (Type D).
Figure 4. Regression results for Example 2 (Type D).
Mathematics 13 02206 g004
Figure 5. Relationship between ( ρ , v ) and the fraction of SVs and errors.
Figure 5. Relationship between ( ρ , v ) and the fraction of SVs and errors.
Mathematics 13 02206 g005
Figure 6. Relationship between ( ρ , v ) and ZoneWidth.
Figure 6. Relationship between ( ρ , v ) and ZoneWidth.
Mathematics 13 02206 g006
Figure 7. Relationship between ( ρ , v ) and RMSE and R 2 .
Figure 7. Relationship between ( ρ , v ) and RMSE and R 2 .
Mathematics 13 02206 g007
Figure 8. STPISVR performance under varying values of noise standard deviation of ϵ p .
Figure 8. STPISVR performance under varying values of noise standard deviation of ϵ p .
Mathematics 13 02206 g008
Figure 9. Performance of four algorithms under different training sample sizes.
Figure 9. Performance of four algorithms under different training sample sizes.
Mathematics 13 02206 g009
Figure 10. Comparison between STPISVR and TPISVR under different training sample sizes.
Figure 10. Comparison between STPISVR and TPISVR under different training sample sizes.
Mathematics 13 02206 g010
Figure 11. Regression results for the Motocycle dataset.
Figure 11. Regression results for the Motocycle dataset.
Mathematics 13 02206 g011
Table 1. List of notations.
Table 1. List of notations.
NotationDescription
R The real space
H The feature space
| · | The cardinality of a set
· 1 The L1 norm: sum of absolute values of a vector’s components
· 2 The L2 norm: Euclidean norm of a vector
IThe index set of samples: I = { 1 , 2 , , | I | }
nThe dimension of samples
x p the pth sample, x p R n × 1 , p = 1 , 2 , , | I |
XThe samples matrix: X = ( x 1 , x 2 , , x | I | ) T R | I | × n  
y p The response value of x p , p = 1 , 2 , , | I |
YThe response vector, Y = ( y 1 , y 2 , , y | I | ) T R | I | × 1
eThe vector of opportune dimensions with all components equal to 1
w , w d , w u The normal vectors
φ ( · ) The mapping function from a original input space R n to a feature space H
φ ( X ) The mapping matrix: φ ( X ) = ( φ ( x 1 ) , φ ( x 2 ) , , φ ( x | I | ) ) T R | I | × n
k ( u , v ) The kernel function: k ( u , v ) = φ ( u ) T φ ( v ) , u , v are all random n-dimensional vector
KThe kernel matrix: K = φ ( X ) φ ( X ) T R | I | × | I | with k p q = k ( x p , x q ) , p , q I
K ( X , x ) The column vector: K ( X , x ) = ( k ( x 1 , x ) , k ( x 2 , x ) , , k ( x | I | , x ) ) T R | I | × 1
ξ , ξ + , ξ d , ξ u The slack variables
θ , θ d , θ u The bias terms
ρ , ρ d , ρ u , v d , v u The regularization parameters
α , α + Lagrange multiplier vectors, the dual variables in SVR or SSVR
α p The pth element of α , p I
α q + The qth element of α q + I
α d , α u Lagrange multiplier vectors, the dual variables in TPISVR or STPISVR
α p d The pth element of α d , p I
α q u The qth element of α u , q I
ϵ The insensitive tube parameter
Note: The symbol T denotes the transpose of a matrix or vector.
Table 2. Results for Example 1.
Table 2. Results for Example 1.
Type of NoiseEvaluation IndexSVRSSVRTPISVRSTPISVR
Type ARMSE0.1085 ± 0.00530.1063 ± 0.00260.1110 ± 0.00430.1028 ± 0.0041
Noisy test R 2 0.9010 ± 0.01460.9102 ± 0.01020.8965 ± 0.01430.9111 ± 0.0133
Num-SVs107.60 ± 7.806018.80 ± 1.0328154.10 ± 3.634746.60 ± 1.5776
Tr-time (s)0.4377 ± 0.04350.2509 ± 0.02200.8406 ± 0.02530.3270 ± 0.0213
Te-time (s)0.0116 ± 0.00130.0028 ± 0.00080.0186 ± 0.00290.0060 ± 0.0012
Noiseless testRMSE0.0447 ± 0.00500.0377 ± 0.00350.0487 ± 0.00490.0366 ± 0.0049
R 2 0.9815 ± 0.00380.9844 ± 0.00120.9781 ± 0.00340.9873 ± 0.0042
Te-time (s)0.0117 ± 0.00140.0021 ± 0.00040.0172 ± 0.00060.0051 ± 0.0005
Type BRMSE0.1281 ± 0.00400.1194 ± 0.00480.1261 ± 0.00450.1171 ± 0.0038
Noisy test R 2 0.8645 ± 0.01430.8820 ± 0.01530.8648 ± 0.01690.8866 ± 0.0151
Num-SVs214.50 ± 11.365215.30 ± 0.9487164.30 ± 4.347445.50 ± 1.7159
Tr-time (s)0.4605 ± 0.04230.2753 ± 0.03490.7615 ± 0.03240.3232 ± 0.0113
Te-time (s)0.0268 ± 0.00380.0026 ± 0.00050.0198 ± 0.00320.0059 ± 0.0009
Noiseless testRMSE0.0548 ± 0.00690.0277 ± 0.00550.0476 ± 0.00510.0140 ± 0.0036
R 2 0.9719 ± 0.00730.9927 ± 0.00290.9787 ± 0.00560.9981 ± 0.0010
Te-time (s)0.0244 ± 0.00250.0022 ± 0.00060.0207 ± 0.00250.0048 ± 0.0004
Type A RMSE0.1089 ± 0.00490.1076 ± 0.00300.1115 ± 0.00450.1070 ± 0.0040
Noisy test R 2 0.9003 ± 0.01190.9033 ± 0.01110.8957 ± 0.01450.9065 ± 0.0143
Num-SVs173.00 ± 8.326717.00 ± 0.4714138.50 ± 3.0277101.50 ± 1.9579
Tr-time (s)0.4255 ± 0.04430.9490 ± 0.05580.8805 ± 0.04840.8482 ± 0.0442
Te-time (s)0.0198 ± 0.00320.0016 ± 0.00050.0170 ± 0.00250.0125 ± 0.0016
Noiseless testRMSE0.0448 ± 0.00650.0385 ± 0.00330.0503 ± 0.00470.0379 ± 0.0058
R 2 0.9809 ± 0.00390.9838 ± 0.00130.9766 ± 0.00370.9864 ± 0.0034
Te-time (s)0.0191 ± 0.00230.0019 ± 0.00020.0167 ± 0.00110.0059 ± 0.0006
Note: indicates the results obtained using grid search alone for parameter tuning.
Table 3. Results for Example 2.
Table 3. Results for Example 2.
Type of NoiseEvaluation IndexSVRSSVRTPISVRSTPISVR
Type CRMSE0.0518 ± 0.00270.0520 ± 0.00220.0508 ± 0.00250.0505 ± 0.0023
Noisy test R 2 0.7950 ± 0.01630.7950 ± 0.01700.8040 ± 0.01140.8070 ± 0.0104
Num-SVs133.10 ± 5.83954.30 ± 0.483043.0 ± 0.666726.0 ± 0.6667
Tr-time (s)0.2779 ± 0.02430.1143 ± 0.02020.5414 ± 0.04200.1721 ± 0.0946
Te-time (s)0.0096 ± 0.00120.0008 ± 0.00050.0036 ± 0.00070.0020 ± 0.0004
Noiseless testRMSE0.0146 ± 0.00290.0135 ± 0.00200.0097 ± 0.00210.0061 ± 0.0020
R 2 0.9795 ± 0.00770.9827 ± 0.00560.9910 ± 0.00360.9962 ± 0.0021
Te-time (s)0.0093 ± 0.00100.0007 ± 0.00040.0035 ± 0.00110.0020 ± 0.0002
Type DRMSE0.0455 ± 0.00250.0447 ± 0.00240.0447 ± 0.00240.0445 ± 0.0025
Noisy test R 2 0.8350 ± 0.02220.8410 ± 0.02130.8410 ± 0.02030.8419 ± 0.0222
Num-SVs78.90 ± 4.40836.20 ± 0.4216117.50 ± 3.274578.70 ± 0.6749
Tr-time (s)0.2892 ± 0.01540.5111 ± 0.03220.5069 ± 0.01830.1587 ± 0.0174
Te-time (s)0.0052 ± 0.00050.0007 ± 0.00030.0081 ± 0.00110.0059 ± 0.0005
Noiseless testRMSE0.0116 ± 0.00200.0088 ± 0.00280.0087 ± 0.00310.0085 ± 0.0025
R 2 0.9872 ± 0.00430.9923 ± 0.00550.9924 ± 0.00440.9928 ± 0.0037
Te-time (s)0.0052 ± 0.00050.0006 ± 0.00010.0073 ± 0.00080.0060 ± 0.0005
Table 4. Results for 10 benchmark datasets.
Table 4. Results for 10 benchmark datasets.
DatasetEvaluation IndexSVRSSVRTPISVRSTPISVR
D1RMSE0.7107 ± 0.08680.7121 ± 0.10000.7073 ± 0.08250.7034 ± 0.0548
(398 × 8) R 2 0.4721 ± 0.17050.4519 ± 0.13730.4765 ± 0.08460.4897 ± 0.606
Num-SVs196.20 ± 5.165132.60 ± 3.0258266.70 ± 5.5187188.80 ± 4.4920
Tr-time (s)0.0215 ± 0.00740.3734 ± 0.05960.9445 ± 0.02600.5137 ± 0.1133
Te-time (s)0.0017 ± 0.00140.0003 ± 0.00020.0019 ± 0.00030.0015 ± 0.0002
D2RMSE0.3801 ± 0.21370.3741 ± 0.16540.3940 ± 0.17200.3712 ± 0.1163
(159 × 16) R 2 0.7813 ± 0.15430.7958 ± 0.09440.7761 ± 0.14260.7983 ± 0.1179
Num-SVs79.60 ± 3.025852.80 ± 2.8206106.50 ± 4.143337.10 ± 2.5582
Tr-time (s)0.0269 ± 0.00790.0868 ± 0.01250.5812 ± 0.02440.1352 ± 0.0806
Te-time (s)0.0010 ± 0.00060.0006 ± 0.00100.0006 ± 0.00050.0003 ± 0.0002
D3RMSE0.3911 ± 0.09280.3812 ± 0.10290.3556 ± 0.10250.3426 ± 0.0896
(506 × 14) R 2 0.8412 ± 0.05160.8422 ± 0.06310.8516 ± 0.08530.8554 ± 0.0583
Num-SVs273.20 ± 5.370761.50 ± 2.9533307.10 ± 5.3635226.20 ± 3.3928
Tr-time (s)0.1777 ± 0.02150.8818 ± 0.05021.1177 ± 0.05440.8554 ± 0.0583
Te-time (s)0.0020 ± 0.00050.0009 ± 0.00050.0029 ± 0.00040.0018 ± 0.0004
D4RMSE0.3384 ± 0.06100.3372 ± 0.05330.4223 ± 0.05040.3902 ± 0.0465
(1030 × 9) R 2 0.8783 ± 0.04910.8795 ± 0.04590.8173 ± 0.04890.8424 ± 0.0413
Num-SVs603.70 ± 7.3944156.90 ± 4.0947634.60 ± 12.1546485.6 ± 4,8808
Tr-time (s)0.8718 ± 0.09049.4158 ± 0.35004.7170 ± 0.15323.1539 ± 0.2175
Te-time (s)0.0110 ± 0.00090.0036 ± 0.00090.0122 ± 0.00180.0103 ± 0.0009
D5RMSE0.1058 ± 0.08750.1108 ± 0.10470.1012 ± 0.13130.1013 ± 0.1299
(60 × 10) R 2 0.9566 ± 0.05740.9536 ± 0.08630.9659 ± 0.03870.9657 ± 0.0361
Num-SVs36.90 ± 1.852911.60 ± 0.843339.00 ± 2.357023.80 ± 0.8756
Tr-time (s)0.0070 ± 0.00130.0102 ± 0.00190.5196 ± 0.04630.0805 ± 0.0021
Te-time (s)0.0007 ± 0.00040.0004 ± 0.00070.0003 ± 0.00040.0002 ± 0.0002
D6RMSE0.0878 ± 0.06860.0906 ± 0.08590.1816 ± 0.28150.1824 ± 0.1871
(209 × 7) R 2 0.9880 ± 0.00460.9845 ± 0.01410.9549 ± 0.05360.9546 ± 0.0234
Num-SVs30.90 ± 1.523918.90 ± 1.2649122.30 ± 19.3861103.40 ± 1.2293
Tr-time (s)0.0664 ± 0.03300.1188 ± 0.01670.6926 ± 0.13720.1327 ± 0.0395
Te-time (s)0.0012 ± 0.00150.0004 ± 0.00050.0006 ± 0.00040.0004 ± 0.0002
D7RMSE0.5786 ± 0.19700.6029 ± 0.24140.5576 ± 0.21190.5362 ± 0.1878
(107 × 4) R 2 0.5454 ± 0.19820.5429 ± 0.15410.5555 ± 0.27200.5569 ± 0.2134
Num-SVs73.30 ± 2.633110.40 ± 1.173849.10 ± 2.936433.70 ± 2.0248
Tr-time (s)0.0180 ± 0.00610.0215 ± 0.00610.5677 ± 0.04460.0861 ± 0.0026
Te-time (s)0.0009 ± 0.00070.0003 ± 0.00050.0004 ± 0.00050.0003 ± 0.0002
D8RMSE0.4619 ± 0.11200.4616 ± 0.12380.4531 ± 0.10120.4398 ± 0.0918
(133 × 2) R 2 0.6967 ± 0.34890.6969 ± 0.21390.7055 ± 0.15600.8256 ± 0.2867
Num-SVs74.80 ± 2.406013.40 ± 0.699266.80 ± 1.636435.40 ± 0.8756
Tr-time (s)0.0138 ± 0.02670.0555 ± 0.01590.5382 ± 0.02090.1007 ± 0.0094
Te-time (s)0.0008 ± 0.00060.0004 ± 0.00040.0007 ± 0.00040.0004 ± 0.0002
D9RMSE0.1008 ± 0.02110.1391 ± 0.02230.1764 ± 0.07370.1692 ± 0.0663
(308 × 6) R 2 0.9846 ± 0.00330.9705 ± 0.00540.9450 ± 0.03960.9613 ± 0.0405
Num-SVs73.20 ± 3.794761.10 ± 1.7029156.00 ± 1.825785.00 ± 2.9814
Tr-time (s)0.0445 ± 0.04250.3344 ± 0.05640.8192 ± 0.04300.2916 ± 0.0124
Te-time (s)0.0008 ± 0.00050.0006 ± 0.00050.0010 ± 0.00040.0007 ± 0.0004
D10RMSE0.6999 ± 0.07800.6848 ± 0.07420.6681 ± 0.10200.6429 ± 0.0711
(536 × 8) R 2 0.4870 ± 0.08340.5089 ± 0.07610.5343 ± 0.08080.5428 ± 0.0547
Num-SVs475.20 ± 2.529816.90 ± 1.1005288.40 ± 2.7568238.50 ± 1.3540
Tr-time (s)0.0378 ± 0.01620.5449 ± 0.05051.3134 ± 0.13160.5132 ± 0.0426
Te-time (s)0.0019 ± 0.00100.0005 ± 0.00040.0023 ± 0.00050.0020 ± 0.0002
D8 RMSE0.4911 ± 0.11150.4759 ± 0.10570.4673 ± 0.10270.4464 ± 0.1179
(133 × 2) R 2 0.5718 ± 0.45100.6161 ± 0.34900.6189 ± 0.57900.6572 ± 0.5867
Num-SVs102.20 ± 1.316011.00 ± 0.471443.20 ± 0.788832.50 ± 0.8498
Tr-time (s)0.3333 ± 0.04310.3548 ± 0.03230.5215 ± 0.02830.5054 ± 0.0165
Te-time (s)0.0017 ± 0.00150.0004 ± 0.00050.0006 ± 0.00040.0005 ± 0.0003
Note: indicates the results obtained using grid search alone for parameter tuning.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qu, S.; Guo, Y.; De Leone, R.; Huang, M.; Li, P. An Efficient Sparse Twin Parametric Insensitive Support Vector Regression Model. Mathematics 2025, 13, 2206. https://doi.org/10.3390/math13132206

AMA Style

Qu S, Guo Y, De Leone R, Huang M, Li P. An Efficient Sparse Twin Parametric Insensitive Support Vector Regression Model. Mathematics. 2025; 13(13):2206. https://doi.org/10.3390/math13132206

Chicago/Turabian Style

Qu, Shuanghong, Yushan Guo, Renato De Leone, Min Huang, and Pu Li. 2025. "An Efficient Sparse Twin Parametric Insensitive Support Vector Regression Model" Mathematics 13, no. 13: 2206. https://doi.org/10.3390/math13132206

APA Style

Qu, S., Guo, Y., De Leone, R., Huang, M., & Li, P. (2025). An Efficient Sparse Twin Parametric Insensitive Support Vector Regression Model. Mathematics, 13(13), 2206. https://doi.org/10.3390/math13132206

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop