Next Article in Journal
Assessment of Tree Species Classification by Decision Tree Algorithm Using Multiwavelength Airborne Polarimetric LiDAR Data
Next Article in Special Issue
Observer-Based Robust H Control for Stochastic Markov Jump Delay Systems Through Dual Adaptive Sliding Mode Approach
Previous Article in Journal
SG-LPR: Semantic-Guided LiDAR-Based Place Recognition
Previous Article in Special Issue
An Explainable Data-Driven Optimization Method for Unmanned Autonomous System Performance Assessment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Twin Extreme Learning Machine Based on Soft Truncated Capped L1-Norm Loss Function

School of Mathematics and Information Sciences, North Minzu University, Yinchuan 750021, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(22), 4533; https://doi.org/10.3390/electronics13224533
Submission received: 13 October 2024 / Revised: 11 November 2024 / Accepted: 13 November 2024 / Published: 19 November 2024
(This article belongs to the Special Issue Advanced Control Strategies and Applications of Multi-Agent Systems)

Abstract

:
Currently, most researchers propose robust algorithms from different perspectives for overcoming the impact of outliers on a model, such as introducing loss functions. However, some loss functions often fail to achieve satisfactory results when the outliers are large. Therefore, the capped loss has become a better choice for researchers. The majority of researchers directly set an upper bound on the loss function, which reduces the impact of large outliers, but also introduces non-differentiable regions. To avoid this shortcoming, we propose a robust twin extreme learning machine based on a soft-capped L 1 -normal loss function (SCTELM). It uses a soft capped L 1 -norm loss function. This not only overcomes the shortcomings of the hard capped loss function, but also improves the robustness of the model. Simultaneously, to improve the learning efficiency of the model, the stochastic variance-reduced gradient (SVRG) optimization algorithm is used. Experimental results on several datasets show that the proposed algorithm can compete with state-of-the-art algorithms in terms of robustness.

1. Introduction

Extreme learning machine (ELM) [1,2] was proposed by Huang et al. on a single hidden layer feedforward network. Compared with traditional training methods, ELM has advantages in learning speed and generalization performance. This is because ELM differs from traditional methods, that is, the input consists of randomly generated weights and bias values of the hidden layer nodes, allowing for the quick and effective determination of output weights. Additionally, The ultimate aim of ELM is to reduce the training error and the norm of the weights, which enhances ELM’s generalization performance in feedforward neural networks. Nowadays, ELM is widely used in ship detection [3], online visual tracking [4], image quality assessment [5], and other fields.
In recent years, many excellent scholars have proposed different algorithms based on ELM and achieved good performance. However, most of the ELM-based algorithms are more sensitive to outliers and affect the performance of the model. In recent years, the robust learning algorithm has become a new direction of machine learning. In order to overcome the influence of outliers on ELM, researchers have proposed some robust algorithms based on ELM from different perspectives [6,7,8,9,10,11,12]. From the perspective of loss function and based on expected penalty and correntropy, Wu et al. [7] proposed an asymmetric non-convex bounded loss function (AC-loss) and obtained a robust ELM learning framework of L 1 -norm robust regularized extreme learning machine with asymmetric AC-loss ( L 1 -ACELM). Ren and Yang [11] proposed a new hybrid loss function using pinball loss and least squares loss. Then, based on the hybrid loss function, a robust ELM framework with a hybrid loss function was proposed. Luo et al. [12] proposed that the correntropy loss function induced by the p-order Laplace kernel can be used to replace the L 2 loss function to reduce the influence of outliers during learning.
In recent years, the  L 1 -norm, called the sparse rule operator, has been given attention and applied in more and more fields, and it is considered to be robust to outliers [13]. Zhang and Luo et al. [14] took advantage of the robustness of L 1 -norm to introduce the L 1 loss function into ELM to solve the outlier problem. Although  L 1 -norm performs well in improving the robustness of the algorithm, most existing algorithms based on L 1 -norm may not obtain a satisfactory result when the outliers are large. Therefore, more and more scholars have tried to use the capped L 1 -norm to achieve some good results [15,16,17,18,19]. Following is a brief review of some representative works, Wang et al. [16] proposed a new capped L 1 -norm twin SVM. Wang and Chen et al. [17] proposed a capped L 1 -norm sparse representation method (CSR) to remove outliers and ensure the quality of the graph. Li et al. [18] proposed a new robust capped L 1 -norm twin support vector machine (R-CTSVM+), which adopted the capped L 1 regularization distance to ensure the robustness of the model.
In recent years, inspired by SVM and GEPSVM, Jayadeva et al. [20] proposed twin SVM (TSVM). Compared with SVM, TWSVM shows great improvement in computational complexity and computational speed because it improves the speed of work by solving a large quadratic programming problem into a pair of small quadratic programming problems, which is theoretically four times faster than the traditional SVM [20,21]. Due to the excellent performance of TSVM, scholars have proposed many variants of TSVM. For example, Kumar et al. and Shao et al. proposed least squares TSVM (LSTSVM) [22] and twin bounded SVM (TBSVM) [23] based on TSVM. LSTSVM simplifies the solution of a pair of complex QPPs in TSVM to only solving two linear equations, which greatly improves the operation speed but loses the sparsity, and TBSVM considers the minimization of the structural risk of TSVM by introducing a regulaization term. These improvements make TSVM closer to perfection, while the recognition ability and adaptability are also strengthened. Wan et al. [24] proposed twin extreme learning machine inspired by TWSVM. He combined twin support vector machine (TWSVM) with extreme learning machine (ELM), so TELM has the advantages of both TWSVM and ELM. At the same time, compared with ELM, TELM has faster learning speed.
Most researchers use the hard capped strategy to enhance the robustness of models by reducing the impact of noisy data, but it has inherent shortcomings, such as inadvertently ignoring key data and introducing non-differentiable regions. To address this, this paper proposes a soft capped loss function based on the L 1 -norm (SC-loss) and introduces SC-loss into TELM to construct the SCTELM model. SC-loss is a soft capped loss function that reaches the upper limit more smoothly than hard capped and can be applied to various machine learning problems, effectively reducing the impact of large outliers on the model. We also employ stochastic variance-reduced gradient (SVRG) to further improve the learning rate. A large number of experimental results show that SCTELM is more competitive in terms of accuracy and learning efficiency.
The main contributions of this paper are as follows:
  • This paper proposes an efficient and reliable learning framework based on TELM, namely robust twin extreme learning machine with the L 1 -norm loss function based on the soft capped strategy (SCTELM).
  • This paper proposes an L 1 -norm loss with soft capped, which can reach the upper bound more smoothly to reduce the impact of large outliers on the model.
  • Experimental results on various datasets show that our proposed SCTELM is competitive with other algorithms in terms of accuracy and learning efficiency.
The rest of this paper is organized as follows, and Section 2 briefly reviews some properties of ELM, TWSVM and TELM. In Section 3, some details of our proposed SC-loss and SCTELM are described and the related theoretical analysis is given. Experimental results on multiple datasets are presented in Section 4. Section 5 concludes the paper.

2. Related Work

In this section, we will give a brief overview of extreme learning machine (ELM), twin support vector machine (TWSVM) and twin extreme learning machine (TELM), along with an introduction to the symbols covered in this paper.

2.1. Notations

In this paper, uppercase letters are used to represent matrices while lowercase letters represent vectors. All vectors are column vectors. The symbol ( . ) T denotes transpose, e represents an all-one column vector of any dimension and I stands for the identity matrix.
The model being discussed in this article is a binary classification model implemented in R n . The sample data comprise l 1 positive samples labeled with +1, and  l 2 negative samples labeled with −1, and  l 1 + l 2 = l . Let the matrix A R l 1 × n , B R l 2 × n represent the positive and negative classes, respectively.

2.2. ELM

The extreme learning machine (ELM) is a generalized single-layer feedforward network (SLFN) proposed by Huang et al. It employs the random initialization of the connection weights and biases between the hidden layer and the output layer. The output function of ELM is
f ( x ) = h ( x ) β
The ELM algorithm is shown in Figure 1, where β is the weight between the hidden layer and the output layer, and  h ( x ) = ( h 1 ( x ) , h 2 ( x ) , h 3 ( x ) , , h L ( x ) ) R 1 × L is the random mapping of the hidden layer for the input pattern x = ( x 1 , x 2 , x 3 , , x n ) T . L is the number of hidden nodes of ELM and h j ( x ) = g ( i = 1 L x i ω i j + b j ) is the activation function for example e x e x + 1 .
To obtain the optimal β , it is necessary to ensure that the training error is minimized. The optimal solution can be achieved by minimizing the norm of the difference between H β and the labeled sample T, which involves minimizing both the norm β of the training error and output weights.
min H β T a n d min β
where H is the output matrix of the hidden layer and T is the sample label matrix, H = ( h ( x 1 ) T , h ( x 2 ) T , h ( x 3 ) T , , h ( x n ) T ) T , T = ( t 1 , t 2 , t 3 , , t n ) T .
Thus, the optimal solution of Equation (2) can be deduced as
β = H T
where H is the Moore–Penrose generalized inverse matrix of the matrix H.

2.3. TWSVM

The TWSVM algorithm is used to find two non-parallel hyperplanes, one for each class, such that the sample points of one class are as close to this hyperplane as possible, while the sample points of the other class are as far away as possible. Finally, whether a point belongs to this class is judged by judging whether it is close to this type of hyperplane. The two hyperplanes are written, respectively, as follows:
f 1 ( x ) = ω 1 T x + b 1 = 0 a n d f 2 ( x ) = ω 2 T x + b 2 = 0
TWSVM is composed of two quadratic programming problems instead of a QPP problem, so TSVM produces two QPP problems of smaller size, and it is faster than the traditional SVM in computing time. The TWSVM classifier is obtained by solving the following two quadratic programming problems:
min ω 1 , b 1 1 2 ( A ω 1 + e 1 b 1 ) T ( A ω 1 + e 1 b 1 ) + C 1 e 2 T ξ , s . t ( B ω 1 + e 2 b 1 ) + ξ e 2 , ξ 0 .
and
min ω 2 , b 2 1 2 ( B ω 2 + e 2 b 2 ) T ( B ω 2 + e 2 b 2 ) + C 2 e 1 T ξ , s . t ( A ω 2 + e 1 b 2 ) + ξ e 1 , ξ 0 .
where C 1 0 and C 2 0 are regularization parameters; and ξ is the slack vectors. After using the Lagrange multipliers α 0 , β 0 , and the Karush–Kuhn–Tucker (K.K.T.) conditions, the duals of the QPPs in (5) and (6) are defined as follows:
min α 1 2 α T G ( H T H ) 1 G T α e 2 T α , s . t 0 α C 1 e 2 .
min β 1 2 β T H ( G T G ) 1 H T β e 2 T β , s . t 0 β C 2 e 1 .
where H = [ A , e 1 ] and G = [ B , e 2 ] . After solving (7) and (8), the proximal hyperplanes are given as follows:
[ w 1 , b 1 ] T = ( H T H + ϵ I ) 1 G T α , [ w 2 , b 2 ] T = ( G T G + ϵ I ) 1 H T β .
where ϵ I is a regularization term and ϵ > 0 . After we obtain ( w 1 , b 1 ) , ( w 2 , b 2 ) , we can classify the new sample point x according to the following decision function:
y = a r g m i n { | f 1 ( x ) | w 1 , | f 2 ( x ) | w 2 }

2.4. TELM

The concept of TSVM is followed by TELM, which also determines two non-parallel hyperplanes in ELM:
f 1 ( x ) = h ( x ) β 1 = 0 a n d f 2 ( x ) = h ( x ) β 2 = 0
Similar to TWSVM, the primal problem of TELM can be expressed as
min β 1 , ξ 1 2 U β 1 2 2 + C 1 e 2 T ξ s . t V β 1 + ξ e 2 , ξ 0 .
min β 2 , η 1 2 V β 2 2 2 + C 2 e 1 T η s . t U β 2 + η e 1 , η 0 .
where C 1 , C 2 > 0 are trade-off parameters; ξ  and  η  are non-negative slack vectors; and U = h ( x 1 ) T , h ( x 2 ) T , , h ( x l 1 ) T T R l 1 × L and V = h ( x 1 ) T , h ( x 2 ) T , , h ( x l 2 ) T T R l 2 × L correspond to the output matrices of the hidden layer for the positive and negative classes, respectively. Based on the the Karush–Kuhn–Tucker (K.K.T.) conditions, training such a TELM is equivalent to solving the following dual optimization problems:
max α e 2 T α 1 2 α T V ( U T U ) 1 V T α , s . t 0 α C 1 e 2 ,
max γ e 2 T γ 1 2 γ T U ( V T V ) 1 U T γ , s . t 0 γ C 2 e 1 .
where α and γ are the vectors of the Lagrange multiplier. Simultaneously,  β 1 = ( U T U + ϵ I ) 1 V T α  and β 2 = ( V T V + ϵ I ) 1 U T γ  can be obtained. When a new sample vector x is obtained, the positive and negative classes of x can be divided according to the following decision function:
f ( x ) = a r g m i n { | h ( x ) β 1 | , | h ( x ) β 2 | }

3. Main Contributions

In this section, we introduce a capped L 1 -norm loss into TELM, thereby establishing the proposed SCTELM in this paper. We optimize it using the SVRG algorithm and further conduct a convergence analysis of the model to investigate its stability.

3.1. Capped SC-Loss Function

To alleviate the non-differentiable region that may be introduced by hard capped, we propose the soft-truncated SC-loss function L θ as follows:
Definition 1.
Given a vector u, the  L θ function is defined as
L θ ( u ) = 1 1 1 + θ | u |
where θ 0 is the adaptive parameter of the loss function. For example, as shown in Figure 2, the capped L 1 -norm ( ϵ = 1 ) loss is non-differentiable at u = 1 , and our proposed L θ -loss can avoid this problem, thus making it more convenient for us when dealing with it.
The following are some properties of the L θ - l o s s :
Theorem 1.
L θ is a symmetric function.
Proof of Theorem 1.
L θ ( u ) = 1 1 1 + θ | u | = 1 1 1 + θ | u | = L θ ( u )
   □
Theorem 2.
L θ is a non-negative function.
Proof of Theorem 2.
L θ ( u ) = 1 1 1 + θ | u | = θ | u | 1 + θ | u | 0
   □
Theorem 3.
L θ is a bounded function.
Proof of Theorem 3.
lim u L θ ( u ) = 1 1 1 + θ | u | = 1
   □
Remark 1.
Here, 1 is the maximum value of L θ ( u ) , which determines the last term of the loss function and is therefore robust.

3.2. Linear SCTELM

Based on (11), (12) and our proposed L θ -loss function, we can obtain the formula for SCTELM, expressed as follows:
SCTELM1: 
min β 1 1 2 U β 1 2 + C 1 i = 1 l 2 ( 1 1 1 + θ | 1 + h ( x i ) β 1 | )
SCTELM2: 
min β 2 1 2 V β 2 2 + C 2 j = 1 l 2 ( 1 1 1 θ | 1 + h ( x j ) β 2 | )
for SCTELM1, substituting β 1 = ( U T U + ϵ 1 I ) 1 V T α into U β 1 2 gives
U β 1 2 = ( U β 1 ) T ( U β 1 ) = [ U ( U T U + ϵ 1 I ) 1 V T α ] T [ U ( U T U + ϵ 1 I ) 1 V T α ] = α T V ( U T U + ϵ 1 I ) 1 U T U ( U T U + ϵ 1 I ) 1 V T α
Let D 1 = ( U T U + ϵ I ) 1 U T U ( U T U + ϵ I ) 1 ; there is
U β 1 2 = ( U β 1 ) T ( U β 1 ) = α T V D 1 V T α = i = 1 l 2 α i h ( x i ) D 1 k = 1 l 2 h ( x k ) T α k = i = 1 l 2 k = 1 l 2 α i α k h ( x i ) D 1 h ( x k ) T
Substituting β 1 = ( U T U + ϵ I ) 1 V T α into h ( x i ) β 1 gives
h ( x i ) β 1 = h ( x i ) ( U T U + ϵ I ) 1 V T α
Let E 1 = ( U T U + ϵ I ) 1 ; there is
h ( x i ) β 1 = h ( x i ) E 1 V T α = h ( x i ) E 1 k = 1 l 2 h ( x k ) T α k = k = 1 l 2 α k h ( x i ) E 1 h ( x k ) T
Thus, (20) can be expressed as
min α i 1 2 i = 1 l 2 k = 1 l 2 α i α k h ( x i ) D 1 h ( x k ) T + C 1 i = 1 l 2 1 1 1 + θ | 1 k = 1 l 2 α k h ( x i ) E 1 h ( x k ) T |
For a given sample x i , its objective function is denoted by
f i ( α ) = 1 2 k = 1 l 2 α i α k h ( x i ) D 1 h ( x k ) T + C 1 1 1 1 + θ | 1 k = 1 l 2 α k h ( x i ) E 1 h ( x k ) T |
If 1 k = 1 l 2 α j h ( x i ) E 1 h ( x k ) T 0 , there is
f i ( α ) = 1 2 k = 1 l 2 α i α k h ( x i ) D 1 h ( x k ) T + C 1 1 1 1 + θ ( j = 1 l 2 α k h ( x i ) E 1 h ( x k ) T 1 ) = 1 2 k = 1 l 2 α i α k h ( x i ) D 1 h ( x k ) T + C 1 θ ( k = 1 l 2 α k h ( x i ) E 1 h ( x k ) T 1 ) 1 + θ ( k = 1 l 2 α k h ( x i ) E 1 h ( x k ) T 1 )
Let ξ i ( α ) = 1 k = 1 l 2 α k h ( x i ) E 1 h ( x k ) T , there is
f i ( α ) = 1 2 k = 1 l 2 α i α k h ( x i ) D 1 h ( x k ) T C 1 θ ξ i ( α ) 1 θ ξ i ( α )
and the gradient of f i ( α ) with respect to α can be represented by
f i ( α ) = f i ( α ) α = V D 1 V T α l 2 + C 1 θ V E 1 h ( x i ) T ( 1 θ ξ i ( α ) ) 2
If 1 k = 1 l 2 α k h ( x i ) E 1 h ( x k ) T > 0 , we then have
f i ( α ) = 1 2 k = 1 l 2 α i α k h ( x i ) D 1 h ( x k ) T + C 1 1 1 1 + θ ( 1 k = 1 l 2 α k h ( x i ) E 1 h ( x k ) T ) = 1 2 k = 1 l 2 α i α k h ( x i ) D 1 h ( x k ) T + C 1 θ ( 1 k = 1 l 2 α k h ( x i ) E 1 h ( x k ) T ) 1 + θ ( 1 k = 1 l 2 α k h ( x i ) E 1 h ( x k ) T ) = 1 2 k = 1 l 2 α i α k h ( x i ) D 1 h ( x k ) T + C 1 θ ξ i ( α ) 1 + θ ξ i ( α )
and the gradient of f i ( α ) with respect to α can be represented by
f i ( α ) = f i ( α ) α = V D 1 V T α l 2 C 1 θ V E 1 h ( x i ) T ( 1 + θ ξ i ( α ) ) 2
Combining (30) and (32), the gradient of f i ( α ) with respect to α can be represented by
f i ( α ) = V D 1 V T α l 2 C 1 θ V E 1 h ( x i ) T ( 1 + s g n ( ξ i ( α ) ) θ ξ i ( α ) ) 2 s g n ( ξ i ( α ) )
Then, the average gradient over l samples can be computed as
R l 2 = 1 l 2 i = 1 l 2 f i ( α ) = V D 1 V T α l 2 C 1 θ V E 1 l 2 i = 1 l 2 h ( x i ) T ( 1 + s g n ( ξ i ( α ) ) θ ξ i ( α ) ) 2 s g n ( ξ i ( α ) )
The same can be said for the following:
f j ( γ ) = U D 2 U T γ l 1 C 2 θ U E 2 h ( x j ) T ( 1 + s g n ( η j ( γ ) ) θ η j ( γ ) ) 2 s g n ( η j ( γ ) )
R l 1 = U D 2 U T γ l 1 C 2 θ U E 2 l 1 j = 1 l 1 h ( x j ) T ( 1 + s g n ( η j ( γ ) ) θ η j ( γ ) ) 2 s g n ( η j ( γ ) )
where D 2 = ( V T V + ϵ 2 I ) 1 V T V ( V T V + ϵ 2 I ) 1 , η j ( γ ) = 1 m = 1 l 1 γ m h ( x j ) D 2 h ( x m ) T , E 2 = ( V T V + ϵ 2 I ) 1 . From the above equation, we can obtain α , γ ; when we obtain a new sample vector x, we can make predictions according to the following decision function:
f ( x ) = a r g m i n { | h ( x ) ( U T U + ϵ 2 I ) 1 V T α | , | h ( x ) ( V T V + ϵ 2 I ) 1 U T γ | }

3.3. Nonlinear SCTELM

Sometimes, we will encounter datasets that are not linearly separable. In such a case, we can extend SCTELM to the nonlinear case by considering the surfaces generated by the following two kernels:
K E L M ( x T , G T ) β 1 = 0 , K E L M ( x T , G T ) β 2 = 0 ,
where G = ( A ; B ) , and K E L M is the kernel function of ELM, which can be expressed as K E L M = h ( x i ) h ( x j ) . From this, the original problem of TELM can be expressed as follows:
min β 1 , ξ 1 2 K E L M ( A T , G T ) β 1 2 2 + C 1 e 2 T ξ s . t K E L M ( B T , G T ) β 1 + ξ e 2 , ξ 0 .
min β 2 , η 1 2 K E L M ( B T , G T ) β 2 2 2 + C 2 e 1 T η s . t K E L M ( A T , G T ) β 2 + η e 1 , η 0 .
From the KKT condition, we can obtain β 1 = ( R T R + ϵ I ) 1 S T α  and  β 2 = ( S T S + ϵ I ) 1 R T γ , where R = K E L M ( A T , G T )  and  S = K E L M ( B T , G T ) , and α a n d γ are Lagrange multipliers. The original problem for SCTELM then becomes the following:
SCTELM1: 
min β 1 1 2 K E L M ( A T , G T ) β 1 2 + C 1 i = 1 l 2 ( 1 1 1 + θ | 1 + h ( x i ) β 1 | )
SCTELM2: 
min β 2 1 2 K E L M ( B T , G T ) β 2 2 + C 2 j = 1 l 2 ( 1 1 1 θ | 1 + h ( x j ) β 2 | )
Following the way linear SCTELM is computed, we can obtain the following, according to SCTELM1:
f i ( α ) = V D 3 S T α l 2 C 1 θ S E 3 h ( x i ) T ( 1 + s g n ( ξ i ( α ) ) θ ξ i ( α ) ) 2 s g n ( ξ i ( α ) )
R l 2 = = S D 3 S T α l 2 C 1 θ S E 3 l 2 i = 1 l 2 h ( x i ) T ( 1 + s g n ( ξ i ( α ) ) θ ξ i ( α ) ) 2 s g n ( ξ i ( α ) )
and according to SCTELM2:
f j ( γ ) = R D 4 R T γ l 1 C 2 θ R E 4 h ( x j ) T ( 1 + s g n ( η j ( γ ) ) θ η j ( γ ) ) 2 s g n ( η j ( γ ) )
R l 1 = R D 4 R T γ l 1 C 2 θ R E 4 l 1 j = 1 l 1 h ( x j ) T ( 1 + s g n ( η j ( γ ) ) θ η j ( γ ) ) 2 s g n ( η j ( γ ) )
where D 3 = ( R T R + ϵ 2 I ) 1 R T R = R ( R T R + ϵ 2 I ) 1 , E 3 = ( R T R + ϵ 2 I ) 1 , ζ i ( α ) = 1 k = 1 l 2 α k h ( x i ) E 3 h ( x k ) T , D 4 = ( R T R + ϵ 2 I ) 1 R T R = R ( R T R + ϵ 2 I ) 1 , E 4 = ( R T R + ϵ 2 I ) 1 , ρ j ( γ ) = 1 m = 1 l 1 γ m h ( x j ) E 4 h ( x m ) T . In the same way, from the above equation, we can obtain α , γ . When we obtain a new sample vector x, we can make predictions according to the following decision function:
f ( x ) = a r g m i n { | h ( x ) ( R T R + ϵ 2 I ) 1 S T α | , | h ( x ) ( S T S + ϵ 2 I ) 1 R T γ | }

3.4. Convergence Analysis

Consider a fixed-stage n in Algorithm 1. Let ω = arg min α f ( α ) . Assume that m is sufficiently large and, therefore,
η = 1 P τ ( 1 2 K τ ) T + 2 K τ 1 2 K τ < 1
Then, we can determine that the expectation on SVRG has geometric convergence, as follows:
E [ f ( ω ˜ n ) ] E [ f ( ω ) ] + η E [ f ( ω ˜ n 1 ) f ( ω ) ]
where K > P > 0 a n d τ < 1 K . For a detailed proof, see the work of Johnson and Zhang (2013) [25].
Algorithm 1 SVRG for SCTELM
Input: Training data A R l 1 × n and B R l 2 × n ; Parameters θ , C i , ( i = 1 , 2 ) , ε i , ( i = 1 , 2 ) , update frequency m and learning rate τ .
Output:  ω ˜ n .
Initialize  ω ˜ 0 ;
  • FOR n = 1, 2, 3, ⋯, N;
  •     
  • ω ˜ = ω ˜ n 1 ;
  •     
  • Calculate R l 1 a n d R l 2 according to (34) and (36) for SCTELM and (44) and (46) for Nonlinear SCTELM;
  •     
  • FOR t = 1, 2, 3, ⋯, T;
  •     
  • Randomly pick i t { 1 , 2 , 3 , , l } ;
  •     
  • Update weight ω t = ω t 1 τ ( f i t ( ω t 1 ) f i t ( ω ˜ ) + R l ( ω ˜ ) ) according to (33) and (35) for SCTELM and (43) and (45) for Nonlinear SCTELM;
  •     
  • END
  •     
  • ω ˜ n = ω m ;
 END

3.5. Computational Complexity Analysis

In this section, we briefly analyze the complexity of the proposed algorithm. Here, we only compute the linear SCTELM, and similarly, the nonlinear SCTELM, which is well known to be determined by the computational cost and the number of iterations. First, we analyze the former; the computational complexity of D 1 and D 2 in (33) and (35) is O ( l × L 2 + L 3 ) , while the computational complexity of E 1 and E 2 is O ( l × L 2 + L 3 ) , so the overall complexity of (33) and (35) is O ( l × L 2 + L 3 ) . Similarly, the complexity of (34) and (36) is O ( L 3 ) . In Algorithm 1, the outer loop is executed N times and the inner loop is executed T times, so the total computational cost of the algorithm is O ( N T ( l × L 2 + L 3 ) ) . While L is the number of neurons, there is usually L l , so the complexity of Algorithm 1 is O ( N T l × L 2 ) . It can be seen that the complexity of the SCTELM algorithm is mainly affected by the number of samples, and the convergence speed of the SVRG algorithm is faster than that of other algorithms, so the convergence speed of the SCTELM algorithm is faster than that of other algorithms.
SVRG reduces the variance of the stochastic gradient by introducing an estimate of the global gradient, thus improving the convergence speed. This means that in each iteration, SVRG can use fewer samples (mini-batch) to achieve the same accuracy as traditional SGD. This variance reduction property makes the algorithm more efficient when dealing with large-scale data, thus improving the scalability. Each iteration of SVRG needs to calculate the global gradient, which increases a certain computational overhead compared with only calculating the gradient of local samples. However, due to its faster convergence speed, the same performance can usually be achieved in a smaller number of iterations, thus reducing the overall computation time on large-scale datasets.

4. Numerical Experiments

In this section, to evaluate the robustness of SCTELM, we systematically compare it with other advanced algorithms in multiple datasets, including ELM [2], LELM [26], CTSVM [18],CHELM [27], TELM [24], and FRTELM [28]. For ease of observation, the bold font in the table below indicates the optimum. After dataset testing, we also performed classification visualization to further embody the performance of our proposed model. All experiments were performed using the Windows operating system with a 12th generation Intel (R) core (TM), i7-6700HQ @ 2.10 GHz, and 8.00 GB RAM, and the environment was implemented on a PC running a 64-bit XP operating system. All codes were run by MATLAB R2021b.

4.1. Experimental Setup

We can evaluate the performance of an algorithm based on different metrics, but here, we will use accuracy (ACC), which measures the proportion of instances that the model correctly predicted. ACC is defined as follows:
A C C = T P + T N T P + F P + T N + F N
where TP and TN represent true positive and true negative, respectively, which represent the correct prediction of positive and negative samples; FP and FN represent false positive and false negative, respectively, which represent the incorrect prediction of positive and negative samples. Among them, a higher ACC value represents a more accurate classification and a better model performance.
In our experiment design, the training and testing samples were randomly selected from the dataset, and we added outliers of 0%, 10%, 20%, and 30% to the training set, respectively, and used these contaminated training samples to verify the robustness of the model. In this experiment, due to the large number of parameters, we used the random search method to find the best parameters. The parameters selected for the proposed model are as follows: parameter C i is selected from the set { 10 i | i = 5 , 4 , 3 , 2 , 1 , 0 , 1 , 2 , 3 , 4 , 5 } , and parameter ε i is selected from the set { 10 i | i = 5 , 4 , 3 , 2 , 1 , 0 , 1 , 2 , 3 , 4 , 5 } .

4.2. Description of the Datasets

To verify the performance of our proposed SCTELM, we will numerically simulate it using various datasets, including eight benchmark datasets from UCI and two artificial datasets.
Among them, the UCI dataset includes the following: Australian (Australian Credit Approval), Ionosphere, German, Pima, Vote (Congressional Voting Records), WDBC (Breast Cancer Wisconsin), Spect (SPECTF Heart) and QSAR (QSAR biodegradation). We will use these UCI datasets to compare the performance of our proposed algorithm with other algorithms. See Table 1 for relevant information on these datasets.
In terms of artificial datasets, we generated a two moons dataset containing two categories, 1000 positive and 1000 negative, and a dataset containing three categories. See Figure 3 for the detailed dataset, where the positive and negative classes are represented by green and orange spheres. And the three-classification dataset contains 900 samples, and each class has 300 samples. See Figure 4 for details.
We performed a 10-fold cross-validation on the selected datasets, which means that we randomly split one of the datasets into ten parts, nine of which were used as the training set and the remaining one as the test set. This process was repeated ten times, and the average of the ten results was taken as the final performance metric to reduce the risk of overfitting and underfitting. At the same time, to obtain more objective experimental results, we normalized all the datasets to make the data in the interval [0, 1]. Considering that we want to verify the robust performance of the model, we will successively increase the noise ratio for each dataset. If the accuracy of classification does not change much with the increase in noise, it indicates that the algorithm has good robust performance.

4.3. Experimental Results on Artificial Datasets

In order to verify the advantages of the proposed model over other original models, only ELM and TELM will be compared in this section. Observing Table 2, it can be found that when the noise ratio is zero, the accuracy of SCTELM is higher than that of TELM and ELM, which indicates that SCTELM performs better than TELM and ELM. When the noise ratio is 0.1, the accuracy of the model is all reduced, which indicates that noise does affect the performance of the model. When the noise ratio is 0.2, the accuracy of all the models slightly decreases, but the accuracy of SCTELM is higher than that of TELM and ELM. When the noise ratio is 0.3, the accuracy of SCTELM decreases slightly, while the accuracy of ELM and TELM decreases more, which clearly illustrates the robustness of SCTELM.
To make a more intuitive comparison, the line chart of the proposed model and other models under different noises is drawn in Figure 5. It can be seen from Figure 5 that the corresponding line of SCTELM is always higher than that of the other models, which indicates that the performance of SCTELM is better than that of the other two models. When the noise gradually increases, the accuracy of all the models shows a decreasing trend, which indicates that noise does have a large impact on the performance of the model. It can be seen from the figure that when the noise rises from 0.2 to 0.3, the accuracy decline rate of ELM and TELM is faster than that of SCTELM, which also precisely demonstrates the superior robustness of the proposed model.
To extend SCRELM to the multi-class case, we adopted an artificial dataset with three classes. Generally speaking, either one-to-one or one-to-many strategy is used when using models such as ELM or TELM for multi-classification. Here, we use the more commonly used one-to-one strategy for classification. The strategy decomposes a three-class classification problem into three two-class classification problems, each of which is responsible for solving a pair of classes. The final classification results are determined by voting technology. The final results of classification are given in Table 3. As expected, SCTELM shows its amazing ability and outperforms ELM and TELM in the final classification accuracy.

4.4. Experimental Results on UCI Datasets

In this section, to validate the classification performance algorithms, we ran them on eight datasets. We preliminarily verified the robustness of the SCTELM algorithm on the previous artificial datasets. Here, to further verify its robustness, we will add 20% and 30% Gaussian noise to all datasets. See Table 4, Table 5 and Table 6 for details. However, the choice to set 20% or 30% noise in the SCTELM algorithm experiment can, on the one hand, simulate a data bias that may exist in the real environment to evaluate the performance of the model in the face of noisy data. A clearer understanding of the anti-noise performance of the SCTELM algorithm can be obtained by comparing the experimental results under different noise levels. For example, if the model performs well under 30% noise, it is robust to noise. On the other hand, the increase in noise level will increase the complexity and randomness of the data, making the distribution difference between the training data and the test data larger. Whether the SCTELM model can maintain good generalization performance when the noise is high reflects its ability to adapt to unseen data. All in all, this setting is useful for assessing how well the model generalizes on test sets or real-world application data.

4.4.1. Experimental Results on UCI Datasets Without Outliers

From Table 4, we can clearly find that the classification performance of our proposed SCTELM is better than the other five algorithms on all datasets. In addition, we also find that TELM and CTSVM are faster than the other algorithms on most datasets, this is because they solve one large QPPs instead of two small QPPs. Our proposed algorithm also inherits this advantage. It can be clearly seen in Table 4 that the learning efficiency of SCTELM is ahead of other algorithms in most datasets.

4.4.2. Experimental Results on UCI Datasets with Outliers

We verified the superiority of SCTELM in the previous experiments. To further verify the robustness of SCTELM, we added a certain proportion of outliers in each dataset and conduct new experiments. In this experiment, we chose to add 20% and 30% Gaussian noise, respectively, to test the robustness of the model. Table 5 and Table 6 present the experimental results of the six models with 20% and 30% noise introduced, respectively. From Table 5 and Table 6, we can intuitively find that the accuracy of each model decreases after the introduction of noise, which indicates that noise pollution indeed has a great impact on model classification. At the same time, we can also find that the accuracy of the SCTELM model still maintains its superiority, and it can be concluded that the robustness of our proposed SCTELM model is significantly better than the robustness of other models.
In order to improve our intuition, we drew Figure 6, Figure 7, Figure 8 and Figure 9 according to the noise ratio and accuracy. From the figure, we can clearly see that the accuracy decreases when the noise increases, and we can also intuitively see that the “slope” of SCTELM is small. It further shows that SCTELM has strong robustness.

4.5. Statistical Analysis

In this section, to analyze the significant differences in the six presented algorithms on these eight UCI datasets, we conducted the Friedman test [29], which is considered as a non-parametric test with low data distribution requirements, strong processing power, and easy interpretation of the results. In this test, the null hypothesis H 0 is as follows: there is no significant difference between the six algorithms; and if the null hypothesis is rejected, then the Nemeny test will be performed [29]. The average ranking and average accuracy of these six algorithms on the eight datasets are shown in Table 4, Table 5 and Table 6.
First, we can make a statistical comparison according to the noise-free case, and according to the Friedman statistic formula, it can be obtained as follows:
χ F 2 = 12 N k ( k + 1 ) [ j R j 2 k ( k + 1 ) 2 4 ] = 35.57
where N is the number of datasets, k is the number of algorithms, and R j is the average ranking of the jth algorithm over all datasets. In this paper, k = 7 and N = 8.
Second, we can obtain the χ F 2 -distribution with ( k 1 ) degrees of freedom, as follows:
F F = ( N 1 ) χ F 2 N ( k 1 ) χ F 2 = 20.03
Similarly, we can obtain that the χ F 2 values of noise 0.2 and 0.3 are 32.41 and 40.07, respectively, and the F F values are 9.12 and 20.48, respectively. For α = 0.05 , we can obtain F α = ( 6 , 42 ) = 2.32 . Clearly, all F F has F F > F α . Therefore, it can be said that there are significant differences between the six algorithms.
In the following, to further compare the six algorithms, the Nemenyi post hoc test will be performed. Referring to the table, we obtain q α = 0.05 = 2.85 , which gives us the critical difference (CD):
C D = q α = 0.05 k ( k + 1 ) 6 N = 2.85 × 7 × 8 6 × 8 = 3.078
If the difference between the rankings of the two algorithms calculated is larger than the critical value, it represents a significant difference between the two algorithms. From Table 4, we can obtain the difference between SCTELM and the other five algorithms when there is no noise, and the calculation results are as follows:
D ( E L M S C T E L M ) = 6.75 1.13 = 5.62 > 3.078 D ( T E L M S C T E L M ) = 4 1.13 = 2.87 < 3.078 D ( L E L M S C T E L M ) = 4.5 1.13 = 3.37 > 3.078 D ( C T S V M S C T E L M ) = 4.88 1.13 = 3.75 > 3.078 D ( C H E L M S C T E L M ) = 4.63 1.13 = 3.5 > 3.078 D ( F R T E L M S C T E L M ) = 2.13 1.13 = 1 < 3.078
The notation D (A-B) represents the difference in average rankings between the two algorithms. We can conclude that SCTELM significantly outperforms ELM, LELM, CTSVM and CHELM on the noise-free dataset, while showing no significant difference compared to TELM and FRTELM. According to Table 4 and Table 5, it is clear that in datasets with 20% and 30% Gaussian noise, SCTELM shows a significant difference compared to ELM, TELM and LELM, but there is little difference compared to CTSVM and CHELM. To make the conclusions clear, the post hoc detection is visualized below in Figure 10.
In Figure 10, the abscissa represents the average ranking of each algorithm, and the ordinate represents the algorithms to be compared. If two algorithms overlap with respect to the abscissa, then there is no significant difference between them. From Figure 10, we can easily draw this conclusion.

4.6. Parameter Analysis

To make it more convenient for other researchers to use SCTELM, the performance of SCTELM under different parameters will be further studied from the perspective of parameters, so we chose four representative datasets Australian, German, Spect and Ionosphere). Starting from the parameters C1, C2 and θ , the sensitivity of the proposed algorithm to these parameters is analyzed. At the same time, the other parameters are fixed for experiments. The experimental results are shown in Figure 11 and Figure 12.
For the parameters C1 and C2, we observe that on these datasets, the ACC values show large fluctuations as the parameters change, and the fluctuations are particularly significant on some datasets. For example, on the German dataset, when C1 and C2 are changed, the fluctuation amplitude of ACC values is close to 60%, which indicates that this dataset is very sensitive to parameter changes. In contrast, the fluctuations of datasets like Australian are relatively small, about 30%. Through further analysis, it is found that the ACC value rises significantly as the parameter C1 decreases and reaches its maximum value at the value of C1 equal to 10 4 , indicating that a smaller C1 is more beneficial to improve the classification accuracy. Therefore, parameter C1 is suitable for selecting a smaller value to optimize the algorithm performance. On the other hand, the effect of changing the parameter C2 on the results shows a linear trend. As the value of C2 increases, the ACC value also gradually rises, and the best classification effect is achieved when the value of C2 is 10 4 . This indicates that appropriately increasing C2 is helpful to further improve the classification performance of the algorithm.
Compared with parameters C1 and C2, parameter θ shows relatively stable characteristics in the experiment, especially on the Australian dataset; its maximum fluctuation range is less than 2%, showing extremely strong stability. However, on the Ionosphere dataset, the parameter θ fluctuates by nearly 10%, indicating that there is still some sensitivity of θ on some datasets. Further analysis of the experimental results shows that the influence of the parameter θ on the classification accuracy shows a nonlinear trend. Setting θ in [ 10 3 , 10 3 ] , the algorithm performs best, indicating that the θ value in this range can effectively optimize the classification performance.

4.7. Convergence Analysis Experiments

To verify the convergence of the proposed SCTELM algorithm, experiments will be carried out on the Austrasta, German, Spect and Ionosphere datasets in this section. The experiments are similar to those carried out in Section 4.4.2, while the relevant parameters are set to establish the optimal parameters. The experimental results are shown in Figure 13.
In Figure 13, it is easy to see that the convergence condition of SCTELM steadily decreases with the increase in the number of iterations, and it converges to a stable value after a finite number of iterations, so it can be seen that the proposed SCTELM algorithm can converge within a finite number of iterations.

5. Conclusions

This paper proposes a novel robust twin extreme learning machine that employs a newly designed soft capped L 1 -norm loss function, designated as SCTELM for brevity. In contrast to the conventional hard capped methodology, the proposed soft capped approach effectively circumvents the potential non-differentiability issues associated with hard capped, whilst retaining the beneficial aspects of the hard capped technique. The objective of this approach is to mitigate the impact of outliers on the model, thereby increasing its robustness. Furthermore, the SCTELM only requires the resolution of a pair of relatively small quadratic programming problems (QPPs), which significantly reduces the complexity associated with the resolution of large QPPs in the past. Consequently, this enhances the learning efficiency of the model. The proposed model demonstrates enhanced stability and increased computational efficiency in the presence of data anomalies. The experimental results on noise-free and noisy datasets indicate that the SCTELM model exhibits heightened robustness compared to traditional methods. This is particularly evident in datasets containing outliers, where the “slope” of the SCTELM model exhibits a more gradual incline, and the learning efficiency is superior to other methods.
Our research is of considerable importance within the academic field. This paper introduces a novel loss function based on soft capped, which enables the construction of a robust twin extreme learning machine. This provides a new perspective and methodology for the field of machine learning and has the potential to encourage subsequent researchers to pursue innovative avenues in the design of loss functions and learning algorithms, thereby promoting the further advancement of both theoretical and practical developments. Furthermore, in the context of real-world data, such as the credit card approval dataset from the University of California, Irvine (UCI), users can be classified into two distinct categories: those who are approved for credit and those who are not. It is evident that SCTELM has gained significant advantages in terms of robustness to outliers in comparison to alternative algorithms. To illustrate, WDBC represents the Diagnostic Wisconsin Breast Cancer Database, and the SCTELM approach has also yielded significant benefits. It follows, therefore, that our research has significant practical implications for the fields of finance and medical treatment. The model is capable of facilitating the provision of real-time analysis and decision support to end users. The enhancement of accuracy and stability in the model enables enterprises and healthcare facilities to effectively address decision-making challenges in intricate contexts. The model may be utilized in a multitude of domains, including, although not limited to, image processing, signal processing, and natural language processing. These fields often have to contend with data noise and uncertainty, which robust models can help to overcome, thereby improving the accuracy and reliability of results. In conclusion, our study offers a significant contribution to the field of machine learning, providing a valuable perspective on robustness research and its potential for practical applications, which will inform future research and developments.
It should be noted that the algorithm is not without its limitations. At the commencement of each cycle, the SVRG algorithm is required to calculate the full gradient of the entire dataset. Although this will facilitate convergence more rapidly than the conventional SGD algorithm, the associated computational cost may be considerable in the context of large-scale datasets, potentially leading to a reduction in the rate of algorithmic execution. In future work, it would be worthwhile to consider employing periodic random sampling in place of calculating the full gradient each time, or to investigate combining the SVRG algorithm with an adaptive learning rate optimization algorithm. This would allow the learning rate to be adaptively changed in accordance with the gradient.

Author Contributions

Conceptualization, Z.X. and J.M.; Methodology, Z.X. and J.M.; Software, Z.X.; Verification, Z.X. and B.W.; Resource acquisition. G.Y. and B.W.; Data management, G.Y.; Writing—first draft preparation, Z.X.; Writing—reviewing and editing, B.W., J.M. and G.Y.; Visualization, Z.X., B.W. and G.Y.; Supervising, G.Y. and B.W.; Project Management, G.Y., B.W. and J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural Science Foundation of Ningxia Province (No. 2024AAC05055, No. 2023AAC02053), in part by the Fundamental Research Funds for the Central Universities of North Minzu University (No. 2023ZRLG01, No. 2021JCYJ07), in part by the National Natural Science Foundation of China (No. 62366001, No. 12361062), in part by the Key Research and Development Program of Ningxia (Introduction of Talents Project) (No. 2022BSB03046), and in part by the Postgraduate Innovation Project of North Minzu University (YCX24280).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Alzaqebah, A.; Aljarah, I.; Al-Kadi, O. A hierarchical intrusion detection system based on extreme learning machine and nature-inspired optimization. Comput. Secur. 2023, 124, 102957. [Google Scholar] [CrossRef]
  2. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks, Budapest, Hungary, 25–29 July 2004; pp. 985–990. [Google Scholar]
  3. Zheng, G.; Hua, W.; Qiu, Z.; Gong, Z. Detecting Water Depth from Remotely Sensed Imagery Based on ELM and GA-ELM. J. Indian Soc. Remote Sens. 2021, 49, 947–957. [Google Scholar] [CrossRef]
  4. Yang, H.; Zhao, Y.; Zhao, Y.; Chen, N. Drivers’ visual interaction performance of on-board computer under different heat conditions: Based on ELM and entropy weight. Sustain. Cities Soc. 2022, 81, 103835. [Google Scholar] [CrossRef]
  5. Wang, H.; Li, C.; Guan, T.; Zhao, S. No-reference stereoscopic image quality assessment using quaternion wavelet transform and heterogeneous ensemble learning. Displays 2021, 69, 102058. [Google Scholar] [CrossRef]
  6. Zhan, W.; Wang, K.; Cao, J. Elastic-net based robust extreme learning machine for one-class classification. Signal Process. 2023, 211, 109101. [Google Scholar] [CrossRef]
  7. Wu, Q.; Wang, F.; An, Y.; Li, K. L1-Norm Robust Regularized Extreme Learning Machine with Asymmetric C-Loss for Regression. Axioms 2023, 12, 204. [Google Scholar] [CrossRef]
  8. Lu, X.; Ming, L.; Liu, W.; Li, H.X. Probabilistic regularized extreme learning machine for robust modeling of noise data. IEEE Trans. Cybern. 2017, 48, 2368–2377. [Google Scholar] [CrossRef]
  9. Zhang, Y.; Wu, Q.; Hu, J. An adaptive learning algorithm for regularized extreme learning machine. IEEE Access 2021, 9, 20736–20745. [Google Scholar] [CrossRef]
  10. Chen, K.; Lv, Q.; Lu, Y.; Dou, Y. Robust regularized extreme learning machine for regression using iteratively reweighted least squares. Neurocomputing 2017, 230, 345–358. [Google Scholar] [CrossRef]
  11. Ren, Z.; Yang, L. Robust extreme learning machines with different loss functions. Neural Process. Lett. 2019, 49, 1543–1565. [Google Scholar] [CrossRef]
  12. Luo, L.; Wang, K.; Lin, Q. Robust Extreme Learning Machine Based on p-order Laplace Kernel-Induced Loss Function. Int. J. Adv. Comput. Sci. Appl. 2024, 14, 1281–1291. [Google Scholar] [CrossRef]
  13. Meng, D.; Zhao, Q.; Xu, Z. Improve robustness of sparse PCA by L1-norm maximization. Pattern Recognit. 2012, 45, 487–497. [Google Scholar] [CrossRef]
  14. Zhang, K.; Luo, M. Outlier-robust extreme learning machine for regression problems. Neurocomputing 2015, 151, 1519–1527. [Google Scholar] [CrossRef]
  15. Jiang, W.; Nie, F.; Huang, H. Robust dictionary learning with capped L1-norm. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015; pp. 3590–3596. [Google Scholar]
  16. Wang, C.; Ye, Q.; Luo, P.; Ye, N.; Fu, L. Robust capped L1-norm twin support vector machine. Neural Netw. 2019, 114, 47–59. [Google Scholar] [CrossRef]
  17. Chen, M.; Wang, Q.; Chen, S.; Li, X. Capped L1-norm sparse representation method for graph clustering. IEEE Access 2019, 7, 54464–54471. [Google Scholar] [CrossRef]
  18. Li, Y.; Sun, H.; Yan, W.; Cui, Q. R-CTSVM+: Robust capped L1-norm twin support vector machine with privileged information. Inf. Sci. 2021, 574, 12–32. [Google Scholar] [CrossRef]
  19. Nie, F.; Wang, X.; Huang, H. Multiclass capped Lp-Norm SVM for robust classifications. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 2415–2421. [Google Scholar]
  20. Jayadeva; Khemchandani, R.; Chandra, S. Twin Support Vector Machines for Pattern Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 905–910. [Google Scholar] [CrossRef] [PubMed]
  21. Ding, S.; Hua, X.; Yu, J. An overview on nonparallel hyperplane support vector machine algorithms. Neural Comput. Appl. 2014, 25, 975–982. [Google Scholar] [CrossRef]
  22. Kumar, M.A.; Gopal, M. Least squares twin support vector machines for pattern classification. Expert Syst. Appl. 2009, 36, 7535–7543. [Google Scholar] [CrossRef]
  23. Shao, Y.H.; Zhang, C.H.; Wang, X.B.; Deng, N.Y. Improvements on twin support vector machines. IEEE Trans. Neural Netw. 2011, 22, 962–968. [Google Scholar] [CrossRef]
  24. Wan, Y.; Song, S.; Huang, G.; Li, S. Twin extreme learning machines for pattern classification. Neurocomputing 2017, 260, 235–244. [Google Scholar] [CrossRef]
  25. Johnson, R.; Zhang, T. Accelerating stochastic gradient descent using predictive variance reduction. In Proceedings of the 26th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; Volume 1, pp. 315–323. [Google Scholar]
  26. Ma, J.; Wen, Y.; Yang, L. Lagrangian supervised and semi-supervised extreme learning machine. Appl. Intell. 2019, 49, 303–318. [Google Scholar] [CrossRef]
  27. Ren, Z.; Yang, L. Correntropy-based robust extreme learning machine for classification. Neurocomputing 2018, 313, 74–84. [Google Scholar] [CrossRef]
  28. Ma, J.; Yang, L.; Sun, Q. Capped L1-norm distance metric-based fast robust twin bounded support vector machine. Neurocomputing 2020, 412, 295–311. [Google Scholar] [CrossRef]
  29. Demšar, J.; Schuurmans, D. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
Figure 1. ELM algorithm diagram.
Figure 1. ELM algorithm diagram.
Electronics 13 04533 g001
Figure 2. Capped L θ -loss ( θ = 0.5 , 3 , 10 ) versus hard capped L 1 -norm loss ( ϵ = 1 ) .
Figure 2. Capped L θ -loss ( θ = 0.5 , 3 , 10 ) versus hard capped L 1 -norm loss ( ϵ = 1 ) .
Electronics 13 04533 g002
Figure 3. Two moons dataset distribution.
Figure 3. Two moons dataset distribution.
Electronics 13 04533 g003
Figure 4. Three-class dataset distribution.
Figure 4. Three-class dataset distribution.
Electronics 13 04533 g004
Figure 5. Comparison of algorithms on artificial datasets.
Figure 5. Comparison of algorithms on artificial datasets.
Electronics 13 04533 g005
Figure 6. The accuracy of the six algorithms under different noises: (a) Australian datasets; (b) WDBC datasets.
Figure 6. The accuracy of the six algorithms under different noises: (a) Australian datasets; (b) WDBC datasets.
Electronics 13 04533 g006
Figure 7. The accuracy of the six algorithms under different noises: (a) QSAR datasets; (b) German datasets.
Figure 7. The accuracy of the six algorithms under different noises: (a) QSAR datasets; (b) German datasets.
Electronics 13 04533 g007
Figure 8. The accuracy of the six algorithms under different noises: (a) Vote datasets; (b) Spect datasets.
Figure 8. The accuracy of the six algorithms under different noises: (a) Vote datasets; (b) Spect datasets.
Electronics 13 04533 g008
Figure 9. The accuracy of the six algorithms under different noises: (a) Ionosphere datasets; (b) Pima datasets.
Figure 9. The accuracy of the six algorithms under different noises: (a) Ionosphere datasets; (b) Pima datasets.
Electronics 13 04533 g009
Figure 10. Visualization of critical difference: (a) without noise; (b) with 20% Gaussian noise; and (c) with 30% Gaussian noise.
Figure 10. Visualization of critical difference: (a) without noise; (b) with 20% Gaussian noise; and (c) with 30% Gaussian noise.
Electronics 13 04533 g010
Figure 11. Histogram of sensitivity of parameters C1, C2 on four datasets: (a) Australian; (b) German; (c) Spect; (d) Ionosphere.
Figure 11. Histogram of sensitivity of parameters C1, C2 on four datasets: (a) Australian; (b) German; (c) Spect; (d) Ionosphere.
Electronics 13 04533 g011
Figure 12. Histogram of sensitivity of parameters θ on four datasets: (a) Australian; (b) German; (c) Spect; (d) Ionosphere.
Figure 12. Histogram of sensitivity of parameters θ on four datasets: (a) Australian; (b) German; (c) Spect; (d) Ionosphere.
Electronics 13 04533 g012
Figure 13. Convergence analysis of SCTELM on four UCI datasets: (a) Australian; (b) German; (c) Spect; (d) Ionosphere.
Figure 13. Convergence analysis of SCTELM on four UCI datasets: (a) Australian; (b) German; (c) Spect; (d) Ionosphere.
Electronics 13 04533 g013aElectronics 13 04533 g013b
Table 1. Characteristics of the UCI datasets.
Table 1. Characteristics of the UCI datasets.
DatasetsSamplesAttributesPositive SamplesNegative Samples
Australian69014307383
VOTE43216167265
WDBC56930357212
Spect2674421255
QSAR105541699356
Ionosphere35134225126
German100024700300
Pima7688500268
Table 2. Performance on artificial datasets.
Table 2. Performance on artificial datasets.
ELMTELMSCTELM
Noise RatioACC (%)ACC (%)ACC (%)
092.5393.8295.91
0.190.1690.5094.64
0.285.3186.7391.36
0.376.6477.5583.45
Table 3. Experimental results of multi-classification artificial datasets.
Table 3. Experimental results of multi-classification artificial datasets.
ELMTELMSCTELM
 ACC (%)ACC (%)ACC (%)
Class 163.7477.8385.17
Class 279.3171.1784.33
Class 372.8373.0083.17
Final result71.8574.0084.22
Table 4. Experimental results on UCI datasets without noise.
Table 4. Experimental results on UCI datasets without noise.
ELMTELMLELMCTSVMCHELMFRTELMSCTELM
DatasetsACC (%)ACC (%)ACC (%)ACC (%)ACC (%)ACC (%)ACC (%)
Times (s)Times (s)Times (s)Times (s)Times (s)Times (s)Times (s)
Australian84.3785.7184.5785.6484.9385.7386.32
1.4311.9830.7930.3243.3681.1730.601
WDBC95.3197.3296.5497.3196.8697.6798.21
1.35460.9240.7340.8472.7360.5140.525
QSAR83.9485.6385.3185.1986.1786.3786.51
1.3740.5611.3511.1693.4830.4371.659
German76.5977.8078.6276.5776.8877.6378.55
1.7091.0932.0181.6875.0210.7291.486
Vote95.1995.5296.3595.5196.4797.0397.62
1.2440.8210.4650.6564.0760.5190.339
Spect81.3782.7182.6882.2581.9682.9483.62
0.7700.8310.7691.1321.9820.4920.250
Ionosphere86.3388.0090.6788.6087.6591.0891.37
0.5890.3610.3630.3992.5610.2290.264
Pima76.1276.3175.9776.9276.7877.1377.83
1.8541.0081.1051.3173.5611.3470.539
Avg. ACC84.9086.1386.3486.0085.9687.0787.50
Avg. rank6.754.004.454.884.632.131.13
Table 5. Experimental results on UCI datasets with 20% Gaussian noise.
Table 5. Experimental results on UCI datasets with 20% Gaussian noise.
ELMTELMLELMCTSVMCHELMFRTELMSCTELM
DatasetsACC (%)ACC (%)ACC (%)ACC (%)ACC (%)ACC (%)ACC (%)
Times (s)Times (s)Times (s)Times (s)Times (s)Times (s)Times (s)
Australian75.7476.3475.9777.4978.5379.8181.88
1.3710.9830.8930.7816.3681.4690.330
WDBC85.4983.7284.1486.7286.9589.7492.86
1.4781.1241.7151.1958.9691.6490.791
QSAR64.7366.8165.2965.7365.2969.5273.29
2.3144.5392.0592.99410.2372.7841.728
German68.9367.3868.1968.7169.7171.3772.30
2.5392.3481.8181.4344.3201.6470.965
Vote91.7293.5791.2991.0394.2794.5495.86
0.9090.4950.9650.9383.8671.6520.439
Spect70.0972.4970.1880.6380.3480.6381.77
0.7591.0950.8191.1515.1841.5250.375
Ionosphere78.3781.5479.7479.3579.5283.3983.82
0.8892.5390.7510.8044.2641.2570.350
Pima65.7165.4066.7471.7472.7872.3773.34
1.8040.9430.9051.0322.5531.5610.711
Avg. ACC75.1075.9175.1977.6878.4280.2581.89
Avg. rank6.005.135.504.503.632.001.00
Table 6. Experimental results on UCI datasets with 30% Gaussian noise.
Table 6. Experimental results on UCI datasets with 30% Gaussian noise.
ELMTELMLELMCTSVMCHELMFRTELMSCTELM
DatasetsACC (%)ACC (%)ACC (%)ACC (%)ACC (%)ACC (%)ACC (%)
Times (s)Times (s)Times (s)Times (s)Times (s)Times (s)Times (s)
Australian64.8163.6263.9467.3268.0174.8475.74
1.7461.4931.5131.6416.4681.4260.494
WDBC62.5363.0363.5374.8374.5779.4180.36
1.711.5481.5851.8358.3611.4571.451
QSAR62.7963.7262.5366.3666.7669.3670.04
2.8564.7923.2832.18610.5933.4291.994
German67.2967.9869.5262.9368.7169.3769.80
3.0942.4292.0433.1844.3021.8271.029
Vote91.6190.7892.3391.9793.6793.9194.10
0.7490.7650.6491.0983.6251.3720.372
Spect62.9265.9764.4278.3478.5778.8779.54
0.9630.9610.9131.2414.2151.5000.321
Ionosphere72.7267.0472.7975.0574.0175.1975.59
0.8590.9790.8110.9723.6831.0170.350
Pima63.9164.2262.9467.8868.5870.6372.11
1.6731.2630.7340.3712.1021.5270.729
Avg. ACC68.5768.3069.0073.0974.1176.4777.16
Avg. rank6.135.885.384.253.382.001.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, Z.; Wei, B.; Yu, G.; Ma, J. Robust Twin Extreme Learning Machine Based on Soft Truncated Capped L1-Norm Loss Function. Electronics 2024, 13, 4533. https://doi.org/10.3390/electronics13224533

AMA Style

Xu Z, Wei B, Yu G, Ma J. Robust Twin Extreme Learning Machine Based on Soft Truncated Capped L1-Norm Loss Function. Electronics. 2024; 13(22):4533. https://doi.org/10.3390/electronics13224533

Chicago/Turabian Style

Xu, Zhendong, Bo Wei, Guolin Yu, and Jun Ma. 2024. "Robust Twin Extreme Learning Machine Based on Soft Truncated Capped L1-Norm Loss Function" Electronics 13, no. 22: 4533. https://doi.org/10.3390/electronics13224533

APA Style

Xu, Z., Wei, B., Yu, G., & Ma, J. (2024). Robust Twin Extreme Learning Machine Based on Soft Truncated Capped L1-Norm Loss Function. Electronics, 13(22), 4533. https://doi.org/10.3390/electronics13224533

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop