Next Article in Journal
Freshness-Keeping Strategy of Logistics Service Providers: The Role of the Interaction between Blockchain and Overconfidence
Previous Article in Journal
FSSSA: A Fuzzy Squirrel Search Algorithm Based on Wide-Area Search for Numerical and Engineering Optimization Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Objective Models for Sparse Optimization in Linear Support Vector Machine Classification

1
Department of Computer Engineering, Modelling, Electronics and Systems Engineering, University of Calabria, 87036 Rende, Italy
2
Department of Civil Engineering, University of Calabria, 87036 Rende, Italy
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(17), 3721; https://doi.org/10.3390/math11173721
Submission received: 26 June 2023 / Revised: 18 August 2023 / Accepted: 27 August 2023 / Published: 29 August 2023
(This article belongs to the Section Engineering Mathematics)

Abstract

:
The design of linear Support Vector Machine (SVM) classification techniques is generally a Multi-objective Optimization Problem (MOP). These classification techniques require finding appropriate trade-offs between two objectives, such as the amount of misclassified training data (classification error) and the number of non-zero elements of the separator hyperplane. In this article, we review several linear SVM classification models in the form of multi-objective optimization. We put particular emphasis on applying sparse optimization (in terms of minimization of the number of non-zero elements of the separator hyperplane) to Feature Selection (FS) for multi-objective optimization linear SVM. Our primary purpose is to demonstrate the advantages of considering linear SVM classification techniques as MOPs. In multi-objective cases, we can obtain a set of Pareto optimal solutions instead of one optimal solution in single-objective cases. The results of these linear SVMs are reported on some classification datasets. The test problems are specifically designed to challenge the number of non-zero components of the normal vector of the separator hyperplane. We used these datasets for multi-objective and single-objective models.

1. Introduction

In most machine learning problems, several objectives are aggregated as one objective function. Therefore, the design of machine learning systems can generally be considered a Multi-objective Optimization Problem (MOP) [1]. In the multi-objective optimization form of classification problem, appropriate trade-offs must be found between several objective functions, for example, between model complexity and accuracy, sensitivity and specificity, the sum of distances of misclassified points to the separating hyperplanes and the distance between the two bounding planes that generate the separating plane or the number of misclassified training data and the number of non-zero elements of separating hyperplane [1,2]. In various research, it has been shown that multi-objective machine learning algorithms are more powerful in improving generalization and knowledge extraction ability compared to single-objective learning, especially in topics such as Feature Selection, sparsity, and clustering [1,3].
Optimization algorithms, when there are a large number of variables or constraints, could account for most of the computation time. So far, various sparse matrices that arise in optimization have been investigated [4]. In many fields of linear systems, such as engineering problems, science, and signal and image processing, a search for sparse solutions is required. Mathematical optimization plays an essential role in the development of numerical algorithms for searching the sparsity in solutions [5].
Support vector machines (SVMs) use a hyperplane to separate samples into one of two classes. It is mentioned in [6] that it is convenient to combine the SVM problem with a set theory for set-based particle swarm optimization (SBPSO) to be used to find the optimal separator hyperplane. This method is called SBPSO-SVM [6].
In many MOPs, conflicting objective functions must be optimized [7,8]. These problems are used when the optimal decision to adopt two or more objectives is interdependent, for example, in economics, logistics, and many engineering and scientific problems [6]. In this case, the optimization problem has no single solution representing the optimal solution for all objectives simultaneously [9,10,11]. In MOPs, a solution with the most appropriate trade-off between objectives is found in which no objective is improved without worsening at least one other objective [12,13]. This solution is known as the Pareto optimal solution [14]. The set of all Pareto optimal solutions is known as the Pareto set or Pareto frontier [15,16].
Although single-objective machine learning problems have been well studied [17,18,19,20,21,22,23], there are fewer studies on multi-objective machine learning problems. Multi-objective machine learning is an approach to determining an appropriate trade-off between generally conflicting objectives [24,25,26]. In multi-objective machine learning approaches, the main advantage is that you can obtain a deeper insight into the learning problem by analyzing the produced Pareto frontier [27,28,29,30,31,32,33,34,35,36,37,38,39,40,41]. In some multi-objective approaches, two objectives are simultaneously considered: minimizing the classification error and the norm of the weight vectors [42].
This article presents multi-objective classification problems to obtain Pareto-optimal solutions (Pareto frontier). In these multi-objective optimization problems, one objective is used to minimize the classification error, and another objective is used to minimize the number of non-zero elements of the separator hyperplane.
The rest of the article is organized as follows. In Section 2, some basic concepts and notations, including binary classification, support vector machine classification methods, sparse optimization, and multi-objective optimization problems, are given. In Section 3, multi-objective reformulation of support vector machine models is presented. The results of several numerical experiments are presented in Section 4. Conclusions are devoted to Section 5.

2. Basic Concepts and Notations

To make a more accessible understanding of this article, some basic concepts and notations are presented in this section. First, we briefly describe binary classification, and then we will focus on Support Vector Machines classification methods and some models for sparse optimization. Some concepts of multi-objective optimization problems will also be discussed.

2.1. Binary Classification

Data mining algorithms predict to which category of the target variables each case belongs. This activity is called binary classification [43]. The goal of binary classification is to assign a new object to one of two classes from certain sets of classes based on the feature values of this object [44,45].
Suppose that we have two classes of individuals in the form of two finite sets A , B R n , such that A B = . In binary classification, we want to classify an input vector x R n as a member of the class denoted by A or that by B . For binary classification, the training set is defined as follows [46,47]:
T = x i , y i x i R n ,   y i ± 1   a n d   i = 1 ,   ,   m
with the two classes A and B labelled by +1 and −1, respectively. The function f : R n ± 1 , is in the following form that determines the class membership of a given vector x [46,47]:
f x = + 1 ,   i f   x A 1 ,   i f   x B
Assume that there are two finite point sets A and B in R n that consist of m and k points, respectively. They are associated with the matrices A R m × n and B R k × n , where each point set is represented as a row of the corresponding matrix. In the SVM method, we want to construct the separating hyperplane P as follows [46,47]:
P = x   x R n ,   x T w = γ
with normal vector w R n [46,47].
The separating plane P determines two open halfspaces as follows:
  • P 1 = x   x R n ,   x T w > γ ,
  • P 2 = x   x R n ,   x T w < γ .
P 1 is intended to have most of the points belonging to A and P 2 is intended to have most of the points belonging to B .
Therefore, we want to satisfy the following inequalities to the possible extent, where e is a vector of ones by the appropriate dimension:
A w > e γ ,   B w < e γ
The problem can be equivalently stated as follows [46,47]:
A w > e γ + e ,   B w < e γ e
As we will see next, using Feature Selection in SVM means suppressing as many of the components of vector w as possible [46,47].

2.2. Support Vector Machine Classification Methods

In the Support Vector Machine (SVM) classification methods, in addition to minimizing the error function, we also want to maximize the distance between the two bounding planes (referred to as the separation margin) that generate the separating hyperplane [48,49]. The standard formulation of SVM is the following, where variables y i and z l represent the classification error associated with the points of A and B , respectively:
M i n                   C i = 1 m 1 y i + l = 1 m 2 z l + w 2 2 s . t             a i T w + γ + 1 y i ,               i = 1 ,   ,   m 1 b l T w γ + 1 z l ,                   l = 1 ,   ,   m 2 y i 0 ,   z l 0
Positive parameter C defines the trade-off between the two objectives: minimizing the classification error and maximizing the separation margin [50,51].
Since in feature selection, the goal is suppressing as many elements of w as possible, replaced l 2 -norm with l 1 -norm and a feature selection term introduced as the following form [52,53]:
M i n             C i = 1 m 1 y i + l = 1 m 2 z l + w 1 s . t       a i T w + γ + 1 y i ,         i = 1 ,   ,   m 1 b l T w γ + 1 z l ,           l = 1 ,   ,   m 2 y i 0 ,   z l 0

2.3. Sparse Optimization

In sparse SVM, in addition to maintaining satisfactory classification accuracy, the goal is to control the number of non-zero components of the normal vector to the separating hyperplane [54]. Therefore, the following two objectives should be minimized [46]:
  • Classification error (the number of misclassified training data).
  • The number of non-zero elements of the normal vector of the separator hyperplane (vector w).
Feature selection in SVM as a special case of sparse optimization states the following problem [46,54,55]:
M i n                   C i = 1 m 1 y i + l = 1 m 2 z l + w 0 s . t     a i T w + γ + 1 y i ,         i = 1 ,   ,   m 1 b l T w γ + 1 z l ,       l = 1 ,   ,   m 2 y i 0 ,   z l 0
where . 0 is the l 0 -pseudo-norm, which counts the number of non-zero components of any vector. The l 0 -pseudo-norm is a nonconvex discontinuous function, so problems with this norm lead to cardinality-constrained problems that are hard to solve (NP-hard problems) [43,44]. In many applications, the l 0 -pseudo-norm is replaced by the l 1 -norm (and l 2 -norm), model (7) (and model (6)), which is more tractable [46,47].
The use of k -norms has attracted much attention in recent years, which has led to several ways to deal with the cardinality-constrained problem with l 0 -pseudo-norm [48,55,56].
In the next, at first, we define the k -norm, and then we introduce two models to ensure sparsity using the k -norm.
Definition 1
 ( k -norm) [54,57]. The sum of k largest component of the vector X is called the k -norm of vector X:
x k = x i 1 + x i 2 + + x i k   w h e r e ,   x i 1 x i 2 x i n
The k-norm is intermediate between . 1 and . and it is a polyhedral norm. This norm enjoys the following fundamental property linking x k to x 0   1 k n :
x 0 k     x 1 x k = 0
The following problem based on k -norm proposed to sparse optimization for Feature Selection in the SVM model in [54]:
M i n   C i = 1 m 1 y i + l = 1 m 2 z l + e T u + v + σ e T w + + w u v T w + w s . t a i T w + w + γ + 1 y i ,       i = 1 ,   ,   m 1 b l T w + w γ + 1 z l ,         l = 1 ,   ,   m 2 y i 0 ,   z l 0 ,   w + ,   w 0 ,   0 u ,   v e
where w = w + w ,   w + ,   w 0 . As mentioned in [54], ( u v ) is the subdifferential of w k at point 0 and u + v T e = k . This model is called SVM0.
Additionally, based on k -norm the following problem proposed in [58,59,60] for Sparse Optimization:
M i n     C i = 1 m 1 y i + l = 1 m 2 z l 1 w 1 k = 1 n w k s . t a i T w + γ + 1 y i ,                 i = 1 ,   ,   m 1 b l T w γ + 1 z l ,             l = 1 ,   ,   m 2 y i 0 ,   z l 0
This model is called BM-SVM.

2.4. Multi-Objective Optimization Problem

A Multi-objective optimization problem (MOP) is given as follows [61]:
M i n i m i z e             f x = f 1 x ,   ,   f p x   s . t :                                             x   X
where X   R n is the set of constraint, and f k   : R n R , k = 1 ,   ,   p , are the continuous objective functions. If at least two objective functions are conflicting in (13) then no single x X would generally minimize every f k at the same time. Therefore, it is necessary to some new notions introduce for optimality in MOP [61].
Definition 2 (Dominance Vector).
The vector f x 1 dominates vector f x 2 , and we say x 1   dominates x 2 , if and only if f k x 1   f k x 2 for all k = 1 ,   ,   p and for at least one i     1 ,   ,   p this inequality be established in the strict form f i x 1 < f i x 2  [61].
Definition 3 (Pareto Optimality and Pareto frontier).
Supposed that x ^ X be a feasible solution of MOP (13). This feasible solution is called Pareto optimal if there is no other x X such that x   dominates x ^ . The set of all Pareto optimal solutions is called the Pareto set or Pareto frontier [61].
In the ε-constraint method, one of the objective functions is optimized, while the rest of the objective functions are considered in the form of constraints. Several versions of the ε-constraint method have been proposed to try to improve its performance [62].
Several methods have been proposed to construct the Pareto frontier of MOPs, but in this article, we will use the modified algorithm introduced in [63,64] based on the ε-constraint method.
In [63,64], a modified algorithm based on the ε-constraint method is proposed, which systematically generates Pareto optimal solutions.
In this algorithm at the first phase, the following single-objective optimization problems are solved for k = 1 ,   ,   p [60]:
M i n i m i z e     f k x   s . t :                   x   X
Let x 1 * ,   ,   x p * be the optimal solutions to these problems, respectively. Then, the restricted region is defined as follows for k = 1 ,   ,   p [63]:
  x X :   f k x k * f k x max i = 1 ,   , p ; i k   f k x i *
In the second phase, the steps’ lengths Δ x j are determined in the region (15), for j = 1 ,   ,   p , and then the following single-objective optimization problems are solved [63]:
M i n i m i z e         f k x s . t :                         f j x Δ x j ,                         j = 1 ,   ,   n ,   j k , x   X
It is proved in [63,64] that If x * is an optimal solution of (16), then it will be a Pareto optimal solution of multi-objective optimization.
In the next section, we will present some of the single-objective SVM models in the form of multi-objective optimization problems.

3. Multi-Objective Support Vector Machine

It has been shown in [1] that multi-objective machine learning methods are more powerful compared to single-objective forms in dealing with different machine learning topics. Additionally, a major advantage of the multi-objective machine learning approach is that by analyzing the Pareto frontier, one can gain a deeper insight into the learning problem [1]. Support vector machines have been investigated in [2] in the form of multi-objective optimization problems, and an approach to design SVM on a real-world pattern recognition task has been made [2].
Here, we reformulate several linear SVM models as multi-objective models. Our primary purpose is to demonstrate the advantages of considering these single-objective models as MOP models. In multi-objective form, we can obtain a set of Pareto-optimal solutions instead of an optimal solution in a single-objective form [58,59,60,61], and then the decision maker can choose one of these solutions [58,59,60].
In this section, we will reformulate l 1 (model (7)), l 2 (model (6)), SVM0 (model (11)), and BM-SVM (model (12)) models into MOPs.
The MOP reformulations of the l 1 -norm and l 2 -norm (models (6) and (7) in Section 2.2) are as follows, respectively:
M i n                     f 1 = i = 1 m 1 y i + l = 1 m 2 z l M i n                     f 2 = w 1 s . t                 a i T w + γ + 1 y i ,       i = 1 ,   ,   m 1 b l T w γ + 1 z l ,               l = 1 ,   ,   m 2 y i 0 ,   z l 0
M i n             f 1 = i = 1 m 1 y i + l = 1 m 2 z l M i n                 f 2 = w 2 2 s . t       a i T w + γ + 1 y i ,           i = 1 ,   ,   m 1 b l T w γ + 1 z l ,           l = 1 ,   ,   m 2 y i 0 ,   z l 0
The BM-SVM model (model (12) in Section 2.3) is reformulated as the following MOP:
M i n         f 1 = i = 1 m 1 y i + l = 1 m 2 z l M i n       f 2 = 1 w 1 k = 1 n w k s . t   a i T w + γ + 1 y i ,       i = 1 ,   ,   m 1 b l T w γ + 1 z l ,       l = 1 ,   ,   m 2 y i 0 ,   z l 0
The SVM0 Model (model (11) in Section 2.3) is reformulated as the following MOP ( w = w + w ,   w + ,   w 0 and as mentioned in [54], ( u v ) is the subdifferential of w k at point 0 and u + v T e = k ):
M i n   f 1 = i = 1 m 1 y i + l = 1 m 2 z l M i n   f 2 = e T u + v + σ e T w + + w u v T w + w s . t   a i T w + w + γ + 1 y i ,       i = 1 ,   ,   m 1 b l T w + w γ + 1 z l ,       l = 1 ,   ,   m 2 y i 0 ,   z l 0 ,   w + ,   w 0 ,   0 u ,   v e
To solve these MOPs, we can use the modified algorithm based on the ε-constraint method, which was introduced in Section 2.4 and in [63,64].

4. Numerical Experiments

The results of models mentioned in the previous sections on some numerical experiments are presented in this section. To compare the results, all these models are solved as single-objective and multi-objective forms. To solve the test problems, we used “GlobalSolve” in the Global Optimization package in MAPLE version 18.01. The Global Optimization Toolbox uses global search algorithms that systematically search the entire feasible region for a global extremum [63]. The algorithms in the Global Optimization toolbox are global search methods, which systematically search the entire feasible region for a global extremum [65]. The global solver minimizes a merit function and considers a penalty term for the constraints. In this method, the global search phase is followed by a series of local searches to refine solutions. This solver is designed to search the specified region for a general solution, especially in non-convex optimization problems [66].
We solved all single-objective models (models (6), (7), (11), and (12)) for C = 1 and C = 10 . However, only the results of C = 10 have been reported because the error of some of these models for C = 1 was not equal to zero.
We have implemented all the multi-objective models to obtain 100 Pareto optimal solutions. That is, the algorithm ends after 100 repetitions, and this is the stopping criterion of the algorithm. Since the second objective functions are different in models (17) to (20), we have used the projection of Pareto solutions in the objective function space of model (17) to better compare the Pareto optimal solutions of these models.
Since the minimization of the number of non-zero components of the normal vector of the separator hyperplane and the minimization of the classification error at the same time are two goals of different SVM models, the test problems are specifically designed to challenge the number of non-zero components of the normal vector of the separator hyperplane.
Test Problem 1.
The number of samples is 14, and the number of features is 3 in this test problem. Suppose that we have two sets as follows:
A =   1.7 ,   4 ,   1.5 ,   2 ,   5 ,   1 , 2.5 ,   3.5 ,   1.4 ,   2.8 ,   4 ,   1.2 , 3 ,   5.5 ,   1.6 ,   2.5 ,   5.3 ,   1.3 , 1.5 ,   1.5 ,   0.8 ,     B = 3.8 ,   8 ,   2 ,   [ 5 ,   4.1 ,   1.9 ] , 6 ,   6 ,   2 ,   4.2 ,   6.1 ,   1.8 , [ 3.2 ,   6 ,   2 ] ,   [ 3.5 ,   5.8 ,   2.4 ] ,   [ 4 ,   6.5 ,   3 ] .
The single-objective models all provide the correct set separator (that is, the error of all these models is zero). The vector w returned by BM-SVM and SVM0 methods has just one non-zero component, but l 1 and l 2 return a vector w where components are all non-zero. The results of these single-objective models are depicted in Table 1 and Figure 1.
We used the dataset for our MOP models to obtain 100 Pareto solutions. We have considered 6 Pareto solutions out of 100 Pareto solutions obtained for each MOP for further investigation. In Figure 2, Figure 3, Figure 4 and Figure 5, we have considered a suitable viewing angle for each specific sample (6 Pareto solutions) to have a better view of the separating hyperplanes for MOP models. Additionally, in Table 2, Table 3, Table 4 and Table 5, the results obtained for the same Pareto optimal solutions are displayed.
In Table 2 for the l 1 MOP, the value of w 1 gradually decreases in the solutions while the error value increases.
For example, in the first and second Pareto solutions, a smaller value for the w 1 (with an error value equal to zero) has been achieved compared to the results of the single-objective l 1 , presented in Table 1. In the sixth Pareto solution, one of the components of the vector w is equal to zero, but the error has increased.
In Table 3 for the l 2 MOP model, in the first and second Pareto solutions, a smaller value for the w 1 has been achieved (with an error value equal to zero) compared to the results of the single-objective l 2 problem, presented in Table 1.
In Table 4 for the BM-SVM MOP model, in the third Pareto solution, two components of the vector w are non-zero (with an error value equal to zero) while compared to the results of the single-objective model, presented in Table 1, a smaller value for the w 1 has been achieved.
For the SVM0 MOP model, as shown in Figure 5 and Table 5, in the third Pareto solution, two components of the vector w are non-zero (with an error value equal to zero) while compared to the results of the single-objective model, presented in Table 1, a smaller value for the w 1 has been achieved.
The projection of all Pareto solutions (in the space of Error (Vertical axis) and l 1 norm (Horizontal axis)) obtained from multi-objective models (BM-SVM, SVM0, l 1 , l 2 ) are shown in Figure 6. Additionally, the run time (second) of l 1 , l 2 , BM-SVM and SVM0 multi-objective models, respectively, are 227.906, 269.468, 901.235, and 1236.515 for obtaining 100 Pareto optimal solutions. The lowest run time was related to model l 1 , but as the results of the previous tables, models BM-SVM and SVM0 have performed better in terms of the minimum number of non-zero components of the normal vector of the separator hyperplane.
Test Problem 2.
The number of samples is 12, and the number of features is 4 in this test problem. Suppose that we have two sets as follows:
A = 1.5 ,   4.2 ,   1 ,   2 ,   1.9 ,   4.6 ,   1.5 ,   1.5 ,   1.8 ,   4.5 ,   1.6 ,   1.9 , 1.5 ,   4.3 ,   1.2 ,   1.8 ,   1.2 ,   4.5 ,   1.6 ,   1.6 ,   [ 1.7 ,   4.5 ,   1.4 ,   2 ]
B = 2.2 ,   6 ,   3 ,   2.1 ,   2.6 ,   5 ,   2 ,   2.3 ,   4 ,   4.7 ,   1.7 ,   2.5 , [ 3.2 ,   4.5 ,   2.1 ,   2.3 ] , [ 3.5 ,   5.3 ,   2.5 ,   3.1 ] ,   [ 2.1 ,   5.6 ,   2.5 ,   3.2 ]
The results are shown in Table 6. The single-objective models all provide the correct set separator (that is, the error of all these models is zero). The vector w returned by BM-SVM and SVM0 methods has just one non-zero component, but l 1 and l 2 return a vector w where all components are non-zero.
Pareto optimal solutions obtained from MOP models are depicted in Figure 7. In this figure, the horizontal axis represents the value of l 1 norm of vector w , and the vertical axis represents the error level. To clarify the discussion, in Figure 8a, the Pareto frontier of the BM-SVM multi-objective model is displayed in the space of the objective functions of this model for Test Problem 2. In Figure 8b the projection of this Pareto frontier in the space of Error (vertical axis) and l 1 -norm (horizontal axis) is displayed.
We have considered only three Pareto solutions out of the 100 Pareto optimal solutions obtained for each MOP model that seemed more interesting for consideration. The results are displayed in Table 7, Table 8, Table 9 and Table 10.
For the BM-SVM MOP model, as shown in Table 7, for all Pareto solutions that are considered, three components of vector w is equal to zero, and in each solution, the smaller value for w 1 has been achieved, but the errors are not equal to zero.
As shown in Table 8 for the SVM0 MOP model, in the first Pareto solution, two components of vector w is equal to zero, and in the two other Pareto solutions, three components of vector w is equal to zero, but the error value is non-zero.
As shown in Table 9, in the l 1 MOP model, for all Pareto solutions, one component of the vector w is equal to zero but with non-zero error.
As shown in Table 10 for the l 2 MOP model, in all Pareto solutions which are considered, all components of the vector w are non-zero.
The run time (second) of l 1 , l 2 , BM-SVM and SVM0 multi-objective models, respectively, are 885.031, 418.578, 133.594, and 546.472 for obtaining 100 Pareto optimal solutions. The lowest run time was related to model BM-SVM. Additionally, as the results of the previous tables, models BM-SVM and SVM0 have performed better in terms of the minimum number of non-zero components of the normal vector of the separator hyperplane.
Test Problem 3.
The number of samples is 8, and the number of features is 5 in this test problem. Suppose that we have two sets as follows:
A = [ 2.3 ,   3.5 ,   1 ,   2.7 ,   1 ] ,   [ 2.8 ,   3.6 ,   1.5 ,   2.5 ,   1.1 ] ,   [ 2 ,   4.9 ,   1.6 ,   2.4 ,   1.2 ] ,   [ 2.5 ,   3.9 ,   1.8 ,   2 ,   1.3 ] , B = [ 3.1 ,   5.6 ,   3 ,   3.1 ,   2 ] ,   [ 3.6 ,   4.6 ,   2 ,   3.3 ,   2.1 ] ,   [ 4 ,   5 ,   1.7 ,   2.9 ,   2.2 ] ,   [ 3.2 ,   4.2 ,   2.3 ,   2.5 ,   2.4 ]
All single-objective models provide the correct separator. The vector w returned by BM-SVM and SVM0 has just one non-zero component, but the l 1 and l 2 return a vector w where components are all non-zero. The results are depicted in Table 11.
Pareto solutions obtained from MOP models are shown in Figure 9. We have considered only three Pareto solutions that seemed more interesting out of the 100 Pareto optimal solutions obtained for each MOP model. The results are shown in Table 12, Table 13, Table 14 and Table 15.
For BM-SVM MOP, as shown in Table 12, for the first and second Pareto solutions, four components of vector w are equal to zero, and the error values are zero. Additionally, compared to the results of the single-objective model, a smaller value for the w 1 has been achieved.
For the SVM0 MOP model, as shown in Table 13, for the first Pareto solution, four components of vector w are equal to zero, and the error value is equal to zero, for the second and third Pareto solutions, while the error value is non-zero, four components of vector w are equal to zero.
For the l 1 MOP model, as shown in Table 14, one component of vector w is equal to zero, with a non-zero error.
For the l 2 MOP model, as shown in, for all Pareto solutions, all components of vector w are non-zero.
The run time (second) of l 1 , l 2 , BM-SVM, and SVM0 multi-objective models, respectively, are 102.500, 134.188, 239.953, and 576.343 for obtaining 100 Pareto optimal solutions. The lowest run time was related to l 1 model. Additionally, as the results of the previous tables, models BM-SVM and SVM0 have performed better in terms of the minimum number of non-zero components of the normal vector of the separator hyperplane.

5. Conclusions

The design of linear Support Vector Machine (SVM) classification techniques is generally a multi-objective optimization that requires finding appropriate trade-offs between several objectives, such as misclassified training data (classification error) and the number of non-zero elements of the separator hyperplane. We proposed multi-objective binary classification problems to show the advantages of considering these problems for sparse optimization in linear SVM classification techniques. The results of the proposed classification methods in single-objective and multi-objective forms are reported on several datasets. The results showed that by using multi-objective models, we can choose a more appropriate separating hyperplane. By using multi-objective models (especially BM-SVM and SVM0 multi-objective models), separator hyperplanes have been obtained with the minimum possible error and, at the same time, the minimum number of non-zero components of the normal vector.

Author Contributions

Conceptualization, B.P. (Behzad Pirouz) and B.P. (Behrouz Pirouz); methodology, B.P. (Behzad Pirouz) and B.P. (Behrouz Pirouz); software, B.P. (Behzad Pirouz); validation, B.P. (Behzad Pirouz); formal analysis, B.P. (Behzad Pirouz); investigation, B.P. (Behzad Pirouz); resources, B.P. (Behrouz Pirouz); data curation, B.P. (Behzad Pirouz); writing—original draft preparation, B.P. (Behzad Pirouz) and B.P. (Behrouz Pirouz); writing—review and editing, B.P. (Behzad Pirouz) and B.P. (Behrouz Pirouz); visualization, B.P. (Behrouz Pirouz); All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The datasets analyzed during the current study are available in the Manuscript.

Acknowledgments

We are grateful to Manlio Gaudioso.

Conflicts of Interest

The authors have no relevant financial or non-financial interests to disclose. The authors declare that they have no conflict of interest.

References

  1. Jin, Y.; Sendhoff, B. Pareto-Based Multiobjective Machine Learning: An Overview and Case Studies. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 2008, 38, 397–415. [Google Scholar] [CrossRef]
  2. Suttorp, T.; Igel, C. Multi-Objective Optimization of Support Vector Machines. In Multi-Objective Machine Learning; Jin, Y., Ed.; Studies in Computational Intelligence; Springer: Berlin/Heidelberg, Germany, 2006; Volume 16. [Google Scholar] [CrossRef]
  3. Zoltan, Z.; Kalmar, Z.; Szepesvari, C. Multi-criteria reinforcement learning. Proc. Int. Conf. Mach. Learn. 1998, 98, 197–205. [Google Scholar]
  4. Coleman, T.F. (Ed.) Large Sparse Numerical Optimization; Springer: Berlin/Heidelberg, Germany, 1984. [Google Scholar] [CrossRef]
  5. Zhao, Y.B. Sparse Optimization Theory and Methods; CRC Press: Boca Raton, FL, USA, 2018. [Google Scholar] [CrossRef]
  6. Van Zyl, J.P.; Engelbrecht, A.P. Set-Based Particle Swarm Optimisation: A Review. Mathematics 2023, 11, 2980. [Google Scholar] [CrossRef]
  7. Zitzler, E. Evolutionary Algorithms for Multiobjective Optimization: Methods and Applications; Shaker: Ithaca, NY, USA, 1999; Volume 63. [Google Scholar]
  8. Collette, Y.; Siarry, P. Multiobjective optimization: Principles and case studies. In Decision Engineering; Springer: Berlin/Heidelberg, Germany, 2003. [Google Scholar]
  9. Das, I.; Dennis, J.E. A closer look at drawbacks of minimizing weighted sums of objectives for Pareto set generation in multicriteria optimization problems. Struct. Optim. 1997, 14, 63–69. [Google Scholar] [CrossRef]
  10. Teixeira, R.A.; Braga, A.P.; Takahashi, R.H.C.; Saldanha, R.R. Improving generalization of MLPs with multi-objective optimization. Neurocomputing 2000, 35, 189–194. [Google Scholar] [CrossRef]
  11. Deb, K. Multi-objective optimisation using evolutionary algorithms: An introduction. In Multi-Objective Evolutionary Optimisation for Product Design and Manufacturing; Springer: London, UK, 2011; pp. 3–34. [Google Scholar]
  12. Sawaragi, Y.; Nakayama, H.; Tanino, T. Theory of multiobjective optimization. In Mathematics in Science and Engineering; Elsevier: Amsterdam, The Netherlands, 1985; Volume 176. [Google Scholar]
  13. Chankong, V.; Haimes, Y. Optimization-based methods for multiobjective decision-making: An overview. Large Scale Syst. 1983, 5, 1–33. [Google Scholar]
  14. Coello Coello, C.A.; Van Veldhuizen, D.A.; Lamont, G.B. Evolutionary Algorithms for Solving Multi-Objective Problems; Kluwer Academic Publishers: Dordrecht, The Netherlands, 2002. [Google Scholar]
  15. Pareto, V.; Bonnet, A. Manuel D’économie Politique; Giard, V., Brière, E., Eds.; Bibliothèque Internationale D’économie Politique: Paris, France, 1909. [Google Scholar]
  16. Lopez-Ibanez, M.; Dubois-Lacoste, J.; Stutzle, T.; Birattari, M. The Irace Package, Iterated Race for Automatic Algorithm Configuration; Technical Report TR/IRIDIA/2011-004, IRIDIA; Université Libre de Bruxelles: Bruxelles, Belgium, 2011. [Google Scholar]
  17. Lang, M.; Kotthaus, H.; Marwedel, P.; Weihs, C.; Rahnenführer, J.; Bischl, B. Automatic model selection for high-dimensional survival analysis. J. Stat. Comput. Simul. 2015, 85, 62–76. [Google Scholar] [CrossRef]
  18. Jones, D.R. A taxonomy of global optimization methods based on response surfaces. J. Glob. Optim. 2001, 21, 345–383. [Google Scholar] [CrossRef]
  19. Thornton, C.; Hutter, F.; Hoos, H.H.; Leyton-Brown, K. August. Auto-WEKA: Combined selection and hyperparameter optimization of classification algorithms. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Ser. KDD ’13, Chicago, IL, USA, 11–14 August 2013; ACM: New York, NY, USA, 2013; pp. 847–855. [Google Scholar]
  20. Koch, P.; Bischl, B.; Flasch, O.; Bartz-Beielstein, T.; Weihs, C.; Konen, W. Tuning and evolution of support vector kernels. Evol. Intell. 2012, 5, 153–170. [Google Scholar] [CrossRef]
  21. Jin, Y. Surrogate-assisted evolutionary computation: Recent advances and future challenges. Swarm Evol. Comput. 2011, 1, 61–70. [Google Scholar] [CrossRef]
  22. Horn, D.; Demircioğlu, A.; Bischl, B.; Glasmachers, T.; Weihs, C. A comparative study on large scale kernelized support vector machines. Adv. Data Anal. Classif. 2018, 12, 867–883. [Google Scholar]
  23. Jin, Y. (Ed.) Multi-Objective Machine Learning; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2006; Volume 16. [Google Scholar]
  24. Everson, R.M.; Fieldsend, J.E. Multi-class {ROC} analysis from a multi-objective optimisation perspective. Pattern Recognit. Lett. 2006, 27, 918–927. [Google Scholar]
  25. Graning, L.; Jin, Y.; Sendhoff, B. Generalization improvement in multi-objective learning. In Proceedings of the 2006 IEEE International Joint Conference on Neural Network Proceedings, Vancouver, BC, Canada, 16–21 July 2006; IEEE: New York, NY, USA, 2006; pp. 4839–4846. [Google Scholar]
  26. Law, M.H.; Topchy, A.P.; Jain, A.K. Multiobjective data clustering. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 27 June–2 July 2004; IEEE: New York, NY, USA, 2004; Volume 2, p. II. [Google Scholar]
  27. Liu, G.P.; Kadirkamanathan, V. Learning with multi-objective criteria. In Proceedings of the 1995 Fourth International Conference on Artificial Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; pp. 53–58. [Google Scholar]
  28. Bi, J. Multi-objective programming in SVMs. In Proceedings of the 20th International Conference on Machine Learning, Washington, DC, USA, 21–24 August 2003; pp. 35–42. [Google Scholar]
  29. Igel, C. Multi-objective model selection for support vector machines. In Evolution Multi-Criterion Optimization Lecture Notes in Computer Science; Springer: New York, NY, USA, 2005; Volume 3410, pp. 534–546. [Google Scholar]
  30. Nakayama, H.; Asada, T. Support vector machines formulated as multi-objective linear programming. In Proceedings of the ICOTA, Hong Kong, 15–17 December 2001; pp. 1171–1178. [Google Scholar]
  31. Bernado-Manssilla, E.; Garrell-Guii, J. MOLeCS: Using multiobjective evolutionary algorithms for learning. In Proceedings of the EMO 2001 Lecture Notes in Computer Science, Zurich, Switzerland, 7–9 March 2001; Springer: New York, NY, USA, 2001; Volume 1993, pp. 696–710. [Google Scholar]
  32. Zhang, Y.; Rockett, P.I. Evolving optimal feature extraction using multi-objective genetic programming: A methodology and preliminary study on edge detection. In Proceedings of the Genetic and Evolutionary Computation Conference, Washington, DC, USA, 25–29 June 2005; pp. 795–802. [Google Scholar]
  33. Ishibuchi, H.; Nakashima, T.; Murata, T. Three-objective genetics-based machine learning for linguistic rule extraction. Inf. Sci. 2001, 136, 109–133. [Google Scholar]
  34. Van Moffaert, K.; Nowé, A. Multi-objective reinforcement learning using sets of Pareto dominating policies. J. Mach. Learn. Res. 2014, 15, 3483–3512. [Google Scholar]
  35. Cordon, O.; Herrera, F.; del-Jesus, M.; Villar, P. A multi-objective genetic algorithm for feature selection and granularity learning in fuzzy-rule based classification systems. In Proceedings of 9th IFSA World Congress and 20th NAFIPS International Conference, Vancouver, BC, Canada, 25–28 July 2001; Volume 3, pp. 1253–1258. [Google Scholar]
  36. Oliveira, L.S.; Sabourin, R.; Bortolozzi, F.; Suen, C.Y. Feature selection for ensembles: A hierarchical multi-objective genetic algorithm approach. In Proceedings of the Seventh International Conference on Document Analysis and Recognition, Edinburgh, UK, 3–6 August 2003; IEEE: New York, NY, USA; pp. 676–680. [Google Scholar]
  37. Handl, J.; Knowles, J. Exploiting the Tradeoff—The Benefits of Multiple Objectives in Data Clustering, Evolutionary Multi-Criterion Optimization Lecture Notes in Computer Science; Springer: New York, NY, USA, 2005; Volume 3410, pp. 547–560. [Google Scholar]
  38. Jin, Y.; Sendhoff, B. Alleviating catastrophic forgetting via multiobjective learning. In Proceedings of International Joint Conference on Neural Network, Vancouver, BC, Canada, 16–21 July 2006; pp. 6367–6374. [Google Scholar]
  39. Kokshenev, I.; Braga, A.P. An efficient multi-objective learning algorithm for RBF neural network. Neurocomputing 2010, 73, 2799–2808. [Google Scholar]
  40. Torres, L.C.B.; Castro, C.L.; Braga, A.P. A computational geometry approach for Pareto-optimal selection of neural networks. In Proceedings of the International Conference on Artificial Neural Networks (ICANN), Lausanne, Switzerland, 11–14 September 2012; pp. 100–107. [Google Scholar]
  41. Teixeira, R.; Braga, A.P.; Saldanha, R.; Takahashi, R.H.; Medeiros, T.H. The usage of golden section in calculating the efficient solution in artificial neural networks training by multi-objective optimization. In Proceedings of the International Conference on Artificial Neural Networks (ICANN), Porto, Portugal, 9–13 September 2007; Volume 4668, pp. 289–298. [Google Scholar]
  42. Chankong, V.; Haimes, Y.Y. Multiobjective Decision Making: Theory and Methodology; Elsevier: Amsterdam, The Netherlands; North-Holland: New York, NY, USA, 1983; Volume 8. [Google Scholar]
  43. Nisbet, R.; Elder, J.; Miner, G.D. Handbook of Statistical Analysis and Data Mining Applications; Academic Press: Cambridge, MA, USA, 2009. [Google Scholar]
  44. Yang, X.S. Introduction to Algorithms for Data Mining and Machine Learning; Academic Press: Cambridge, MA, USA, 2019. [Google Scholar]
  45. Zhao, Y.; Cen, Y. Data Mining Applications with R; Academic Press: Cambridge, MA, USA, 2013. [Google Scholar]
  46. Rinaldi, F. Mathematical Programming Methods for Minimizing the Zero-Norm over Polyhedral Sets. 2009, Sapienza, University of Rome. Available online: https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.726.794&rep=rep1&type=pdf (accessed on 18 August 2023).
  47. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar]
  48. Haykin, S. Neural Networks: A Comprehensive Foundation, 2nd ed.; Prentice-Hall: New York, NY, USA, 1999. [Google Scholar]
  49. Bradley, P.S.; Mangasarian, O.L. Feature selection via concave minimization and support vector machines, Machine Learning Proceedings of the Fifteenth International Conference (ICML 1998), Madison, WI, USA, 24–27 July 1998; Shavlik, J., Ed.; Morgan Kaufmann: San Francisco, CA, USA, 1998; pp. 82–90. [Google Scholar]
  50. Bennett, K.P.; Blue, J.A. A support vector machine approach to decision trees. In Proceedings of the IEEE International Joint Conference on Neural Networks Proceedings. IEEE World Congress on Computational Intelligence (Cat. No.98CH36227), Anchorage, AK, USA, 4–9 May 1998; Volume 3, pp. 2396–2401. [Google Scholar]
  51. Vapnik, V. The Nature of Statistical Learning Theory, 2nd ed.; Springer Science & Business Media: Berlin/Heidelberg, Germany, 1999. [Google Scholar]
  52. Rinaldi, F.; Schoen, F.; Sciandrone, M. Concave programming for minimizing the zero-norm over polyhedral sets. Comput. Optim. Appl. 2010, 46, 467–486. [Google Scholar] [CrossRef]
  53. Bishop, C.M. Neural Networks for Pattern Recognition; Oxford University Press: New York, NY, USA, 1995. [Google Scholar]
  54. Gaudioso, M.; Gorgone, E.; Hiriart-Urruty, J.B. Feature selection in SVM via polyhedral k-norm. Optim. Lett. 2020, 14, 19–36. [Google Scholar] [CrossRef]
  55. Mangasarian, O.L. Machine learning via polyhedral concave minimization. In Applied Mathematics and Parallel Computing: Festschrift for Klaus Ritter; Physica-Verlag HD: Heidelberg, Germany, 1996; pp. 175–188. [Google Scholar]
  56. Gaudioso, M.; Gorgone, E.; Labbé, M.; Rodríguez-Chía, A.M. Lagrangian relaxation for SVM feature selection. Comput. Oper. Res. 2017, 87, 137–145, ISSN 0305-0548. [Google Scholar] [CrossRef]
  57. Gaudioso, M.; Giallombardo, G.; Miglionico, G.; Bagirov, A.M. Minimizing nonsmooth DC functions via successive DC piecewise affine approximations. J. Glob. Optim. 2018, 71, 37–55. [Google Scholar]
  58. Pirouz, B.; Gaudioso, M. New Mixed Integer Fractional Programming Problem for Sparse Optimization. In Proceedings of the ODS 2021: International Conference on Optimization and Decision Sciences, Rome, Italy, 19 September 2021; Available online: http://www.airoconference.it/ods2021/images/ODS2021_Conference_Program_web_v4.pdf (accessed on 18 August 2023).
  59. Pirouz, B.; Gaudioso, M. A Multi-Objective Programming Problem for Sparse Optimization with application in SVM feature selection. In Proceedings of the ODS 2022: International Conference on Optimization and Decision Sciences, Firenze, Italy, 25 August 2024. [Google Scholar]
  60. Pirouz, B.; Gaudioso, M. New mixed integer fractional programming problem and some multi-objective models for sparse optimization. Soft Comput. 2023, 1–12. [Google Scholar] [CrossRef]
  61. Ehrgott, M. Multicriteria Optimization, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2005. [Google Scholar]
  62. Mavrotas, G. Effective implementation of the ε-constraint method in multi-objective mathematical programming problems. Appl. Math. Comput. 2009, 213, 455–465. [Google Scholar] [CrossRef]
  63. Pirouz, B.; Khorram, E. A Computational Approach Based on the epsilon-Constraint Method in Multi-Objective Optimization Problems. In Advances and Applications in Statistics; Pushpa Publishing House: Allahabad, India, 2016. [Google Scholar] [CrossRef]
  64. Pirouz., B.; Ramezani Paschapari, J. A Computational Algorithm Based on Normalization for Constructing the Pareto Front of Multiobjective Optimization Problems. In Proceedings of the 5th International Conference on Industrial and Systems Engineering, Mashhad, Iran, 4 September 2019. [Google Scholar]
  65. Pinter, J.D.; Linder, D.; Chin, P. Global optimization toolbox for maple: An introduction with illustrative applications. Optim. Methods Softw. 2006, 21, 565–582. [Google Scholar] [CrossRef]
  66. Maplesoft. Available online: https://www.maplesoft.com/support/help/maple/view.aspx?path=GlobalOptimization%2FGlobalSolve#info (accessed on 18 August 2023).
Figure 1. The result of Separator Hyperplanes in single-objective models for Test Problem 1. (a) Separator Hyperplanes of BM-SVM. (b) Separator Hyperplanes of SVM0. (c) Separator Hyperplanes of l 1 . (d) Separator Hyperplanes of l 2 .
Figure 1. The result of Separator Hyperplanes in single-objective models for Test Problem 1. (a) Separator Hyperplanes of BM-SVM. (b) Separator Hyperplanes of SVM0. (c) Separator Hyperplanes of l 1 . (d) Separator Hyperplanes of l 2 .
Mathematics 11 03721 g001
Figure 2. Some results of Separator Hyperplanes in l 1 MOP model for Test Problem 1.
Figure 2. Some results of Separator Hyperplanes in l 1 MOP model for Test Problem 1.
Mathematics 11 03721 g002
Figure 3. Some results of Separator Hyperplanes in l 2 MOP model for Test Problem 1.
Figure 3. Some results of Separator Hyperplanes in l 2 MOP model for Test Problem 1.
Mathematics 11 03721 g003
Figure 4. Some results of Separator Hyperplanes in BM-SVM MOP model for Test Problem 1.
Figure 4. Some results of Separator Hyperplanes in BM-SVM MOP model for Test Problem 1.
Mathematics 11 03721 g004
Figure 5. Some results of Separator Hyperplanes in SVM0 MOP model for Test Problem 1.
Figure 5. Some results of Separator Hyperplanes in SVM0 MOP model for Test Problem 1.
Mathematics 11 03721 g005
Figure 6. The projection of Pareto solutions (in the space of error and l_1 norm) obtained from multi-objective models (BM-SVM, SVM0, l 1 and l 2 models) for the dataset of Test Problem 1.
Figure 6. The projection of Pareto solutions (in the space of error and l_1 norm) obtained from multi-objective models (BM-SVM, SVM0, l 1 and l 2 models) for the dataset of Test Problem 1.
Mathematics 11 03721 g006
Figure 7. The projection of Pareto solutions (in the space of Error and l_1 norm) obtained from multi-objective models for the dataset of Test Problem 2.
Figure 7. The projection of Pareto solutions (in the space of Error and l_1 norm) obtained from multi-objective models for the dataset of Test Problem 2.
Mathematics 11 03721 g007
Figure 8. The Pareto frontier of the BM-SVM multi-objective model and its projection for Test Problem 2. (a) The Pareto solutions in the space of objective functions of the BM-SVM model for Test Problem 2. (b) The projection of Pareto solutions in the space of Error and l 1 -norm for Test Problem 2.
Figure 8. The Pareto frontier of the BM-SVM multi-objective model and its projection for Test Problem 2. (a) The Pareto solutions in the space of objective functions of the BM-SVM model for Test Problem 2. (b) The projection of Pareto solutions in the space of Error and l 1 -norm for Test Problem 2.
Mathematics 11 03721 g008
Figure 9. The projection of Pareto solutions (in the space of Error and l_1 norm) obtained from multi-objective models for the dataset of Test Problem 3.
Figure 9. The projection of Pareto solutions (in the space of Error and l_1 norm) obtained from multi-objective models for the dataset of Test Problem 3.
Mathematics 11 03721 g009
Table 1. The results of single-objective models for Test Problem 1.
Table 1. The results of single-objective models for Test Problem 1.
Methodw_1w_2w_3 w 1 Error ValueCorrectness
BM-SVM00 9.9998 9.99980100.00%
SVM0 Model00 10 100100.00%
l 1 Model 0.7500 0.5000 4.0000 5.25000100.00%
l 2 Model 1.8265 1.6276 1.9541 5.40820100.00%
Table 2. The results of l 1 MOP model for Test Problem 1.
Table 2. The results of l 1 MOP model for Test Problem 1.
Pareto Solutionw_1w_2w_3 w 1 Error ValueCorrectness
1 0.7400 0.4933 3.9470 5.1803 0 100.00 %
2 0.7048 0.4699 3.7590 4.9336 0 100.00 %
3 0.4405 0.2937 2.3494 3.0835 0.8253 92.86 %
4 0.8068 0.1502 1.0164 1.9734 1.3569 85.72 %
5 0.7739 0.3108 0.0254 1.1101 2.9710 64.29 %
6 0.7405 0.2462 0 0.9867 3.4620 50.00 %
Table 3. The results of l 2 multi-objective model for Test Problem 1.
Table 3. The results of l 2 multi-objective model for Test Problem 1.
Pareto Solutionw_1w_2w_3 w 1 Error ValueCorrectness
1 1.0146 0.7756 3.5230 5.3132 0100.00%
2 1.7660 1.5736 1.8893 5.2289 0100.00%
3 1.0196 0.9085 1.0908 3.0189 0.9055 92.86%
4 0.9476 0.7882 0.9616 2.6974 1.0317 85.72%
5 0.7108 0.4092 0.7411 1.8611 1.6263 78.57%
6 0.6313 0.3030 0.3473 1.2816 2.7562 57.14%
Table 4. The results of the BM-SVM multi-objective model for Test Problem 1.
Table 4. The results of the BM-SVM multi-objective model for Test Problem 1.
Pareto Solutionw_1w_2w_3 w 1 Error ValueCorrectness
1 0.7500 0.5000 4.0000 5.250000100.00%
2 0.0460 0 9.7375 9.78350100.00%
3 00 8.5929 8.59290100.00%
4 6.5340 006.53400.737492.86%
500 5.0000 5.00001.474885.71%
600 4.0000 4.00001.600078.57%
Table 5. The results of the SVM0 multi-objective model for Test Problem 1.
Table 5. The results of the SVM0 multi-objective model for Test Problem 1.
Pareto Solutionw_1w_2w_3 w 1 Error ValueCorrectness
1 0.7500 0.5000 4.0000 5.25000100.00%
2 5.0000 0 2.5000 7.50000100.00%
3 00 9.8603 9.86030100.00%
4 00 7.2591 7.25910.548192.86%
5 6.5625 006.56251.238692.86%
600 5.0000 5.00002.109078.57%
Table 6. The results of single-objective models for Test Problem 2.
Table 6. The results of single-objective models for Test Problem 2.
Methodw_1w_2w_3w_4 w 1 Error ValueCorrectness
BM-SVM 10.00 00010.000100.00%
SVM0 Model 10.00 00010.000100.00%
l 1 Model 1.7886 0.4878 0.3252 0.4878 3.08940100.00%
l 2 Model 1.3223 0.8264 0.6612 0.6612 3.47110100.00%
Table 7. The results of some Pareto solutions of BM-SVM multi-objective model for the dataset of Test Problem 2.
Table 7. The results of some Pareto solutions of BM-SVM multi-objective model for the dataset of Test Problem 2.
Pareto Solutionw_1w_2w_3w_4 w 1 Error ValueCorrectness
100 4.8920 04.89203.402075.00%
200 2.2222 02.22227.678866.67%
300 5.3441 05.344118.371058.33%
Table 8. The results of some Pareto solutions of the SVM0 multi-objective model for the dataset of Test Problem 2.
Table 8. The results of some Pareto solutions of the SVM0 multi-objective model for the dataset of Test Problem 2.
Pareto Solutionw_1w_2w_3w_4 w 1 Error ValueCorrectness
1 1.6949 2.0339 003.72880100.00%
2 7.2008 0007.20080.559891.67%
300 4.6410 04.64101.679583.33%
Table 9. The results of some Pareto solutions of l 1 multi-objective model for the dataset of Test Problem 2.
Table 9. The results of some Pareto solutions of l 1 multi-objective model for the dataset of Test Problem 2.
Pareto Solutionw_1w_2w_3w_4 w 1 Error ValueCorrectness
1 0.7279 0 0.9160 0.4826 2.12651.098466.67%
2 0.6897 0 0.8314 0.2927 1.81381.827358.33%
3 0.8121 0 0.7514 0 1.56362.527350.00%
Table 10. The results of some Pareto solutions of l 2 multi-objective model for the dataset of Test Problem 2.
Table 10. The results of some Pareto solutions of l 2 multi-objective model for the dataset of Test Problem 2.
Pareto Solutionw_1w_2w_3w_4 w 1 Error ValueCorrectness
1 0.9504 0.3632 0.5770 0.5682 2.45880.600083.33%
2 0.6877 0.4267 0.6012 0.4909 2.20651.200066.66%
3 0.5831 0.3852 0.5331 0.4354 1.93681.800058.33%
Table 11. The results of Test Problem 3 for single-objective models.
Table 11. The results of Test Problem 3 for single-objective models.
Methodw_1w_2w_3w_4w_5 w 1 Error ValueCorrectness
BM-SVM0000 8.3334 8.33340100.00%
SVM0 Model0000 10.00 10.000100.00%
l 1 Model 0.1892 0.0946 0.2270 0.4541 1.3623 2.32730100.00%
l 2 Model 0.6114 0.2620 0.4367 0.4367 0.9607 2.70740100.00%
Table 12. The results of some Pareto solutions of BM-SVM multi-objective model for the dataset of Test Problem 3.
Table 12. The results of some Pareto solutions of BM-SVM multi-objective model for the dataset of Test Problem 3.
Pareto Solutionw_1w_2w_3w_4w_5 w 1 Error ValueCorrectness
10000 7.1575 7.15750100.00%
2 6.7742 00006.77420100.00%
30000 4.0209 4.02090.861387.50%
Table 13. The results of some Pareto solutions of the SVM0 multi-objective model for the dataset of Test Problem 3.
Table 13. The results of some Pareto solutions of the SVM0 multi-objective model for the dataset of Test Problem 3.
Pareto Solutionw_1w_2w_3w_4w_5 w 1 Error ValueCorrectness
10000 9.91479.91470100.00%
2 3.548400003.54840.935587.50%
30000 2.50002.50001.952875.00%
Table 14. The results of some Pareto solutions of l 1 multi-objective model for the dataset of Test Problem 3.
Table 14. The results of some Pareto solutions of l 1 multi-objective model for the dataset of Test Problem 3.
Pareto Solutionw_1w_2w_3w_4w_5 w 1 Error ValueCorrectness
1 0.6347 0.2952 0.0858 0 1.1541 2.16990.206475.00%
2 0.6302 0.2470 0 0.0039 1.1801 2.06120.436962.50%
3 0.3035 0.0437 0 0 1.3887 1.73591.096350.00%
Table 15. The results of some Pareto solutions of l 2 multi-objective model for the dataset of Test Problem 3.
Table 15. The results of some Pareto solutions of l 2 multi-objective model for the dataset of Test Problem 3.
Pareto Solutionw_1w_2w_3w_4w_5 w 1 Error ValueCorrectness
1 0.2549 0.1032 0.2135 0.7962 1.2810 2.64890100.00%
2 0.2120 0.0556 0.3538 0.6104 1.2298 2.46160100.00%
3 0.5363 0.2690 0.3751 0.3867 0.8051 2.37210.277487.50%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pirouz, B.; Pirouz, B. Multi-Objective Models for Sparse Optimization in Linear Support Vector Machine Classification. Mathematics 2023, 11, 3721. https://doi.org/10.3390/math11173721

AMA Style

Pirouz B, Pirouz B. Multi-Objective Models for Sparse Optimization in Linear Support Vector Machine Classification. Mathematics. 2023; 11(17):3721. https://doi.org/10.3390/math11173721

Chicago/Turabian Style

Pirouz, Behzad, and Behrouz Pirouz. 2023. "Multi-Objective Models for Sparse Optimization in Linear Support Vector Machine Classification" Mathematics 11, no. 17: 3721. https://doi.org/10.3390/math11173721

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop