Next Article in Journal
An Optimal Stacked Ensemble Deep Learning Model for Predicting Time-Series Data Using a Genetic Algorithm—An Application for Aerosol Particle Number Concentrations
Previous Article in Journal
Development of a Test-Bench for Evaluating the Embedded Implementation of the Improved Elephant Herding Optimization Algorithm Applied to Energy-Based Acoustic Localization
Previous Article in Special Issue
Structured (De)composable Representations Trained with Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Minimal Complexity Support Vector Machines for Pattern Classification †

Kobe University, Kobe 657-8501, Japan
This paper is an extended version of our paper published in Abe, S. Minimal Complexity Support Vector Machines. In Artificial Neural Networks in Pattern Recognition. ANNPR 2020. Lecture Notes in Computer Science, Vol 12294; Schilling, F.P., Stadelmann, T., Eds.; Springer: Cham, Switzerland, 2020; pp. 89–101.
Computers 2020, 9(4), 88; https://doi.org/10.3390/computers9040088
Submission received: 15 September 2020 / Revised: 23 October 2020 / Accepted: 30 October 2020 / Published: 4 November 2020
(This article belongs to the Special Issue Artificial Neural Networks in Pattern Recognition)

Abstract

:
Minimal complexity machines (MCMs) minimize the VC (Vapnik-Chervonenkis) dimension to obtain high generalization abilities. However, because the regularization term is not included in the objective function, the solution is not unique. In this paper, to solve this problem, we discuss fusing the MCM and the standard support vector machine (L1 SVM). This is realized by minimizing the maximum margin in the L1 SVM. We call the machine Minimum complexity L1 SVM (ML1 SVM). The associated dual problem has twice the number of dual variables and the ML1 SVM is trained by alternatingly optimizing the dual variables associated with the regularization term and with the VC dimension. We compare the ML1 SVM with other types of SVMs including the L1 SVM using several benchmark datasets and show that the ML1 SVM performs better than or comparable to the L1 SVM.

1. Introduction

In the support vector machine (SVM) [1,2], training data are mapped into the high dimensional feature space, and in that space, the separating hyperplane is determined so that the nearest training data of both classes are maximally separated. Here, the distance between a data sample and the separating hyperplane is called margin. Thus, the SVM is trained so that the minimum margin is maximized.
Motivated by the success of SVMs in real world applications, many SVM-like classifiers have been developed to improve the generalization ability. The ideas of extensions lie in incorporating the data distribution (or margin distribution) to the classifiers because the SVM only considers the data that are around the separating hyperplane. If the distribution of one class is different from that of the other class, the separating hyperplane with the same minimum margin for both classes may not be optimal.
To cope with this, one approach proposes kernels based on the Mahalanobis distance [3,4,5,6,7,8,9,10]. Another approach reformulates the SVM such that the margin is measured by the Mahalanobis distance [11,12,13,14,15].
Yet another approach controls the overall margins instead of the minimum margin [16,17,18,19,20,21,22]. In [16], only the average margin is maximized. While the architecture is very simple, the generalization ability is inferior to that of the SVM [17,19]. To improve the generalization ability, the equality constraints were added in [19], but this results in the least squares SVM (LS SVM).
In [17], a large margin distribution machine (LDM) was proposed, in which the average margin is maximized and the margin variance is minimized. The LDM is formulated based on the SVM with inequality constraints. While the generalization ability is better than that of the SVM, the problem is that the LDM has one more hyperparameter than the SVM does. To cope with this problem, in [20,21], the unconstrained LDM (ULDM) was proposed, which has the equal number of hyperparameters and which has the generalization ability comparable to that of the LDM and the SVM.
The generalization ability of the SVM can be analyzed by the VC (Vapnik-Chervonenkis) dimension [1] and the generalization ability will be improved if the radius-margin ratio is minimized, where the radius is the radius of the minimum hypersphere that encloses all the training data in the feature space. The minimum hypersphere changes if the mapping function is changed during model selection, where all the parameter values including the value for the mapping function are determined, or feature selection is performed during training a classifier. So, minimization of the radius-margin ratio can be utilized in selecting the optimal mapping function [23,24]. In [25,26], feature selection is carried out during training by minimizing the radius-margin ratio.
If the center of the hypersphere is assumed to be at the origin, the hypersphere can be minimized for a given feature space as discussed in [27,28]. The minimal complexity machine (MCM) was derived based on this assumption. In the MCM, the VC dimension is minimized by minimizing the upper bound of the soft-margin constraints for the decision function. Because the regularization term is not included, the MCM is trained by linear programming. The quadratic version of the MCM (QMCM) tries to minimizes the VC dimension directly [28]. The generalization performance of the MCM is shown to be superior to that of the SVM, but according to our analysis [29], the solution is non-unique and the generalization ability is not better than that of the SVM. The problem of non-uniqueness of the solution is solved by adding the regularization term in the objective function of the MCM, which is a fusion of the MCM and the linear programming SVM (LP SVM) called MLP SVM.
In [30], to improve the generalization ability of the standard SVM (L1 SVM), we proposed fusing the MCM and the L1 SVM, which is the minimal complexity L1 SVM (ML1 SVM). We also proposed ML1 v SVM, whose component is more similar to the L1 SVM. By the computer experiment using RBF (radius basis function) kernels, the proposed classifiers generalize better than or comparable to the L1 SVM.
In this paper we discuss the ML1 SVM and ML1 v SVM more in detail, propose their training methods, prove their convergence, and demonstrate the effectiveness of the proposed classifiers using polynomial kernels in addition to RBF kernels.
The ML1 SVM is obtained by adding the upper bound on the decision function and the upper bound minimization term in the objective function of the L1 SVM. We show that this corresponds to minimizing the maximum margin. We derive the dual form of the ML1 SVM with one set of variables associated with the soft-margin constraints and the other set, upper-bound constraints. We then decompose the dual ML1 SVM into two subproblems: one for the soft-margin constraints, which is similar to the dual L1 SVM, and the other for the upper-bound constraints. These subproblems include neither the bias term nor the upper bound. Thus, for a convergence check, we derive the exact KKT (Karush-Kuhn-Tucker) conditions that do not include the bias term and the upper bound. The second subproblem is different from the first subproblem in that it includes the inequality constraint on the sum of dual variables. To remove this, we change the inequality constraint into two equality constraints and obtain ML1 v SVM.
We consider training the ML1 SVM and ML1 v SVM optimizing the first and the second subprograms, alternatingly. Because the ML1 SVM and ML1 v SVM are very similar to the L1 SVM, we discuss the training method based on Sequential Minimum Optimization (SMO) fused with Newton’s method [31]. We also show the convergence proof of the proposed training methods. By computer experiments using polynomial kernels, in addition to RBF kernels, we compare the ML1 SVM and ML1 v SVM with other classifiers including the L1 SVM.
In Section 2, we summarize the architectures of L1 SVM and the MCM. In Section 3, we discuss the architecture of the ML1 SVM and derive the dual form and the optimality conditions of the solution. Then, we discuss the architecture of the ML1 v SVM. In Section 4, we discuss the training method that is an extension of SMO fused with Newton’s method [31] and the working set selection method. We also show the convergence proof of the proposed method. In Section 5, we evaluate the generalization ability of the ML1 SVM and other SVM-like classifiers using two-class and multiclass problems.

2. L1 Support Vector Machines and Minimal Complexity Machines

In this section, we briefly explain the architectures of the L1 SVM and the MCM [27]. Then we discuss the problem of non-unique solutions of the MCM and one approach to solving the problem [29].

2.1. L1 Support Vector Machines

Let M training data and their labels be { x i , y i } ( i = 1 , , M ) , where x i is an n-dimensional input vector and y i = 1 for Class 1 and 1 for Class 2. The input space is mapped into the l-dimensional feature space by the mapping function ϕ ( x ) and in the feature space the separating hyperplane is constructed. The decision boundary given by the decision function f ( x ) is given by
f ( x ) = w ϕ ( x ) + b = 0 ,
where w is the l-dimensional coefficient vector and b is the bias term.
The primal form of the L1 SVM is given by
minimize Q ( w , b , ξ ) = 1 2 w 2 + C i = 1 M ξ i
subject   to y i ( w ϕ ( x i ) + b ) + ξ i 1 , ξ i 0 for i = 1 , , M ,
where ξ = ( ξ 1 , , ξ M ) , ξ i is the slack variable for x i , and C is the margin parameter that determines the trade-off between the maximization of the margin and minimization of the classification error. Inequalities (3) are called soft-margin constraints.
The dual problem of (2) and (3) is given by
maximize Q ( α ) = i = 1 M α i 1 2 i , j = 1 M α i α j y i y j K ( x i , x j )
subject   to i = 1 M y i α i = 0 , C α i 0 for i = 1 , , M ,
where K ( x i , x j ) = ϕ ( x i ) ϕ ( x j ) , K ( x i , x j ) is the kernel function, α = ( α 1 , , α M ) , α i is the Lagrange multiplier for the ith inequality constraint, and x i associated with positive α i ( > 0 ) is called a support vector.
The decision function given by (1) is rewritten as follows:
f ( x ) = i S α i y i K ( x i , x ) + b ,
where S is the set of support vector indices.

2.2. Minimal Complexity Machines

The VC (Vapnik-Chervonenkis) dimension is a measure for estimating the generalization ability of a classifier and lowering the VC dimension leads to realizing a higher generalization ability. For an SVM-like classifier with the minimum margin δ min , the VC dimension D is bounded by [1]
D 1 + min R 2 δ min 2 , l ,
where R is the radius of the smallest hypersphere that encloses all the training data.
In training the L1 SVM, both R and l are not changed. In the LS SVM, where ξ i are replaced with ξ i 2 in (2) and the inequality constraints, with equality constraints in (3), although both R and l are not changed by training, the second term in the objective function works to minimize the square sums of y i f ( x i ) 1 . Therefore, like the LDM and ULDM, this term works to condense the margin distribution in the direction orthogonal to the separating hyperplane.
In [27], first the linear MCM, in which the classification problem is linearly separable, is derived by minimizing the VC-dimension, i.e., R / δ min in (7). In the input space, R and δ min are calculated, respectively, by
R = min i = 1 , M ( x i , 1 ) ,
δ min = min i = 1 , M ( w , b ) ( x i , 1 ) ( x i , 1 ) ,
where 1 is added to x i to make the separating hyperplane passes through the origin in the augmented space. Equation (8) shows that the minimum hypersphere is also assumed to pass through the origin in the augmented space. After some derivations, the linear MCM is derived. The nonlinear MCM is derived as follows:
minimize Q ( α , h , ξ , b ) = h + C i = 1 M ξ i
subject   to h y i j = 1 M α j K i j + b + ξ i 1 for i = 1 , M ,
where h is the upper bound of the soft-margin constraints and K i j = K ( x i , x j ) . Here, the mapping function ϕ ( x ) in (1) is [32]
ϕ ( x ) = ( K 11 , , K 1 M ) ,
and w = α . The MCM can be solved by linear programming.
Because the upper bound h in (11) is minimized in (10), the separating hyperplane is determined so that the maximum distance between the training data and the separating hyperplane is minimized. This is a similar idea to that of the LS SVM, LDM, and ULDM.
The MCM is derived based on the minimization of the VC dimension, and thus considers maximizing the minimum margin. However, the MCM does not explicitly include the term related to the margin maximization. This makes the solution non-unique and unbounded.
To make the solution unique under the condition that the extended MCM still is a linear programming problem, in [29] we proposed the minimal complexity LP SVM (MLP SVM), which fuses the MCM an LP SVM:
minimize Q ( α , h , ξ , b ) = C h h + i = 1 M ( C α | α i | + C ξ i )
subject   to h y i j = 1 M α j K i j + b + ξ i 1 for i = 1 , M ,
where C h is the positive parameter and C α = 1 . Deleting C h h in (13) and upper bound h in (14), we obtain the LP SVM. Setting C h = 1 and C α = 0 in (13), we obtain the MCM. Compared to the MCM and the LP SVM, the number of hyperparameters increases by 1.
According to the computer experiment, in general, the MLP SVM is better than the MCM and LP SVM in generalization abilities. However, it is inferior to the L1 SVM [29].

3. Minimal Complexity L1 Support Vector Machines

In this section, we discuss the ML1 SVM, which consists of two optimization subprograms and derive the KKT conditions that the optimal solution of the ML1 SVM must satisfy. Then, we discuss a variant of the ML1 SVM, whose two subprograms are more similar.

3.1. Architecture

The LDM and ULDM maximize the average margin and minimize the average margin variance and LS SVM makes data condense around the minimum margin. The idea of the MCM to minimize the VC dimension results in minimizing the maximum distance of the data from the separating hyperplane. Therefore, these classifiers have the idea in common: condense data as near as possible to the separating hyperplane.
In [29], we proposed fusing the MCM and the LP SVM. Similar to this idea, here we fuse the MCM given by (10) and (11) and the L1 SVM given by (2) and (3):
minimize Q ( w , b , h , ξ ) = C h h + 1 2 w 2 + C i = 1 M ξ i
subject   to y i ( w ϕ ( x i ) + b ) + ξ i 1 , ξ i 0 for i = 1 , , M ,
h y i ( w ϕ ( x i ) + b ) for i = 1 , , M ,
h 1 ,
where ξ = ( ξ 1 , , ξ M ) , C h is the parameter to control the volume that the training data occupy, and h is the upper bound of the constraints. The upper bound defined by (11) is redefined by (17) and (18), which exclude ξ i . This makes the KKT conditions for the upper bound simpler. We call (17) the upper bound constraints and the above classifier minimum complexity L1 SVM (ML1 SVM). The right-hand side of (17) shows the margin with the minimum margin normalized to 1 if the solution is obtained. Therefore, because h is the upper bound of the margins and h is minimized in (15), the ML1 SVM maximizes the minimum margin and minimizes the maximum margin simultaneously.
If we use (12), we can directly solve (15) to (18). However, sparsity of the solution will be lost. Therefore, in a way similar to solving the L1 SVM, we solve the dual problem of (15) to (18).
Introducing the nonnegative Lagrange multipliers α i , β i , and η , we obtain
Q ( w , b , h , ξ , α , β , η ) = C h h + 1 2 w 2 + C i = 1 M ξ i i = 1 M α i y i ( w ϕ ( x i ) + b ) 1 + ξ i i = 1 M β i ξ i i = 1 M α M + i h y i ( w ϕ ( x i ) + b ) ( h 1 ) η ,
where α = ( α 1 , , α M , α M + 1 , , α 2 M ) and β = ( β 1 , , β M ) .
For the optimal solution, the following KKT conditions are satisfied:
Q ( w , b , h , ξ , α , β , η ) w = 0 ,
Q ( w , b , h , ξ , α , β , η ) h = 0 ,
Q ( w , b , h , ξ , α , β , η ) b = 0 ,
Q ( w , b , h , ξ , α , β , η ) ξ = 0 ,
α i ( y i ( w ϕ ( x i ) + b ) 1 + ξ i ) = 0 α i 0 for i = 1 , , M ,
α M + i h y i ( w ϕ ( x i ) + b ) = 0 , α M + i 0 for i = 1 , , M ,
β i ξ i = 0 , β i 0 , ξ i 0 for i = 1 , , M ,
( h 1 ) η = 0 , h 1 , η 0 ,
where 0 is the zero vector whose elements are zero. Equations (24) to (27) are called KKT complementarity conditions.
Substituting (19) into (20) to (23) gives, respectively,
w = i = 1 M ( α i α M + i ) y i ϕ ( x i ) ,
i = 1 M α M + i = C h η ,
i = 1 M ( α i α M + i ) y i = 0 ,
α i + β i = C for i = 1 , , M .
Substituting (28) to (31) into (19), we obtain the following dual problem:
maximize Q ( α ) = i = 1 M ( α i α M + i ) 1 2 i , j = 1 M ( α i α M + i ) ( α j α M + j ) y i y j K ( x i , x j )
subject   to i = 1 M y i ( α i α M + i ) = 0 ,
C h i = 1 M α M + i ,
C α i 0 , C h α M + i 0 for i = 1 , , M .
The dual L1 SVM given by (4) and (5) is obtained by deleting variables α M + i and C h from the above optimization problem.
For the solution of (32) to (35), x i associated with positive α i or α M + j is a support vector. However, from (28), the decision function is determined by the support vectors that satisfy α i α M + i 0 .
We consider decomposing the above optimization problem into two subproblems: (1) optimizing α i ( i = 1 , , M ) while fixing α M + i ( i = 1 , , M ) and (2) optimizing α M + i ( i = 1 , , M ) while fixing α i ( i = 1 , , M ) . To make this possible, we eliminate the interference between α i and α M + i in (33) by
i = 1 M y i α i = 0 , i = 1 M y i α M + i = 0 .
Then the optimization problem given by (32) to (35) is decomposed into the following two subproblems:
Subproblem 1: Optimization of α i
maximize Q ( α 0 ) = i = 1 M ( α i α M + i ) 1 2 i , j = 1 M ( α i α M + i ) ( α j α M + j ) y i y j K ( x i , x j )
subject   to i = 1 M y i α i = 0 ,
C α i 0 for i = 1 , , M ,
where α 0 = ( α 1 , , α M ) .
Subproblem 2: Optimization of α M + i
maximize Q ( α M ) = i = 1 M ( α i α M + i ) 1 2 i , j = 1 M ( α i α M + i ) ( α j α M + j ) y i y j K ( x i , x j )
subject   to i = 1 M y i α M + i = 0 ,
C h i = 1 M α M + i ,
C h > α M + i 0 for i = 1 , , M ,
where α M = ( α M + 1 , , α 2 M ) . Here we must notice that α M + i C h . If α M + i = C h , from (41), at least
j = 1 , , M , y j y i α M + j = C h
is satisfied. This contradicts (42).
We solve Subproblems 1 and 2 alternatingly until the solution converges. Subproblem 1 is very similar to the L1 SVM and can be solved by the SMO (Sequential minimal optimization). Subproblem 2, which includes the constraint (42) can also be solved by a slight modification of the SMO.

3.2. KKT Conditions

To check the convergence of Subproblems 1 and 2, we use the KKT complementarity conditions (24) to (27). However, variables h and b, which are included in the KKT conditions, are excluded from the dual problem given by (32) to (35). Therefore, as with the L1 SVM [33], to make an accurate convergence test, the exact KKT conditions that do not include h and b need to be derived.
We rewrite (24) as follows:
α i ( y i b y i F i + ξ i ) = 0 for i = 1 , , M ,
where
F i = y i j = 1 M y j ( α j α M + j ) K ( x i , x j ) .
We can classify the conditions of (45) into the following three cases:
  • α i = 0 .
    Because y i b y i F i + ξ i 0 and ξ i = 0 ,
    y i b y i F i , i . e . , b F i if y i = 1 ; b F i if y i = 1 .
  • C > α i > 0 .
    Because β i > 0 , ξ i = 0 is satisfied. Therefore,
    b = F i .
  • α i = C .
    Because β i = 0 , ξ i 0 is satisfied. Therefore,
    y i b y i F i or b F i if y i = 1 ; b F i if y i = 1 .
Then the KKT conditions for (45) are simplified as follows:
F i ¯ b F i ˜ for i = 1 , , M ,
where
F i ˜ = F i if ( y i = 1 , α i = 0 ) , C > α i > 0 or ( y i = 1 , α i = C ) ,
F i ¯ = F i if ( y i = 1 , α i = 0 ) , C > α i > 0 or ( y i = 1 , α i = C ) .
To detect the violating variables, we define b low and b up as follows:
b low = max i F i ˜ , b up = min i F i ¯ .
If the KKT conditions are satisfied,
b up b low .
The bias term is estimated to be
b e = 1 2 ( b up + b low ) ,
where b e is the estimate of the bias term using (24).
Likewise, using (46), (25) becomes
α M + i h + y i F i y i b 1 = 0 for i = 1 , , M .
Then the conditions for (25) are rewritten as follows:
  • α M + i = 0 .
    From
    h + y i F i y i b 1 0 ,
    we have
    y i b h y i F i 1 , i . e . , b h F i 1 if y i = 1 ; b + h F i + 1 if y i = 1 .
  • C h > α M + i > 0 .
    y i b h = y i F i 1 , i . e . , b h = F i 1 if y i = 1 ; b + h = F i + 1 if y i = 1 .
The KKT conditions for (25) are simplified as follows:
if y i = 1 , F i ¯ + 1 b F i ˜ + 1 , if y i = 1 , F i ¯ + 1 b + F i ˜ + 1 for i = 1 , , M ,
where b = b + h , b + = b h , and
F i ˜ = F i + 1 if y i = 1 ,
F i ¯ = F i + 1 if y i = 1 , C h > α M + i > 0 ,
F i ˜ + = F i 1 if y i = 1 , C h > α M + i > 0 ,
F i ¯ + = F i 1 if y i = 1 .
To detect the violating variables, we define b low , b low + , b up , and b up + as follows:
b low = max i F i ˜ , b low + = max i F i ˜ + , b up = min i F i ¯ , b up + = min i F i ¯ + .
In general, the distributions of Classes 1 and 2 data are different. Therefore, the upper bounds of h for Classes 1 and 2 are different. This may mean that either of b up ( F i ¯ ) and b low + ( F i ˜ + ) may not exist. However, because of (41), both classes have at least one positive α M + i each, and because of (53), the values of h for both classes can be different. This happens because we separate (33) into two equations as in (36). Then, if the KKT conditions are satisfied, both of the following inequalities hold
b up b low , b up + b low + .
From the first inequality, the estimate of h, h e for Class 2, is given by
h e = b e + 1 2 ( b up + b low ) .
From the second inequality, the estimate of h, h e + for Class 1, is given by
h e + = b e 1 2 ( b up + + b low + ) .

3.3. Variant of Minimal Complexity Support Vector Machines

Subproblem 2 of the ML1 SVM is different from Subproblem 1 in that the former includes the inequality constraint given by (42). This makes the solution process more complicated. In this section, we consider making the solution process similar to that of Subproblem 1.
Solving Subproblem 2 results in obtaining h e + and h e . We consider assigning separate variables h + and h for Classes 1 and 2 instead of a single variable h. Then the complementarity conditions for h + and h are
( h + 1 ) η + = 0 , h + 1 , η + 0 , ( h 1 ) η = 0 , h 1 , η 0 ,
where η + and η are the Lagrange multipliers associated with h + and h , respectively. To simplify Subproblem 2, we assume that η + = η = 0 . This and (41) make (41) and (42) two equality constraints. Then the optimization problem given by (40) to (43) becomes
maximize Q ( α M ) = i = 1 M α i 1 2 i , j = 1 M ( α i α M + i ) ( α j α M + j ) y i y j K ( x i , x j )
subject   to y i = 1 , i = 1 M α M + i = C h , y i = 1 , i = 1 M α M + i = C h ,
C h α M + i 0 for i = 1 , , M .
Here, (41) is not necessary because of (67). We call the above architecture ML1 v SVM.
For the solution of the ML1 SVM, the same solution is obtained by the ML1 v SVM with the C h value given by
C h = i = 1 , , M , y i = 1 α M + i = i = 1 , , M , y i = 1 α M + i .
However, the reverse is not true, namely, the solution of the ML1 v SVM may not be obtained by the ML1 SVM. As the C h value becomes large, the value of η becomes positive for the ML1 SVM, but for the ML1 v SVM, the values of α M + i are forced to become larger.

4. Training Methods

In this section we extend the training method for the L1 SVM that fuses SMO and Newton’s method [31] for training the ML1 SVM. The major part of the training method consists of calculation of corrections by Newton’s method and the working set selection method. The training method of the ML1 v SVM is similar to that of the L1 SVM. Therefore, we only explain the difference of the methods in Section 4.3.

4.1. Calculating Corrections by Newton’s Method

In this subsection, we discuss the corrections by Newton’s method for two subprograms.

4.1.1. Subprogram 1

First we discuss optimization of Subproblem 1. We optimize the variables α i in the working set { α i | i W , i { 1 , , M } } , where W includes working set indices, fixing the remaining variables, by
maximize Q ( α W ) = i W α i
1 2 i W , j = 1 , , M ( α i α M + i ) ( α j α M + j ) y i y j K ( x i , x j )
subject   to i W y i α i = i W y i α i ,
C α i 0 , for i W .
Here α W = ( , α i , ) , i W .
We can solve the above optimization problem by the method discussed in [31]. We select α s in the working set and solve (71) for α s :
α s = i s , i = 1 M y s y i α i .
Substituting (73) into (70), we eliminate the equality constraint. Let α W = ( , α i , ) ( i s , i W ) . Now because Q ( α W ) is quadratic, we can express the change of Q ( α W ) , Δ Q ( α W ) , as a function of the change of α W , Δ α W , by
Δ Q ( α W ) = Q ( α W ) α W Δ α W + 1 2 Δ α W 2 Q ( α W ) α W 2 Δ α W .
Then, neglecting the bounds, Δ Q ( α W ) has the maximum at
Δ α W = 2 Q ( α ) α W 2 1 Q ( α W ) α W ,
where
Q ( α W ) α i = Q ( α W ) α i + Q ( α W ) α s α s α i = y i ( F i F s ) for i W ,
2 Q ( α W ) α i α j = 2 Q ( α W ) α i α j + 2 Q ( α W ) α s α j α s α i + 2 Q ( α W ) α i α s α s α j + 2 Q ( α W ) α s α s α s α i α s α j = y i y j K ( x i , x j ) + K ( x i , x s ) + K ( x s , x j ) K ( x s , x s ) for i , j W .
Here, the partial derivative of Q with substitution of (73) is calculated by the chain rule without substitution.
We assume that 2 Q ( α ) / α W 2 is positive definite. If not, we avoid matrix singularity adding a small value to the diagonal elements.
Then from (73) and (75), we obtain the correction of α s :
Δ α s = i W y s y i Δ α i .
For α i ( i W ) , if
α i = 0 , Δ α i < 0 or α i = C , Δ α i > 0 ,
we delete these variables from the working set and repeat the procedure for the reduced working set. Let Δ α i be the maximum or minimum correction of α i that is within the bounds. Then if α i + Δ α i < 0 , Δ α i = α i . Moreover, if α i + Δ α i > C , Δ α i = C α i . Otherwise Δ α i = Δ α i . Then we calculate
r = min i W Δ α i Δ α i ,
where r ( 0 < r 1 ) is the scaling factor.
The corrections of the variables in the working set are given by
α W new = α W old + r Δ α W .

4.1.2. Subprogram 2

We optimize the variables α M + i in the working set { α M + i | i W , i { 1 , , M } } , fixing the remaining variables, by
maximize Q ( α M ) = i W α M + i
1 2 i W , j = 1 , , M ( α i α M + i ) ( α j α M + j ) y i y j K ( x i , x j )
subject   to i W y i α M + i = i W y i α M + i ,
C h i = 1 M α M + i = η 0 ,
C h > α M + i 0 for i W .
We select α M + s in the working set W and solve (83) for α M + s :
α M + s = i = 1 , i s M y s y i α M + i .
Then similar to Subproblem 1, we substitute (86) into (82) and eliminate α M + s . The partial derivatives in (75) are as follows:
Q ( α W ) α M + i = Q ( α W ) α M + i + Q ( α W ) α M + s α M + s α M + i = y i ( F i F s ) for i W ,
2 Q ( α W ) α M + i α M + j = 2 Q ( α W ) α M + i α M + j + 2 Q ( α W ) α M + s α M + j α M + s α M + i + 2 Q ( α W ) α M + i α M + s α M + s α M + j = y i y j ( K ( x i , x j ) + K ( x i , x s ) + K ( x s , x j ) K ( x s , x s ) ) for i , j W .
From (86), we obtain the correction of α M + s :
Δ α M + s = i W y s y i Δ α M + i .
Now we consider the constraint (84). If η = 0 , the sum of corrections needs to be zero. This is achieved if
y s = y i for i W .
Namely, we select the working set from the same class.
For α M + i ( i W ) , if
α M + i = 0 , Δ α M + i < 0 ,
we delete these variables from the working set and repeat the procedure for the reduced working set. Let Δ α M + i be the maximum or minimum correction of α M + i that is within the bounds. Then if α M + i + Δ α M + i < 0 , Δ α M + i = α M + i . Otherwise Δ α M + i = Δ α M + i . Then we calculate
r = min i W Δ α M + i Δ α M + i ,
where r ( 0 < r 1 ) is the scaling factor.
If r i W Δ α M + i > η > 0 , we further calculate
r = η r i W Δ α M + i .
Then the corrections of the variables in the working set are given by
α W new = α W old + r Δ α W ,
where r is replaced by r r if r i W α M + i > η > 0 .

4.2. Working Set Selection

At each iteration of training, we optimize either Subproblem 1 ( α i ) or Subproblem 2 ( α M + i ). To do so, we define the most violating indices as follows:
V 1 = b low b up , V 2 = b low + b up + , V 3 = b low b up .
We consider that training is converged when
max i V i τ ,
where τ is a small positive parameter.
If (96) is not satisfied, we correct variables associated with V i where i is determined by
arg max i V i .
According to the conditions of (96) and (97), we determine the variable pair as follows:
  • If V 1 is the maximum in (97), we optimize Subproblem 1 ( α i ). Let the variable pair associated with b up and b low be α i min and α i max , respectively.
  • If η = 0 and either V 2 or V 3 is the maximum in (97), or if η 0 and either V 2 or V 3 exceeds τ but not both, we optimize Subproblem 2 ( α M + i belonging to either Class 1 or 2). Let the variable pair associated with b up k and b low k ( k is either + or ) be α M + i min k and α M + i max k , respectively.
  • If η 0 and both V 2 and V 3 exceed τ , we optimize Subproblem 2 ( α M + i selected from Classes 1 and 2). Let the variable pair be α M + i min and α M + i max + . This is to make the selected variables correctable as will be shown in Section 4.4.2.
For the ML1 v SVM, η = 0 . Therefore, Case (3) is not necessary.
Because the working set selection strategies for α i and α M + i are the same, in the following we only discuss the strategy for α i .
In the first order SMO, at each iteration step, α i min and α i max that violate the KKT conditions the most are selected for optimization. This guarantees the convergence of the first order SMO [33]. In the second order SMO, the pair of variables that maximize the objective function are selected. However, to reduce computational burden, fixing α i min , the variable that maximizes the objective function is searched [34]:
i 2 nd = arg max i V KKT Δ Q ( α i , α i min ) ,
where V KKT is the set of indices that violate the KKT conditions:
V KKT = { i | b up < F i ˜ τ or b low < F i ¯ + τ } for i { 1 , , M } } .
We call the pair of variables that are determined either by the first or the second order SMO, SMO variables.
Because the second order SMO accelerates training for a large C value [35], in the following we use the second order SMO. However, for a substantially large C value, training speed of the second order SMO slows down because variables need to be updated many times to reach to the optimal values. To speed up convergence in such a situation, in addition to the SMO variables, we add variables that are selected in the previous steps as SMO variables, into the working set.
We consider that a loop is detected when at least one of the current SMO variables has already appeared as an SMO variable at a previous step. When a loop is detected, we pick up the loop variables that are the SMO variables in the loop. To avoid obtaining an infeasible solution by adding loop variables to the working set, we restrict loop variables to be unbounded support vectors, i.e., α i C (This happens only for Subprogram 1).
Because we optimize Subprograms 1 and 2 alternatingly, loop variables may include those for Subprograms 1 and 2. Therefore, in optimizing Subprogram 1, we need to exclude the loop variables for Subproblem 2; and vice versa. In addition, in optimizing Subproblem 2 with η = 0 , we need to exclude variables belonging to the unoptimized class, in addition to those for Subproblem 1.
If no loop is detected, the working set includes only the SMO variables. If a loop is detected, the working set consists of the SMO variables and the loop variables. For the detailed procedure, please refer to [31].

4.3. Training Procedure of ML1 SVM

In the following we show the training procedure of the ML1 SVM.
  • (Initialization) Select α i and α j in the opposite classes and set α i = α j = C , α k = 0 , k i , j , k = 1 , , M , α M + i = α M + j = a C h , and α M + k = 0 , k i , j , k = 1 , , M , where a 0.5 .
  • (Corrections) If Pr1 (Program 1), calculate partial derivatives (76) and (77) and calculate corrections by (75). Then, modify the variables by (81). Else, if Pr2, calculate partial derivatives (87) and (88) and calculate corrections by (75). Then, modify the variables by (94).
  • (Convergence Check) Update F i and calculate b up , b low , b up + , b low + , b up , and b low . If (96) is satisfied, stop training. Otherwise if V 1 is the maximum, select Pr1, otherwise, Pr2. Calculate the SMO variables.
  • (Loop detection and working set selection) Do loop detection and working set selection shown in the previous section and go to Step 2.
In Subproblem 2 of the ML1 v SVM, data for Class 1 and Class 2 can be optimized separately. Therefore, because the data that are optimized belong to the same class, the sum of corrections is zero. Thus, in Step 1, a = 1 . In Step 3, if Pr2 is optimized, the variables associated with V 2 or V 3 are optimized, not both.

4.4. Convergence Proof

Convergence of the first order SMO for the L1 SVM is proved in [33]. Similarly, we can prove that the training procedure discussed in Section 4.3 converges to the unique solution. In the following we prove the convergence for the ML1 SVM. The proof for the ML1 v SVM is evident from the discussion.
Subprograms 1 and 2 are quadratic programming problems and thus have unique maximum solutions. Therefore, it is sufficient to prove that the objective function increases by the first order SMO. For the second order SMO, the increase of the objective function is also guaranteed because the variable pair that gives the largest increase of the objective function is selected. By combining SMO with Newton’s method, if some variables are not correctable, they are deleted from the working set. By this method, in the worst case, only the SMO variables remain in the working set. Therefore, we only need to show that the first order SMO converges for the ML1 SVM.

4.4.1. Convergence Proof for Subprogram 1

From (48), α i max satisfies
C > α i max 0 for y i max = 1 , C α i max > 0 for y i max = 1 .
Moreover, from (49), α i min satisfies
C α i min > 0 for y i min = 1 , C > α i min 0 for y i min = 1 .
Because the KKT conditions are not satisfied, F i max > F i min . Moreover, we set α s = α i min . Then from (76) and (77),
Q ( α i max ) α i max = y i max ( F i max F i min ) ,
2 Q ( α i max ) α i max 2 = ϕ ( x i max ) ϕ ( x i min ) 2 0 ,
where in (103), if the equality holds, we replace zero with a small negative value to avoid zero division in (75).
Then from (102), (103), and (75), the signs of the corrections Δ α i min and Δ α i max are given by
  • Δ α i max > 0 and Δ α i min < 0 for y i min = y i max = 1 ,
  • Δ α i max < 0 and Δ α i min > 0 for y i min = y i max = 1 ,
  • Δ α i max < 0 and Δ α i min < 0 for y i min = 1 , y i max = 1 ,
  • Δ α i max > 0 and Δ α i min > 0 for y i min = 1 , y i max = 1 .
From (100) and (101), the above corrections are all possible. For instance, in (1) Δ α i max > 0 and C > α i max 0 for y i max = 1 . Therefore, C α i max + Δ α i max > 0 .
Because the corrections are not zero, the objective function increases.

4.4.2. Convergence Proof for Subprogram 2

From (57) and (58), α M + i min and α M + i max satisfy
C h > α M + i min > 0 , C h > α M + i max 0 ,
respectively. Likewise, from (59) and (60), α M + i min + and α M + i max + satisfy
C h > α M + i min + 0 , C h > α M + i max + > 0 ,
respectively.
We set α M + s = α M + i min k . Then from (87) and (88),
Q ( α M + i max k ) α M + i max k = y i max k ( F i max k F i min k ) , 2 Q ( α M + i max k ) ( α M + i max k ) 2 0 ,
If we correct α M + i max and α M + i min ( y i max = y i min = 1 ) , Δ α M + i max > 0 and Δ α M + i min < 0 . From (104), this correction is possible.
Likewise, if we correct α M + i max + and α M + i min + ( y i max + = y i min + = 1 ) , Δ α M + i max + < 0 and Δ α M + i min + > 0 . From (105), this correction is possible.
If both V 2 and V 3 are larger than τ , we select α M + i min and α M + i max + . Then from (104) and (105), these variables are correctable whether they be increased or decreased.
Because the corrections are not zero, the objective function increases.

5. Computer Experiments

First we analyze the behavior of the ML1 SVM and ML1 v SVM using a two-dimensional iris dataset and then to examine the superiority of the proposed classifiers over the L1 SVM, we evaluate their generalization abilities and computation time using two-class and multiclass problems. All the programs used in the performance evaluation were coded in Fortran and tested using Windows machines.

5.1. Analysis of Behaviors

We analyzed the behaviors of the ML1 SVM and ML1 v SVM using the iris dataset [36], which is frequently used in the literature. The iris dataset consists of 150 data with three classes and four features. We used Classes 2 and 3 and Features 3 and 4. For both classes, the first 25 data were used for training and the remaining data were used for testing. We used linear kernels: K ( x , x ) = x x .
We evaluated the h + and h values for the change of the C h value from 0.1 to 2000 fixing C to 1000. Figure 1 shows the result for the ML1 v SVM. Both h + and h values are constant for C h = 0.1 to 100 and they decrease as the C h value is increased. For the ML1 SVM, the h + and h values are constant for the change of C h and are the same as those of the ML1 SVM with C h = 0.1 to 100. For C h = 2000 , η = 1960.00 . Thus, y i = 1 , i = 1 , , M α M + i = 20.00 . For C h = 10,000, η = 9732.75 . Thus, y i = 1 , i = 1 , , M α M + i = 133.63 . This means that C h value is too small to obtain the solution comparable to that of the ML1 v SVM. Therefore, the ML1 SVM is insensitive to the C h value compared to the ML1 v SVM.
Figure 2 shows the h + and h values for C = 10 . For the ML1 v SVM, the h + and h values become smaller than 1 for C h larger than 100. This means that the C h value is so large that the solution is no longer valid. However, for the ML1 SVM, the h + and h values are larger than 1 and thus valid solutions are obtained for C h = 0.1 to 2000. This is because y i = 1 , i = 1 , , M α M + i = 58.96 even at C h = 2000 .
Figure 3 shows the decision boundary of the ML1 v SVM for C = 1000 and C h = 0.1 , 1000. For C h = 0.1 , the decision boundary is almost parallel to the x 1 axis. However, for C h = 1000 , the decision boundary rotates in the clockwise direction to make the data more condensed to the decision boundary. The accuracy for the test data is 92% for C h = 0.1 and is increased to 94% for C h = 1000 .
From the above experiment we confirmed that the solutions of the ML1 SVM and the ML1 v SVM are the same for small C h values but for large C h values both are different and in extreme cases the solution of the ML1 v SVM may be infeasible.

5.2. Performance Comparison

In this section, we compare the generalization performance of the ML1 SVM and ML1 v SVM with the L1 SVM, MLP SVM [29], LS SVM, and ULDM [20] using two-class and multiclass problems. Our main goal is to show that the generalization abilities of the ML1 SVM and ML1 v SVM are better than the generalization ability of the L1 SVM.

5.2.1. Comparison Conditions

We determined the hyperparameter values using the training data by fivefold cross-validation, trained the classifier with the determined hyperparameter values, and evaluated the accuracy for the test data.
We trained the ML1 SVM, ML1 v SVM, and L1 SVM by SMO combined with Newton’s method. We trained the MLP SVM by the simplex method and the LS SVM and ULDM by matrix inversion.
We used RBF kernels:
K ( x , x ) = exp ( γ x x 2 / m ) ,
where γ is the parameter to control the spread of the radius and m is the number of inputs to normalize the kernel, and polynomial kernels including linear kernels:
K ( x , x ) = ( P + x x ) d ,
where P = 0 and d = 1 for linear kernels and P = 1 and d = 2 , 3 , for polynomial kernels.
In cross-validation, we selected the γ values for RBF kernels from { 0.01 , 0.1 , 0.5 , 1 , 5 , 10 , 15 , 20 , 50 , 100 , 200 } , the d values for polynomial kernels from { 1 , 2 , , 8 } , and the C and C h values from { 0.1 , 1 , 10 , 50 , 100 , 500 , 1000 , 2000 } . For the ULDM, the C value was selected from { 0.1 , 1 , 10 , 100 , 1000 , 10 4 , 10 6 , 10 8 } [21]. The value of τ in (99) was set to 0.01 for the ML1 SVM, ML1 v SVM, and L1 SVM.
For the L1 SVM, LS SVM, and ULDM, we determined the γ ( d ) and C values by grid search. For the ML1 SVM, ML1 v SVM, and MLP SVM, we need to determined the value of C h in addition to γ ( d ) and C values. (For the MLP SVM we evaluated the performance using only RBF kernels). To shorten computation time in such a situation, first we determined the γ ( d ) and C values with C h = 1 ( C h = 0.1 for the MLP SVM) by grid search and then we determined the C h value by line search fixing the γ ( d ) and C values with the determined values.
After model selection, we trained the classifier with the determined hyperparameter values and calculated the accuracy for the test data. For two-class problems, which have multiple sets of training and test pairs, we calculated the average accuracies and their standard deviations, and performed Welch’s t test with the confidence level of 5%.

5.2.2. Two-Class Problems

Table 1 lists the two-class problems used [37], which were generated using the datasets from the UC Irvine Machine Learning Repository [38]. In the table, the numbers of input variables, training data, test data, and training and test data pair sets are listed. The table also includes the maximum average prior probability shown in the accuracy (%). The numeral in parentheses shows that for the test data. According to the prior probabilities, the class data are relatively well balanced and there are not much differences between training and test data.
Table 2 shows the evaluation results using RBF kernels. In the table, for each classifier and each classification problem, the average accuracy and the standard deviation are shown. For each problem the best average accuracy is shown in bold and the worst, underlined. The “+” and “−” symbols at the accuracy show that the ML1 SVM is statistically better and worse than the classifier associated with the attached symbol, respectively. For instance, the ML1 SVM is statistically better than the ULDM for the flare-solar problem. The “Average” row shows the average accuracy of the 13 problems for each classifier and “B/S/W” denotes the number of times that the associated classifier show the best, the second best, and the worst accuracy. The “W/T/L” row denotes the number of times that the ML1 SVM is statistically better than, comparable to, and worse than the associated classifier.
According to the “W/T/L” row, the ML1 SVM is statistically better than the MLP SVM but is slightly inferior to ULDM. It is comparable to the remaining classifiers. From the “Average” measure, the ULDM is the best, the ML1 v SVM, the second, the L1 SVM and the ML1 SVM, the third. From the “B/S/W” measure, the ULDM is the best and the LS SVM is the second best.
Table 3 shows the results for polynomial kernels. For each problem the best average accuracy is shown in bold and the worst, underlined. We do not include the MLP SVM for comparison. From the table, the ML1 SVM is statistically comparable to the L1 SVM, slightly better than ML1 v SVM, and better than the LS SVM and ULDM. From the average accuracies, the L1 SVM is the best, ML1 SVM, the second best, and the ULDM, the worst.
The ML1 v SVM is statistically inferior to ML1 SVM for the cancer and splice problems. We study why this happened. For the second file of the breast-cancer problem, the accuracy for the test data is 67.53% by the ML1 v SVM, which is by 7.79% lower than by the ML1 SVM. The parameter values are selected as d = 2 , C = 1 , and C h = 10 . For the C h value higher than or equal to 50, the accuracy is 75.32%, which is the same as that by ML1 SVM. Therefore, model selection did not work properly. For the 17th file of the splice problem, the accuracy for the ML1 v SVM is 83.68% with d = 1 , and C = C h = 1 . This is caused by C h = 1 when the d and C values are determined by grid search. If C h = 0.1 , by model selection d = 2 , C = C h = 0.1 are obtained, and the accuracy is 87.36%, which is better than 87.22% by the ML1 SVM. Therefore, in this case also, the low accuracy was obtained by the problem of model selection.
The ULDM performs worse than the ML1 SVM for banana and thyroid problems. For the banana problem by the ULDM, the average accuracy is by 6.66% lower than by the ML1 SVM. This was caused by mal-selection of the parameter ranges. By setting C = { 10 8 , 10 10 , 10 12 , 10 14 , 10 16 } and d = { 6 , 7 , 8 , 9 , 10 , 11 , 12 , 13 , 14 , 15 } , the average accuracy and standard deviation are 88.35 ± 1.10%, which are statistically comparable to those by the ML1 SVM. However, for the thyroid problem, the change of the parameter ranges does not improve the average accuracy much.
According to the experiment for the two-class problems, generally the accuracies using the RBF kernels are better than those using polynomial kernels, but for both RBF and polynomial kernels, the ML1 SVM, ML1 v SVM, and L1 SVM perform well, while the LS SVM and ULDM do not for the polynomial kernels.

5.2.3. Multiclass Problems

We used pairwise (one-vs-one) classification for multiclass problems. To resolve unclassifiable regions occurred by pairwise classification, we used fuzzy classification, introducing the membership function for each decision function [2].
Table 4 shows the ten multiclass problems used for performance evaluation. Unlike the two-class problems, for each problem there is one training dataset and one test dataset. The numeral problem [2] is to classify numerals in the Japanese license plates, and the thyroid problem [38] is a medical diagnosis problem. The blood cell problem [2] classifies white blood cells labeled according to the maturity of the growing stage. Three hiragana problems [2] are to classify hiragana characters in the Japanese license plates. The satimage problem [38] classifies lands according to the satellite images. The USPS [39] and MNIST [40,41] problems treat numeral classification and the letter problem [38] treats alphabets. Because the original training dataset of the MNIST problem is too large especially for the LS SVM and ULDM, we switched roles of the training data and the test data.
Except for the thyroid problem, the class data are relatively well balanced. For the thyroid data, almost all data belong to one class. Moreover, the classification accuracy of a classifier smaller than 92.41% is meaningless.
Table 5 lists the accuracies using the RBF kernels for the test data. For each problem, the best accuracy is shown in bold, and the worst, underlined. For the MLP SVM, the accuracies for the thyroid, MNIST, and letter problems were not available. The “Average” row shows the average accuracy of the associated classifier for the ten problems and “B/S/W” shows the numbers of times that the best, the second best, and the worst accuracies are obtained. Among the ten problems, the accuracies of the ML1 SVM and ML1 v SVM are better than or equal to those of the L1 SVM for nine and eight problems, respectively. In addition, the best average accuracy is obtained for the ML1 v SVM, the second best, the ML1 SVM, and the third best, the L1 SVM. This is very different from the two-class problems where the ML1 SVM and ML1 v SVM are comparable to the L1 SVM.
Table 6 shows the accuracies of the test data using polynomial kernels. For each problem the best accuracy is shown in bold and the worst, underlined. From the average accuracy, the ML1 v SVM performs best, LS SVM performs the second best, and the L1 SVM and ULDM the worst. The difference between the LS SVM and ML1 SVM is very small. Improvement of the ML1 SVM and ML1 v SVM over the L1 SVM is larger than that using the RBF kernels. For all the 10 problems, they are better than or equal to the L1 SVM. However, as seen from the Average rows in Table 5 and Table 6, the accuracies using polynomial kernels are in general worse than those using RBF kernels. The reason for this is not clear but the ranges of parameter values in model selection might not be well tuned for polynomial kernels.
In the previous experiment we evaluated RBF kernels and polynomial kernels separately, but we can choose the best kernel from RBF and polynomial kernels. If we have cross-validation results for both kernels, we can select the better one that has the higher accuracy. Table 7 shows the accuracies by cross-validation for RBF and polynomial kernels. For each problem and for each classifier, the better accuracy is shown in bold. The last row of the table shows the average accuracies. The average accuracies for the polynomial kernels are worse than those for the RBF kernels for all the classifiers. For each classifier, the number of problems that the polynomial kernels perform better or equal is one to two. Moreover, if we select RBF kernels when both average accuracies are the same, selecting kernels with the better average accuracy results in improving the accuracy for the test dataset as seen from Table 5 and Table 6. However, employing polynomial kernels in addition to RBF kernels does not improve the accuracy significantly.

5.3. Training Time Comparison

First, we examine the complexity of computation for each classifier excluding the MLP SVM. We trained the ML1 SVM, ML1 v SVM, and L1 SVM by SMO combined with Newton’s method. Therefore, the complexity of computation for the subproblem with working set size W is O ( W 3 ) . Because the ML1 SVM and ML1 v SVM solve two quadratic programming programs, each having the same number of variables, M, the complexity of computation is the same with that of the L1 SVM. Therefore, the three classifiers are considered to be trained in comparable time.
Because the matrices associated with the LS SVM and ULDM are positive definite, they can be solved by iterative methods such as stochastic gradient methods [17]. However, we trained the LS SVM and ULDM by Cholesky factorization to avoid the inaccuracy caused by insufficient convergence. Therefore, the complexity of computation of both methods is O ( M 3 ) .
The purpose of this section is to confirm that the ML1 SVM, ML1 v SVM, and L1 SVM can be trained in comparable time.
Excluding that of the MLP SVM, we compared the time for training and testing a classifier using a Windows machine with 3.2 GHz CPU and 16 GB memory. For the two-class problems, we set the parameter values with the frequently selected values and trained the classifier using a training dataset and tested the trained classifier using the associated test dataset. For the multiclass problems, we trained a classifier with the parameter values obtained by cross-validation and tested the trained classifier with the test dataset.
Because the tendency is similar we only show the results using RBF kernels. Table 8 shows the parameter values selected for the two-class and multiclass problems. In the table Thyroid (m) denotes the multiclass thyroid problem. For each problem, the γ values in bold, the γ and C values in bold, and γ , C, and C h values in bold show that they appear more than once.
From the table, it is clear that the ML1 SVM, ML1 v SVM, and L1 SVM selected the same γ and C values frequently and the ML1 SVM and ML1 v SVM selected the same γ , C, and C h values 10 times out of 23 problems.
Table 9 sows the CPU time for training and test with the optimized parameter values listed in Table 8. For each problem the shortest CPU time is shown in bold and the longest, underlined. The CPU time for the two-class problem is that per file. For the multiclass problems, we used fuzzy pairwise classification. Therefore, the training time includes that for determining n ( n 1 ) / 2 decision functions where n is the number of classes. For example, for the letter problem, 329 decision functions need to be determined. The last row of the table shows the numbers of times that the associated classifier are the fastest (B), the second fastest (S), and the slowest (W).
From the table, the ML1 SVM and ML1 v SVM show comparable computation time. Moreover, except for the hiragana-50, hiragana-13, hiragana-105, USPS, MNIST, and letter problems, computation time for the ML1 SVM and ML1 v SVM is comparable to that of the L1 SVM. For these problems, the ML1 SVM and ML1 v SVM are much slower than L1 SVM. Analyzing the convergence process, we found that for these problems, monotonicity of the objective function value was sometimes violated. To improve convergence, we need to clarify why it happens and to find a way to speed up training in such a situation. However we leave this problem in the future study.

5.4. Discussions

As discussed in Section 3.3, the ML1 SVM and ML1 v SVM are equivalent for a small C h value but for a large C h value they are different. The computer experiments in Section 5.1 also revealed that the ML1 SVM is insensitive to the change of a C h value. However, the difference of the generalization performance between the ML1 SVM and ML1 v SVM is not large for the two-class and multiclass problems.
The execution time for the ML1 and ML1 v SVM was sometimes longer than that for the L1 SVM. This will cause problem in model selection. While we leave the discussions of speeding up training, for the model selection, this problem may be alleviated for the ML1 v SVM. If h + < 1 or h < 1 is satisfied, the solution is infeasible. Therefore, we can skip cross-validation at the current C h value and the larger ones.
We used line search to speed up cross-validation. If grid search was used, in fivefold cross-validation we needed to train the ML1 SVM or ML1 v SVM 3520 ( 11 × 8 × 8 × 5 ) times, instead of 480 ( ( 11 × 8 + 8 ) × 5 ) times. The speed up ratio is estimated to be 7.3. We evaluated the difference between the grid search and line search for the heart problem. We measured the execution time of cross-validation, training the classifier with the determined parameter values, and testing the classifier using the test data. For the ML1 SVM, the speed up ratio by line search was 35.7, and the average accuracy with the standard deviation of the grid search was 82.78 ± 3.45%, which was slightly lower. By the ML1 v SVM, the speed up ratio was 40.0, and the average accuracy with the standard deviation was 82.85 ± 3.31%, which was also lower than that by line search. Therefore, because model selection slowed down very much and the improvement of the average accuracy was not obtained at least for the heart problem, grid search will not be a good selection.
To speed up model selection of the ML1 SVM or ML1 v SVM, we may use the L1 SVM considering that the same values were selected frequently for the γ ( d ) and C values (see Table 8 for γ values). For the multiclass problems, four problems do not have the same values. To check whether the idea works, we performed model selection of C h values fixing the values of γ and C determined by model selection of the L1 SVM for the hiragana-13, hiragana-105, USPS, and letter problems. Among the four problems, the ML1 SVM and ML1 v SVM performed worse than the L1 SVM for the letter problem. Moreover, the resulting average accuracies of the ML1 SVM and ML1 v SVM for ten problems were 97.14% and 97.16%, respectively, which were lower than by the original model selection by 0.03% (see Table 5) but were still better than the accuracy of the L1 SVM. If we switch back the roles of the training and test data for the MNIST problem, the selected parameter values for the L1 SVM were the same. The accuracies for the test data were 98.55%, 98.77%, and 98.78% for the L1 SVM, ML1 SVM, and ML1 v SVM, respectively.
For the polynomial kernels, different kernel parameters were selected for six problems: the numeral (only for the ML1 v SVM), blood cell, hiragana-50, hiragana-13, hiragana-105, and USPS problems. For each problem we determined the C h value by cross-validation fixing the values of d and C determined by the L1 SVM. The accuracies for the test datasets were all better than or equal to those by the L1 SVM. The average accuracies of the ML1 SVM or ML1 v SVM for all the ten problems were 96.53% and 96.64%, respectively, which were lower than by the original model selection by 0.19% and 0.12%, respectively. For the MNIST problem with the switched training and test data, the selected parameter values for the L1 SVM were the same. Moreover, the accuracies for the test data were 98.17%, 98.23%, and 98.34% for the L1 SVM, ML1 SVM, and ML1 v SVM, respectively.

6. Conclusions

The minimal complexity machine (MCM) minimizes the VC dimension and generalizes better than the standard support vector machine (L1 SVM). However, according to our previous analysis, the solution of the MCM is non-unique and unbounded.
In this paper, to solve the problem of the MCM and to improve the generalization ability of the L1 SVM, we fused the MCM and the L1 SVM, namely, we introduced minimizing the upper bound of the absolute decision function values to the L1 SVM. This corresponds to minimizing the maximum margin. Converting the original classifier into dual one, we derived two subproblems: the first subproblem corresponds to the L1 SVM and the second subproblem corresponds to minimizing the upper bound. We further modified the second subproblem by converting the inequality constraint into two equality constraints: one for optimizing the variables associated with the positive class and the other for the negative class. We call this architecture ML1 v SVM and the original architecture, ML1 SVM.
We derived the exact KKT conditions for the first and second subproblems that exclude the bias term and the upper bound and discussed training the two subproblems alternatingly, fusing sequential minimal optimization (SMO) and Newton’s method.
According to computer experiments of the two-class problems using RBF kernels, the average accuracy of the ML1 SVM is statistically comparable to that of the ML1 v SVM and L1 SVM. Using polynomial kernels, the ML1 SVM is statistically comparable to the L1 SVM but is slightly better than ML1 v SVM.
For the multiclass problems using RBF kernels, the ML1 v SVM and ML1 SVM generalize better than the L1 SVM and the ML1 v SVM performs best, and the ML1 SVM, the second, among six classifiers tested. Using polynomial kernels, the ML1 v SVM performs best, the LS SVM the second best, ML1 SVM the third, and the L1 SVM worst.
Therefore, the idea of minimizing the VC dimension for the L1 SVM worked to improve the generalization ability of the L1 SVM.
Execution time for the ML1 SVM and ML1 v SVM is comparable to that for the L1 SVM for most of the problems tested, but in some problems, execution time is much longer. In the future study, we would like to clarify the reason and propose a fast training method in such cases. Another study will be to consider robustness for outliers by the soft upper bound, instead of the hard upper bound.

Funding

This work was supported by JSPS KAKENHI Grant Number 19K04441.

Conflicts of Interest

The author declares no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

References

  1. Vapnik, V.N. Statistical Learning Theory; John Wiley & Sons: New York, NY, USA, 1998. [Google Scholar]
  2. Abe, S. Support Vector Machines for Pattern Classification, 2nd ed.; Springer: London, UK, 2010. [Google Scholar]
  3. Abe, S. Training of Support Vector Machines with Mahalanobis Kernels. In Artificial Neural Networks: Formal Models and Their Applications (ICANN 2005)—Proceedings of Fifteenth International Conference, Part II, Warsaw, Poland; Duch, W., Kacprzyk, J., Oja, E., Zadrożny, S., Eds.; Springer-Verlag: Berlin, Germany, 2005; pp. 571–576. [Google Scholar]
  4. Wang, D.; Yeung, D.S.; Tsang, E.C.C. Weighted Mahalanobis Distance Kernels for Support Vector Machines. IEEE Trans. Neural Netw. 2007, 18, 1453–1462. [Google Scholar] [CrossRef]
  5. Shen, C.; Kim, J.; Wang, L. Scalable Large-Margin Mahalanobis Distance Metric Learning. IEEE Trans. Neural Netw. 2010, 21, 1524–1530. [Google Scholar] [CrossRef] [Green Version]
  6. Liang, X.; Ni, Z. Hyperellipsoidal Statistical Classifications in a Reproducing Kernel Hilbert Space. IEEE Trans. Neural Netw. 2011, 22, 968–975. [Google Scholar] [CrossRef]
  7. Fauvel, M.; Chanussot, J.; Benediktsson, J.; Villa, A. Parsimonious Mahalanobis kernel for the classification of high dimensional data. Pattern Recognit. 2013, 46, 845–854. [Google Scholar] [CrossRef] [Green Version]
  8. Reitmaier, T.; Sick, B. The responsibility weighted Mahalanobis kernel for semi-supervised training of support vector machines for classification. Inf. Sci. 2015, 323, 179–198. [Google Scholar] [CrossRef] [Green Version]
  9. Jiang, H.; Ching, W.K.; Yiu, K.F.C.; Qiu, Y. Stationary Mahalanobis kernel SVM for credit risk evaluation. Appl. Soft Comput. 2018, 71, 407–417. [Google Scholar] [CrossRef]
  10. Sun, G.; Rong, X.; Zhang, A.; Huang, H.; Rong, J.; Zhang, X. Multi-Scale Mahalanobis Kernel-Based Support Vector Machine for Classification of High-Resolution Remote Sensing Images. Cogn. Comput. 2019. [Google Scholar] [CrossRef]
  11. Lanckriet, G.R.G.; Cristianini, N.; Bartlett, P.; Ghaoui, L.E.; Jordan, M.I. Learning the Kernel Matrix with Semidefinite Programming. J. Mach. Learn. Res. 2004, 5, 27–72. [Google Scholar]
  12. Shivaswamy, P.K.; Jebara, T. Ellipsoidal Kernel Machines. In Proceedings of the Eleventh International Conference on Artificial Intelligence and Statistics (AISTATS 2007), San Juan, Puerto Rico, 21–24 March 2007. [Google Scholar]
  13. Xue, H.; Chen, S.; Yang, Q. Structural Regularized Support Vector Machine: A Framework for Structural Large Margin Classifier. IEEE Trans. Neural Netw. 2011, 22, 573–587. [Google Scholar] [CrossRef]
  14. Peng, X.; Xu, D. Twin Mahalanobis distance-based support vector machines for pattern recognition. Inf. Sci. 2012, 200, 22–37. [Google Scholar] [CrossRef]
  15. Ebrahimpour, Z.; Wan, W.; Khoojine, A.S.; Hou, L. Twin Hyper-Ellipsoidal Support Vector Machine for Binary Classification. IEEE Access 2020, 8, 87341–87353. [Google Scholar] [CrossRef]
  16. Pelckmans, K.; Suykens, J.; Moor, B.D. A Risk Minimization Principle for a Class of Parzen Estimators. In Advances in Neural Information Processing Systems 20; Platt, J., Koller, D., Singer, Y., Roweis, S., Eds.; Curran Associates, Inc.: New York, NY, USA, 2008; pp. 1137–1144. [Google Scholar]
  17. Zhang, T.; Zhou, Z.H. Large Margin Distribution Machine. In Proceedings of the Twentieth ACM SIGKDD Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 313–322. [Google Scholar]
  18. Zhu, Y.; Wu, X.; Xu, J.; Zhang, D.; Zuo, W. Radius-margin based support vector machine with LogDet regularizaron. In Proceedings of the 2015 International Conference on Machine Learning and Cybernetics (ICMLC), Guangzhou, China, 12–15 July 2015; Volume 1, pp. 277–282. [Google Scholar]
  19. Abe, S. Improving Generalization Abilities of Maximal Average Margin Classifiers. In Artificial Neural Networks in Pattern Recognition, Proceedings of the 7th IAPR TC3 Workshop (ANNPR 2016), Ulm, Germany, 28–30 September 2016; Schwenker, F., Abbas, H.M., Gayar, N.E., Trentin, E., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 29–41. [Google Scholar]
  20. Abe, S. Unconstrained Large Margin Distribution Machines. Pattern Recognit. Lett. 2017, 98, 96–102. [Google Scholar] [CrossRef] [Green Version]
  21. Abe, S. Effect of Equality Constraints to Unconstrained Large Margin Distribution Machines. In IAPR Workshop on Artificial Neural Networks in Pattern Recognition; Lecture Notes in Computer Science; Pancioni, L., Schwenker, F., Trentin, E., Eds.; Springer: Cham, Switzerland, 2018; Volume 11081, pp. 41–53. [Google Scholar]
  22. Zhang, T.; Zhou, Z. Optimal Margin Distribution Machine. IEEE Trans. Knowl. Data Eng. 2020, 32, 1143–1156. [Google Scholar] [CrossRef] [Green Version]
  23. Burges, C.J.C. A tutorial on support vector machines for pattern recognition. Data Min. Knowl. Discov. 1998, 2, 121–167. [Google Scholar] [CrossRef]
  24. Duarte, E.; Wainer, J. Empirical comparison of cross-validation and internal metrics for tuning SVM hyperparameters. Pattern Recognit. Lett. 2017, 88, 6–11. [Google Scholar] [CrossRef]
  25. Du, J.Z.; Lu, W.G.; Wu, X.H.; Dong, J.Y.; Zuo, W.M. L-SVM: A radius-margin-based SVM algorithm with LogDet regularization. Expert Syst. Appl. 2018, 102, 113–125. [Google Scholar] [CrossRef]
  26. Wu, X.; Zuo, W.; Lin, L.; Jia, W.; Zhang, D. F-SVM: Combination of Feature Transformation and SVM Learning via Convex Relaxation. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 5185–5199. [Google Scholar] [CrossRef] [Green Version]
  27. Jayadeva. Learning a hyperplane classifier by minimizing an exact bound on the VC dimension. Neurocomputing 2015, 149, 683–689. [Google Scholar] [CrossRef] [Green Version]
  28. Jayadeva; Soman, S.; Pant, H.; Sharma, M. QMCM: Minimizing Vapnik’s bound on the VC dimension. Neurocomputing 2020, 399, 352–360. [Google Scholar] [CrossRef]
  29. Abe, S. Analyzing Minimal Complexity Machines. In Proceedings of the International Joint Conference on Neural Networks, Budapest, Hungary, 14–19 July 2019. [Google Scholar]
  30. Abe, S. Minimal Complexity Support Vector Machines. In Artificial Neural Networks in Pattern Recognition; Lecture Notes in Computer Science; Schilling, F.P., Stadelmann, T., Eds.; Springer: Cham, Switzerland, 2020; Volume 12294, pp. 89–101. [Google Scholar]
  31. Abe, S. Fusing Sequential Minimal Optimization and Newton’s Method for Support Vector Training. Int. J. Mach. Learn. Cybern. 2016, 7, 345–364. [Google Scholar] [CrossRef] [Green Version]
  32. Abe, S. Sparse Least Squares Support Vector Training in the Reduced Empirical Feature Space. Pattern Anal. Appl. 2007, 10, 203–214. [Google Scholar] [CrossRef] [Green Version]
  33. Keerthi, S.S.; Gilbert, E.G. Convergence of a generalized SMO algorithm for SVM classifier design. Mach. Learn. 2002, 46, 351–360. [Google Scholar] [CrossRef] [Green Version]
  34. Fan, R.E.; Chen, P.H.; Lin, C.J. Working Set Selection Using Second Order Information for Training Support Vector Machines. J. Mach. Learn. Res. 2005, 6, 1889–1918. [Google Scholar]
  35. Barbero, A.; Dorronsoro, J.R. Faster Directions for Second Order SMO. In Artificial Neural Networks—ICANN 2010; Lecture Notes in Computer Science; Diamantaras, K., Duch, W., Iliadis, L.S., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; Volume 6353, pp. 30–39. [Google Scholar]
  36. Bezdek, J.C.; Keller, J.M.; Krishnapuram, R.; Kuncheva, L.I.; Pal, N.R. Will the real iris data please stand up? IEEE Trans. Fuzzy Syst. 1999, 7, 368–369. [Google Scholar] [CrossRef] [Green Version]
  37. Rätsch, G.; Onoda, T.; Müller, K.R. Soft Margins for AdaBoost. Mach. Learn. 2001, 42, 287–320. [Google Scholar] [CrossRef]
  38. Asuncion, A.; Newman, D.J. UCI Machine Learning Repository. 2007. Available online: https://archive.ics.uci.edu/ml/index.php (accessed on 23 October 2020).
  39. USPS Dataset. Available online: https://www.kaggle.com/bistaumanga/usps-dataset (accessed on 23 October 2020).
  40. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  41. LeCun, Y.; Cortes, C. The MNIST Database of Handwritten Digits. Available online: http://yann.lecun.com/exdb/mnist/ (accessed on 23 October 2020).
Figure 1. Upper bounds h + and h for the change of the C h value with C = 1000 .
Figure 1. Upper bounds h + and h for the change of the C h value with C = 1000 .
Computers 09 00088 g001
Figure 2. Upper bounds h + and h for the change of the C h value with C = 10 .
Figure 2. Upper bounds h + and h for the change of the C h value with C = 10 .
Computers 09 00088 g002
Figure 3. Separating hyperplanes for the different C h values with C = 1000 .
Figure 3. Separating hyperplanes for the different C h values with C = 1000 .
Computers 09 00088 g003
Table 1. Benchmark datasets for two-class problems.
Table 1. Benchmark datasets for two-class problems.
ProblemInputsTraining DataTest DataSetsPrior (%)
Banana2400490010054.62 (55.21)
Breast cancer92007710070.59 (71.19)
Diabetes846830010065.03 (65.22)
Flare-solar966640010055.25 (55.27)
German2070030010069.92 (70.18)
Heart1317010010055.53 (55.60)
Image18130010102057.40 (56.81)
Ringnorm20400700010050.27 (50.50)
Splice60100021752051.71 (52.00)
Thyroid51407510069.51 (70.25)
Titanic3150205110067.83 (67.69)
Twonorm20400700010050.52 (50.01)
Waveform21400460010066.90 (67.07)
Table 2. Accuracies of the test data for the two-class problems using RBF kernels.
Table 2. Accuracies of the test data for the two-class problems using RBF kernels.
ProblemML1 SVMML1 v SVML1 SVMMLP SVMLS SVMULDM
Banana89.18 ± 0.7089.10 ± 0.7089.17 ± 0.7289.07 ± 0.7389.17 ± 0.6689.13 ± 0.72
Cancer73.03 ± 4.4573.12 ± 4.4373.03 ± 4.5172.81 ± 4.5973.13 ± 4.6873.82 ± 4.44
Diabetes76.17 ± 2.2576.33 ± 1.9476.29 ± 1.7376.05 ± 1.7476.19 ± 2.0076.50 ± 1.94
Flare-solar66.98 ± 2.1466.99 ± 2.1666.99 ± 2.1266.62 ± 3.1066.25 + ± 1.9866.34 + ± 1.94
German75.91 ± 2.0375.97 ± 2.2175.95 ± 2.2475.63 ± 2.5776.10 ± 2.1076.15 ± 2.29
Heart82.84 ± 3.2682.96 ± 3.2582.82 ± 3.3782.52 ± 3.2782.49 ± 3.6082.70 ± 3.66
Image97.29 ± 0.4497.29 ± 0.4797.16 ± 0.4196.47 + ± 0.8797.52 ± 0.5497.15 ± 0.68
Ringnorm98.12 ± 0.3697.97 ± 1.1198.14 ± 0.3597.97 + ± 0.3798.19 ± 0.3398.16 ± 0.35
Splice89.05 ± 0.8388.99 ± 0.8388.89 ± 0.9186.71 + ± 1.2788.98 ± 0.7089.13 ± 0.60
Thyroid95.32 ± 2.4195.37 ± 2.5095.35 ± 2.4495.12 ± 2.3895.08 ± 2.5595.29 ± 2.34
Titanic77.37 ± 0.8177.40 ± 0.7977.39 ± 0.7477.41 ± 0.7777.39 ± 0.8377.40 ± 0.85
Twonorm97.36 ± 0.2897.38 ± 0.2597.38 ± 0.2697.13 + ± 0.2997.43 ± 0.2797.43 ± 0.25
Waveform89.72 ± 0.7389.67 ± 0.7589.76 ± 0.6689.39 + ± 0.5390.05 ± 0.5990.24 ± 0.50
Average (B/S/W)85.26 (1/3/1)85.27 (3/3/1)85.26 (1/2/0)84.84 (1/0/9)85.23 (3/4/3)85.34 (6/2/0)
W/T/L0/13/00/13/05/8/01/11/11/10/2
For each problem the best average accuracy is shown in bold and the worst underlined. The “+” and “−” symbolsatthe accuracy show that the ML1 SVM is statistically better and worse than the classifier associated with the attached symbol, respectively.
Table 3. Accuracies of the test data for the two-class problems using polynomial kernels.
Table 3. Accuracies of the test data for the two-class problems using polynomial kernels.
ProblemML1 SVMML1 v SVML1 SVMLS SVMULDM
Banana88.97 ± 0.6989.01 ± 0.6289.01 ± 0.6288.07 + ± 1.0082.31 + ± 2.49
Cancer72.90 ± 5.1071.66 + ± 4.8672.84 ± 5.2572.75 ± 4.6172.75 ± 4.71
Diabetes76.19 ± 1.7576.22 ± 1.6776.29 ± 1.7576.39 ± 1.9176.29 ± 1.66
Flare-solar67.30 ± 2.0167.26 ± 2.1267.19 ± 2.1766.46 + ± 1.9267.09 ± 1.97
German75.62 ± 2.1675.71 ± 2.2375.79 ± 2.2675.70 ± 2.0575.23 + ± 1.92
Heart82.77 ± 3.6282.93 ± 3.2582.85 ± 3.4683.60 ± 3.3983.22 ± 3.48
Image96.59 ± 0.5196.62 ± 0.5196.74 ± 0.4797.01 ± 0.4395.35 + ± 0.55
Ringnorm93.29 ± 1.0293.31 ± 0.9493.39 ± 0.9592.43 + ± 0.8594.71 ± 0.73
Splice87.27 ± 0.7986.00 + ± 1.4787.67 ± 0.6886.22 + ± 0.7187.62 ± 0.67
Thyroid95.09 ± 2.5894.99 ± 2.5995.04 ± 2.6891.37 + ± 3.4189.99 + ± 3.56
Titanic77.60 ± 0.7277.63 ± 0.6677.61 ± 0.6877.52 ± 0.7477.59 ± 0.74
Twonorm97.30 ± 0.4297.25 ± 0.4397.42 ± 0.3497.47 ± 0.2497.14 + ± 0.51
Waveform89.14 ± 0.7389.16 ± 0.8689.13 ± 0.7689.70 ± 0.7190.00 ± 0.62
Average (B/S/W)84.62 (3/0/2)84.44 (2/2/2)84.69 (3/7/0)84.21 (4/1/3)83.79 (2/3/5)
W/T/L2/11/00/13/05/4/45/5/3
For each problem the best average accuracy is shown in bold and the worst underlined. The “+” and “−” symbolsatthe accuracy show that the ML1 SVM is statistically better and worse than the classifier associated with the attached symbol, respectively.
Table 4. Benchmark datasets for the multiclass problems.
Table 4. Benchmark datasets for the multiclass problems.
ProblemInputsClassesTraining DataTest DataPrior (%)
Numeral121081082010.00 (10.00)
Thyroid2133772342892.47 (92.71)
Blood cell13123097310012.92 (12.90)
Hiragana-5050394610461012.90 (5.64)
Hiragana-131338837583566.29 (6.29)
Hiragana-10510538837583566.29 (6.29)
Satimage3664435200024.17 (23.50)
USPS256107291200716.38 (17.89)
MNIST7841010,00060,00011.35 (11.23)
Letter162616,00040004.05 (4.20)
Table 5. Accuracies of the test data using RBF kernels for the multiclass problems.
Table 5. Accuracies of the test data using RBF kernels for the multiclass problems.
ProblemML1 SVMML1 v SVML1 SVMMLP SVMLS SVMULDM
Numeral99.7699.7699.7699.2799.1599.39
Thyroid97.2697.2397.2695.3995.27
Blood cell93.4593.5593.1693.3694.2394.32
Hiragana-5099.1199.2299.0098.9699.4898.96
Hiragana-1399.8999.9499.7999.9099.8799.89
Hiragana-105100.00100.00100.00100.00100.00100.00
Satimage91.8591.8591.9091.1091.9592.25
USPS95.4295.3795.2795.1795.4795.42
MNIST96.9696.9596.5596.9997.03
Letter98.0398.0397.8597.8897.75
Average (B/S/W)97.17 (4/1/0)97.19 (4/1/0)97.05 (3/0/3)—(1/1/3)97.04 (3/3/1)97.03 (4/1/3)
For each problem, the best accuracy is shown in bold, and the worst underlined.
Table 6. Accuracies of the test data using polynomial kernels for the multiclass problems.
Table 6. Accuracies of the test data using polynomial kernels for the multiclass problems.
ProblemML1 SVMML1 v SVML1 SVMLS SVMULDM
Numeral99.6399.6399.6399.0299.27
Thyroid97.3897.3897.3894.6694.66
Blood cell94.3293.7792.1394.3994.55
Hiragana-5098.9299.0598.8199.2498.76
Hiragana-1399.7599.7799.6499.8999.44
Hiragana-105100.00100.0099.99100.00100.00
Satimage89.2589.2589.2590.3088.85
USPS94.9295.2794.4295.4295.37
MNIST96.1496.3995.8596.8496.54
Letter96.9097.1096.1097.5396.10
Average (B/S/W)96.72 (3/1/0)96.76 (3/4/0)96.32 (2/1/5)96.73 (7/1/2)96.35 (2/2/5)
For each problem the best accuracy is shown in bold and the worst underlined.
Table 7. Cross-validation accuracies using RBF and polynomial kernels for the multiclass problems.
Table 7. Cross-validation accuracies using RBF and polynomial kernels for the multiclass problems.
ProblemML1 SVMML1 v SVML1 SVMLS SVMULDM
RBFPolyRBFPolyRBFPolyRBFPolyRBFPoly
Numeral99.6399.5199.6399.6399.6399.5199.6399.5199.5199.51
Thyroid97.6198.2097.6198.2097.5998.2095.9795.7195.8194.75
Blood cell94.8794.7794.9394.7494.8794.6494.8394.8794.8094.70
Hiragana-5099.7099.6399.7299.6599.5799.4699.6799.6399.6399.59
Hiragana-1399.8199.7699.8099.7799.6499.5199.8699.9099.8399.63
Hiragana-10599.9599.8699.9499.8699.8999.7999.9899.9299.9599.89
Satimage92.6089.8392.5689.9092.4789.8392.7289.8592.3689.11
USPS98.3798.0198.3798.2498.2797.6498.4498.4298.4698.31
MNIST97.5696.8697.5897.1497.3196.6897.6097.4897.6597.14
Letter97.6496.9797.7297.1097.4395.5297.8397.3797.7396.18
Average97.7797.3497.7997.4297.6797.0897.6597.2797.5796.88
For each problem and for each classifier, the better accuracy is shown in bold.
Table 8. Selected parameter values for the RBF kernels. For two-class problems, most frequently selected values ( γ , C, C h ) are shown.
Table 8. Selected parameter values for the RBF kernels. For two-class problems, most frequently selected values ( γ , C, C h ) are shown.
DataML1 SVMML1 v SVML1 SVMLS SVMULDM
Banana10, 1, 150, 1, 120, 150, 1050, 10 4
B. cancer0.5, 1, 10.5, 1, 10.5, 15, 110, 10
Diabetes0.1, 50, 10.1, 1, 10.1, 5000.5, 15, 100
Flare-solar0.01, 50, 10.01, 50, 10.01, 500.01, 100.01, 0.1
German0.1, 1, 10.1, 1, 10.1, 10.1, 110, 100
Heart0.01, 500, 10.01, 500, 10.01, 1000.01, 100.01, 10 8
Image100, 10, 1100, 50, 1100, 5050, 5015, 10 8
Ringnorm100, 0.1, 150, 1, 150, 150, 0.150, 10
Splice10, 10, 105, 10, 110, 1010, 1010, 10 4
Thyroid5, 50, 15, 500, 15, 50100, 150, 10
Titanic0.01, 50, 10.01, 50, 10.01, 500.01, 100.01, 10 4
Twonorm0.01, 50, 10.01, 50, 10.01, 10.01, 500.01, 1000
Waveform5, 1, 150, 1, 115, 120, 150, 100
Numeral5, 10, 15, 10, 15, 101, 10015, 10 4
Thyroid (m)10, 2000, 110, 2000, 1010, 200050, 2000200, 10 8
Blood cell5, 100, 1005, 100, 505, 1005, 5005, 10 8
Hiragana-505, 100, 505, 100, 505, 10010, 10010, 10 6
Hiragana-1350, 50, 50050, 50, 5015, 100015, 200020, 10 8
Hiragana-10520, 10, 50020, 10, 5010, 1015, 200010, 10 6
Satimage200, 10, 10200, 10, 10200, 10200, 10200, 10 4
USPS10, 50, 200010, 50, 200010, 1005, 5005, 10 8
MNIST20, 10, 50020, 10, 50020, 1010, 5010, 10 6
Letter100, 10, 10100, 10, 100200, 1050, 5050, 10 6
For each problem, the γ values in bold, the γ and C values in bold, and γ , C, and C h values in bold show that they appear more than once.
Table 9. Computation time using the RBF kernels (in seconds).
Table 9. Computation time using the RBF kernels (in seconds).
DataML1 SVMML1 v SVML1 SVMLS SVMULDM
Banana0.0960.0530.0670.1920.254
B. cancer0.0060.0050.0050.0100.018
Diabetes0.0250.0260.0290.1190.222
Flare-solar0.0570.0590.0550.3410.693
German0.0590.0550.0590.4180.783
Heart0.0040.0040.0050.0100.013
Image0.3060.3540.3278.2421.7
Ringnorm0.2260.1410.1300.3620.420
Splice5.291.798.523.778.52
Thyroid0.0030.0020.0020.0050.008
Titanic0.0170.0160.0170.0280.031
Twonorm0.2440.2500.3360.4220.484
Waveform0.1220.1560.1060.2680.334
Numeral0.1250.1250.0471.171.14
Thyroid (m)1.531.520.9386211452
Blood cell2.913.080.73433.149.7
Hiragana-5047.397.78.67244268
Hiragana-133482959.67740920
Hiragana-10595092048.517791997
Satimage27.428.919.7292693
USPS51363435.510891996
MNIST614366701435832311,372
Letter1390212343930366544
B/S/W4/11/07/7/015/4/10/1/10/0/22
For each problem the shortest CPU time is shown in bold and the longest underlined.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Abe, S. Minimal Complexity Support Vector Machines for Pattern Classification. Computers 2020, 9, 88. https://doi.org/10.3390/computers9040088

AMA Style

Abe S. Minimal Complexity Support Vector Machines for Pattern Classification. Computers. 2020; 9(4):88. https://doi.org/10.3390/computers9040088

Chicago/Turabian Style

Abe, Shigeo. 2020. "Minimal Complexity Support Vector Machines for Pattern Classification" Computers 9, no. 4: 88. https://doi.org/10.3390/computers9040088

APA Style

Abe, S. (2020). Minimal Complexity Support Vector Machines for Pattern Classification. Computers, 9(4), 88. https://doi.org/10.3390/computers9040088

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop