Next Article in Journal
Advanced Control and Fault Detection Strategies for District Heating and Cooling Systems—A Review
Previous Article in Journal
A Computer Simulation Research of Two Types of Cardiac Physiological Pacing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interclass Interference Suppression in Multi-Class Problems

1
Harbin Institute of Technology, Harbin 150001, China
2
Science and Technology on Thermal Energy and Power Laboratory, Wuhan 430205, China
3
Systems Engineering Research Institute, China State Shipbuilding Corporation Limited, Beijing 100000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(1), 450; https://doi.org/10.3390/app11010450
Submission received: 13 November 2020 / Revised: 25 December 2020 / Accepted: 28 December 2020 / Published: 5 January 2021
(This article belongs to the Section Robotics and Automation)

Abstract

:
Multi-classifiers are widely applied in many practical problems. But the features that can significantly discriminate a certain class from others are often deleted in the feature selection process of multi-classifiers, which seriously decreases the generalization ability. This paper refers to this phenomenon as interclass interference in multi-class problems and analyzes its reason in detail. Then, this paper summarizes three interclass interference suppression methods including the method based on all-features, one-class classifiers and binary classifiers and compares their effects on interclass interference via the 10-fold cross-validation experiments in 14 UCI datasets. Experiments show that the method based on binary classifiers can suppress the interclass interference efficiently and obtain the best classification accuracy among the three methods. Further experiments were done to compare the suppression effect of two methods based on binary classifiers including the one-versus-one method and one-versus-all method. Results show that the one-versus-one method can obtain a better suppression effect on interclass interference and obtain better classification accuracy. By proposing the concept of interclass inference and studying its suppression methods, this paper significantly improves the generalization ability of multi-classifiers.

1. Introduction

Classification tasks exist widely in real-world applications, such as computer vision [1], fault diagnosis [2,3], human action recognition [4], face recognition [5], image recognition [6], material science [7], big data analysis, [8,9] etc. Many of them are classification tasks that include more than two classes, which are called multi-class problems [10]. Multi-class problems are usually more difficult to solve than binary classification problem because multi-class problems need to distinguish more classes. Multi-class problems are quite common in the real world [11,12,13,14,15,16,17]. For example, an image recognition device needs to distinguish many different kinds of images. An industrial fault diagnosis system [18,19] needs to diagnose what fault occurs in the machine. A sentiment analysis system [20,21] needs to classify different attitudes of people according to the information in social media such as Twitter and Facebook. Improving the classification accuracy of multi-class problems has great significance in actual applications [22].
Multi-class problems can be addressed in many ways. Among the well-recognized classifiers, some classifiers can handle multi-class problems directly, such as decision tree [23,24], rough set-based classifiers [25,26,27,28], neural networks [29,30], naive Bayes [31], K-nearest neighbor (KNN) [32] and so forth. Others can only deal with binary classification problems like support vector machine (SVM) [33,34]. Although there are some studies on how to use binary classifiers to solve multi-class problems, these studies usually focus on the situations where some classifiers cannot solve multi-class problems by themselves. For these classifiers, current studies focus on how to address multi-class problems by decomposing multi-class problems into binary classification problems. When binary classifiers are used to address multi-class problems, the most common strategies are currently one-versus-one (OVO) strategies [35] and one-versus-all (OVA) strategies [36,37]. For the OVO strategy, Galar et al. [38] developed a distance-based combination strategy, which weights the competence of the outputs of the base classifiers depending on the closeness of the query instance to each one of the classes. Galar et al.’s method reduced the effect of the non-competent classifiers, enhancing the results obtained by the state-of-the-art combinations for the OVO strategy. Galar et al. [39] proposed a dynamic classifier selection strategy for the OVO scheme that tries to avoid non-competent classifiers. For the OVA approach, Dinh et al. [40] proved new fast learning rates for OVA multi-class plug-in classifiers trained either from exponentially strongly mixing data or from data generated by a converging drifting distribution. They reported that their results retain the optimal learning rate in the independently identically distributed case in contrast to previous works for least squares SVMs under the binary-class setting. In literature [41], Rebetez et al. used an ensemble of binary OVA neural network classifiers and reported that the performances of their methods are comparable to lazy learning methods that require the whole dataset.
Currently, the generalization ability of multi-classifiers remains a longstanding challenge [42]. Many scholars have carried out relevant research on it. Eiadon et al. [43] decomposed the classes into subsets by embedding a structure of binary trees and put forward a novel splitting criterion based on minimizing generalization errors and greedy search procedures across the classes. Lei et al. [44] established data-dependent error bounds in terms of complexities of a linear function class defined on a finite set induced by training examples, for which they showed tight lower and upper bounds, applied the results to several prominent multi-class learning machines and exhibited a tight dependency on the number of classes. Kantavat et al. [45] proposed new methods for support vector machine (SVM) using tree architecture for multi-class classification and reported that their proposed methods run much faster than the traditional techniques but still provide comparable accuracy. Dhifli et al. [46] introduced a novel multi-class classification method for the open-set problem and proved the efficiency of their approach in classifying novel instances from known as well as unknown classes through experiments on benchmark datasets and synthetic datasets.
From the literature above, it can be noticed that many researchers focus on improving the generalization ability of classifiers. However, these researchers merely focus on generalized classification problems. There is little research specifically for multi-class problems. Researchers usually regard multi-class problems as ordinary classification problems to improve their generalization ability. Few researchers focus on the unique characteristics of multi-class problems. Although there are some studies on how to use binary classifiers to solve multi-class problems, these studies usually focus on the situations where some classifiers cannot solve multi-class problems by themselves. The differences between common binary classifiers and multi-classifiers are not studied sufficiently. This is the problem that this paper deals with.
This paper observes that there exists a special phenomenon in multi-class problems, which are quite different from binary classification problems. The features that can significantly discriminate a certain class from others are often deleted in the process of feature selection because they cannot discriminate among other classes. By contrast, the features that can discriminate among all classes are often reserved after feature selection. Although there are small errors in the training set, there may be large errors in test sets because reserved features cannot reflect the essence of a certain class. This phenomenon is an inherent problem in feature selection of multi-class problems and becomes more serious as the number of classes increases. This phenomenon significantly decreases the generalization ability of multi-classifiers. To improve the generalization ability of multi-classifiers better, this phenomenon must be eliminated or suppressed. Therefore, this paper elaborates this special phenomenon in multi-class problems, names it interclass interference, analyzes its reason systematically, designs its suppression methods and compares their suppression effects on interclass interference through 10-fold cross-validation experiments in 14 UCI datasets. The main contributions of this paper are listed as follows.
Firstly, this paper observes a special phenomenon in the feature selection of multi-classifiers and names it interclass inference. To the best of our knowledge, this is the first time that the concept of interclass inference is proposed. Secondly, the reasons for interclass inference are analyzed. The essence of interclass inference is revealed in this paper. The negative influence of interclass inference on the generalization ability of multi-class classifiers is also analyzed in detail. Thirdly, this paper summarizes the possible methods for suppressing interclass inference and compares their effects on interclass inference suppression.
The rest of this paper is organized as follows. In Section 2, the concept of interclass interference in multi-classifiers is proposed and its reason is analyzed systematically. In Section 3, suppression methods of interclass interference are summarized. In Section 4, interclass interference suppression algorithms used for the comparison experiments in this paper are designed. Section 5 shows the results of our comparison experiments. Section 6 concludes the paper.

2. Interclass Interference of Multi-Class Problems

Interclass interference is an inherent problem in feature extraction of multi-class problems. The removal of key features of some classes as redundant features is the direct cause of interclass interference in multi-class problems.
In this section, this paper will use an example of multi-class vibration fault diagnosis of steam turbine shown in Table 1 to introduce the concept of interclass interference and reveal its essence. According to expert knowledge, the vibration signal of 0.4f~0.6f, 1f and 2f, namely the features a 1 , a 2 and a 3 , are the essential features that can significantly discriminate oil film whirl, unbalance and misalignment from other faults, respectively [47,48,49]. Note that f denotes the rotation frequency in Table 1.
In machine learning, feature selection is an essential step in classification, which can reduce dimensionality. In Table 1, a 1 is the feature that can significantly distinguish the oil film whirl fault from other two fault classes including unbalance and misalignment. Meanwhile, the fact that attribute a 1 is the intrinsic characteristics of oil film whirl also meets the physical laws [50,51]. However, feature a 1 cannot distinguish unbalance and misalignment faults. In the feature selection process, features that can distinguish all classes have the highest priorities to be preserved. Feature a 1 may be deleted because it cannot distinguish between unbalance and misalignment fault. If feature a 1 is deleted, the generalization ability of this multi-classifier will be bad because the essential feature of the oil film whirl is deleted. This phenomenon is called interclass inference in this paper. In the following part of this section, this paper will use a rough set for feature selection, elaborate the interclass inference and analyze the bad influence of interclass inference in detail.
Rough set is a popular feature selection method. This paper uses rough sets for feature selection because the attribute reduction of rough sets removes unnecessary and unimportant features and has great advantages in feature selection. In rough set theory, data are usually stored in the form of a decision table < U , A = C D , V , f > , where U is the universe, A is the set of attributes, C is the set of conditional attributes, D is the set of decision attributes, V is the set of all attributive values and f is the information function. For any conditional attribute subset B C , there exits an equivalence relation I N D ( B ) , and it is defined by I N D ( B ) = { ( x , y ) U × U | f ( x , a ) = f ( y , a ) , a B } . The set of all equivalence classes is denoted as U / B . For every x U , the equivalence classes of x , denoted by [ x ] B , are defined by [ x ] B = { y U | ( x , y ) I N D ( B ) } . Let X be a subset of U ; then, the B-lower approximation B _ ( X ) and B-upper approximation B ¯ ( X ) of X are defined by B _ ( X ) = { x U | [ x ] B X } and B ¯ ( X ) = { x U | [ x ] B X } , respectively. The B-positive region P O S B ( D ) in the relation I N D ( D ) is defined by P O S B ( D ) = X U / D B _ ( X ) . The dependency degree γ B ( D ) of U / D on B is defined by γ B ( D ) = | P O S B ( D ) | / | U | , where | F | denotes the cardinality of set F . If B is a subset of C such that γ B ( D ) = γ C ( D ) , then B is a reduct of C [50]. Currently, there are many attribute reduction algorithms, among which one of the most common-used ones is the dependency degree-based algorithm proposed in [51]. Therefore, this paper used it for feature selection.
According to the feature selection algorithm proposed in the literature [51], a feature subset { a 2 , a 3 } is obtained after attribute reduction. In this feature subset { a 2 , a 3 } , two rules namely ( a 2 = l o w ) ^ ( a 3 = m e d i u m ) ( d = o i l   f i l m   w h i r l ) and ( a 2 = l o w ) ^ ( a 3 = l o w ) d = o i l   f i l m   w h i r l , are needed to diagnose the fault of oil film whirl. The support coefficients of two rules are both 1 / 10 . If the essential feature of oil film whirl, namely feature a 1 is used, only one rule ( a 1 = h i g h ) ( d = o i l   f i l m   w h i r l ) is needed to diagnose the fault of oil film whirl, and the support coefficient of this rule increases to 2 / 10 . The rule obtained by using essential features and non-essential features can both realize the perfect classification of the existing instances of oil film whirl in Table 1. However, we can obtain shorter rules and larger support coefficients based on the essential feature a 1 . In general, it can reflect the characteristics of data better and contribute to better generalization ability if there are fewer rules and larger support coefficients. Thus, the rule ( a 1 = h i g h ) ( d = o i l   f i l m   w h i r l ) is a more direct reflection of the oil film whirl compared with the previous two rules and accords with well expert knowledge. However, in the process of feature selection, the essential feature a 1 of oil film whirl is deleted. Obviously, the deletion of essential features deviates from the intrinsic characteristics and thus seriously decreases the generalization ability of multi-classifiers. In this paper, the deletion of essential features that can significantly discriminate a certain class from the other classes in the feature selection of multi-classifiers is called interclass interference.
Next, we will analyze the reason for interclass interference, namely the reason why essential features are deleted. We performed feature selection through the algorithm in [50]. First, we calculate the significance of conditional attribute a 1 , a 2 and a 3 in Table 1: we can obtain that γ a 1 ( d ) = 2 10 , γ a 2 ( d ) = 3 10 and γ a 3 ( d ) = 0 . The feature a 2 has the greatest significance, so it is selected first. Then, on the basis of attribute a 2 , we calculate the significance of a 1 and a 3 . We obtain γ { a 1 , a 2 } ( d ) = 6 10 and γ { a 2 , a 3 } ( d ) = 8 10 . The feature a 3 has a greater significance, so it is selected. Because of γ { a 2 , a 3 } ( d ) = γ { a 1 , a 2 , a 3 } ( d ) , the feature selection ends and the selected feature subset is { a 2 , a 3 } . In this process, the reason for the deletion of attribute a 1 is not because it has no contribution to classification. The essential features of the imbalance a 2 and misalignment a 3 are capable of classifying not only the instances of imbalance and misalignment but also the existing instances of oil film whirl correctly. After selecting feature a 2 and a 3 , the essential feature a 1 becomes redundant in terms of the existing instances of oil film whirl and thus is deleted.
From the above analysis, the essence of interclass inference can be revealed. The phenomenon of interclass interference occurs in the feature selection process of multi-class problems. Usually, features with the ability to distinguish all classes are prioritized in feature selection. The features that can significantly discriminate a certain class from others are called essential features of a certain class in this paper. Some essential features not only have strong classification ability for corresponding classes but also have classification ability for other classes to some extent. Thus, they are often reserved. By contrast, other essential features, which can distinguish a certain class significantly but cannot discriminate among other classes well, are often deleted as reductant features in feature selection. Once essential features are deleted, the phenomenon of interclass inference occurs. Interclass interference becomes more serious as the number of attributes increases. Although there are small training errors, there are usually large test errors because reserved features cannot reflect the essence of a certain class. The phenomenon of interclass inference seriously decreases the generalization ability of multi-classifiers. Therefore, interclass interference must be eliminated or suppressed in order to improve the generalization performance of multi-classifiers.

3. Interclass Interference Suppression Methods

3.1. All-Features-Based Approach

In the feature selection process of multi-classifiers, the essential features of a certain class are often deleted as redundant features, which causes interclass inference. Therefore, an intuitive solution to interclass inference suppression is to forgo feature selection and retain all features. This idea is intuitive and can eliminate interclass inference completely. However, this method goes against the conventional way that is widely adopted in the machine learning field, and it also seriously increases the computational cost. In the field of machine learning, many studies have shown that feature selection is an important way to improve the generalization performance of learning machines. As a result, for many machine learning methods, feature selection has become an indispensable part in the learning process [52]. Feature selection can greatly reduce the dimension of input space by removing those unimportant or redundant features, thus reducing the complexity of functions implemented by machine learning methods. Feature selection can generally improve the generalization performance of machine learning methods. Therefore, reserving all attributes are not very suitable to suppress the interclass inference.

3.2. One-Class Classifier Based Approach

One-class classifier [53] only uses the information of a certain classifier, so there is no interclass interference problem. Therefore, methods based on one-class classifiers can completely eliminate the interclass interference in multi-classifiers.
The typical example of a one-class classifier is the one-class support vector machine (SVM) [54]. One-class SVM uses a kernel function to map the original normal data to a high-dimensional space, where one-class SVM tries to find a hyperplane that enables the normal data to be as far from the origin as possible. Let the distance between the hyperplane and the origin be ρ . Then, the samples whose distance from the origin is smaller than ρ is detected as abnormal samples [54,55]. If there are m features and N training samples in the training set, let x i ( i = 1 , , N ) denote the training data, then one-class SVM can be denoted by the following optimization problem.
min 1 2 w 2 ρ + 1 v N i = 1 N ξ i s . t . ( w · φ ( x i ) ) ρ ξ i ξ i 0 , i = 1 , , N
where ξ i is the slack variable, v ( 0 , 1 ) is the error rate, φ ( . ) is a nonlinear mapping that is usually realized by a kernel function.
Research studies have shown that a one-class classifier can only use the information of the target class when defining the classification boundary, unlike multi-classifiers that can use the information of other classes. Therefore, in general, the performance of a one-class classifier is difficult to make comparable to that of multi-classifiers [56]. Thus, one-class classifier-based approaches are not very suitable to suppress the interclass inference, either.

3.3. Binary-Classifier-Based Approach

There is also interclass interference in binary classifiers; that is, the essential features of a certain class are deleted as redundant features. Nevertheless, compared with multi-classifiers, the interclass interference of the binary classifiers is significantly reduced. Therefore, it is possible to solve multi-class problems by constructing binary classifiers to suppress interclass interference. Currently, two popular ways of solving multi-class problems through binary classifiers are one-versus-one (OVO) strategies [35] and one-versus-all (OVA) strategies [40].
For an M-class problem, the OVA strategy requires M binary classifiers, each of them distinguishes one class from the rest. Taking a three-class classification problem as an example, there are three classes: Class 1, Class 2 and Class 3 in the problem. OVA strategy uses three binary classifiers to solve this multi-class problem. The first binary classifier distinguishes the samples from belonging to class 1 and not belonging to class 1. The second binary classifier distinguishes the samples from belonging to class 2 and not belonging to class 2. The third binary classifier distinguishes the samples from belonging to class 3 and not belonging to class 3. This method is simple and direct in the construction of binary classifiers. However, the samples processed by each of the binary classifiers constructed by OVA are usually class-imbalanced. Unlike the OVA strategy, the OVO strategy builds binary classifiers between any two classes of the original multi-class problem. For an M-class problem, OVO needs to construct M(M-1)/2 binary classifiers, which are usually much larger than the number of classifiers required by OVA.
Although OVO requires more classifiers, each classifier built by OVO only needs to distinguish any two classes in the original multi-class problem. Thus, each classifier processes fewer instances, and the problems to be learned by each classifier are usually simpler compared with the original multi-class problem and the OVA strategy. In addition, the class-imbalanced problem existing in OVA strategy does not exist in the OVO strategy. Many practical applications have shown that the OVO construction method usually achieves the best performance when dealing with multi-class problems.
To sum up, although the all-features-based approach and the one-class classifier-based approach can eliminate interclass interference, their classification performances are usually not good. The binary classifier-based approach is a good choice to suppress the interclass inference.

4. Design of Interclass Interference Suppression Algorithms

In this section, this paper designs interclass interference suppression algorithms in order to perform the comparison experiments between the interclass inference methods better. The all-features-based approach and one-classifier based approach are easy to implement, so this section focuses on the binary-classifier-based approach only. The binary classifier-based approach is discussed in detail, and corresponding algorithms for the comparison experiments are presented in this section.

4.1. Construction of Binary Classifier

The binary-classifier-based approach decomposes multi-class problems into several binary classification problems. OVA and OVO are two commonly used strategies.
Firstly, OVA strategy is presented. For a given M-class problem, OVA strategy uses the instances belonging to class i ( i = 1 , 2 , , M ) and the instances not belonging to class i to construct M binary classifiers. In this paper, other classes except class i are called the negative class of class i , denoted as class i ¯ . Algorithm 1 shows the detailed procedure of experiments based on OVA in this paper.
Secondly, the OVO strategy is presented. For a given M-class problem, OVO strategy uses to construct M(M-1)/2 binary classifiers respectively. Algorithm 2 shows the detailed procedure of experiments based on OVO strategy in this paper.
For OVO and OVA method, there are three decision strategies. The first one is based on the voting of classification results. The second one is based on the voting of a certain coefficient. The third one is based on the voting of the support coefficient. For voting based on the support coefficient, the OVA method can use the support coefficient of either the positive class or the negative class.
Algorithm 1 Binary classifier construction procedure based on OVA strategy
Input: the original dataset
Output: M binary classifiers; the corresponding feature subset i R e d u , rule set i R u l e s e t and support coefficient i S u p p of M binary classifiers; the average number of attributes a v e N u m _ R e d u ; the overall feature subsets u s e d _ R e d u ; the overall number of rules N u m ; average support coefficient a v e S u p p ; the average rule length a v e L e n ; the overall number of rules N u m ¯ for negative class; average support coefficient a v e S u p p ¯ for negative class and the average rule length a v e L e n ¯ for negative class.
Begin
i = 0 ;
while i M ; // M is the number of classes in multi-class problems
1. Mark all instances not belonging to class i as class i ¯ and obtain the dataset i _ D a t a s e t   of i-th classifier.  // class i ¯ is denoted as the negative class of class i here.
2. Obtain the feature subset i _ R e d u of i _ D a t a s e t via attribute reduction  and denote the number of attributes in feature subset as i N u m _ R e d u .
3. Extract rules from feature subset, and denote the rule set of i-th classifier as i R u l e s e t
4. Compute the support coefficient of each rule in rule set i R u l e s e t and denote the set of support coefficients as i S u p p .
5. Compute the rule number i N u m , the sum of support coefficient i S u m _ S u p p and the sum of rule length i S u m _ L e n of the class i .
6. Compute the rule number i N u m ¯ , the sum of support coefficient i S u m _ S u p p ¯ and the sum of rule length i S u m _ L e n ¯ of the class i ¯ .
7. i = i + 1 ;
end {while}
a v e N u m _ R e d u s u m i ( i N u m _ R e d u ) / M   // the average number of attributes
u s e d _ R e d u i i R e d u   // used feature subset for M-class problem
N u m s u m i ( i N u m )   // the number of rules for M-class problem
a v e S u p p s u m i ( i S u m _ S u p p ) / N u m   // average support coefficient of rules
a v e L e n s u m i ( i S u m _ L e n ) / N u m   // average rule length
N u m ¯ s u m i ( i N u m ¯ )    // the number of rules for negative class
a v e S u p p ¯ s u m i ( i S u m _ S u p p ¯ ) / N u m ¯ // average support coefficient of negative class
a v e L e n ¯ s u m i ( i S u m _ L e n ¯ ) / N u m ¯   // average rule length for negative class
End
Algorithm 2 Binary classifier construction procedure based on OVO strategy
Input: the original dataset
Output: M ( M 1 ) / 2 binary classifiers; the corresponding feature subset i j R e d u ; rule set i j R u l e s e t ; support coefficient i j S u p p of M binary classifiers; the average number of attributes a v e N u m _ R e d u ; the overall feature subsets u s e d _ R e d u ; the equivalent overall number of rules N u m ; average support coefficient a v e S u p p ; average rule length a v e L e n .
Begin
for each class i in the original class problem ( 1 i M 1 )
for each class i in the original class problem ( i + 1 j M )
1. Search all instances of class i and class j in the original dataset and construct new dataset i j _ D a t a s e t ;
2. Perform feature selection for dataset i j _ D a t a s e t through rough set, obtain the feature subset i j Re d u and denote the number of attributes in feature subset i j Re d u as i j N u m _ Re d u .
3. Extract rules and obtain the rule set i j R u l e s e t and calculate the rule’s support set i j S u p p .
4. For class i , calculate the rule number i N u m , the sum of the rule support i S u m _ S u p p and the sum of the rule length i S u m _ L e n . Then store i N u m , i S u m _ S u p p and i S u m _ L e n in the hash table S u m _ N u m , S u m _ S u p p and S u m _ L e n respectively with the key being class i .
5. For class j , calculate the rule number j N u m , the sum of support coefficient j S u m _ S u p p and the sum of rule length j S u m _ L e n . Then store j N u m , j S u m _ S u p p , j S u m _ L e n in the above hash table S u m _ N u m , S u m _ S u p p and S u m _ L e n respectively with the key being class j .
end {for}
end {for}
a v e N u m _ Re d u s u m i , j ( i j N u m _ Re d u ) ( M ( M 1 ) / 2 ) .
u s e d _ Re d u i , j i j Re d u . // actually used reduction subset
N u m s u m i ( S u m _ N u m ( M 1 ) ) . // equivalent overall rule number
a v e S u p p s u m i ( S u m _ S u p p ) s u m i ( S u m _ N u m ) . // average support coefficient
a v e L e n s u m i ( S u m _ L e n ) s u m i ( S u m _ N u m ) . // average rule length
End

4.2. Multi-Classifier Unified Collaborative Decision Algorithm

Algorithms for constructing binary classifiers based on OVA and OVO have been elaborated in Algorithm 1 and Algorithm 2. In order to classify new instances of multi-class problems, it is necessary to design a collaborative decision algorithm for these binary classifiers. Algorithm 3 gives a unified classifier collaborative decision algorithm. In Algorithm 3, the classification decision scoring indexes of each classifier can be classified in terms of support coefficients, certain coefficients and classification decision results. Different scoring indicators are chosen to get different collaborative decision-making strategies. It is necessary to explain that, for the binary classifiers based on OVA, the decision-making algorithm can make decisions based on either the classification decision score of each category or the negative class of each category. Meanwhile, we can still use the unified storage structure and algorithm described in Figure 1 to classify decision; the difference is that the classification decision scoring index value needs to be negative, and the algorithm is based on the absolute minimum principle of categorical decision scoring of negative categories to classify new model instances. The classification decision based on each class of negative classes is usually based on the majority of the class rules, so the rules usually have good statistical properties.
Algorithm 3 Multi-classifier unified collaborative decision algorithm.
Input: two types of classifiers for M class problems ( i Re d u , i R u l e s e t and i S u p p ) and new instance x
Output: the final classification result of x
Begin
for the i classifier in all binary classifiers do
1. According to the relevant classification decision method, the classifier is used to classify x ;
2. According to the classification decision scoring index, the classification decision score of the classifier is stored in the corresponding classification decision of the unified structure of Figure 1 through the accumulative method.
end{for}
Selecting the largest category of cumulative classification decision score from the structure as the final classification result of x
End
The binary-classifier-based approach decomposes multi-class problems into several binary classification problems. The OVO strategy needs M(M-1)/2 binary classifiers, and the OVA strategy needs M classifiers. The classification of a new instance is made by voting of these binary classifiers. The OVO strategy has three voting methods, including classification results, certain coefficient and support coefficient of each binary classifier. The OVA method has one method besides the above three methods, namely voting based on the support coefficient of negative classes of each binary classifier. The consumed score of each class is stored in the structure shown in Figure 1 after selecting one voting method. The final classification decision is made in terms of the class with the highest consumed score.

5. Experiments

5.1. Configurations of Experiments

In Section 4, binary classifiers construction algorithms based on OVO and OVA are presented, and three different classifier collaborative decision strategies can be selected for each algorithm. In order to verify the effect of these methods on interclass inference, this paper carries out 10-fold cross-validation experiments [57] of 14 UCI datasets [58]. Table 2 summarizes the information of these datasets. In this section, this paper compares the classification accuracy, the number of attributes, the number of rules, the length of rules and the support coefficient of various classification algorithms to compare the effect of these methods on interclass inference.

5.2. Comparison Among Different Interclass Interference Suppression Methods

In this section, this paper compares the OVA method and the OVO method with conventional multi-classifiers. This paper uses the algorithm in the literature [51] for feature selection and to classify through LEM2 algorithm [59], a common-used rule-based classification algorithm. For the conventional algorithm, feature selection is made once only. For the OVA and OVO methods, we need to construct M and M ( M 1 ) / 2 classifiers, respectively, and we need to make feature selection for each constructed classifier. The classification accuracy obtained from the various methods is given in Table 3. OVA_V, OVA_C, OVA_PS and OVA_NS represent the strategies of voting based on classification results, voting based on certain coefficients of rule, voting based on the support coefficient of positive classes and voting based on the support coefficient of negative classes for the OVA method, respectively. Meanwhile, OVO_V, OVO _C and OVO _S represent the strategies of voting based on classification results, voting based on certain coefficients of rule and voting based on the support coefficient for the OVO method, respectively. This paper performs 10-fold cross-validation to select the parameters of these algorithms and compares the interclass inference suppression effects. Specifically, the data are randomly divided into ten equal parts. Each unique part is selected as the test set, and the other nine parts are used as the training set. Thus, ten experiments are performed in total. The classification accuracy of 10-fold cross-validation is the mean value of the ten experiments. Corresponding 10-fold cross-validation classification accuracies of these methods are shown in Table 3. From Table 3, the following phenomena can be observed.
(1)
Based on the OVA and OVO methods, the classification accuracy is significantly improved compared with the conventional algorithm. This shows that the binary-classifier-based method can effectively suppress the interclass interference in multi-classifiers.
(2)
The OVO method obtains the best classification accuracy among OVO, OVA and conventional algorithms. Compared with conventional algorithms, the OVA method can also improve the classification accuracy, but the improvement is not obvious. This shows that the OVO method can suppress the interclass inference better than the OVA method. The reason for this is that the OVA method usually causes the class-imbalanced problem in classification.
(3)
In all kinds of decision strategies, the strategy of voting based on classification results obtains the optimal accuracy, and the strategy of voting based on certain coefficient of rules obtains the suboptimal accuracy. The strategy of voting based on the support coefficient of rules is the worst. For the OVA method, classification based on negative classes obtains a worse classification accuracy than that based on positive classes. This shows that voting based on classification results can be chosen as the optimal collaborative decision strategy.
(4)
The OVO-based binary classification method requires building more classifiers than the OVA-based binary classification method and thus usually costs more computational time. Literature [60] points out the computational burden of the two methods. For a multi-class problem with M classes, OVO requires M ( M 1 ) / 2 base binary classifiers and the computational complexity can be regarded as O ( M 2 ) . By contrast, OVA requires M base binary classifiers, and the computational complexity can be regarded as O ( M ) . Although the OVO-based binary classification method costs more time, OVO can obtain better classification accuracy and thus better interclass inference suppression performance than the OVA method and is more suitable for the case where users require high classification accuracy.
Next, we will analyze the reasons for the performance improvement through some basic evaluation indexes of classifiers. Table 4 and Table 5 give the number of selected features, rule number, rule length and rule support coefficient of various methods. From Table 4 and Table 5, we can obtain the following results:
(1)
The number of attributes and rules obtained by the OVO interclass interference suppression method is significantly smaller than that of the conventional algorithm. Meanwhile, the rule length is shorter, and the support coefficient is larger. Obviously, OVO method can reflect the intrinsic characteristics better and obtain better generalization ability.
(2)
Compared with the OVO approach, the above-mentioned indexes obtained by the OVA approach are similar to the conventional multi-classifiers, which explains why the performance improvement is not obvious to some extent. In addition, although the negative-classes-based classification obtains larger support coefficients and shorter rules than the positive-classes-based classification, its classification accuracy is not ideal since it is an indirect decision-making strategy.

5.3. Performance of the All-Features-Based Approach

When dealing with multi-class problems, deletion of essential features leads to interclass interference. If all features are reserved, interclass inference can be eliminated. Thus, this paper evaluates the effect of this method on interclass interference through experiments.
Table 6 shows the classification accuracy, number of rules, rule length and the support coefficient of rules obtained by the all-features-based method and the OVO method. In Table 6, CV_A represents the all-features-based methods and OVO_VA represents the method that reserves all features and uses OVO strategy and makes classification by voting based on the classification results of each classifier. By comparing it with Table 3, Table 4 and Table 5, the following phenomena are observed.
(1)
By reserving all features, the classification accuracy of multi-class problems is obviously improved. Further comparison shows that all-features-based method obtains fewer rules and a larger support coefficient than the conventional algorithm. This shows that each class can be expressed by its own essential features and that the intrinsic characteristics of data can be reflected better when reserving all features. This can explain the why all-features-based method obtains better classification accuracy to some extent.
(2)
The OVO_VA method and OVO_V method has few differences in classification accuracy. Thus, reserving all features has little effect on the classification accuracy of OVO method. This is because the OVO method itself can make use of the essential features to express the knowledge and suppress interclass interference.
(3)
The all-features-based method still obtains worse classification accuracy than the OVO_V method shown in Table 3. Compared with the conventional algorithm and the all-features-based algorithm, the OVO method extracts simpler and clearer classification knowledge and stronger obtained rules. Thus, the OVO method obtains better generalization ability.
Compared with conventional algorithms, the all-features-based algorithm can effectively suppress interclass interference and achieve better generalization ability in multi-class problems. However, the method also preserves redundant features, leading to more computational cost. Table 7 shows the number of attributes obtained by the conventional method, the all-features-based method and the OVO method when dealing with multi-class problems. It is observed that the number of attributes used by the all-features-based method is significantly higher than that of the conventional method. Although the number of attributes obtained by the OVO method is also larger than by the conventional algorithm, it uses much fewer attributes than all-features-based method. This shows that the OVO method can not only effectively suppress the interclass interference of multi-class problems but also delete redundant features and avoid the additional cost caused by these features. Therefore, when dealing with multi-class problems, the OVO method is a better choice to suppress interclass inference than the all-features-based method.

5.4. Comparison Among the One-Class Classifier and Binary Classifier-Based Approach

The method based on one-class classifiers can completely eliminate the class interference existing in multi-classifiers. However, a one-class classifier only uses the information of the target class in the definition of the classification boundary. Many studies show that the performance of one-class classifiers is usually worse than that of multi-classifiers. In order to confirm that the OVO method is the best choice to suppress interclass interference, we used a binary support vector machine (SVM) based on the OVO method and one-class SVM [61] to carry out a comparison experiment. In the experiment, we did not carry out feature selection. Radial basis function was used for binary SVM and one-class SVM. We first compared the classification accuracy of binary and one-class SVM under different parameters to determine their optimal parameters and obtained their best classification accuracy. Then, the best classification accuracy of the two methods was compared to evaluate the performance of binary and one-class SVM. Results are shown in Table 8.
From Table 8, we can draw the following conclusions. When we take C = 100 in the experiment, SVM based on the OVO method acquires the best performance. When we take v = 0.01, one-class SVM obtains the better performance. By comparing the best performance obtained by the two methods, it is observed that the classification accuracy of the binary SVM is significantly better than one-class SVM, which is consistent with the previous research results of one-class classifiers. The performance of one-class classifiers is usually worse than binary classifiers, and it is difficult to implement one-class classification algorithm in many multi-classifiers. Therefore, the interclass interference suppression method based on binary classifiers is the best choice to suppress the interclass interference of multi-classifiers.

6. Conclusions

The generalization ability of classifiers is a crucial problem in pattern recognition. In multi-classifiers, there is a special phenomenon that essential features with the ability to discriminate a certain class from others are often deleted in feature selection. This phenomenon seriously decreases the classification accuracy of multi-classifiers. To address this problem, this paper called this phenomenon interclass inference, analyzed its reasons in detail and summarized three interclass inference suppression methods including all-features-based method, the one-class classifier-based method and the binary-classifier-based method. By comparing the three methods in 14 UCI datasets through 10-fold cross-validation, the following conclusions can be drawn.
Firstly, the concept of interclass inference is proposed and the essence of interclass inference is revealed. The deletion of essential features leads to interclass inference.
Secondly, the three methods can all improve the classification accuracy to some extent when they are compared with the conventional algorithm. This shows that suppressing the interclass inference is an effective method to improve the generalization ability of multi-classifiers. The binary classifier-based method can suppress the interclass inference best and obtain the best generalization ability.
Thirdly, this paper compares the suppression effects of two binary classifier-based methods, including one-versus-one (OVO) and one-versus-all (OVA) on interclass inference. OVO method obtains better classification accuracy than the OVA method. Thus, it is the best method to suppress the interclass inference. By studying the interclass inference and its suppression methods, the generalization ability of multi-classifiers is significantly improved.

Author Contributions

Conceptualization, M.B. and J.L.; writing, N.J.; methodology, M.B., X.L., Y.W. and R.C.; supervision, D.Y. and J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China under Grant No. 51976042, National Science and Technology Major Project of China under Grant No. 2017-I-0007-0008, National Science and Technology Major Project of China under Grant No. 2017-V-0005-0055, National Key R&D Program of China No. 2017YFB0902100 and Science and Technology on Thermal Energy and Power Laboratory Foundation No. TPL2017CA010.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the anonymous reviewers for their valuable suggestions to refine this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sugaya, Y.; Sugibuchi, K.; Omachi, S. Effectiveness of Integration of Multiple Classification Methods within the AdaBoost Framework (Special Issue on Computer Vision and Applications). IIEE J. Trans. Image Electron. Vis. Comput. 2015, 3, 174–184. [Google Scholar]
  2. Zhao, W.; Wang, Z.; Ma, J.; Li, L. Fault Diagnosis of a Hydraulic Pump Based on the CEEMD-STFT Time-Frequency Entropy Method and Multiclass SVM Classifier. Shock Vib. 2016, 8. [Google Scholar] [CrossRef]
  3. Hang, J.; Zhang, J.; Cheng, M. Application of multi-class fuzzy support vector machine classifier for fault diagnosis of wind turbine. Fuzzy Sets Syst. 2016, 297, 128–140. [Google Scholar] [CrossRef]
  4. Zhang, B.; Yang, Y.; Chen, C.; Yang, L.; Han, J.; Shao, L. Action Recognition Using 3D Histograms of Texture and A Multi-class Boosting Classifier. IEEE Trans. Image Process. 2017, 26, 4648–4660. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Nan, D.; Xu, Z.; Bian, S.Q. Face Recognition Based on Multi-classifier Weighted Optimization and Sparse Representation. Int. J. Signal Process. Image Process. Pattern Recognit. 2013, 6, 423–436. [Google Scholar]
  6. Qian, H.; Mao, Y.; Xiang, W.; Wang, Z. Recognition of human activities using SVM multi-class classifier. Pattern Recognit. Lett. 2010, 31, 100–111. [Google Scholar] [CrossRef]
  7. Wick, P.; Louw-Gaume, A.E.; Kucki, M.; Krug, H.F.; Kostarelos, K.; Fadeel, B.; Dawson, K.A.; Salvati, A.; Vazquez, E.; Ballerini, L.; et al. Classification framework for graphene-based materials. Angew. Chem. Int. Ed. 2014, 53, 7714–7718. [Google Scholar] [CrossRef] [Green Version]
  8. De Melo, G.; Varde, A.S. Scalable Learning Technologies for Big Data Mining. In Proceedings of the 20th International Conference on Database Systems for Advanced Applications, DASFAA, Hanoi, Vietnam, 20–23 April 2015; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  9. Varde, A.S.; Ma, S.; Maniruzzaman, M.D.; Brown, D.C.; Rundensteiner, E.A.; Sisson, R.D., Jr. Comparing mathematical and heuristic approaches for scientific data analysis. AI EDAM Artif. Intell. Eng. Des. Anal. Manuf. 2008, 22, 53. [Google Scholar] [CrossRef] [Green Version]
  10. Burnap, P.; Colombo, G.; Amery, R.; Hodorog, A.; Scourfield, J. Multi-class machine classification of suicide-related communication on Twitter. Online Soc. Netw. Media 2017, 2, 32–44. [Google Scholar] [CrossRef]
  11. Tang, L.; Tian, Y.; Pardalos, P.M. A novel perspective on multiclass classification: Regular simplex support vector machine. Inf. Sci. 2019, 480, 324–338. [Google Scholar] [CrossRef]
  12. Karthikeyan, D.; Varde, A.S.; Wang, W. Transfer learning for decision support in Covid-19 detection from a few images in big data. In Proceedings of the IEEE Big Data Conf., Atlanta, GA, USA, 10 December2020. [Google Scholar]
  13. Ang, J.H.; Guan, S.U.; Tan, K.C.; Al-Mamun, A. Interference-less neural network training. Neurocomputing 2008, 71, 3509–3524. [Google Scholar] [CrossRef]
  14. Basavaraju, P.; Varde, A.S. Supervised learning techniques in mobile device apps for Androids. ACM Sigkdd Explor. Newsl. 2017, 18, 18–29. [Google Scholar] [CrossRef]
  15. Har-Peled, S.; Roth, D.; Zimak, D. Constraint Classification: A New Approach to Multiclass Classification. International Conference on Algorithmic Learning Theory; Springer: Berlin/Heidelberg, Germany, 2002; pp. 365–379. [Google Scholar]
  16. Amit, Y.; Fink, M.; Srebro, N.; Ullman, S. Uncovering Shared Structures in Multiclass Classification. In Proceedings of the 24th International Conference on Machine Learning, Corvallis, OR, USA, 20–24 June 2007; pp. 17–24. [Google Scholar]
  17. Tewari, A.; Bartlett, P.L. On the consistency of multiclass classification methods. J. Mach. Learn. Res. 2007, 8, 1007–1025. [Google Scholar]
  18. Bai, M.; Liu, J.; Chai, J.; Zhao, X.; Yu, D. Anomaly detection of gas turbines based on normal pattern extraction. Appl. Eng. 2020, 166, 114664. [Google Scholar] [CrossRef]
  19. Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
  20. Gandhe, K.; Varde, A.S.; Du, X. Sentiment Analysis of Twitter Data with Hybrid Learning for Recommender Applications. In Proceedings of the 2018 9th IEEE Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON), New York, NY, USA, 8–10 November 2018; IEEE: New York, NY, USA, 2018; pp. 57–63. [Google Scholar]
  21. Aly, M. Survey on multiclass classification methods. Neural Netw. 2005, 19, 1–9. [Google Scholar]
  22. Pawara, P.; Okafor, E.; Groefsema, M.; He, S.; Schomaker, L.; Wiering, M.A. One-vs-One classification for deep neural networks. Pattern Recognit. 2020, 108, 107528. [Google Scholar] [CrossRef]
  23. Katuwal, R.; Suganthan, P.N.; Zhang, L. An ensemble of decision trees with random vector functional link networks for multi-class classification. Appl. Soft Comput. 2018, 70, 1146–1153. [Google Scholar] [CrossRef]
  24. Wu, K.; Zheng, Z.; Tang, S. BVDT: A Boosted Vector Decision Tree Algorithm for Multi-Class Classification Problems. Int. J. Pattern Recognit. Artif. Intell. 2017, 31, 1750016. [Google Scholar] [CrossRef]
  25. Vluymans, S.; Fernández, A.; Saeys, Y.; Cornelis, C.; Herrera, F. Dynamic affinity-based classification of multi-class imbalanced data with one-versus-one decomposition: A fuzzy rough set approach. Knowl. Inf. Syst. 2017, 56, 1–30. [Google Scholar] [CrossRef]
  26. Vluymans, S.; Cornelis, C.; Herrera, F.; Saeys, Y. Multi-label classification using a fuzzy rough neighborhood consensus. Inf. Sci. 2018, 433, 96–114. [Google Scholar] [CrossRef]
  27. Liu, J.; Bai, M.; Jiang, N.; Yu, D. Structural risk minimization of rough set-based classifier. Soft Comput. 2020, 24, 2049–2066. [Google Scholar] [CrossRef]
  28. Liu, J.; Bai, M.; Jiang, N.; Yu, D. A novel measure of attribute significance with complexity weight. Appl. Soft Comput. 2019, 82, 105543. [Google Scholar] [CrossRef]
  29. Melin, P.; Amezcua, J.; Valdez, F.; Castillo, O. A new neural network model based on the LVQ algorithm for multi-class classification of arrhythmias. Inf. Sci. 2014, 279, 483–497. [Google Scholar] [CrossRef]
  30. Zhang, Z.L.; Luo, X.G.; García, S.; Herrera, F. Cost-Sensitive back-propagation neural networks with binarization techniques in addressing multi-class problems and non-competent classifiers. Appl. Soft Comput. 2017, 56, 357–367. [Google Scholar] [CrossRef]
  31. Tang, B.; Kay, S.; He, H. Toward optimal feature selection in naive Bayes for text categorization. IEEE Trans. Knowl. Data Eng. 2016, 28, 2508–2521. [Google Scholar] [CrossRef] [Green Version]
  32. Zhang, S.; Li, X.; Zong, M.; Zhu, X.; Cheng, D. Learning k for knn classification. ACM Trans. Intell. Syst. Technol. 2017, 8, 1–19. [Google Scholar] [CrossRef] [Green Version]
  33. Li, K.; Wu, Y.; Nan, Y.; Li, P.; Li, Y. Hierarchical multi-class classification in multimodal spacecraft data using DNN and weighted support vector machine. Neurocomputing 2017, 259, 55–65. [Google Scholar] [CrossRef]
  34. Wang, G.S.; Ren, Q.H.; Su, Y.Z. The Interference Classification and Recognition Based on SF-SVM Algorithm. In Proceedings of the 2017 IEEE 9th International Conference on Communication Software and Networks (ICCSN), GuangZhou, China, 6–8 May 2017; IEEE: New York, NY, USA, 2017; pp. 835–841. [Google Scholar]
  35. Knerr, S.; Personnaz, L.; Dreyfus, G. Single-Layer Learning Revisited: A Stepwise Procedure for Building and Training a Neural Network. Neurocomputing; Springer: Berlin/Heidelberg, Germany, 1990; pp. 41–50. [Google Scholar]
  36. Clark, P.; Boswell, R. Rule Induction with CN2: Some Recent Improvements. European Working Session on Learning; Springer: Berlin/Heidelberg, Germany, 1991; pp. 151–163. [Google Scholar]
  37. Anand, R.; Mehrotra, K.; Mohan, C.K.; Ranka, S. Efficient classification for multiclass problems using modular neural networks. IEEE Trans. Neural Netw. 1995, 6, 117–124. [Google Scholar] [CrossRef]
  38. Galar, M.; Fernández, A.; Barrenechea, E.; Herrera, F. DRCW-OVO: Distance-based relative competence weighting combination for One-versus-one strategy in multi-class problems. Pattern Recognit. 2015, 48, 28–42. [Google Scholar] [CrossRef]
  39. Galar, M.; Fernández, A.; Barrenechea, E.; Bustince, H.; Herrera, F. Dynamic classifier selection for One-versus-one strategy: Avoiding non-competent classifiers. Pattern Recognit. 2013, 46, 3412–3424. [Google Scholar] [CrossRef]
  40. Dinh, V.; Ho, L.S.T.; Cuong, N.V.; Nguyen, D.; Nguyen, B.T. Learning from Non-Iid Data: Fast Rates for the One-Versus-All Multiclass Plug-in Classifiers. In Proceedings of the International Conference on Theory and Applications of Models of Computation; Springer: Cham, Germany, 2015; pp. 375–387. [Google Scholar]
  41. Rebetez, J.; Perez-Uribe, A. Indoor Activity Recognition by Combining One-vs.-All Neural Network Classifiers Exploiting Wearable and Depth Sensors. In Proceedings of the International Conference on Artificial Neural Networks: Advences in Computational Intelligence; Springer: Berlin/Heidelberg, Germany, 2013; pp. 216–223. [Google Scholar]
  42. Shen, X.; Wang, L. Generalization error for multi-class margin classification. Electron. J. Stat. 2007, 1, 307–330. [Google Scholar] [CrossRef] [Green Version]
  43. Eiadon, M.; Pipanmaekaporn, L.; Kamonsantiroj, S. Mining discriminative class codes for multi-class classification based on minimizing generalization errors. First International Workshop on Pattern Recognition. Int. Soc. Opt. Photonics 2016, 10011, 100111D. [Google Scholar]
  44. Lei, Y.; Dogan, Ü.; Zhou, D.; Kloft, M. Generalization error bounds for extreme multi-class classification. CoRR 2017, abs/1706.09814. [Google Scholar]
  45. Kantavat, P.; Kijsirikul, B.; Songsiri, P.; Fukui, K.-I.; Numao, M. Efficient Decision Trees for Multi–Class Support Vector Machines Using Entropy and Generalization Error Estimation. Int. J. Appl. Math. Comput. Sci. 2018, 28, 705–717. [Google Scholar] [CrossRef] [Green Version]
  46. Dhifli, W.; Diallo, A.B. Galaxy-X: A Novel Approach for Multi-class Classification in an Open Universe. arXiv 2015, arXiv:1511.00725. [Google Scholar]
  47. Liu, S.; Chen, J.; Wang, F.; Feng, Y.X.; Hong-Bo, G.U.; Han, J.F.; Wu, J.B. Analysis and Treatment of Oil Whirl on 1000MW Ultra-supercritical Unit. Turbine Technol. 2010, 52, 373–376. [Google Scholar]
  48. Dong, X. Vibration Analysis and Experiment Research on Misalignment of Rotor System. Master’s Thesis, Northeastern University, Shenyang, China, 2010. [Google Scholar]
  49. Huang, Y. An Analysis of Rotor Unbalance. Power Equip. 2009, 23, 164–169. [Google Scholar]
  50. Pawlak, Z.; Skowron, A. Rudiments of rough sets. Inf. Sci. 2007, 177, 3–27. [Google Scholar] [CrossRef]
  51. Liu, J.; Hu, Q.; Yu, D. Weighted Rough Set Learning: Towards a Subjective Approach. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining; Springer: Berlin/Heidelberg, Germany, 2007; pp. 696–703. [Google Scholar]
  52. Xu, F.; Zhang, Y.J. Integrated patch model: A generative model for image categorization based on feature selection. Pattern Recognit. Lett. 2007, 28, 1581–1591. [Google Scholar] [CrossRef]
  53. Tax, D.; Juszczak, P. Kernel Whitening for one-Class Classification. Int. J. Pattern Recognit. Artif. Intell. 2003, 17, 333–347. [Google Scholar] [CrossRef] [Green Version]
  54. Tax, D.; Duin, R. Support Vector Data Description. Mach. Learn. 2004, 54, 45–66. [Google Scholar] [CrossRef] [Green Version]
  55. Bai, M.; Liu, J.; Ma, Y.; Zhao, X.; Long, Z.; Yu, D. Long short-term memory network-based normal pattern group for fault detection of three-shaft marine gas turbine. Energies 2021, 14, 13. [Google Scholar] [CrossRef]
  56. Yu, H. SVMC: Single-class Classification with Support Vector Machines. In Proceedings of the 18th International Joint Conference on Artificial Intelligence, Acapulco, Mexico, 9–15 August 2003; pp. 567–572. [Google Scholar]
  57. Ma, S.; Li, H.; Zhang, H.; Xie, X. Reverberation Level Recognition by Formants Based on 10-fold Cross Validation of GMM. International Forum on Digital TV and Wireless Multimedia Communications; Springer: Singapore, 2017; pp. 161–171. [Google Scholar]
  58. Asuncion, A.; Newman, D. Uci Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2007; p. 1994. [Google Scholar]
  59. Grzymala-Busse, J.W. LERS—A System for Learning from Examples Based on Rough Sets; Intelligent Decision Support; Springer: Dordrecht, The Netherland, 1992; pp. 3–18. [Google Scholar]
  60. Rocha, A.; Goldenstein, S.K. Multiclass from Binary: Expanding One-Versus-All, One-Versus-One and Ecoc-Based Approaches. IEEE Trans. Neural Netw. Learn. Syst. 2013, 25, 289–302. [Google Scholar] [CrossRef] [PubMed]
  61. Manevitz, L.M.; Yousef, M. One-class SVMs for Document Classification. J. Mach. Learn. Res. 2001, 2, 139–154. [Google Scholar]
Figure 1. Store Structure of Consumed Score for the M-Class Problem.
Figure 1. Store Structure of Consumed Score for the M-Class Problem.
Applsci 11 00450 g001
Table 1. Decision table for vibration fault diagnosis of a steam turbine.
Table 1. Decision table for vibration fault diagnosis of a steam turbine.
Fault InstancesConditional Attributes (C)Decision Attributes (D)
U 0.4f~0.6f ( a 1 )1f ( a 2 )2f ( a 3 )Fault ( d )
x 1 LowHighLowUnbalance
x 2 LowHighMediumUnbalance
x 3 LowHighHighUnbalance
x 4 LowMediumLowUnbalance
x 5 LowMediumMediumUnbalance
x 6 LowLowHighMisalignment
x 7 LowMediumMediumMisalignment
x 8 LowMediumHighMisalignment
x 9 HighLowLowOil Film Whirl
x 10 HighLowMediumOil Film Whirl
Table 2. Datasets used for experiments.
Table 2. Datasets used for experiments.
No.NameSizeConditional
Attribute Number
Number
of Classes
1zoo101167
2lymphography148184
3wine178133
4flags194288
5autos205236
6machine20978
7images210197
8glass21496
9audiology2266924
10heart303135
11solar323103
12soybean6833519
13vehicle846184
14anneal898385
Table 3. Classification accuracy of OVA and OVO approach.
Table 3. Classification accuracy of OVA and OVO approach.
DatasetConventional AlgorithmOVAOVO
OVA_VOVA_COVA_PS1vR_NSOVO_VOVO_COVO_S
zoo0.94000.95090.94090.95090.93090.95090.95090.9009
lymphography0.81860.82520.83240.83190.83190.82570.82570.7786
wine0.93890.94440.93890.94410.92190.95490.95490.9157
flags0.59370.58840.58370.59840.59820.64950.62870.4700
autos0.74380.77310.78330.76360.75900.76330.75860.5221
machine0.65520.62670.65050.64570.64100.68860.68860.5074
images0.86670.86670.87140.84290.85240.86670.86670.6238
glass0.69550.67710.67680.66750.57810.68230.67770.5799
audiology0.76150.74470.74900.73160.77080.78000.78000.5002
heart0.52160.52460.52460.52440.50510.55440.54430.5841
solar0.86710.86090.86070.85780.84830.87610.87310.7865
soybean0.84620.89600.90330.88430.88140.92540.92980.4802
vehicle0.66310.68550.70200.68430.69620.71750.72100.6192
anneal1.00001.00001.00001.00001.00000.99890.99890.9232
Mean0.77940.78320.78700.78050.77250.80240.79990.6565
Table 4. Numbers of attributes and rules obtained by different methods.
Table 4. Numbers of attributes and rules obtained by different methods.
DatasetNumbers of AttributesNumbers of Rules
Conventional
Algorithm
OVAOVOConventionOVAOVO
PN
zoo4.90002.45711.085712.200010.000020.70007.4500
lymphography6.00003.70002.100034.900033.350039.250015.7333
wine4.00003.56672.500013.500012.500016.50009.1500
flags8.80005.16252.757173.800076.2500144.250038.9857
autos9.20005.56672.960051.100052.350082.950028.6400
machine6.70003.71251.725037.000035.000069.100017.7286
images6.30003.21431.704828.300027.650053.650016.2500
glass6.80005.16673.046732.400029.900050.600018.5400
audiology13.30002.97161.102358.100055.3000122.400028.7445
heart9.80009.18008.010098.200096.3000135.500059.8000
solar9.00007.90006.566752.100052.100063.400037.5500
soybean11.30003.04211.1234115.700091.0500151.150028.5278
vehicle14.200011.52508.8500181.5000181.9000207.800091.6333
anneal3.00001.20001.00007.00005.00008.00005.5250
Mean 8.09294.88323.180856.842954.189383.232128.8756
Table 5. Length and support coefficient of rules obtained by different methods.
Table 5. Length and support coefficient of rules obtained by different methods.
DatasetRule LengthRule Support Coefficient
Conventional
Algorithm
OVAOVOConventionOVAOVO
PNPN
zoo2.30882.09041.11081.04020.08230.10240.36790.4726
lymphography2.58432.52642.37252.04720.04570.04710.09190.1438
wine2.18452.17821.61841.51690.10700.10270.17550.2150
flags2.82972.75841.83701.60130.01750.01550.07390.1219
autos2.82482.66711.97571.70140.02460.02100.08310.1202
machine2.89342.95942.04991.69840.03340.02660.12630.2206
images2.59502.54121.60861.37160.04230.03700.13730.2281
glass3.10333.10452.15251.94600.03350.03140.14330.1732
audiology3.27762.97911.48811.07090.02230.01980.25220.4143
heart4.83494.95663.87693.58630.01550.01250.06570.0649
solar3.31483.49203.08692.93500.02720.02220.04420.0503
soybean4.16963.86032.19251.22800.01050.01120.16760.3356
vehicle4.42204.54943.94273.63000.00870.00690.02660.0287
anneal1.71431.20001.00001.00000.14290.20000.58880.4526
Mean3.07552.99022.16521.88380.04380.04690.16740.2173
Table 6. Classification accuracy obtained by the method based on all features.
Table 6. Classification accuracy obtained by the method based on all features.
DatasetClassification AccuracyNumber of RulesRule LengthSupport Coefficient
CV_AOVO_VACV_AOVO_VACV_AOVO_VACV_AOVO_VA
zoo0.95180.95189.30007.35002.22391.05440.11100.4798
lymphography0.83190.845225.600012.90002.95512.21180.07440.2199
wine0.94970.955210.10008.15002.12061.60220.19610.3247
flags0.61870.613457.100031.34293.38451.98940.02660.1994
autos0.76400.793348.200023.88002.89191.89570.02850.1761
machine0.65050.674337.200016.70002.89341.75710.03350.2673
images0.88570.885725.000013.51672.59771.46000.04880.2854
glass0.69550.667732.300018.02003.12291.99130.03360.1792
audiology0.79720.824140.700024.61333.63481.11280.03620.4837
heart0.52160.528398.800058.60004.84743.59520.01520.0731
solar0.85800.867251.600036.35003.28502.99840.02770.0531
soybean0.91650.937061.700022.66674.06351.26610.02070.4276
vehicle0.67600.7151177.286.26674.48253.68810.00920.0381
anneal1.00001.00005.00005.00001.20001.00000.20000.5000
Mean0.79410.804248.557126.09693.12171.97300.06150.2648
Table 7. Numbers of attributes obtained by the method based on all features and other methods.
Table 7. Numbers of attributes obtained by the method based on all features and other methods.
DatasetConventional AlgorithmCV_AOVO_V
zoo4.9000168.7000
lymphography6.0000188.1000
wine4.0000135.6000
flags8.80002814.9000
autos9.20002312.5000
machine6.700077.0000
images6.30001911.4000
glass6.800097.0000
audiology13.30006923.8000
heart9.80001310.0000
solar9.0000108.9000
soybean11.30003524.9000
vehicle14.20001816.1000
anneal3.0000384.9000
Mean8.092922.571411.7000
Table 8. Comparison between classification accuracies obtained by binary SVM and one-class SVM.
Table 8. Comparison between classification accuracies obtained by binary SVM and one-class SVM.
DatasetBinary SVM (OVO)One-Class SVM
C = 1C = 100C = 1000ν = 0.001ν = 0.01ν = 0.3
zoo0.92090.96090.96090.94180.87180.8218
lymphography0.78520.84570.84570.72330.71000.7171
wine0.46110.48920.48920.50620.52290.5173
flags0.37660.37680.37680.22260.26370.2639
autos0.34690.34690.34690.04330.17050.0690
machine0.60310.60310.60310.15330.16290.1819
images0.30480.31900.31900.29050.29520.2952
glass0.66750.70000.68550.54610.55060.5240
audiology0.50450.81980.81980.39740.43340.4557
heart0.54120.54120.54120.12180.12180.1218
solar0.88550.86400.84830.79250.81390.6469
soybean0.93980.94280.93550.85360.86530.8404
vehicle0.28950.29890.29890.30020.30130.3013
anneal0.92540.95990.95990.85080.85300.8408
Mean0.61090.64770.64510.48170.49550.4712
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, J.; Bai, M.; Jiang, N.; Cheng, R.; Li, X.; Wang, Y.; Yu, D. Interclass Interference Suppression in Multi-Class Problems. Appl. Sci. 2021, 11, 450. https://doi.org/10.3390/app11010450

AMA Style

Liu J, Bai M, Jiang N, Cheng R, Li X, Wang Y, Yu D. Interclass Interference Suppression in Multi-Class Problems. Applied Sciences. 2021; 11(1):450. https://doi.org/10.3390/app11010450

Chicago/Turabian Style

Liu, Jinfu, Mingliang Bai, Na Jiang, Ran Cheng, Xianling Li, Yifang Wang, and Daren Yu. 2021. "Interclass Interference Suppression in Multi-Class Problems" Applied Sciences 11, no. 1: 450. https://doi.org/10.3390/app11010450

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop