Next Article in Journal
Low-Velocity Impacts on a Polymeric Foam for the Passive Safety Improvement of Sports Fields: Meshless Approach and Experimental Validation
Previous Article in Journal
Refractive Index Sensing through Surface Plasmon Resonance in Light-Diffusing Fibers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Incorporating Grey Total Influence into Tolerance Rough Sets for Classification Problems

1
College of Management & College of Tourism, Fujian Agriculture and Forestry University, Fuzhou 350002, China
2
Department of Business Administration, Chung Yuan Christian University, Taoyuan 32023, Taiwan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2018, 8(7), 1173; https://doi.org/10.3390/app8071173
Submission received: 7 June 2018 / Revised: 12 July 2018 / Accepted: 13 July 2018 / Published: 18 July 2018

Abstract

:
Tolerance-rough-set-based classifiers (TRSCs) are known to operate effectively on real-valued attributes for classification problems. This involves creating a tolerance relation that is defined by a distance function to estimate proximity between any pair of patterns. To improve the classification performance of the TRSC, distance may not be an appropriate means of estimating similarity. As certain relations hold among the patterns, it is interesting to consider similarity from the perspective of these relations. Thus, this study uses grey relational analysis to identify direct influences by generating a total influence matrix to verify the interdependence among patterns. In particular, to maintain the balance between a direct and a total influence matrix, an aggregated influence matrix is proposed to form the basis for the proposed grey-total-influence-based tolerance rough set (GTI-TRS) for pattern classification. A real-valued genetic algorithm is designed to generate the grey tolerance class of a pattern to yield high classification accuracy. The results of experiments showed that the classification accuracy obtained by the proposed method was comparable to those obtained by other rough-set-based methods.

1. Introduction

Rough set theory [1,2] can effectively deal with the formulation of vague concepts [3,4,5,6,7,8,9,10]. Traditional rough-set-based methods use discretization methods to deal with quantitative attributes. Given that no one method of discretization is optimal, the tolerance rough set (TRS) has proven effective in dealing with problems involving numerical attributes [11,12,13]. Several TRS-based classifiers (TRSCs) use the traditional TRS with a simple distance measure to estimate similarity between any pair of patterns, viewing each category in a classification problem as a concept [11,12,13,14,15,16,17,18,19,20,21]. To expand the applicability of TRSCs, instead of the simple distance measure, several measures such as flows [22] and relationships [19,23] among patterns have been proposed to improve classification performance.
Direct relationships have been effectively measured in grey TRSC [19] through grey relational analysis (GRA) from the viewpoint of relationships obtained between patterns. However, direct as well as indirect relationships can exist between patterns. The widely used Decision-Making Trial and Evaluation Laboratory (DEMATEL) can effectively verify interdependencies among patterns or variables [20,21,22]. Furthermore, the total influence matrix plays an important role in the DEMATEL, and can be used to indicate direct/indirect influences among patterns. This is why the DEMATEL has been widely applied to various decision problems [24,25]. This motivated us to use the total influence matrix to realize direct/indirect relationships when constructing a TRS-based classifier. This article thus proposes a grey-total-influence-based TRS (GTI-TRS) for pattern classification. Furthermore, a genetic algorithm is used to determine the parameters required to construct the proposed classifier with high classification accuracy.
The remainder of this paper is organized as follows. Section 2 introduces a traditional similarity measure for the TRS and its computation steps. In Section 3, we present the proposed grey total influence matrix and GTI-TRS for pattern classification. In Section 4, we provide a genetic algorithm to construct the proposed GTI-TRS-based classifier (GTI-TRSC). Some real-world datasets were used to determine the classification accuracy of the proposed method. The experiments and results described in Section 5 show that the proposed GTI-TRSC can perform well compared to rough-set-based methods considered. Section 6 contains a discussion of the results and the conclusions of this study.

2. Tolerance Rough Sets

The rough set is briefly introduced in Section 2.1. TRS with a similarity measure is described in Section 2.2. In Section 2.3, we detail the classification procedure for the TRSC.

2.1. Rough Set Theory

Uncertainty and vagueness can be handled by rough set theory. Let S = (U, A ⋃ {d}) be a decision table, where U, A, and D are nonempty finite sets. U is the universe of objects, A is a set of conditional attributes, and dA is a decision attribute. An information function a: UVa can be defined for aA, where Va is the set of values of a called the domain of a. An indiscernibility relation Ind(P) is defined for any PA:
Ind ( P ) = { ( x i , x j ) U 2 | a ( x i ) = a ( x j ) , a P }  
where xi and xj are indiscernible if (xi, xj) belongs to Ind(P). Ind(P) is called the P-indiscernibility relation, and its equivalence classes are denoted by [x]P = {yU | a(x) = a(y), ∀aP}. A P-definable set denotes any finite union of elementary sets [26]. In a classification problem, a concept X is composed of elements with the same class label such that XU/{d}.
Sometimes, XU is not P-definable. If X is a vague concept, a pair of precise concepts, P-upper ( P ¯ X) and P-lower approximations ( P ¯ X), can be used to approximate X [1,8]:
P ¯ X = { x | x U ,   [ x ] P X ϕ }  
P ¯ X = { x | x U , [ x ] P X }  
It is clear that P ¯ X ⊆ P ¯ X. Elements that certainly belong to X constitute P ¯ X, whereas those possibly belonging to X constitute P ¯ X. The tuple 〈 P ¯ X, P ¯ X〉 is called a rough set, and P ¯ X and P ¯ X are singleton approximations. When P ¯ X = P ¯ X, X is definable because X is precise with respect to P. In contrast, P ¯ X ≠ P ¯ X means that X is undefinable.
A boundary region BNDP(X) defined for a vague concept X is as follows:
B N D P ( X ) = P ¯ X P ¯ X  
A rough membership function defines the degree of inclusion of x within X with respect to P:
μ X P ( x )   = | [ x ] P X | | [ x ] P |  
where μ X P ( x ) ∊ [0, 1], and |[x]P| denotes the cardinality of [x]P.

2.2. Traditional Similarity Measure

Let m and n denote the numbers of patterns and attributes, respectively. xi and xj (1 ≤ i, jm) are some objects in U. A simple distance function can be defined by a similarity measure Sa(xi, xj) to measure the closeness between a(xi) and a(xj) as in [14,15]:
S a ( x i , x j )   = 1 | a ( x i ) a ( x j ) | d m a x  
where a(xi) and a(xj) are attribute values of xi and xj, respectively, in Va, and dmax is the maximum value among |a(xi) − a(xj)|. dmax can be replaced by (maxa − mina) [17], where maxa and mina are the maximum and minimum values, respectively, of the domain interval of a [12]. Then, xi and xj are similar with respect to a when Sa(xi, xj) ≥ ta, where ta ∊ [0, 1] is the similarity threshold with respect to a. The tolerance relation Ra has a relation with Sa(xi, xj):
a(xi) Ra a(xj) ⇔ Sa(xi, xj) ≥ ta
This means that xi and xj are similar with respect to attribute a when a(xi)Ra a(xj). SA(xi, xj), an overall similarity measure, can be further defined as:
S A ( x i , x j )   = a A S a ( x i , x j ) | A |  
where |A| is the number of attributes in A, and |A| = n here. Kim and Bang [14] used xi τA xj to denote the above similarity between objects xi and xj with respect to all attributes A. As a result, the tolerance relation τA can be related to SA as
xi τA xjSA(xi, xj) ≥ tA
where tA ∊ [0, 1] is a similarity threshold based on A.
Patterns that have a tolerance relation with xi form a tolerance class TC(xi) of xi:
TC(xi) = {xjU|xi τA xj}
X can be approximated by the lower ( τ A _ X) and upper approximations ( τ A ¯ X). For subset approximations, τ A _ X and τ A ¯ X are defined as [26]:
τ A _ X = { T C ( x ) | x U , T C ( x ) X }  
τ A ¯ X = { T C ( x ) | x U , T C ( x ) X φ }  
For concept approximations, τ A _ X and τ A ¯ X are defined as:
τ A _ X = { T C ( x ) | x X , T C ( x ) X }  
τ A ¯ X = { T C ( x ) | x X , T C ( x ) X φ }  
The tuple 〈 τ A _ X, τ A ¯ X〉 is the tolerance rough set. The main difference between these two approximations is associated with objects belonging to U or X.

2.3. Computational Steps of a TRS-Based Classifier

The computational steps of a TRSC [14,15] can be briefly described as follows:
Step 1.
Determine 〈 τ A _ TC(x), τ A ¯ TC(x)〉
With x, τ A _ TC(x) is composed of patterns certainly similar to x, and τ A ¯ TC(x) is composed of patterns possibly similar to x. For subset and concept approximations, τ A _ TC(x) is identical to TC(x), but τ A ¯ TC(x) is not.
Step 2.
Classification using lower approximations
If τ A _ TC(x) = {x}, the classification of x can be left to the next step. If the cardinality of τ A _ TC(x) is at least two, τ A _ TC(x) − {x} is used to determine the relative frequency of the class inclusion of the training patterns in τ A _ TC(x) − {x}. Then, x can be assigned to the class with the highest relative frequency by majority vote. However, if the highest relative frequency is not unique, the classification of x can be left until the next step.
Step 3.
Classification using upper approximations
The boundary region BNDA(TC(x)) ( τ A ¯ TC(x) − τ A _ TC(x)) of x can be used to determine the class label of x. Assume that patterns belonging to class Ci constitute Xi. With y in BNDA(TC(x)) ≠ φ, the rough membership function denoted by μ C i ( y ) defined as:
μ C i ( y ) = | T C ( y ) X i | | T C ( y ) |  
where |TC(y)| denotes the cardinality of TC(y). Then, the average rough membership function of x regarding Ci is computed as:
μ ¯ C i ( x ) = 1 m y BND A ( TC ( x ) ) μ C i ( y )  
where m is the cardinality of BNDA(TC(x)). x can be assigned to a class with the largest degree of average rough membership. However, the class label of x cannot be confirmed if BNDA(TC(x)) = φ.

3. Grey-Total-Influence-Based Tolerance Rough Sets

The proposed GTI-TRS plays an important role in designing the proposed classifier. Thus, related studies with respect to the measurement of total influence are first described in Section 3.1. Three main components constitute the GTI-TRS: the GRA introduced in Section 3.2, the proposed grey total influence presented in Section 3.3, and the GTI-based tolerance relation described in Section 3.4.

3.1. Studies Related to Measuring Total Influence

Pattern classification refers to the problem of partitioning a pattern space into classes and assigning a pattern to one of them [15]. As mentioned above, to improve the classification performance of the TRSC, this study addresses direct relationships measured in the grey TRSC [19]. The main issue addressed is that as a pattern is likely to influence another directly and/or indirectly, indirect relationships among patterns should be studied as well. To develop novel similarity measures for the TRS by means of relationships, this study focuses on ways of leveraging direct relationships among patterns measured by GRA to further generate the total influence, consisting of indirect and direct relationships, for a TRSC. It should be noted that many studies (e.g., [19,27,28,29]) have shown the effectiveness of GRA in measuring relationships among attributes and patterns.
The use of grey theory to measure direct influences among patterns for the DEMATEL has been addressed in recent studies, such as [30,31,32,33,34,35,36]. These studies derived the total influence matrix by representing a direct influence as a grey number [27]. As the proposed method derives the total influence matrix by using GRA to automatically generate direct influences from crisp-valued data, the focus is completely different from that of the aforementioned grey DEMATEL methods.

3.2. Grey Relational Analysis

Unlike statistical correlation measuring the relationship between random variables, GRA explores relationships between data sequences [27,29] by treating one of these sequences as the goal [28]. Assume that n denotes the number of attributes. Let xj = (xj1, xj2, …, xjn) (1 ≤ jm) be a comparative sequence and xi = (xi1, xi2, …, xin) (1 ≤ im) be a reference sequence. The grey relational coefficient ξk(xi, xj) can be used to measure the relationship between these two sequences on attribute k (1 ≤ kn) [37]:
ξ k ( x i , x j )   = Δ min + ρ Δ max Δ i j k + ρ Δ max  
where
Δ m i n = min s min l | x j l x s l | ,   1   s m ,   1   l n  
Δ max = max s max l | x j l x s l | ,   1   s m ,   1   l n  
Δijk = |xikxjk|
where ρ is a discriminative coefficient, commonly specified as 0.5 [29], but this is apparently not an optimal setting. ξk(xi, xj) falls somewhere between zero and one.
The grey relational grade (GRG) ϒ(xi, xj) can be used to measure the overall relationship between xi and xj:
ϒ ( x i , x j ) = k = 1 n w k ξ k ( x i , x j )  
where ϒ(xi, xj) ∊ [0, 1]. wk denotes the relative importance of attribute k, and w1, w2, …, wn satisfy
j = 1 n w j =   1  

3.3. Determining Grey Total Influence

3.3.1. Generating a Direct Influence Matrix Using GRA

The total influence matrix in the DEMATEL can be used to indicate causal relationships among patterns. Prior to obtaining the total influence matrix T = [tij]m×m, a direct influence matrix, Z = [zij]m×m is constructed, where zij (1 ≤ i, jm) represents the extent to which xi influences xj. The values of zero and one represent “no effect” and “very strong effect,” respectively, when zij ranges from zero to one. The higher the value of zij, the more xi is likely to directly influence xj.
In particular, as zij represents the impact of xi on xj, it is reasonable to attribute such an impact to a relationship between xi and xj. This implies that the stronger the relationship between xi and xj, the greater the direct impact of xi on xj. As GRA is an appropriate technique to identify relationships among patterns, this inspired us to determine the grey total influence using GRA to generate the total influence matrix. Compared with the traditional method, the distinctive feature of determining the grey total influence is that it can automatically determine the impact z1v, z2v, …, zuv of x1, x2, …, and xu on xv (1 ≤ u, vm), respectively, at a time by means of GRA, when x1, x2, …, xu act as comparative sequences, and xv is a reference sequence such that zsv = ϒ(xs, xv) (1 ≤ su). To obtain Z, x1, x2, …, and xu act as reference sequences in turn. Thus, this study calls Z a grey direct influence matrix.

3.3.2. Generating a Grey Direct Influence Matrix for Pattern Classification

For a multiclass problem, to verify the performance of a classifier, it is necessary to partition the collected patterns into training (Category 1) and testing (Category 2) data. That is, each pattern (e.g., xi, xj) can be categorized into either Category 1 or 2. As a result, four matrix segments, Z11, Z12, Z21, and Z22 make up a partitioned matrix Z, where each matrix segment represents a relationship between categories in a classification system:
Z = [ Z 11 Z 12 Z 21 Z 22 ] = [ z ( x 11 , x 11 ) z ( x 11 , x 12 ) z ( x 11 , x 1 m 1 ) z ( x 11 , x 21 ) z ( x 11 , x 22 ) z ( x 11 , x 1 m 2 ) z ( x 12 , x 11 ) z ( x 12 , x 12 ) z ( x 12 , x 1 m 1 ) z ( x 12 , x 21 ) z ( x 12 , x 22 ) z ( x 12 , x 1 m 2 ) z ( x 1 m 1 , x 11 ) z ( x 1 m 1 , x 12 ) z ( x 1 m 1 , x 1 m 1 ) z ( x 1 m 1 , x 21 ) z ( x 1 m 1 , x 22 ) z ( x 1 m 1 , x 1 m 2 ) z ( x 21 , x 11 ) z ( x 21 , x 12 ) z ( x 21 , x 1 m 1 ) z ( x 21 , x 21 ) z ( x 21 , x 22 ) z ( x 21 , x 1 m 2 ) z ( x 22 , x 11 ) z ( x 22 , x 12 ) z ( x 22 , x 1 m 1 ) z ( x 22 , x 21 ) z ( x 22 , x 22 ) z ( x 22 , x 1 m 2 ) z ( x 2 m 2 , x 11 ) z ( x 2 m 2 , x 12 ) z ( x 2 m 2 , x 1 m 1 ) z ( x 2 m 2 , x 21 ) z ( x 2 m 2 , x 22 ) z ( x 2 m 2 , x 1 m 2 ) ]
Z11 and Z22 describe the inner impacts of the patterns on those in Categories 1 and 2, respectively, whereas Z12 and Z21, respectively, describe the outer impacts of the imposition of Category 1 on Category 2, and vice versa. Let pattern p in Category 1, represented by x1p = (x1pi1, x1p2,…, x1pn) (1 ≤ pm1), be a reference pattern, and pattern q in Category 2, x2q, represented by (x2q1, x2q2,…, x2qn) (1 ≤ qm2), be a comparative pattern, where m1 and m2 denote the number of patterns in Categories 1 and 2, respectively. Therefore, m1 + m2 = m. Z11, Z12, Z21, and Z22 are derived as follows:
(1)
Z11: z(x1l, x1p) (1 ≤ lm1) is obtained using x11, x12, …, x 1 m 1 as comparative sequences and x1i as a reference sequence, so that z(x1l, x1p) = ϒ(x1l, x1p).
(2)
Z12: z(x1p, x2q) is obtained using x11, x12, …, x 1 m 1 as comparative sequences and x2j as a reference sequence, so that z(x1p, x2q) = ϒ(x1p, x2q).
(3)
Z21: As the testing patterns are unseen by the training patterns, they do not have any impact on the training patterns. Therefore, z(x2q, x1p) is set to zero, so that Z21 = 0.
(4)
Z22: As the testing patterns are unseen, they do not have any impact on themselves. Therefore, z(x2k, x2q) (1 ≤ km2) is set to zero, so that Z22 = 0.
Of course, the symmetry of Z is not required. Note that during the training phase, only training patterns are considered, Z12 = 0.

3.3.3. Generating a Grey Total Influence Matrix

All diagonal elements of Z should first be set to zero [24]. Z is in turn normalized to produce a normalized direct influence matrix X:
X = λZ
where
λ = 1 max i , j { max i j = 1 m z i j , max j i = 1 m z i j }  
Finally, the grey total influence matrix T = [tij]m×m can be further generated by X(IX)−1. Unlike Z, T considers not only direct, but also indirect relationships for each pair of patterns. As Z and T express different prospects regarding the impact of xi on xj, considering zij and tij, it is reasonable to aggregate Z and T into a new hybrid matrix G = [gij]m×m. By balancing Z and T, G is defined as follows:
G = αZ + (1 − α)T
This means that
gij = αzij + (1 − α)tij
where 0 ≤ α ≤ 1, and the relative importance of the two items is measured by α. gij tends to affect direct influence when α > 0.5 and total influence when α < 0.5. The higher the value of gij, the greater the degree to which xi influences xj.

3.4. Grey-Total-Influence-Based Tolerance Relation

The overall relationship index (gij) forms the foundation of the proposed grey-total-influence-based tolerance rough set (GTI-TRS). S A G T I (xi, xj), an overall GTI-based similarity measure, is defined as
S A G T I ( x i , x j )   = g i j  
Let xi τ A G T I xj denote that xi and xj are similar regarding A, where τ A G T I is called a GTI-based tolerance relation with respect to all attributes A. τ A G T I can be related with S A G T I as
x i τ A G T I x j S A G T I ( x i , x j )   t A G T I  
where t A G T I ∊ [0, 1] is a similarity threshold based on A.
GTI-TC(xi), a GTI-based tolerance class of xi, can be generated by considering patterns that have a GTI-based tolerance relation with xi:
GTI-TC ( x i )   =   { x j U | x i τ A G T I x j }  
The higher the value of tij, the more likely it is that xj can be included in GTI-TC(xi). The lower ( τ A G T I _ X) and upper approximations ( τ A G T I ¯ X) of X can be determined by subset and concept approximations by replacing τ A _ , τ A ¯ , and TC(x) with τ A G T I _ , τ A G T I ¯ , and GTI-TC(x), respectively. 〈 τ A G T I _ X, τ A G T I ¯ X〉 is called a GTI-TRS. As shown in Figure 1, by finding the grey total influence, the proposed GTI-TRSC can be set up by merging the proposed classifier with the computational steps of the TRSC.

3.5. Illustrative Example

To explain the generation of a grey direct influence matrix and its total influence matrix during the training and testing phases, a small decision table is shown in Table 1, where x1, x2, and x3 are training patterns, and x4, x5, and x6 are used for testing. In a practical problem, such as bankruptcy prediction, each pattern may be a firm, and conditional attributes may be explanatory financial ratios. Each pattern is composed of four real-valued conditional attributes. Let ρ be 0.5 and wk be ¼ (1 ≤ k ≤ 4).

3.5.1. Training Phase

During the training phase, only Z11 needs to be computed by the training patterns. If we use x1 as the reference pattern, Δmax and Δmin are 3.1 and zero, respectively. For ϒ(x2, x1), ξ1(x2, x1), ξ2(x2, x1), ξ3(x2, x1), and ξ4(x2, x1), ξ1(x2, x1) can be computed as:
ξ 1 ( x 2 , x 1 ) = 0 + 0.5 × 3.1 1.9 + 0.5 × 3.1 = 0.449  
Moreover, ξ2(x2, x1), ξ3(x2, x1), and ξ4(x2, x1) are 0.463, 0.608, and 1.0, respectively. ϒ(x2, x1) can thus be computed as:
ϒ(x2, x1) = ¼(0.449 + 0.463 + 0.608 + 1) = 0.630
In a similar way, ϒ(x3, x1) can be computed as 0.589. Finally, Z11 is as follows:
Z 11 = [ 0 0.712 0.671 0.630 0 0.558 0.589 0.558 0 ]  
During the training phase, the grey direct influence matrix Z is generated as follows:
Z = [ 0 0.712 0.671 0 0 0 0.630 0 0.558 0 0 0 0.589 0.558 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]  
As λ = 1.383 occurs in the first row of Z, the normalized matrix X derived from Z is:
X = [ 0 0.515 0.486 0 0 0 0.456 0 0.403 0 0 0 0.426 0.403 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]  
The grey total influence matrix T can be easily generated by X(IX)−1:
T = [ 2.833 3.253 3.172 0 0 0 2.872 2.632 2.859 0 0 0 2.791 2.851 2.505 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]  

3.5.2. Testing Phase

During the testing phase, Z11 and Z12 need to be generated. For Z12, when we use x4 as the reference pattern, Δmax and Δmin are 27 and zero, respectively. For ϒ(x1, x4), ξ1(x1, x4) can be computed as follows:
ξ 1 ( x 1 , x 4 ) = 0 + 0.5 × 27 0.9 + 0.5 × 27 = 0.938  
Furthermore, ξ2(x1, x4), ξ3(x1, x4), and ξ4(x1, x4) are 0.622, 0.351, and 0.517, respectively. ϒ(x1, x4) can thus be computed as:
ϒ(x1, x4) = ¼(0.938 + 0.622 + 0.351 + 0.517) = 0.607
In a similar manner, ϒ(x2, x4) and ϒ(x3, x4) can be computed as 0.596 and 0.612, respectively. Finally, Z12 is as follows:
Z 12 = [ 0.607 0.572 0.559 0.596 0.574 0.571 0.612 0.594 0.567 ]  
Therefore, the grey direct influence matrix Z is generated as follows:
Z = [ 0 0.712 0.671 0.607 0.572 0.559 0.630 0 0.558 0.596 0.574 0.571 0.589 0.558 0 0.612 0.612 0.567 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]  
As λ = 3.121 occurs in the first row of Z, the normalized matrix X derived from Z is:
X = [ 0 0.228 0.215 0.194 0.183 0.179 0.202 0 0.179 0.191 0.184 0.183 0.189 0.179 0 0.196 0.190 0.182 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]  
The grey total influence matrix T can be easily generated by X(IX)−1:
T = [ 0.118 0.308 0.295 0.334 0.318 0.310 0.272 0.108 0.257 0.315 0.302 0.298 0.260 0.256 0.102 0.315 0.304 0.294 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]  

4. Genetic-Algorithm-Based Learning Algorithm

Basic genetic operations such as selection, crossover, and mutation [37,38,39] are involved in construction of the proposed GTI-TRSC. To construct a GTI-TRSC with high classification accuracy, n + 3 parameters—w1, w2, …, wn, ρ, α, and τ A G —constituting a chromosome were determined by a real-valued genetic algorithm (GA). The related parameters were the probability of crossover Prc, probability of mutation Prm, total number of generations nmax, population size nsize, and the number of elite chromosomes ndel (0 ≤ ndelnsize). The pseudo-code of the learning algorithm is as follows Algorithm 1:
Algorithm 1 The pseudo-code of the learning algorithm
  Set 0 to k; //1 ≤ knmax
  Initialize population (k, nsize);
  Evaluate chromosomes (k, nsize);
  While not satisfying the stopping rule do
  Set k + 1 to k;
  Select (k, nsize); //Select generation k from generation k − 1
  Crossover (k, nsize);
  Mutation (k, nsize);
  Elitist (k, nsize);
  Evaluate chromosome (k, nsize);
  End while
The function of each operation is as follows:
(1)
Initialize population: The most common population size is between 50 and 500. Generate an initial population of nsize chromosome. Each parameter in a chromosome is assigned a real random value ranging from zero to one.
(2)
Evaluate chromosomes: Each chromosome corresponds to a GTI-TRSC that can be generated by the process shown in Figure 1. For each pattern, determine the lower and upper approximations for a GTI-based tolerance class. Furthermore, correct classification serves as a fitness function. Classification accuracy is the number of correct predictions made divided by the total number of predictions made, multiplied by 100 to turn it into a percentage.
(3)
Select: To produce generation k, randomly select two chromosomes from generation k − 1 by a binary tournament and place the one with higher fitness in a mating pool.
(4)
Crossover: Let w i 1 k w i 2 k w i n k ρ i k α i k τ i k   and   w j 1 k w j 2 k w j n k ρ j k α j k τ j k be randomly selected chromosomes (1 ≤ i, jnsize) from generation k. Prc determines whether crossover can be performed on any two real-valued parameters. Two new chromosomes, w i 1 k n e w w i 2 k n e w w i n k n e w ρ i k n e w α i k n e w τ i k n e w   and   w j 1 k n e w w j 2 k n e w w j n k n e w ρ j k n e w α j k n e w τ j k n e w are generated and are added into Pk+1. The related crossover operations are performed as:
w i w k n e w = a w w i w k + ( 1 a w ) w j w k ,   w j w k n e w =   ( 1 a w ) w i w k + a w w j w k ( 1   w n ) ρ i k n e w = b ρ i k +   ( 1 b ) ρ j k ,   ρ i k n e w =   ( 1 b ) ρ i k + b ρ j k α i k n e w = c α i k +   ( 1 c ) α j k ,   α i k n e w =   ( 1 b ) α i k + b α j k τ i k n e w = d τ i k +   ( 1 d ) τ j k ,   τ j k n e w =   ( 1 c ) τ i k + c τ j k
where aw, b, c, and d are random numbers ranging from zero to one.
(5)
Mutation: Prm determines whether a mutation can be performed on each real-valued parameter of a newly generated chromosome. With a mutation, the affected gene is altered by adding a random number selected from a prespecified interval, such as (−0.01, 0.01). A smaller Prm is required to avoid excessive perturbation.
(6)
Elitist strategy: Randomly remove ndel chromosomes from generation k. Insert ndel chromosomes with the maximum fitness from generation k − 1. A smaller ndel is required to generate a smaller perturbation in generation k.
(7)
Stopping rule: When nmax generations have been created, the algorithm reaches the stopping condition.
When the algorithm is terminated, the chromosome with the maximum fitness among all successive generations can be used to examine the generalization capability of the proposed GTI-TRSC.

5. Computer Simulations

As shown in Table 2, the generalization capability of the proposed GTI-TRSC was examined by experiments on some practical datasets available from the UCI Machine Learning Repository (http://www.ics.uci.edu/~mlearn/MLRepository.html). In Section 5.1, the performance of different rough-set-based classification methods is reported. Section 5.2 describes a statistical analysis to compare the different rough-set-based methods considered.

5.1. Evaluating Classification Performance

There is no optimal setting for parameter specifications for genetic algorithms, but we can refer to the principles introduced in [40,41]. Parameter specifications were specified for all experiments as follows: nsize = 50, nmax = 500, ndel = 2, Prc = 0.8, and Prm = 0.01. Five-fold cross-validation (5-CV) was considered for each classification method ten times by means of a distribution-balanced stratified CV (DBSCV) [42]. This study divided all patterns into five disjoint subsets of equal size such that four served as training patterns and one as test data. We iterated this procedure until each subset had been tested.
The classification performance of the proposed GTI-TRSC was compared with that of several representative rough-set-based classification methods: a rule-based method with shortening optimization (RSES-O) using the Rough Set Exploration System (RSES) [3,4,5], a hierarchical version of the lattice machine (HLM) [43,44], a hierarchical form of RSES-O (RSES-H) [44,45], and the Rule Induction with Optimal Neighborhood Algorithm (RIONA) [46]. These classification methods are briefly introduced as follows:
(1)
HLM: The lattice machine generates hypertuples as a model of the data. Some more general hypertuples can be used in the hierarchy that covers objects covered by the hypertuples. The covering hypertuples locate various levels of the hierarchy.
(2)
RSES-O: RSES-O is implemented in RSES. An optimal threshold for the positive region is used to shorten all decision rules with a minimal number of descriptors.
(3)
RSES-H: RSES-H can be obtained by constructing a hierarchy of rule-based classifiers. The levels of the hierarchy are defined by different levels of minimal rule shortening. A new pattern can be classified by a single hierarchy of the classifier.
(4)
RIONA: RIONA is also implemented in RSES. It uses the nearest neighbor method to induce distance-based rules. For a new pattern, the patterns most similar to it can vote for its decision, but patterns that do not match any rule are excluded from voting.
The classification performance of the above methods, reported in [44], is summarized in Table 3.
Variants of TRSC were also considered: TRSC with subset approximations (TRSC-SU), TRSC with concept approximations (TRSC-CO), flow-based TRSC (FTRSC) with subset approximations (FTRSC-SU), flow-based TRSC with concept approximations (FTRSC-CO) [22], Grey-tolerance-rough- set-based classifier (GTRSC) with subset approximations (GTRSC-SU), and GTRSC with concept approximations (GTRSC-CO) [19]. The GTRSC and FTRSC were chosen because it is interesting to investigate whether different measures of similarity or relationship can influence classification accuracy. To implement the basic differences between GTI-TRSC, GTRSC, and FTRSC, GTRSC and FTRSC are briefly introduced as follows:
(1)
GTRSC: Instead of a simple distance measure used to evaluate the proximity of any two patterns, the GRG (grey relational grade) is used here to implement a relationship-based similarity measure that generates a tolerance class for each pattern. As mentioned above, only direct relationships were considered in the GTRSC.
(2)
FTRSC: The FTRSC uses preference information expressed by flows among patterns to measure similarity between patterns. The flow of each pattern is computed by the well-known preference ranking organization methods for enrichment evaluations (PROMETHEE) [47,48].
The results are summarized in Table 3 and Table 4. It is clear that the classification performance of GTI-TRSC was comparable to that of FTRSC and GTRSC. This means that different measures can impose a certain impact on classification results.

5.2. Statistical Analysis

The nonparametric Friedman test [49] was used to statistically analyze the aforementioned classification methods. Using the null hypothesis, whereby the ranks of the classification methods were identical on average, the FF statistic, distributed as an F distribution with k1 − 1 and (k1 − 1) (k2 − 1) degrees of freedom, can be defined as follows [20]:
F F = ( k 2 1 ) χ F 2 k 2 ( k 1 1 ) χ F 2  
where rj, k1, and k2 are the average rank of method j, the number of methods, and the number of datasets considered, respectively. χ F 2 is defined as
χ F 2 = 12 k 2 k 1 ( k 1 + 1 ) [ j = 1 k 1 r j 2 k 1 ( k 1 + 1 ) 2 4 ]  
FF is 14.09 because k1 = 12, k2 = 10, and χ F 2 = 67.13. The null hypothesis was rejected at the 5% level as FF was above the critical value of 1.98 (i.e., F(9, 99)).
Subsequently, the Nemenyi test [50] was used to detect significant differences among the classification methods. The classification accuracies of the two methods were significantly different and the differences in their average ranks were greater than CD:
C D = q α k 1 ( k 1 + 1 ) 6 k 2  
where CD is a critical difference, and CD = 4.89 because q0.10 = 3.03 at the 10% level. We summarize the results as follows:
(1)
GTI-TRSC-SU significantly outperformed TRSC-CO (9.15 − 2 = 7.15), TRSC-SU (9.55 − 2 = 7.55), RSES-H (7.90 − 2 = 5.90), RSES-O (8.80 − 2 = 6.00), RIONA (8.80 − 2 = 6.80), and HLM (9.20 − 2 = 7.2).
(2)
GTI-TRSC-CO significantly outperformed TRSC-CO (9.15 − 1.80 = 7.35), TRSC-SU (9.55 − 1.80 = 7.75), RSES-H (7.90 − 1.80 = 6.10), RSES-O (8.80 − 1.80 = 6.20), RIONA (8.80 − 1.80 = 7.00), and HLM (9.20 − 1.80 = 7.40).
(3)
There was no significant difference between GTI-TRSC and GTRSC for both set and concept approximations. Even so, GTI-TRSC outperformed GTRSC on seven out of ten datasets.
(4)
Although GTI-TRSC did not significantly outperform the FTRSC, the difference between GTI-TRSC-CO and FTRSC-SU was slightly less than CD (6.35 − 1.80 = 4.50). Therefore, it is reasonable to conclude that GTI-TRSC-CO was superior to FTRSC-SU. Even so, it is interesting to investigate the applications that can render GTI-TRSC and FTRSC significantly different.

6. Discussion and Conclusions

From the perspective of numerous or few relationships between any pair of patterns, this study used GRA to identify direct influences among patterns. A total influence matrix was generated to verify the direct/indirect influences among them. The total influence formed the foundation of the proposed GTI-TRS associated with the construction of the TRSC. The idea is that the higher the value of gij, the more similar xi is to xj. We noted that [19] had proposed grey tolerance rough sets (GTRS), which defined an overall similarity measure S A G (xi, xj), so that S A G (xi, xj) = ϒ(xi, xj). Therefore, the main difference between the GTRS and the GTI-TRS is that the former considers only direct relationships, but the latter considers direct as well as indirect relationships using the proposed grey total influence matrix. In particular, a GA was implemented to determine the optimal parameter specification for the proposed GTI-TRSC that cannot be easily determined by users.
Even though parameter specifications for GA are subjective, experimental results showed that the chosen parameters were acceptable. Indeed, the proposed GTI-TRSC is sufficiently simple to implement as a computer program without conforming to any statistical assumptions. Experimental results obtained by the proposed GTI-TRSC are promising. We see that GTI-TRSC-CO and GTI-TRSC-SU produced satisfactory results compared with the other rough-set-based classification methods considered. In particular, these two classifiers were superior in terms of classification performance to the TRSC. However, it should be noted that there is no best classifier [42].
This study has motivated us to investigate the subject further. As mentioned above, the grey DEMATEL has been an important issue for multiple attribute decision making. It is interesting to examine a distinctive version of the grey DEMATEL built on the proposed grey direct influence matrix. First, a novel DEMATEL-based Analytic Network Process (DANP) proposed in [20] was used to avoid agonizing pairwise comparisons for the ANP by replacing the total influence matrix produced by the DEMATEL by directly using the unweighted supermatrix of the ANP. The DANP has gained considerable research attention in recent years due to its convenience [21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52]. It is interesting to explore its applicability to practical problems by incorporating the new version of the grey DEMATEL into DANP.
Second, considering decision-making in a fuzzy environment, the linguistic interpretation of elements of the grey direct influence matrix is challenging. In practical applications, it may be desirable that a linguistic term be generated from numerical data [41]. Linguistic values such as “low influence”, “medium influence”, and “high influence” associated with fuzzy sets can be considered. Such a linguistic grey direct influence matrix can then be incorporated into the DEMATEL for further processing.
Finally, the traditional GRG is implemented using the weighted-average method, where dependency among the attributes is not considered. However, an assumption of additivity may not be realistic in practical applications [53]. Thus, it is more useful to employ a nonadditive grey total influence matrix with a nonadditive GRG [54], and check the resultant impact on the performance of GTI-TRSC. The aforementioned grey DANP, fuzzy-grey-DEMATEL, and nonadditive grey DEMATEL will be investigated in future studies.

Author Contributions

Conceptualization, Supervision, and Methodology: Yi-Chung Hu; Validation and Investigation: Yu-Jing Chiu.

Funding

Research was funded by the Ministry of Science and Technology, Taiwan, under grant MOST 106-2410-H-033-006-MY2.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pawlak, Z. Rough sets. Int. J. Comput. Inf. Sci. 1982, 11, 341–356. [Google Scholar] [CrossRef]
  2. Pawlak, Z. Rough Sets: Theoretical Aspects of Reasoning about Data; Kluwer: Dordrecht, The Netherlands, 1991. [Google Scholar]
  3. Bazan, J.G.; Szczuka, M. RSES and RSESlib-A collection of tools for rough set computation. In Lecture Notes in Computer Science; Figueira, W., Yao, Y., Eds.; Springer: Berlin, Germany, 2001; pp. 106–113. [Google Scholar]
  4. Bazan, J.G.; Szczuka, M. The rough set exploration system. In Lecture Notes in Computer Science; James, F.P., Andrzej, S., Eds.; Springer: Berlin, Germany, 2005; pp. 37–56. [Google Scholar]
  5. Bazan, J.G.; Szczuka, M.; Wroblewski, J. A new version of rough set exploration system. In Lecture Notes in Computer Science; Alpigini, J.J., Peters, J.F., Skowron, A., Zhong, N., Eds.; Springer: Berlin, Germany, 2002; pp. 397–404. [Google Scholar]
  6. Pawlak, Z.; Skowron, A. Rough sets and boolean reasoning. Inf. Sci. 2007, 177, 41–73. [Google Scholar] [CrossRef]
  7. Pokowski, L. Rough Sets: Mathematical Foundations; Physica-Verlag: Heudelberg, Germany, 2002. [Google Scholar]
  8. Walczak, B.; Massart, D.L. Rough set theory. Chemom. Intell. Lab. Syst. 1999, 47, 1–16. [Google Scholar] [CrossRef]
  9. Zhang, X.; Dai, J.; Yu, Y. On the union and intersection operations of rough sets based on various approximation spaces. Inf. Sci. 2015, 292, 214–229. [Google Scholar] [CrossRef]
  10. Shu, W.; Shen, H. Incremental feature selection based on rough set in dynamic incomplete data. Pattern Recognit. 2014, 47, 3890–3906. [Google Scholar] [CrossRef]
  11. Jensen, R.; Shen, Q. Tolerance-based and fuzzy-rough feature selection. In Proceedings of the IEEE International Conference on Fuzzy Systems (FUZZ-IEEE’07), London, UK, 23–26 July 2007; pp. 877–882. [Google Scholar]
  12. Parthaláin, N.M.; Shen, Q. Exploring the boundary region of tolerance rough sets for feature selection. Pattern Recognit. 2009, 42, 655–667. [Google Scholar] [CrossRef] [Green Version]
  13. Stepaniuk, J. Rough Granular Computing in Knowledge Discovery and Data Mining; Springer: Berlin, Germany, 2008. [Google Scholar]
  14. Kim, D.; Bang, S.Y. A handwritten numeral character classification using tolerant rough set. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 923–937. [Google Scholar]
  15. Kim, D. Data classification based on tolerant rough set. Pattern Recognit. 2001, 34, 1613–1624. [Google Scholar] [CrossRef]
  16. Ma, J.; Hasi, B. Remote sensing data classification using tolerant rough set and neural networks. Sci. China Ser. D Earth Sci. 2005, 48, 2251–2259. [Google Scholar] [CrossRef]
  17. Skowron, A.; Stepaniuk, J. Tolerance approximation spaces. Fund. Inform. 1996, 27, 245–253. [Google Scholar]
  18. Yun, O.; Ma, J. Land cover classification based on tolerant rough set. Int. J. Remote Sens. 2006, 27, 3041–3047. [Google Scholar] [CrossRef]
  19. Hu, Y.C. Pattern classification using grey tolerance rough sets. Kybernetes 2016, 45, 266–281. [Google Scholar] [CrossRef]
  20. Yang, Y.P.O.; Shieh, H.M.; Leu, J.D.; Tzeng, G.H. A novel hybrid MCDM model combined with DEMATEL and ANP with applications. Int. J. Oper. Res. 2008, 5, 160–168. [Google Scholar]
  21. Peng, K.H.; Tzeng, G.H. A hybrid dynamic MADM model for problems-improvement in economics and business. Technol. Econ. Dev. Econ. 2013, 19, 638–660. [Google Scholar] [CrossRef]
  22. Hu, Y.C. Flow-based tolerance rough sets for pattern classification. Appl. Soft Comput. 2015, 27, 322–331. [Google Scholar] [CrossRef]
  23. Hu, Y.C. Tolerance rough sets for pattern classification using multiple grey single-layer perceptrons. Neurocomputing 2016, 179, 144–151. [Google Scholar] [CrossRef]
  24. Tzeng, G.H.; Huang, J.J. Multiple Attribute Decision Making: Methods and Applications; CRC Press: Boca Raton, FL, USA, 2011. [Google Scholar]
  25. Lin, C.L.; Hsieh, M.S.; Tzeng, G.H. Evaluating vehicle telematics system by using a novel MCDM techniques with dependence and feedback. Expert Syst. Appl. 2010, 37, 6723–6736. [Google Scholar] [CrossRef]
  26. Grzymala-Busse, J.W.; Siddhaye, S. Rough set approaches to rule induction from incomplete data. In Proceedings of the International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems, Perugia, Italy, 4–9 July 2004; pp. 923–930. [Google Scholar]
  27. Liu, S.; Lin, Y. Grey Information: Theory and Practical Applications; Springer-Lerlag: London, UK, 2006. [Google Scholar]
  28. Hu, Y.C.; Chen, R.S.; Hsu, Y.T.; Tzeng, G.H. Grey self-organizing feature maps. Neurocomputing 2002, 48, 863–877. [Google Scholar] [CrossRef]
  29. Deng, J.L. Control problems of grey systems. Syst. Control Lett. 1982, 1, 288–294. [Google Scholar]
  30. Bai, C.; Sarkis, J. A grey-based DEMATEL model for evaluating business process management critical success factors. Int. J. Prod. Econ. 2013, 146, 281–292. [Google Scholar] [CrossRef]
  31. Liang, H.W.; Ren, J.Z.; Gao, Z.Q.; Gao, S.Z.; Luo, X.; Dong, L.; Scipioni, A. Identification of critical success factors for sustainable development of biofuel industry in China based on grey decision-making trial and evaluation laboratory (DEMATEL). J. Clean. Prod. 2016, 131, 500–508. [Google Scholar] [CrossRef]
  32. Asad, M.M.; Mohajerani, N.S.; Nourseresh, M. Prioritizing Factors Affecting Customer Satisfaction in the Internet Banking System Based on Cause and Effect Relationships. Procedia Econ. Financ. 2016, 36, 210–219. [Google Scholar] [CrossRef]
  33. Rajesh, R.; Ravi, V. Modeling enablers of supply chain risk mitigation in electronic supply chains: A Grey-DEMATEL approach. Comput. Ind. Eng. 2015, 87, 126–139. [Google Scholar] [CrossRef]
  34. Shao, J.; Taisch, M.; Ortega-Mier, M. A grey-DEcision-MAking Trial and Evaluation Laboratory (DEMATEL) analysis on the barriers between environmentally friendly products and consumers: Practitioners’ viewpoints on the European automobile industry. J. Clean. Prod. 2016, 112, 3185–3194. [Google Scholar] [CrossRef]
  35. Su, C.M.; Horng, D.J.; Tseng, M.L.; Chiu, A.S.F.; Wu, K.J.; Chen, H.P. Improving sustainable supply chain management using a novel hierarchical grey-DEMATEL approach. J. Clean. Prod. 2016, 134, 469–481. [Google Scholar] [CrossRef]
  36. Xia, X.Q.; Govindan, K.; Zhu, Q.H. Analyzing internal barriers for automotive parts remanufacturers in China using grey-DEMATEL approach. J. Clean. Prod. 2015, 87, 811–825. [Google Scholar] [CrossRef]
  37. Goldberg, D.E. Genetic Algorithms in Search, Optimization, and Machine Learning; Addison-Wesley: Boston, MA, USA, 1989. [Google Scholar]
  38. Man, K.F.; Tang, K.S.; Kwong, S. Genetic Algorithms: Concepts and Designs; Springer: London, UK, 1999. [Google Scholar]
  39. Rooij, A.J.F.; Jain, L.C.; Johnson, R.P. Neural Network Training Using Genetic Algorithms; World Scientific: Singapore, 1996. [Google Scholar]
  40. Osyczka, A. Evolutionary Algorithms for Single and Multicriteria Design Optimization; Physica-Verlag: New York, NY, USA, 2002. [Google Scholar]
  41. Ishibuchi, H.; Nakashima, T.; Nii, M. Classification and Modeling with Linguistic Information Granules: Advanced Approaches to Linguistic Data Mining; Springer: Heidelberg, Germany, 2004. [Google Scholar]
  42. Zeng, X.; Martinez, T.R. Distribution-balanced stratified cross-validation for accuracy estimation. J. Exp. Theor. Artif. Intell. 2000, 12, 1–12. [Google Scholar] [CrossRef]
  43. Wang, H.; Düntsch, I.; Gediga, G.; Skowron, A. Hyperrelations in version space. Int. J. Approx. Reason. 2004, 36, 223–241. [Google Scholar] [CrossRef]
  44. Skowron, A.; Wang, H.; Wojna, A.; Bazan, J.G. Multimodal classification: Case studies. In Lecture Notes in Computer Science 4100; Peters, J.F., Skowron, A., Eds.; Springer: Berlin, Germany, 2006; pp. 224–239. [Google Scholar]
  45. Skowron, A.; Wang, H.; Wojna, A.; Bazan, J. A hierarchical approach to multimodal classification. In Lecture Notes in Artificial Intelligence 3642; Slezak, D., Yao, J.T., Peters, J.F., Ziarko, W., Hu, X., Eds.; Springer: Berlin, Germany, 2005; pp. 119–127. [Google Scholar]
  46. Bazan, J.G.; Szczuka, M.; Wojna, A.; Wojnarski, M. On the evolution of rough set exploration system. In Lecture Notes in Artificial Intelligence 3066; Tsumoto, S., Słowiński, R., Komorowski, J., Grzymala-Busse, J.W., Eds.; Springer: Heidelberg, Germany, 2004; pp. 592–601. [Google Scholar]
  47. Brans, J.P.; Marechal, B.; Vincke, P. PROMETHEE: A new family of outranking methods in multicriteria analysis. Oper. Res. 1984, 84, 477–490. [Google Scholar]
  48. Brans, J.P.; Vincke, P.; Marechal, B. How to select and how to rank projects: The PROMETHEE method. Eur. J. Oper. Res. 1986, 24, 228–238. [Google Scholar] [CrossRef]
  49. Friedman, M. A comparison of alternative tests of significance for the problem of m rankings. Ann. Math. Stat. 1940, 11, 86–92. [Google Scholar] [CrossRef]
  50. Nemenyi, P.B. Distribution-Free Multiple Comparisons. Ph.D. Thesis, Princeton University, Princeton, NJ, USA, 1963. [Google Scholar]
  51. Hu, Y.C.; Chiu, Y.J.; Hsu, C.S.; Chang, Y.Y. Identifying key factors of introducing GPS-based fleet management systems to the logistics industry. Math. Probl. Eng. 2015. [Google Scholar] [CrossRef]
  52. Hu, J.W.S.; Hu, Y.C.; Yang, T.P. DEMATEL and analytic network process for evaluating stock trade strategies using Livermore’s key price logic. Univers. J. Account. Finance 2017, 5, 18–35. [Google Scholar]
  53. Wang, W.; Wang, Z.; Klir, G.J. Genetic algorithms for determining fuzzy measures from data. J. Intell. Fuzzy Syst. 1998, 6, 171–183. [Google Scholar]
  54. Hu, Y.C. Nonadditive grey single-layer perceptron with Choquet integral for pattern classification problems using genetic algorithms. Neurocomputing 2008, 72, 332–341. [Google Scholar] [CrossRef]
Figure 1. A flowchart for constructing the proposed classifier (GRA: Grey Relational Analysis, GTI-TRS: grey-total-influence-based tolerance rough set).
Figure 1. A flowchart for constructing the proposed classifier (GRA: Grey Relational Analysis, GTI-TRS: grey-total-influence-based tolerance rough set).
Applsci 08 01173 g001
Table 1. An example decision table.
Table 1. An example decision table.
PatternConditional AttributeDecision Attribute
1234
x151.135.214.02.01
x253.037.015.02.01
x350.032.112.02.02
x452.027.039.014.61
x559.030.042.315.02
x656.725.439.011.02
Table 2. Data used in computer simulations.
Table 2. Data used in computer simulations.
Data# Patterns# Attributes# Classes
Australian approval690142
Glass21496
Hepatitis155192
Iris15043
Pima Indian diabetes76882
Sonar208602
Statlog Heart270132
Tic-Tac-Toe95892
Voting435162
Wine178133
Table 3. Classification accuracy of different classification methods. HLM: A hierarchical version of the lattice machine; RSES-O: a rule-based method with shortening optimization; RSES-H: A hierarchical form of RSES-O; RIONA: Rule Induction with Optimal Neighborhood Algorithm; TRSC: Tolerance-rough-set-based classifiers; SU: Subset approximations; CO: Concept approximations.
Table 3. Classification accuracy of different classification methods. HLM: A hierarchical version of the lattice machine; RSES-O: a rule-based method with shortening optimization; RSES-H: A hierarchical form of RSES-O; RIONA: Rule Induction with Optimal Neighborhood Algorithm; TRSC: Tolerance-rough-set-based classifiers; SU: Subset approximations; CO: Concept approximations.
DatasetClassification Methods
HLMRSES-HRSES-ORIONATRSC-SUTRSC-CO
Australian approval92.087.086.485.785.987.1
Glass71.363.461.266.165.768.1
Hepatitis78.781.982.682.083.983.5
Iris94.195.594.994.495.795.2
Diabetes72.673.873.875.474.173.6
Sonar73.775.374.386.174.375.0
Statlog Heart79.084.083.882.382.983.3
TTT95.099.199.093.682.382.3
Voting95.496.596.495.393.494.0
Wine92.691.290.795.493.095.3
Average rank9.207.908.808.809.559.15
Table 4. Classification accuracy of TRSC variants. FTRSC: Flow-based TRSC; GTRSC: Grey-tolerance-rough- set-based classifier.
Table 4. Classification accuracy of TRSC variants. FTRSC: Flow-based TRSC; GTRSC: Grey-tolerance-rough- set-based classifier.
DatasetClassification Methods
FTRSC-SUFTRSC-COGTRSC-SUGTRSC-COGTI-TRSC-SUGTI-TRSC-CO
Australian approval88.087.789.389.191.090.9
Glass69.169.470.169.979.879.7
Hepatitis85.684.386.087.088.789.8
Iris95.796.296.196.396.396.4
Diabetes75.775.976.576.077.981.6
Sonar78.879.583.082.886.787.8
Statlog Heart83.984.184.484.086.986.1
TTT97.397.898.598.598.998.5
Voting96.696.396.096.297.497.7
Wine93.295.197.497.997.998.1
Average rank6.355.604.404.552.001.80

Share and Cite

MDPI and ACS Style

Hu, Y.-C.; Chiu, Y.-J. Incorporating Grey Total Influence into Tolerance Rough Sets for Classification Problems. Appl. Sci. 2018, 8, 1173. https://doi.org/10.3390/app8071173

AMA Style

Hu Y-C, Chiu Y-J. Incorporating Grey Total Influence into Tolerance Rough Sets for Classification Problems. Applied Sciences. 2018; 8(7):1173. https://doi.org/10.3390/app8071173

Chicago/Turabian Style

Hu, Yi-Chung, and Yu-Jing Chiu. 2018. "Incorporating Grey Total Influence into Tolerance Rough Sets for Classification Problems" Applied Sciences 8, no. 7: 1173. https://doi.org/10.3390/app8071173

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop