Next Article in Journal
Deformed Wavelet Transform and Related Uncertainty Principles
Previous Article in Journal
Pressure Arch Effect of Deeply Buried Symmetrically Distributed Triple Tunnels
Previous Article in Special Issue
Effects of Symmetry Restriction on the Antenna Gain Optimized Using Genetic Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Attribute Reduction Algorithm for Incomplete Information Systems Based on a Binary Similarity Matrix

College of Mathematics and System Sciences, Xinjiang University, Urumqi 830046, China
*
Author to whom correspondence should be addressed.
Symmetry 2023, 15(3), 674; https://doi.org/10.3390/sym15030674
Submission received: 8 February 2023 / Revised: 24 February 2023 / Accepted: 1 March 2023 / Published: 7 March 2023
(This article belongs to the Special Issue Algorithms for Optimization 2022)

Abstract

:
With databases growing at an unrelenting rate, it may be difficult and complex to extract statistics by accessing all of the data in many practical problems. Attribute reduction, as an effective method to remove redundant attributes from massive data, has demonstrated its remarkable capability in simplifying information systems. In this paper, we concentrate on reducing attributes in incomplete information systems. We introduce a novel definition of a binary similarity matrix and present a method to calculate the significance of attributes in correspondence. Secondly, We develop a heuristic attribute reduction algorithm using a binary similarity matrix and attribute significance as heuristic knowledge. In addition, we use a numerical example to showcase the practicality and accuracy of the algorithm. In conclusion, we demonstrate through comparative analysis that our algorithm outperforms some existing attribute reduction methods.

1. Introduction

In the age of information explosion, abundant information rapidly gushes towards people like tidewater, but it is too deficient to adopt traditional knowledge to dispose of the information. How to mine, store and process a large amount of complex data is a critical problem that needs to be solved urgently. Extracting useful data information efficiently from incomplete information systems has become a hot topic in information science and technology areas. Attribute reduction, also known as feature selection, can find out as few attributes as possible to keep the classification of information tables and has been applied to many practical problems [1,2,3,4,5,6,7]. Specifically, Chen et al. [8] investigated feature selection by combining multiple feature selection results and pointed out that a combination of filter (i.e., principal component analysis) and wrapper (i.e., genetic algorithms) techniques by the union method is a better choice. Yuan et al. [9] studied mixed attribute reduction to maintain learning ability without decision information based on fuzzy rough sets. Xie et al. [10] considered the weight of each attribute in information systems with the help of a data binning method and information entropy theory, and further designed a novel attribute reduction method by using weighted neighborhood probabilistic rough sets.
So far, there exist two kinds of algorithms for attribute reduction. One is based on a discernibility matrix proposed by Skowron and Rauszer [11]. The central steps involve obtaining a discernibility matrix and transforming the conjunctive normal form to a disjunctive normal form. The attribute reduction can be obtained directly through the disjunctive normal form. Although the attribute reduction algorithm based on discernibility matrix can acquire all reductions simultaneously, storing the discernibility matrix may require a lot of time and storage space, and the transformation from conjunctive normal form to disjunctive normal form may encounter a combinatorial explosion problem [12]. The other is heuristic algorithm, which has low time complexity and is effective for large data sets, but it can obtain only one reduction at one time. For an information system, obtaining the optimal reduction of attributes has been proved to be an NP-hard problem, which motivates many scholars to devote themselves to heuristic attribute reduction algorithms. There are many types of heuristic knowledge for heuristic attribute reduction algorithms, the most common is attribute significance, which can reflect the importance of an attribute for an information system. Specifically, Shen and Jiang [13] adopted information entropy to calculate attribute significance. Later on, Zhou et al. [14] proposed a method to acquire attribute significance with the help of a similarity matrix. Subsequently, Dai et al. [15] put forward a new conditional entropy and provided its application to attribute reduction. Lately, Hu et al. [16] utilized a dependency relationship of weighted neighborhood rough set (WNRS) model to evaluate attribute significance, and designed a greedy search heuristic algorithm based on WNRS. In addition, the discernibility matrix also behaved well in calculating attribute significance. However, it often requires large storage space during the attribute reduction process. To overcome this defect, many scholars considered a binary discernibility matrix, which is a generalization of the discernibility matrix. Instituto and Nacional [17] defined a more ordered matrix based on the binary discernibility matrix, which allows us to reduce the number of candidate attributes. Ma and Zhang [18] proposed a form of generalized binary discernibility matrix, and further built a new algorithm based on a generalized binary discernibility matrix. Li et al. [19] developed an attribute reduction algorithm based on an improved binary discernibility matrix. Compared with the discernibility matrix, binary discernibility matrix can effectively save storage space and obtain information to calculate attribute significance conveniently. Meanwhile, binary discernibility matrix has the advantages of visual representation and being easy to understand.
As mentioned above, both discernibility matrix and binary discernibility matrix have been applied to attribute reduction algorithms, and binary discernibility matrix outperformed discernibility matrix usually. On the other hand, attribute reduction algorithms based on similarity matrix have been studied systemically. Up to now, there have not been any reduction attribute algorithms based on binary similarity matrix. Therefore, in this paper, we concentrate on proposing a binary similarity matrix and further establish a heuristic attribute reduction algorithm based on the binary similarity matrix for incomplete information systems.
The structure of the paper is as follows. In Section 2, we firstly review some basic definitions about attribute reduction and define binary similarity matrix. In Section 3, we propose a method to calculate attribute significance based on binary similarity matrix and develop an attribute reduction algorithm for incomplete decision information systems. In Section 4, the effectiveness and advantages of the algorithm are demonstrated by some numerical examples. Section 5 concludes the paper.

2. Preliminaries

In this section, we recall some basic definitions about incomplete information systems, tolerance relations and binary similarity matrix.
Definition 1 
([20]). Suppose ( U , A , V , f ) is a quadruplet, U = { x 1 , x 2 , , x n } is a finite set of objects, A = { a 1 , a 2 , , a m } is a set of attributes, f: U × A V is an information function and V = a A V a is a set of values, then S = ( U , A , f , V ) is called an information system. If there exist x U , a A such that f ( x , a ) = , “∗” represents a missing value, then S = ( U , A , f , V ) is an incomplete information system.
Sometimes, S = ( U , A , f , V ) can be abbreviated as S = ( U , A ) while there are no confusions. If A = C D and C D = , then ( U , C D ) is called a decision information system, where C and D are termed as a condition attribute set and decision attribute set, respectively. In many practical applications, D often contains just one attribute, denoted by D = { d } .
Definition 2 
([11]). Suppose S = ( U , A ) is an incomplete information system, C A , for any x i , x j U , tolerance relation R C is defined as:
x i R C x j a C , f ( x i , a ) = f ( x j , a ) or f ( x i , a ) = or f ( x j , a ) = .
It is obvious that R C is reflexive and symmetric, but not necessarily transitive, and U / R C = { R C ( x 1 ) , R C ( x 2 ) , , R C ( x | U | ) } is a cover of U, where R C ( x i ) = { x j | x i R C x j } ( x i , x j U ) .
Definition 3 
([21]). In an incomplete decision information system S = ( U , C D ) , U / R C = { R C ( x 1 ) , R C ( x 2 ) , , R C ( x | U | ) } , U / R D = { R D ( x 1 ) , R D ( x 2 ) , , R D ( x | U | ) } , then U C D = { x | R C ( x ) R D ( x ) } is called the consistent part of S. If U C D = U , then S is consistent. Otherwise, S is inconsistent.
Definition 4 
([22] ). Let S = ( U , C D ) be an incomplete decision information system, T C . If T satisfies the following conditions
(1) 
U T D = U C D ,
(2) 
For any T T , U T D U C D ,
then T is called a reduction of C.
In a decision information system S = ( U , C D ) , the attribute reduction algorithm based on similarity matrix is effective and is simple to implement.
Definition 5 
([14]). For an incomplete decision information system S = ( U , C { d } ) , U = { x 1 , x 2 , , x n } , C = { a 1 , a 2 , , a m } , the similarity matrix M = [ m ( i , j ) ] n × n ( i , j = 1 , 2 , , n ) is defined as
m ( i , j ) = { a C | a ( x i ) = a ( x j ) a ( x i ) = a ( x j ) = } , d ( x i ) d ( x j ) min { | σ C ( x i ) | , | σ C ( x j ) | } = 1 , , e l s e .
where σ C ( x i ) = { d ( y ) | y R C ( x i ) } .
Definition 6 
([14]). In an incomplete decision information system S = ( U , C { d } ) , U = { x 1 , x 2 , , x n } , C = { a 1 , a 2 , , a m } , the significance of attribute a k is
S G F ( a k ) = i = 1 n j = 1 n λ i j / c a r d ( m ( i , j ) ) ,
if a k ( a k C ) m ( i , j ) , then λ i j = 1 , otherwise λ i j = 0 , c a r d ( m ( i , j ) ) represents the number of attributes contained in m ( i , j ) .
In Definition 5, all elements in the similarity matrix are attribute sets and all objects need to be distinguished, so a large space is needed to store the similarity matrix. To save storage space, we define a binary similarity matrix.
Definition 7. 
In an incomplete decision information system S = ( U , C { d } ) , U = { x 1 , x 2 , , x n } , C = { a 1 , a 2 , , a m } . For x i , x j U ( x i x j ) , if d ( x i ) d ( x j ) and min { | σ C ( x i ) | , | σ C ( x j ) | } = 1 , then the binary similarity matrix B S M = [ m ( ( i , j ) , a k ) ] ( a k C ) is defined as,
m ( ( i , j ) , a k ) = 1 , a k ( x i ) = a k ( x j ) a k ( x i ) = a k ( x j ) = , 0 , a k ( x i ) a k ( x j ) .
where σ C ( x i ) = { d ( y ) | y R C ( x i ) } .
In a binary similarity matrix, each element is represented by “0” or “1”, and we just need to consider objects that satisfy the constraints. Therefore, the storage space of the binary similarity matrix is small, and the calculation of binary similarity matrix is simple.
Example 1. 
Given an incomplete decision information system S = ( U , C { d } ) (as shown in Table 1), U = { x 1 , x 2 , x 3 , x 4 , x 5 , x 6 , x 7 , x 8 } , C = { a 1 , a 2 , a 3 , a 4 , a 5 } . The binary similarity matrix can be obtained by the following steps.
Firstly, it is easy to obtain object pairs that satisfy condition d ( x i ) d ( x j ) : ( x 1 , x 2 ) , ( x 1 , x 5 ) , ( x 1 , x 8 ) , ( x 2 , x 3 ) , ( x 2 , x 4 ) , ( x 2 , x 6 ) , ( x 2 , x 7 ) , ( x 3 , x 5 ) , ( x 3 , x 8 ) , ( x 4 , x 5 ) , ( x 4 , x 8 ) , ( x 5 , x 6 ) , ( x 5 , x 7 ) , ( x 6 , x 8 ) , ( x 7 , x 8 ) .
Secondly, it is obvious that R C ( x 1 ) = { x 1 } , R C ( x 2 ) = { x 2 , x 3 } , R C ( x 3 ) = { x 2 , x 3 } , R C ( x 4 ) = { x 4 } , R C ( x 5 ) = { x 5 , x 6 , x 7 } , R C ( x 6 ) = { x 5 , x 6 , x 7 } , R C ( x 7 ) = { x 5 , x 6 , x 7 } , R C ( x 8 ) = { x 2 , x 3 , x 6 , x 8 } . Therefore, σ C ( x 1 ) = { 1 } , σ C ( x 2 ) = { 1 , 2 } , σ C ( x 3 ) = { 1 , 2 } , σ C ( x 4 ) = { 1 } , σ C ( x 5 ) = { 1 , 2 } , σ C ( x 6 ) = { 1 , 2 } , σ C ( x 7 ) = { 1 , 2 } , σ C ( x 8 ) = { 1 , 2 } .
Finally, we can have the object pairs satisfying d ( x i ) d ( x j ) and min { | σ C ( x i ) | , | σ C ( x j ) | } = 1 are ( x 1 , x 2 ) , ( x 1 , x 5 ) , ( x 2 , x 4 ) , ( x 4 , x 5 ) , ( x 1 , x 8 ) , ( x 4 , x 8 ) . Therefore, a binary similarity matrix is obtained as shown in Table 2.

3. Heuristic Attribute Reduction Algorithm Based on Binary Similar Matrix

In this part, we apply the binary similarity matrix to define attribute significance and then design a heuristic attribute reduction algorithm with attribute significance as heuristic knowledge.
In order to calculate attribute significance conveniently, we add a bottom row to the binary similarity matrix to describe the number of “1” in the column where a k ( a k C ) is located, representing the number of object pairs ( x i , x j ) belonging to the same tolerance class under attribute a k . Take Table 2 as an example, the extended binary similarity matrix (as shown in Table 3) can be acquired easily.
Definition 8. 
Given an incomplete decision information system S = ( U , C { d } ) with binary similarity matrix B S M , the significance attribute of a k ( a k C ) is defined as C S ( a k ) = t k | B S M | , where t k is the number of “1” in the column a k , and | B S M | is the number of object pairs in the binary similarity matrix.
In Definition 8, if | B S M | is fixed, then the larger C S ( a k ) is, the smaller t k is, indicating less object pairs in terms of tolerance relation induced by attribute a k . In addition, the larger C S ( a k ) is, it indicates the stronger ability of a k to distinguish object pairs, that is, the more significant a k is. Therefore, C S ( a k ) directly reflects the significance of a k .
In order to reduce the time complexity, most attribute reduction algorithms start from core attributes. Next, Theorem 1 shows how to determine the core attributes of an information system according to binary similarity matrix.
Theorem 1. 
Given an incomplete decision system S = ( U , C { d } ) . For objects x i , x j U ( x i x j ) , if there is only one a k C satisfying m ( ( x i , x j ) , a k ) = 0 , then a k is a core attribute, i.e., a k c o r e ( C ) .
Proof. 
According to Definition 7, we have d ( x i ) d ( x j ) and min { | σ C ( x i ) | , | σ C ( x j ) | } = 1 . As min { | σ C ( x i ) | , | σ C ( x j ) | } = 1 , then | σ C ( x j ) | = 1 or | σ C ( x i ) | = 1 . If | σ C ( x j ) | = 1 , then R C ( x j ) R { d } ( x j ) , so x j U C D . If | σ C ( x i ) | = 1 , then R C ( x i ) R { d } ( x i ) , hence x i U C D . Because d ( x i ) d ( x j ) , we have x j R { d } ( x i ) and x i R { d } ( x j ) . In addition, if there is only one a k C such that m ( ( x i , x j ) , a k ) = 0 , it indicates that objects x i and x j satisfy the condition a k ( x i ) a k ( x j ) , so a l ( x i ) = a l ( x j ) ( a l ( C { a k } ) , and x j R C ( x i ) , x i R C ( x j ) . Let C { a k } = E , obviously, x j R E ( x i ) , x i R E ( x j ) . Since U E D = { x | R E ( x ) R { d } ( x ) } , we have x i U E D and x j U E D , which implies U E D U C D . Therefore, a k c o r e ( C ) .    □
According to Theorem 1, it is easy to obtain c o r e ( C ) = { a 1 , a 2 } in Example 1.
Next, we propose a new attribute reduction algorithm based on binary similarity matrix, which takes attribute significance as heuristic knowledge. The main process is as follows.
The binary similarity matrix in our algorithm needs less storage space compared with some existing similarity matrixes. On the other hand, our attribute reduction method is simpler and more direct than previous similarity matrix-based methods. In addition, in some heuristic attribute reduction algorithms, multiple tests are needed to verify whether a reduct is obtained. In our heuristic attribute reduction algorithm, the sign of the end of the algorithm is that the binary similarity matrix does not contain “1”, and multiple tests are not needed. Therefore, the time complexity can be reduced significantly.

4. Numerical Illustration

4.1. Attribute Reduction of Incomplete Decision Tables

To verify the effectiveness of the proposed algorithm, an incomplete decision table (as shown in Table 4) in [14] is adopted. Next, we apply our heuristic algorithm to calculate the attribute reduction of Table 4.
Example 2 
([14]). Given an incomplete decision table S = ( U , C { d } ) (as shown in Table 4), U = { x 1 , x 2 , , x 12 } , C = { a 1 , a 2 , , a 8 } .
Step 1. Firstly, we apply a Boolean matrix to calculate U C D . According to Table 4, the Boolean relation matrices of R a 1 , R a 2 , , R a 8 , and R d can be acquired as below.
M R a 1 = 1 0 0 1 1 0 1 1 1 0 1 1 0 1 1 1 1 1 0 1 0 0 1 0 0 1 1 1 1 1 0 1 0 0 1 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 1 1 1 1 1 0 1 0 0 1 0 1 0 0 1 1 0 1 1 1 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 1 1 0 1 1 1 0 1 1 0 0 0 1 1 0 0 1 0 1 1 0 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 1 1 0 1 1 1 0 1 1 , M R a 2 = 1 0 0 1 1 0 1 0 1 1 1 1 0 1 1 0 0 1 1 0 0 1 0 0 0 1 1 0 0 1 1 0 0 1 0 0 1 0 0 1 1 0 1 0 1 1 1 1 1 0 0 1 1 0 1 0 1 1 1 1 0 1 1 0 0 1 1 0 0 1 0 0 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 1 1 0 1 0 0 1 0 0 1 1 0 1 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 1 1 0 1 0 1 1 1 1 1 0 0 1 1 0 1 0 1 1 1 1 ,
M R a 3 = 1 0 0 1 1 0 1 0 1 1 1 1 0 1 1 1 1 1 1 0 0 1 1 0 0 1 1 1 1 1 1 0 0 1 1 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 1 1 1 1 1 1 0 0 1 1 0 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 1 1 0 1 1 0 1 1 0 1 0 0 1 1 0 1 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 1 1 0 1 0 1 0 1 1 , M R a 4 = 1 0 0 1 1 1 0 1 0 1 1 1 0 1 1 0 0 0 0 1 0 1 1 1 0 1 1 0 0 0 0 1 0 1 1 1 1 0 0 1 1 1 0 1 0 1 1 1 1 0 0 1 1 1 0 1 0 1 1 1 1 0 0 1 1 1 0 1 0 1 1 1 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 ,
M R d = 1 1 0 0 0 0 1 0 0 1 1 1 1 1 0 0 0 0 1 0 0 1 1 1 0 0 1 1 1 1 0 1 1 0 0 0 0 0 1 1 1 1 0 1 1 0 0 0 0 0 1 1 1 1 0 1 1 0 0 0 0 0 1 1 1 1 0 1 1 0 0 0 1 1 0 0 0 0 1 0 0 1 1 1 0 0 1 1 1 1 0 1 1 0 0 0 0 0 1 1 1 1 0 1 1 0 0 0 1 1 0 0 0 0 1 0 0 1 1 1 1 1 0 0 0 0 1 0 0 1 1 1 1 1 0 0 0 0 1 0 0 1 1 1 , i = 1 8 M R a i = 1 0 0 0 0 0 0 0 0 0 1 1 0 1 1 0 0 0 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 0 0 0 0 0 1 1 0 0 0 0 0 1 0 0 0 0 1 1 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 1 0 0 0 1 0 0 0 0 0 0 1 1 0 1 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 1 0 1 0 0 1 0 0 1 1 0 0 0 0 0 1 0 1 0 0 0 0 0 1 0 0 0 0 1 .
According to matrix i = 1 8 M R a i , it is obvious that R C ( x 1 ) = { x 1 , x 11 , x 12 } , R C ( x 2 ) = { x 2 , x 3 } , R C ( x 3 ) = { x 2 , x 3 } , R C ( x 4 ) = { x 4 , x 5 , x 11 } , R C ( x 5 ) = { x 4 , x 5 , x 11 } , R C ( x 6 ) = { x 6 } , R C ( x 7 ) = { x 7 , x 8 , x 12 } , R C ( x 8 ) = { x 7 , x 8 , x 10 } , R C ( x 9 ) = { x 9 } , R C ( x 10 ) = { x 8 , x 10 } , R C ( x 11 ) = { x 1 , x 4 , x 5 , x 11 } , R C ( x 12 ) = { x 1 , x 7 , x 12 } .
According to matrix M R d , we have R { d } ( x 1 ) = { x 1 , x 2 , x 7 , x 10 , x 11 , x 12 } , R { d } ( x 2 ) = { x 1 , x 2 , x 7 , x 10 , x 11 , x 12 } , R { d } ( x 3 ) = { x 3 , x 4 , x 5 , x 6 , x 8 , x 9 } , R { d } ( x 4 ) = { x 3 , x 4 , x 5 , x 6 , x 8 , x 9 } , R { d } ( x 5 ) = { x 3 , x 4 , x 5 , x 6 , x 8 , x 9 } , R { d } ( x 6 ) = { x 3 , x 4 , x 5 , x 6 , x 8 , x 9 } , R { d } ( x 7 ) = { x 1 , x 2 , x 7 , x 10 , x 11 , x 12 } , R { d } ( x 8 ) = { x 3 , x 4 , x 5 , x 6 , x 8 , x 9 } , R { d } ( x 9 ) = { x 3 , x 4 , x 5 , x 6 , x 8 , x 9 } , R { d } ( x 10 ) = { x 1 , x 2 , x 7 , x 10 , x 11 , x 12 } , R { d } ( x 11 ) = { x 1 , x 2 , x 7 , x 10 , x 11 , x 12 } , R { d } ( x 12 ) = { x 1 , x 2 , x 7 , x 10 , x 11 , x 12 } , then U C D = { x | R C ( x ) R { d } ( x ) } = { x 1 , x 6 , x 9 , x 12 } .
Secondly, it is easy to obtain σ C ( x 1 ) = { 0 } , σ C ( x 2 ) = { 0 , 1 } , σ C ( x 3 ) = { 0 , 1 } , σ C ( x 4 ) = { 0 , 1 } , σ C ( x 5 ) = { 0 , 1 } , σ C ( x 6 ) = { 1 } , σ C ( x 7 ) = { 0 , 1 } , σ C ( x 8 ) = { 0 , 1 } , σ C ( x 9 ) = { 1 } , σ C ( x 10 ) = { 0 , 1 } , σ C ( x 11 ) = { 0 , 1 } , σ C ( x 12 ) = { 0 } .
Finally, according to Definition 7, we can obtain a binary similarity matrix (as shown in Table 5).
Step 2. According to Theorem 1, we can obtain c o r e ( C ) = { a 4 , a 6 , a 7 } .
Step 3. Test whether c o r e ( C ) is a reduction of C. Similar to the method of calculating U C D , U c o r e ( C ) D = { x 1 , x 9 , x 12 } U C D = { x 1 , x 6 , x 9 , a n d x 12 } . Thus, we delete the columns where attributes a 4 , a 6 and a 7 are located in Table 5 to obtain Table 6. Initialization: D E = .
Step 4. According to Table 6, we have | B S M | = 20 . Calculate the C S of attributes in Table 6: C S ( a 1 ) = t 1 | B S M | = 12 20 , C S ( a 2 ) = t 2 | B S M | = 12 20 , C S ( a 3 ) = t 3 | B S M | = 13 20 , C S ( a 5 ) = t 5 | B S M | = 16 20 , C S ( a 8 ) = t 8 | B S M | = 12 20 .
Step 5. Obviously, C S of a 5 is the smallest, so delete the column where a 5 is located and the row with value “1” in this column, and obtain Table 7, then D E = { a 5 } . Because Table 7 has “1”, go to Step 4. Calculate the C S of attributes in Table 7: C S ( a 1 ) = 1 4 , C S ( a 2 ) = 2 4 , C S ( a 3 ) = 3 4 , C S ( a 8 ) = 3 4 . It is obvious that C S of a 3 and a 8 are the smallest. If we delete the column where a 3 is located and the row with value “1” in that column to obtain Table 8, then D E = { a 5 , a 3 } . Table 8 has “1”, then go to Step 4 again. Perform Step 4 again. Calculate the C S of attributes in Table 8: C S ( a 1 ) = 0 , C S ( a 2 ) = 0 , C S ( a 8 ) = 1 . Then, delete the column where a 8 is located and the row with value “1” in that column, and we have D E = { a 5 , a 3 , a 8 } . B S M does not have “1”, so then go to Step 6.
Step 6. T = C D E = { a 1 , a 2 , a 4 , a 6 , a 7 } .
In Table 7, select the column where a 8 is located and the row with value “1” in that column to obtain Table 9. Then, according to Table 9, delete the column where a 2 is located and the row with value “1” in that column, and we can obtain T = { a 1 , a 3 , a 4 , a 6 , a 7 } , then delete a 3 and in the last row, we also obtain T = { a 1 , a 2 , a 4 , a 6 , a 7 } .
In order to verify the effectiveness of our algorithm, it is necessary to test whether T is a reduction of C. When T = { a 1 , a 3 , a 4 , a 6 , a 7 } , U T D = { x 1 , x 9 , x 12 } U C D . When T = { a 1 , a 2 , a 4 , a 6 , a 7 } , we have U T D = { x 1 , x 6 , x 9 , x 12 } = U C D , and there are no T T such that U T D = U C D . Thus, T = { a 1 , a 2 , a 4 , a 6 , a 7 } is a reduction of attributes set C.
Moreover, the reduction result in reference [14] is T = { a 1 , a 3 , a 4 , a 5 , a 6 , a 8 } . It is obvious that our result contains five attributes, and their result contains six attributes. What is more, it can be proved that our result is also a reduction of C from the point of literature [14]. The purpose of attribute reduction is to classify objects with as few attributes as possible. From the discussion above, our algorithm is not only effective but also efficient.

4.2. Attribute Reduction of Complete Decision Tables

Since complete decision systems are a special case of incomplete decision systems, a decent attribute reduction algorithm for incomplete decision systems should be also available for complete decision systems. In the following, we demonstrate the effectiveness of the proposed attribute reduction algorithm for complete decision systems.
In a complete decision system S = ( U , A ) , C A , for any x i , x j U , the equivalence relation R C is: x i R C x j a C , f ( x i , a ) = f ( x j , a ) [22]. According to Definition 2 and Definition 7, it is not difficult to find that when a decision system is complete, the construction of the binary similarity matrix is the same as that of the incomplete decision table. Next, we give the other steps of Algorithm 1 to obtain the reduction of Table 10 cited from the literature [22].
Step 1. According to Definition 3, it is easy to obtain U C D = { x 1 , x 2 , x 4 } , and according to Definition 7, the binary similarity matrix is obtained (as shown in Table 11).
Step 2. According to Theorem 1, it is not difficult to obtain c o r e ( C ) = { a 3 } from Table 11.
Step 3. Test whether c o r e ( C ) is a reduction of C. Similar to the method of calculating U C D , U c o r e ( C ) D = U C D = { x 1 , x 2 , x 4 } . Thus, we delete the column where attribute a 3 is located in Table 11 to obtain Table 12. Initialization: D E = .
Algorithm 1 Heuristic attribute reduction algorithm based on binary similarity matrix.
  • Input: An incomplete decision information system S = ( U , C { d } ) .
  • Output: A reduction T of C.
  • Step 1: Calculate U C D and the binary similarity matrix B S M of S.
  • Step 2: Determine c o r e ( C ) according to Theorem 1.
  • Step 3: If U c o r e ( C ) D = U C D , then T = c o r e ( C ) , the algorithm ends. Otherwise, delete the columns where core attributes are located to obtain a new B S M . Initialization: D E = .
  • Step 4: Calculate the C S of attributes in the latest B S M .
  • Step 5: Add attribute a k with the smallest C S to D E , delete the column where a k is located and the row with value “1” in the column. If there is no “1” in B S M , go to Step 6, otherwise, go to Step 4.
  • Step 6: T = C D E , and test whether T is a reduction of attributes. The algorithm ends.
Step 4. According to Table 12, we have | B S M | = 5 . Calculate the C S of attributes in Table 12: C S ( a 1 ) = t 1 | B S M | = 2 5 , C S ( a 2 ) = t 2 | B S M | = 2 5 , C S ( a 4 ) = t 3 | B S M | = 3 5 .
Step 5. Obviously, C S of a 4 is the smallest, so delete the column where a 4 is located and the row with value “1” in this column, and obtain Table 13, then D E = { a 4 } . Because Table 13 has “1”, go to Step 4. Calculate the C S of attributes in Table 13: C S ( a 1 ) = 1 2 , C S ( a 2 ) = 1 2 . From Table 13, it is not difficult to see that regardless of deleting a 1 or a 2 , the Table does not have “1”, then D E = { a 1 , a 4 } or D E = { a 2 , a 4 } .
Step 6. T = { a 2 , a 3 } or T = { a 1 , a 3 } . The algorithm ends.
Table 10 is a complete decision table cited from [22], and the attribute reduction result in [22] is T = { a 2 , a 3 } or T = { a 1 , a 3 } . Our result is the same. So our method is effective. The method in [22] is based on the discernibility matrix. Although all the reductions can be obtained, their method may have a combination explosion problem for large data sets.

4.3. Attribute Reduction of Fuzzy Numerical Decision Tables

The method proposed in this paper is also applicable to fuzzy numerical decision tables (such as Table 14). For a fuzzy decision information system, we can not directly classify the objects according to the information. Liu et al. [22] introduced a threshold as follows,
T H = ( m a x i = 1 | U | V a ( x i ) m i n i = 1 | U | V a ( x i ) ) × 10 % ,
where m a x i = 1 | U | V a ( x i ) is the maximum attribute value of V a , m i n i = 1 | U | V a ( x i ) is the minimum attribute value of V a . A binary relation R is defined as: x i R x j if and only if | V a ( x i ) V a ( x j ) | < T H . It is easy to check that R is reflexive and symmetric, but not transitive.
After classifying objects according to the method above, the corresponding binary similarity matrix can be constructed according to Definition 7, and then the attribute reduction of the decision table can be obtained according to the steps of Algorithm 1.

5. Conclusions

In this paper, we introduce the definition of binary similarity matrix in incomplete decision information systems and define the attribute significance based on binary similarity matrix. What is more, a heuristic attribute reduction algorithm with attribute significance as heuristic knowledge is developed. It has been demonstrated that the algorithm is not only effective, but also efficient. In addition, it can be applied to consistent as well as inconsistent decision information systems. More importantly, the algorithm presented in this paper can hopefully provide more penetration into attribute reduction. In many cases, the data in decision tables may be affected by uncertainty and imprecise factors. In the future, we will apply a binary similarity matrix to hesitant fuzzy numerical decision tables to deal with fuzzy pattern recognition problems.

Author Contributions

Methodology, Y.-L.B.; writing–original draft preparation, Y.-L.B. and Y.Z.; writing–review and editing, Y.-L.B.; supervision, Y.-L.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Scientific Research Program of the Higher Education Institution of Xinjiang (No. XJEDU2022P008).

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Akram, M.; Ali, G.; Alcantud, J.C.R. Attributes reduction algorithms for m-polar fuzzy relation decision systems. Int. J. Approx. Reason. 2022, 140, 232–254. [Google Scholar] [CrossRef]
  2. Al-juboori, A.M.; Alsaeedi, A.H.; Nuiaa, R.R.; Alyasseri, Z.A.A.; Sani, N.S.; Hadi, S.M.; Mohammed, H.J.; Musawi, B.A.; Amin, M.M. A hybrid cracked tiers detection system based on adaptive correlation features selection and deep belief neural networks. Symmetry 2023, 15, 358. [Google Scholar] [CrossRef]
  3. He, J.L.; Qu, L.D.; Wang, Z.H.; Chen, Y.Y.; Luo, D.M. Attribute reduction in an incomplete categorical decision information system based on fuzzy rough sets. Artif. Intell. Rev. 2022, 55, 5313–5348. [Google Scholar] [CrossRef]
  4. Hu, M.; Tsang, E.C.; Guo, Y.T.; Xu, W.H. Fast and robust attribute reduction based on the separability in fuzzy decision systems. IEEE Trans. Cybern. 2021, 52, 5559–5572. [Google Scholar] [CrossRef] [PubMed]
  5. Kalaivanan, K.; Vellingiri, J. Normalized hellinger feature selection and soft margin boosting classification for water quality prediction. Expert Syst. 2022. [Google Scholar] [CrossRef]
  6. Singh, H.; Singh, B.; Kaur, M. An efficient feature selection method based on improved elephant herding optimization to classify high-dimensional biomedical data. Expert Syst. 2022, 39, 13038. [Google Scholar] [CrossRef]
  7. Yang, Q.W.; Gao, Y.L.; Song, Y.J. A tent lévy flying sparrow search algorithm for wrapper-based feature selection: A COVID-19 case study. Symmetry 2023, 15, 316. [Google Scholar] [CrossRef]
  8. Chen, C.W.; Tsai, Y.H.; Chang, F.R.; Lin, W.C. Ensemble feature selection in medical datasets: Combining filter, wrapper, and embedded feature selection results. Expert Syst. 2020, 37, 12553. [Google Scholar] [CrossRef]
  9. Yuan, Z.; Chen, H.M.; Li, T.R.; Yu, Z.; Sang, B.B.; Luo, C. Unsupervised attribute reduction for mixed data based on fuzzy rough sets. Inf. Sci. 2021, 572, 67–87. [Google Scholar] [CrossRef]
  10. Xie, J.J.; Hu, B.Q.; Jiang, H.B. A novel method to attribute reduction based on weighted neighborhood probabilistic rough sets. Int. J. Approx. Reason. 2022, 144, 1–17. [Google Scholar] [CrossRef]
  11. Skowron, A.; Rauszer, C. The discernibility matrices and functions in information systems. Intell. Decis. Support 1992, 21, 331–362. [Google Scholar]
  12. Liu, G.L. Attribute reduction algorithms determined by invariants for decision tables. Cogn. Comput. 2022, 14, 1818–1825. [Google Scholar] [CrossRef]
  13. Shen, Q.; Jiang, Y. Attribute reduction of multi-valued information system based on conditional information entropy. In Proceedings of the 2008 International Conference on Granular Computing, Hangzhou, China, 26–28 August 2008; pp. 562–565. [Google Scholar]
  14. Zhou, J.; E, X.; Li, Y.H.; Wang, Z.; Liu, Z.X.; Bai, X.Y.; Huang, X.Y. A new attribute reduction algorithm dealing with the incomplete information system. In Proceedings of the International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery, Zhangjiajie, China, 10–11 October 2009; pp. 12–19. [Google Scholar]
  15. Dai, J.H.; Wang, W.T.; Tian, H.W.; Liu, L. Attribute selection based on a new conditional entropy for incomplete decision systems. Knowl. Based Syst. 2013, 39, 207–213. [Google Scholar] [CrossRef]
  16. Hu, M.; Tsang, E.C.; Guo, Y.T.; Chen, D.G.; Xu, W.H. A novel approach to attribute reduction based on weighted neighborhood rough sets. Knowl. Based Syst. 2021, 220, 106908. [Google Scholar] [CrossRef]
  17. Instituto; Nacional. A new algorithm for computing reducts based on the binary discernibility matrix. Intell. Data Anal. 2016, 20, 317–337. [Google Scholar] [CrossRef]
  18. Ma, F.M.; Zhang, T.F. Generalized binary discernibility matrix for attribute reduction in incomplete information systems. J. China Univ. Posts Telecommun. 2017, 4, 57–75. [Google Scholar]
  19. Li, J.; Wang, X.; Fan, X.W. Improved binary discernibility matrix attribute reduction algorithm in customer relationship management. Procedia Eng. 2019, 7, 473–476. [Google Scholar] [CrossRef] [Green Version]
  20. Qian, Y.H.; Liang, J.Y.; Li, D.Y.; Wang, F.; Ma, N.N. Approximation reduction in inconsistent incomplete decision tables. Knowl. Based Syst. 2010, 23, 427–433. [Google Scholar] [CrossRef]
  21. Liu, G.L.; Hua, Z.; Chen, Z.H. A general reduction algorithm for relation decision systems and its application. Knowl. Based Syst. 2017, 119, 87–93. [Google Scholar] [CrossRef]
  22. Liu, G.L.; Li, L.; Yang, J.T.; Feng, Y.B.; Zhu, K. Attribute reduction approaches for general relation decision systems. Pattern Recognit. Lett. 2015, 65, 81–87. [Google Scholar] [CrossRef]
Table 1. Incomplete decision table.
Table 1. Incomplete decision table.
a 1 a 2 a 3 a 4 a 5 d
x 1 132021
x 2 1102
x 3 1101
x 4 132101
x 5 3312
x 6 001
x 7 311311
x 8 22
Table 2. Binary similarity matrix.
Table 2. Binary similarity matrix.
a 1 a 2 a 3 a 4 a 5
( x 1 , x 2 ) 10100
( x 1 , x 5 ) 01100
( x 2 , x 4 ) 10111
( x 4 , x 5 ) 01100
( x 1 , x 8 ) 01111
( x 4 , x 8 ) 01111
Table 3. Extended binary similarity matrix 1.
Table 3. Extended binary similarity matrix 1.
a 1 a 2 a 3 a 4
( x 1 , x 2 ) 1010
( x 1 , x 5 ) 0110
( x 2 , x 4 ) 1011
( x 4 , x 5 ) 0110
( x 1 , x 8 ) 0111
( x 4 , x 8 ) 0111
t 2463
Table 4. Incomplete decision Table 2.
Table 4. Incomplete decision Table 2.
a 1 a 2 a 3 a 4 a 5 a 6 a 7 a 8 d
x 1 3211100
x 2 23201310
x 3 23201311
x 4 212011
x 5 2112011
x 6 23213111
x 7 331020
x 8 000201
x 9 321311211
x 10 11000
x 11 21010
x 12 3210230
Table 5. Extended binary similarity matrix 2.
Table 5. Extended binary similarity matrix 2.
a 1 a 2 a 3 a 4 a 5 a 6 a 7 a 8
( x 1 , x 3 ) 00001111
( x 1 , x 4 ) 11111011
( x 1 , x 5 ) 11111011
( x 1 , x 6 ) 00010011
( x 1 , x 8 ) 10011111
( x 1 , x 9 ) 11101011
( x 2 , x 6 ) 11101111
( x 2 , x 9 ) 00001101
( x 3 , x 12 ) 00011100
( x 4 , x 12 ) 11111000
( x 5 , x 12 ) 11111000
( x 6 , x 7 ) 01100011
( x 6 , x 10 ) 01110010
( x 6 , x 11 ) 10110111
( x 6 , x 12 ) 00011010
( x 7 , x 9 ) 11111011
( x 8 , x 12 ) 10011110
( x 9 , x 10 ) 01111010
( x 9 , x 11 ) 11111101
( x 9 , x 12 ) 11111010
t 121213151671512
Table 6. Extended binary similarity matrix 3.
Table 6. Extended binary similarity matrix 3.
a 1 a 2 a 3 a 5 a 8
( x 1 , x 3 ) 00011
( x 1 , x 4 ) 11111
( x 1 , x 5 ) 11111
( x 1 , x 6 ) 00001
( x 1 , x 8 ) 10011
( x 1 , x 9 ) 11111
( x 2 , x 6 ) 11111
( x 2 , x 9 ) 00011
( x 3 , x 12 ) 00010
( x 4 , x 12 ) 11110
( x 5 , x 12 ) 11110
( x 6 , x 7 ) 01101
( x 6 , x 10 ) 01100
( x 6 , x 11 ) 10101
( x 6 , x 12 ) 00010
( x 7 , x 9 ) 11111
( x 8 , x 12 ) 10010
( x 9 , x 10 ) 01110
( x 9 , x 11 ) 11111
( x 9 , x 12 ) 11110
t 1212131612
Table 7. Extended binary similarity matrix 4.
Table 7. Extended binary similarity matrix 4.
a 1 a 2 a 3 a 8
( x 1 , x 6 ) 0001
( x 6 , x 7 ) 0111
( x 6 , x 10 ) 0110
( x 6 , x 11 ) 1011
t 1233
Table 8. Extended binary similarity matrix 5.
Table 8. Extended binary similarity matrix 5.
a 1 a 2 a 8
( x 1 , x 6 ) 001
t 001
Table 9. Extended binary similarity matrix 6.
Table 9. Extended binary similarity matrix 6.
a 1 a 2 a 3
( x 6 , x 10 ) 011
t 011
Table 10. Binary similarity matrix 2.
Table 10. Binary similarity matrix 2.
a 1 a 2 a 3 a 4 d
x 1 01112
x 2 10001
x 3 10111
x 4 10011
x 5 10112
Table 11. Extended binary similarity matrix 7.
Table 11. Extended binary similarity matrix 7.
a 1 a 2 a 3 a 4
( x 1 , x 2 ) 0000
( x 2 , x 3 ) 0011
( x 1 , x 4 ) 0001
( x 2 , x 5 ) 1100
( x 4 , x 5 ) 1101
t 2213
Table 12. Extended binary similarity matrix 8.
Table 12. Extended binary similarity matrix 8.
a 1 a 2 a 4
( x 1 , x 2 ) 000
( x 2 , x 3 ) 001
( x 1 , x 4 ) 001
( x 2 , x 5 ) 110
( x 4 , x 5 ) 111
t 223
Table 13. Extended binary similarity matrix 9.
Table 13. Extended binary similarity matrix 9.
a 1 a 2
( x 1 , x 2 ) 00
( x 2 , x 5 ) 11
t 22
Table 14. Fuzzy numerical decision table.
Table 14. Fuzzy numerical decision table.
a 1 a 2 a 3 a 4 a 5 d
x 1 0.9871 0.4352 0.1987 0.8765 0.5436 0.3656
x 2 0.1131 0.5634 0.2134 0.8643 0.6578 0.3452
x 3 0.8675 0.5897 0.3101 0.8321 0.5784 0.3432
x 4 0.4563 0.6235 0.2567 0.8107 0.6785 0.3672
x 5 0.4786 0.4356 0.2675 0.7896 0.7012 0.3213
x 6 0.5342 0.4765 0.2834 0.8876 0.6132 0.3425
x 7 0.8765 0.6123 0.2457 0.8654 0.5324 0.3531
x 8 1 0.5471 0.3721 0.8743 0.5452 0.3456
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Y.; Bao, Y.-L. A Novel Attribute Reduction Algorithm for Incomplete Information Systems Based on a Binary Similarity Matrix. Symmetry 2023, 15, 674. https://doi.org/10.3390/sym15030674

AMA Style

Zhou Y, Bao Y-L. A Novel Attribute Reduction Algorithm for Incomplete Information Systems Based on a Binary Similarity Matrix. Symmetry. 2023; 15(3):674. https://doi.org/10.3390/sym15030674

Chicago/Turabian Style

Zhou, Yan, and Yan-Ling Bao. 2023. "A Novel Attribute Reduction Algorithm for Incomplete Information Systems Based on a Binary Similarity Matrix" Symmetry 15, no. 3: 674. https://doi.org/10.3390/sym15030674

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop