Next Article in Journal
Spreading Speed in A Nonmonotone Equation with Dispersal and Delay
Previous Article in Journal
An Improved Artificial Bee Colony Algorithm Based on Elite Strategy and Dimension Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Matrix Method for the Optimal Scale Selection of Multi-Scale Information Decision Systems

1
Fujian Province University Key Laboratory of Computational Science, School of Mathematics Sciences, Huaqiao University, Quanzhou 362021, China
2
School of Mathematics Sciences and Statistics, Minnan Normal University, Zhangzhou 363000, China
*
Author to whom correspondence should be addressed.
Mathematics 2019, 7(3), 290; https://doi.org/10.3390/math7030290
Submission received: 22 January 2019 / Revised: 13 March 2019 / Accepted: 17 March 2019 / Published: 21 March 2019

Abstract

:
In multi-scale information systems, the information is often characterized at multi scales and multi levels. To facilitate the computational process of multi-scale information systems, we employ the matrix method to represent the multi-scale information systems and to select the optimal scale combination of multi-scale decision information systems in this study. To this end, we first describe some important concepts and properties of information systems using some relational matrices. The relational matrix is then introduced into multi-scale information systems, and used to describe some main concepts in systems, including the lower and upper approximate sets and the consistence of systems. Furthermore, from the view of the relation matrix, the scale significance is defined to describe the global optimal scale and the local optimal scale of multi-scale information systems. Finally, the relational matrix is used to compute the scale significance and to construct the optimal scale selection algorithms. The efficiency of these algorithms is examined by several practical examples and experiments.

1. Introduction

Granular computing [1,2] originated from fuzzy information granulation is a mathematical method for knowledge representation and data mining. The purpose is to solve complex problems by dividing the massive information into relatively simple blocks according to its respective characteristics and performance. Since the concept of granular computing was put forward, it has become a hot research topic and has been widely used in many practical applications [3,4,5,6,7,8,9,10,11,12,13].
The theory of rough set plays an important role in the promotion and development of granular computing [14,15,16,17,18,19,20]. Pawlak [14] used an information system to study granular computing. If a decision is required, it is usually an information system with decision attributes.
People can process a lot of information at different levels and scales. Based on this view point, multi-scale information systems have been developed and widely studied [21,22,23,24,25,26,27]. Ferone et al. used feature granulation to study feature selection [12,13]. In papers [21,22], Wu and Leung et al. proposed a multi-scale decision information system model. In their model, the objects are granulated with granularity methods from fine to coarse. The optimal scale selection of multi-scale decision tables was also investigated. Gu and Wu [23] presented a formal method of knowledge acquisition measured at different granularity. Wu and Qian [24] measured the uncertainty in incomplete multi-scale information tables with the Dempster-Shafer evidence theory [28]. Shen et al. [25] employed a local method to induce decision rules in multi-scale decision tables. Li and Hu [26] introduced a step-wise method for the optimal scale selection of multi-scale decision tables.
Due to the massive information provided in multi-scale information systems, which leads to much time consuming in the computation of the concepts in the systems, this article aims to employ the Boolean matrix and matrix computation in order to facilitate the knowledge description and the optimal scale selection in multi-scale decision tables. Indeed, the matrix method has had a computation advantage in many backgrounds [29,30,31,32,33,34,35,36,37], such as in classical rough set and Information systems [29,30,31], static and dynamic information systems [32,35,36], covering information systems [33,34], and so on. For multi-scale information systems, it is worthy to use the relational matrix to obtain the optimal scale selection and develop some matrix methods to facilitate the computational process.
The paper is structured as follows. We first introduce a relational matrix to represent the relative concepts in information systems. We then examine some properties of information systems based on the Boolean matrix. In Section 3, we introduce multi-scale information systems and the Boolean matrix representation. In Section 4, we use the relational matrix to define the scale significance, and to investigate the optimal scale selection of multi-scale information systems. In Section 5, we seek the optimal scale of multi-scale information systems by using the scale significance based on the relation matrix. The paper is finally concluded with a summary. In Section 7, the effectiveness of the matrix method is illustrated by experiments.

2. Preliminary Background

2.1. Rough Set and Information Systems

In this section we introduce some basic concepts of rough set and information systems.
Definition 1
([1]). The Pawlak approximate space is defined as ( U , R ) , where U = { x 1 , x 2 , , x n } is a non-empty finite set called the universe, and R is the equivalent relation on U, the lower and the upper approximation sets of X U are defined as R ̲ ( X ) , R ¯ ( X ) , respectively:
R ̲ ( X ) = { x U | [ x ] R X } , R ¯ ( X ) = { x U | [ x ] R X } .
where [ x ] R = { y U | ( x , y ) R } is the equivalent class containing x.
Definition 2
([14]). Let S = ( U , A ) be an information system, where U = { x 1 , x 2 , , x n } is a non-empty finite set called the universe, A = { a 1 , a 2 , , a m } is a non-empty finite set of attributes. For any a A , there is a surjective map a : U V a ie. a ( x ) V a , x U , where V a = { a ( x ) | x U } is called the domain of attribute a. For any B A , the equivalent relation R B On U is determined as
R B = { ( x , y ) U × U | a ( x ) = a ( y ) , x B }
The equivalence class of x U in terms of R B is denoted as [ x ] B = { y U | ( x , y ) R B } .
The lower approximation set and the upper approximate set of X U corresponding to the attribute subset B are represented as R ̲ B ( X ) and R ¯ B ( X ) , respectively,
R ̲ B ( X ) = { x U | [ x ] B X } , R ¯ B ( X ) = { x U | [ x ] B X } .
Definition 3
([14]). Let S = ( U , A d ) be a decision information system, where S = ( U , A ) is an information system, d A is called the decision attribute, and d : U V d is a surjective map, where V d = { d ( x ) | x U } is called the domain of d.
Similarly, the equivalent relation on U induced by R d is given by
R d = { ( x , y ) U × U | d ( x ) = d ( y ) } .
The equivalence class of x U is denoted as [ x ] d = { y U | ( x , y ) R d } .
Let S = ( U , A d ) be a decision information system, we summarize the following two kinds of consistent:
(1) If R A R d , then S is globally consistent;
(2) For x U , if [ x ] A [ x ] d , then S is locally consistent for x.

2.2. Boolean Matrix Characterization of Decision Information Systems

Definition 4.
A non-negative matrix M B = ( r i j ) n × n is called a Boolean matrix, where r i j { 0 , 1 } . Let A = ( a i j ) n × m , B = ( b i j ) n × m , C = ( c i j ) m × l are Boolean matrices, define the following operations:
(1) Order relation A B a i j b i j ;
(2) Meet A B = ( a i j b i j ) n × m ;
(3) Boolean product A C = ( d i j ) n × l , where d i j = k = 1 m a i k c k j ;
(4) Complement A = ( 1 a i j ) n × m .
where a i k c k j = m a x { a i k , c k j } , a i j b i j = m i n { a i j , b i j } .
It can be seen that a i j b i j = a i j × b i j and A B A ( B ) = 0 .
Definition 5
([33]). Let U = { x 1 , x 2 , , x n } and X U , The characteristic function of X is defined as f ( X ) = ( f ( x 1 ) , f ( x 2 ) , , f ( x n ) ) , where
f ( x i ) = 1 , x i X , 0 , x i X .
It can be seen that X Y f ( X ) f ( Y ) f ( X ) ( f ( Y ) ) = 0 .
Definition 6
([33]). Let ( U , R ) be an approximate space, U = { x 1 , x 2 , , x n } , The relation matrix of R is defined as M R = ( r i j ) n × n , where
m i j = 1 , ( x i , x j ) R , 0 , ( x i , x j ) R .
Obviously, by Definitions 5 and 6, we have
M R = f ( [ x 1 ] R ) f ( [ x 2 ] R ) f ( [ x n ] R )
Theorem 1
([30]). Let ( U , R ) be an approximate space, U = { x 1 , x 2 , , x n } , and X U , then
f ( R ̲ ( X ) ) = f ( X ) M R , f ( R ¯ ( X ) ) = f ( X ) M R .
Proof. 
Let f ( R ̲ ( X ) ) = ( f 1 , f 2 , , f n ) , f ( X ) M R = ( g 1 , g 2 , , g n ) , f ( X ) M R = ( h 1 , h 2 , , h n ) .
According to the symmetry of matrix M R and Definition 6, we have M R = M R T = ( f ( [ x 1 ] R ) T , f ( [ x 2 ] R ) T , , f ( [ x n ] R ) T ) and g i = f ( X ) f ( [ x i ] R ) T . Therefore f i = 1 x i R ̲ ( X ) [ x i ] R X [ x i ] R ( X ) = f ( X ) f ( [ x i ] R ) T ) = 0 g i = 0 h i = 1 . Similarly, f i = 0 h i = 0 . Therefore f ( R ̲ ( X ) ) = f ( X ) M R .
By R ¯ ( X ) = R ̲ ( X ) , we have f ( R ¯ ( X ) ) = f ( R ̲ ( X ) ) = f ( R ̲ ( X ) ) = f ( X ) ) M R = f ( X ) M R .  □
Example 1.
Let U = { x 1 , x 2 , x 3 , x 4 , x 5 } ,   U / R = { { x 1 } , { x 2 } , { x 3 , x 4 } , { x 5 } } ,   X = { x 1 , x 2 , x 3 }
f ( R ̲ ( X ) ) = f ( X ) M R = 0 0 0 1 1 1 0 0 0 0 0 1 0 0 0 0 0 1 1 0 0 0 1 1 0 0 0 0 0 1 = 1 1 0 0 0 ,
f ( R ¯ ( X ) ) = f ( X ) M R = 1 1 1 0 0 1 0 0 0 0 0 1 0 0 0 0 0 1 1 0 0 0 1 1 0 0 0 0 0 1 = 1 1 1 1 0 .
Therefore, R ̲ ( X ) = { x 1 , x 2 } , R ¯ ( X ) ) = { x 1 , x 2 , x 3 , x 4 } .
Theorem 2
([29]). Let U be an universe, R 1 , R 2 are equivalent relations on U, then the following conclusions hold:
(1) M R 1 R 2 = M R 1 M R 2 ;
(2) If R 1 R 2 , then M R 1 M R 2 .
Let S = ( U , A d ) be an decision information system, B A , the relation matrix of the equivalent relation R B is denote as M B , the relation matrix corresponding to R d is denoted as M d ,
Theorem 3.
Let S = ( U , A d ) be an decision information system, B A , then M B = ( m i j ) n × n , where
m i j = 1 , a ( x i ) = a ( x j ) , a B , 0 , e l s e .
and M d = ( r i j ) n × n , where
r i j = 1 , d ( x i ) = d ( x j ) , 0 , e l s e .
Proof. 
By Definition 2, ( x i , x j ) R B a ( x i ) = a ( x j ) , a B , by Definition 6, we have m i j = 1 a ( x i ) = a ( x j ) , a B . Similarly r i j = 1 d ( x i ) = d ( x j ) .  □
Theorem 4.
Let S = ( U , A d ) be a decision information system, then the following conclusions are obtained:
(1) S is locally consistent for x U if and only f ( [ x ] A ) f ( [ x ] d ) hold;
(2) S is globally consistent if and only if M A M d is hold.
Proof. 
By Definition 5, S is locally consistent for x U [ X ] A [ x ] d f ( [ x ] A ) f ( [ x ] d ) ;
By Theorem 2, we have S is globally consistent R A R d M A M d .  □
Example 2.
Determining the globally consistent of the following system Table 1.
M A is the relational matrix corresponding to A, M d is the relational matrix corresponding to d. According to Theorem 3, by calculation
M A = 1 0 1 0 0 0 0 0 1 0 0 0 0 1 1 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 1 0 0 0 0 1 , M d = 1 0 1 0 0 1 0 0 1 0 0 0 0 1 1 0 1 0 0 1 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 1 0 1 0 0 1 0 0 1 0 0 0 0 1
Obviously M A M d , by Theorem 4, the ( U , A { d } ) is globally consistent.

3. Relational Matrix in Generalized Multi-Scale Information Systems

The theoretical model of multi-scale decision information systems was first proposed by Wu and Leung [21]. Some scholars have participated in this study [21,22,23,24,25,26,27,38,39,40,41], Li and Hu generalized this model in paper [27]. Among many studies, optimal scale selection is an important subject. Wu, Li et al. [29,38,40] studied the optimal scale selection of multi-scale decision information systems. Gu et al. [39] studied the optimal scale selection of incomplete multi-scale decision information systems. Li and Hu [27] studied the optimal scale selection for generalized multi-scale decision information systems. Li et al. [26] gave an attribute significance of generalized multi-scale decision information systems to search for optimal scale selection. Decision information system models involve a large number of set operations. The matrix itself is a powerful tool. The use of matrices to study multi-scale information systems helps to improve algorithm efficiency and further develop multi-scale information table theory. In this part, we introduce relation matrix into multi-scale information systems to prepare for optimal scale selection.
Definition 7
([21,27]). A multi-scale information system is a tuple S = ( U , A ) , where U = { x 1 , x 2 , , x n } is a nonempty and finite set of objects called the universe, A = { a 1 , a 2 , , a m } is a nonempty and finite set of attributes, and each a j has I j ( j = 1 , 2 , , n ) scales. Then a multi-scale information system S = ( U , A ) can be represented as
S = ( U , { a j k | k = 1 , 2 , , I j , j = 1 , 2 , , m } ) .
where a j k : U V j k is a surjective function and V j k is the domain of a j corresponding to the kth scale, and furthermore, for any j = 1 , 2 , , m , 1 k I j 1 , there exists a surjective g j k , k + 1 : V j k V j k + 1 such that a j k + 1 = g j k , k + 1 . a j k + 1 , i.e., a j k + 1 ( x ) = g j k , k + 1 . a j k + 1 ( x ) , x U , g j k , k + 1 is called the information granularity transformation function.
When I 1 = I 2 = = I j = I , the information system defined above is the multi-scale information system proposed by Wu-Leung in paper [21].
The equivalent relation determined by attribute a j l j ( 1 l j I j ) is
R a j l j = { ( x , y ) U × U | a j l j ( x ) = a j l j ( y ) } .
By the existence of the surjective g j k , k + 1 : V j k V j k + 1 , for any a j U , we have
R a j 1 R a j 2 R a j I j .
For a multi-scale information system S = ( U , A ) , if the attribute a j U is restricted on their l j scale, we call K = ( l 1 , l 1 , , l m ) to be a scale combination of the system S = ( U , A ) , and all the scale combinations of the system S = ( U , A ) are denoted as ∑ and ∑ forms a partial ordered lattice [27].
The information system corresponding to the K = ( l 1 , l 1 , , l m ) scale combination is denoted as
S K = ( U , A K ) = ( U , { a 1 l 1 , a 2 l 2 , , a m l m } ) .
The equivalent relation induced by A K is denoted as R A K , while the relation matrix of R A K is denoted as M A K .
Definition 8
([27]). Let S = ( U , A ) be a multi-scale information system, K , L , K = ( l 1 , l 2 , , l m ) , L = ( h 1 , h 2 , , h m ) , if l i h i ( i = 1 , 2 , , m ) , we called the scale combination to be K is finer than L, denoted as K L , and further if K L and there is at least one of i { 1 , 2 , , m } such that l i < h i , we called K is strictly finer than L, denoted as K < L .
Example 3.
The following example gives a comprehensive evaluation of three courses for eight students, and scores are divided into four criteria, as shown in Table 2.
We use attribute a 1 to evaluate the results of the first course, which divided into four levels, attribute a 2 for the second course, which divided into three levels, and attribute a 3 for the third course, which divided into two levels. Attribute d is called decision attribute, 1 is qualified and 0 is unqualified. According to the scores of each student, the following multi-scale decision-making information system is obtained, as shown in Table 3. In order to facilitate the expression below, we denote excellent by "E", and similarly "G" denotes good, "F" denotes fair, "P" denotes pass, "B" denotes bad, "S" denotes super, "M" denotes middle, "L" denotes low, "Y" denotes yes, "N" denotes no.
There are 24 different scale combinations in multi-scale information system in Table 3. According to the scale relation in Definition 8, we can obtain a partial ordered lattices [27], as shown in Figure 1.
We now discuss some properties of multi-scale information systems using the relation matrix.
Let S = ( U , A ) be a multi-scale information system and ∑ is the set of all scale combinations for the system, K = ( l 1 , l 2 , , l m ) , the decision information system corresponding to the scale combination K is S K = ( U , A K d ) , the relation matrix of R A K is denote as M A K .
Theorem 5.
Let S = ( U , A ) be a multi-scale information system, K = ( l 1 , l 2 , , l m ) , then M A K = ( m i j ) n × n , where
m i j = 1 , a k l k ( x i ) = a k l k ( x j ) , k = 1 , 2 , , m 0 , e l s e .
Proof. 
This conclusion can be obtained from Theorem 3.  □
Theorem 6.
Let S = ( U , A ) be a multi-scale information system and ∑ is the set of all scale combinations for the system, K , L , K < L , X U . Then the following conclusions hold:
(1) For any attribute a j A ( j = 1 , 2 , , m ) , M a j 1 M a j 2 M a j I j ;
(2) M A K M A L ;
(3) M A K = l K M a l ;
(4) f ( R A K ̲ ( X ) ) = ( f ( X ) ) M A K ;
(5) f ( R A K ¯ ( X ) ) = ( f ( X ) ) M A K ;
(6) f ( R A L ̲ ( X ) ) f ( R A K ̲ ( X ) ) ;
(7) f ( R A K ¯ ( X ) ) f ( R A L ¯ ( X ) ) .
Proof. 
(1) By Definition 7, as a result of R a j 1 R a j 2 R a j I j , according to Theorem 2, we have M a j 1 M a j 2 M a j I j ;
(2) K L R A K R A L M A K M A L ;
(3) Suppose K = ( l 1 , l 2 , , l m ) , then A K = ( a l 1 , a l 2 , , a l m ) , R A K = R A l 1 R A l 2 R A l m , by Theorem 2, we have M A K = M A 1 l M A 2 l M A m l = l K M a l ;
(4) (4) and (5) can be directly obtained by Theorem 1;
(6) K < L R A K R A L R A L ̲ ( X ) R A K ̲ ( X ) f ( R A L ̲ ( X ) ) f ( R A K ̲ ( X ) ) ;
(7) K < L R A K R A L R A K ¯ ( X ) R A L ¯ ( X ) f ( R A K ¯ ( X ) ) f ( R A K ¯ ( X ) ) .  □

4. Optimal Scale Selection for Consistent Multi-Scale Decision Information Systems

Definition 9
([21,27]). Let S = ( U , A d ) be a multi-scale decision information system, where S = ( U , A ) is a multi-scale information system, ∑ is the collection of all scale combinations, d A is a decision attribute and d : V V d is a surjective map, where V d = { d ( x ) | x U } is called the domain of d.

4.1. Global Optimal Scale

The decision information system corresponding to the K = ( l 1 , l 1 , , l m ) scale combination is denoted as
S K = ( U , A K d ) = ( U , { a 1 l 1 , a 2 l 2 , , a m l m } d ) .
Obviously, S K = ( U , A K d ) is globally consistent if and only if R A K R d
The relation matrix corresponding to the equivalent relation R A k is M A K , the relation matrix corresponding to the equivalent relation R d is M d .
By Theorem 2, S K = ( U , A K d ) is globally consistent if and only if M A K M d .
Definition 10
([21]). Let S = ( U , A d ) be a multi-scale decision information system, and ∑ is the collection of all scale combinations, For the finer scale K 0 = ( 1 , 1 , , 1 ) , if R A K 0 R d , then we call S to be consistent. If there is a K such that R A K R d , but for any H , K < H , R A H R d does not hold, we call K is a global optimal scale combination.
Theorem 7.
Let S = ( U , A d ) be a multi-scale decision information system, and ∑ the collection of all scale combinations. Then the following conclusions hold:
(1) S is globally consistent if and only if M A K 0 M d ;
(2) S is globally consistent if and only if M A K 0 ( M d ) = 0 ;
(3) A scale combination K is the global optimal scale combination if and only if M A K M d holds, and for any H with K < H , M A H M d does not hold;
(4) Suppose K = ( l 1 , l 2 , , l m ) , let K j = ( l 1 , , l j , , l m ) , then M A K j = M A K M a l j 1 .
Proof. 
(1)S is global consistent R A K 0 R A d M A K 0 M d ;
(2) The result is obvious;
(3) By the definition of the global optimal scale combination, it is easy to be obtained;
(4) As a result of M a l j 1 M a l j , we have M a l j 1 M a l j = M a l j 1 , Therefore, M A K j = M a l 1 a l j 1 a l m = M a l 1 M a l j 1 M a l m = M A K M a l j 1 .  □
Definition 11.
Let S = ( U , A d ) be a multi-scale decision information system and ∑ is the collection of all scale combinations, K = ( l 1 , l 2 , , l m ) . Define the significance of scale combination K as
s i g ( K ) = M A K ( M d )
where A represents the numbers of 1 in the matrix A.
Let S = ( U , A d ) be a multi-scale decision information system, and ∑ is the collection of all scale combinations. By Definition 7, the decision information system corresponding to scale combination K is represented as S K = ( U , A K d ) . Obviously, S K = ( U , A K d ) is globally consistent if s i g ( K ) = 0 .
For the scale combination K = ( l 1 , l 2 , , l m ) , if the decision information system S K = ( U , A K d ) is globally inconsistent, then M A K M d is not true, i.e., s i g ( K ) 0 . The direct subsequent scale of scale combination K is denoted as K j = ( l 1 , , l j 1 , , l m ) ( j = 1 , 2 , , m ) , The global consistence of the decision information system S K j is compared, i.e., the following significance is calculated as
s i g ( K j ) = M A K j ( M d ) = M A K M a j l j 1 ( M d )
It can be seen that if s i g ( K j ) is more smaller, the scale K j = ( l 1 , , l j 1 , , l m ) becomes more significant. This means that this metric combination is selected first, in other words, the information system corresponding to this scale combination is closer to consistent. Based on this idea, a global optimal scale combination selection algorithm of a multi-scale decision system can be described as Algorithm 1.
As Algorithm 1 shows, the algorithm sets up the scale combination variable o p t K and significant variable o p t S i g with the coarsest scale and maximun value initially, and then the scale combination K is put into a queue and starts a iteration. In the iteration, the current scale combination K is take out from the queue and s i g ( K ) is calculated. If the current s i g ( K ) is smaller than o p t S i g , then o p t S i g and o p t K are replaced by the current scale combination K and s i g ( K ) . During the iterating, if the value of o p t S i g become zero then iteration is terminal and optimal scale combination is come out. When the queue is empty and o p t S i g 0 , consequence finer scale combinations are constructed from o p t K and are put into the queue, and starts another iteration.
Algorithm 1: Selecting the global optimal scale combination of a multi-scale decision system
Input: A multi-scale decision system S = ( U , A d ) ;
Output: The global optimal scale combination.
 calculates M d ;
 Let Queue=NULL
 Let o p t K = ( I 1 , I 2 , , I m )               //set the coarsest scale combination initially
 Let o p t S i g = Integer. MAX ̲ VALUE               //set maximum value to o p t S i g initially
 Queue. put( o p t K )
 While (true)
      While (not Queue. empty)
          Let K = Queue. get()              //pick a scale combination K from Queue
          Let s i g ( K ) = M A K ( M d )        //calculate s i g ( K )
          If ( s i g ( K ) o p t S i g )              //choose the minimum s i g ( K ) and its K
               Let o p t S i g = s i g ( K )
               Let o p t K =K
          EndIf
          If ( o p t S i g ==0)       //if the optimal scale combination is found, return.
               return o p t K
          EndIf    
 EndWhile              //loop until Queue. empty
     Queue.clear()       //empty the Queue, and construct next finer scale combination
     Let ( l 1 , l 2 , , l m ) = o p t K
     For each j ( 1 , 2 , , m )               //set finer scale to each attribute, respectively
          If ( l j > 1 )
               Let K = ( l 1 , l 2 , , l j 1 , l j 1 , l j + 1 , , l m )
               Queue. put(K)
          EndIf
     EndFor               //next to search finer scale combination
 EndWhile
Example 4.
In Example 2, we seek the global optimal scale combination of the system corresponding to Table 3.
By Definition 11, the sufficient condition for the decision information system S K = ( U , A K d ) being global consistent is s i g ( K ) = M A K ( M d ) = 0 , where K = ( l 1 , l 2 , , l j , , l m ) .
We apply scale combination relation diagram 1 to illustrate the calculation steps of optimal scale selection in Table 3. We first judge whether the system is consistent or not, and then, according to the definition of scale significant, we look for the optimal scale combination from the scale relation diagram 1 from top to bottom.
(1) For the smallest scale combination K 0 = ( 1 , 1 , 1 ) , according to Theorems 3 and 5, by calculation
M A K 0 = 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 , M d = 0 1 0 0 1 1 0 1 1 0 1 1 0 0 1 0 0 1 0 0 1 1 0 1 0 1 0 0 1 1 0 1 1 0 1 1 0 0 1 0 1 0 1 1 0 0 1 0 0 1 0 0 1 1 0 1 1 0 1 1 0 0 1 0
By Definition 11, s i g ( K 0 ) = M A 0 K ( M d ) = 0 , So, the decision information system S K 0 = ( U , A K 0 d ) is global consistent, by Definition 10, we see that S = ( U , A d ) is global consistent;
(2) For the coarsest scale combination K = ( 4 , 3 , 2 ) , similarly, by calculation, s i g ( 4 , 3 , 2 ) = 8 ;
(3) For each scale combination that belongs to (4,3,2) for the second layer, similarly, by calculation, s i g ( 3 , 3 , 2 ) = 8 ; s i g ( 4 , 2 , 2 ) = 2 ; s i g ( 4 , 3 , 1 ) = 4 , therefore, the scale combination K = ( 4 , 2 , 2 ) is the most important. We continue to search for the optimal scale according to this branch;
(4) For each scale combination that belongs to (4,2,2) for the third layer, by calculation, s i g ( 3 , 2 , 2 ) = 2 ; s i g ( 4 , 1 , 2 ) = 2 , s i g ( 4 , 2 , 1 ) = 0 therefore the decision information system S ( 4 , 2 , 2 ) is global consistent and the scale combination ( 4 , 2 , 1 ) is the global optimal scale combination.
This means that the evaluation of the eight students in Table 3 can be fully decision-making and (4, 2, 2) is an optimal scale combination for maintaining decision-making.

4.2. Local Optimal Scale

Definition 12.
Let S = ( U , A d ) be a consistent multi-scale decision system. and ∑ is the collection of all scale combinations, For x U and the scale combination K , if [ x ] A K [ x ] d , but for any H , K < H , [ x ] A H [ x ] d does not hold, we call K is a local optimal scale combination of S for x.
Theorem 8.
Let S = ( U , A d ) be a consistent multi-scale decision system, and ∑ the collection of all scale combinations. Given a scale combination K , the following conclusion holds:
[ x ] A K [ x ] d f ( [ x ] A K ) f ( [ x ] d ) .
Definition 13.
Let S = ( U , A d ) be a multi-scale decision information system, and ∑ is the collection of all scale combinations, K . then the significance of scale combination K for x i is defined as
s i g ( x i , K ) = f ( [ x i ] A K ) ( f ( [ x i ] d ) )
where f ( [ x i ] A K ) ( f ( [ x i ] d ) ) represents the the numbers of 1 in the boolean vector f ( [ x i ] A K ) ( f ( [ x i ] d ) ) .
If s i g ( x i , K ) = 0 , then f ( [ x ] A K ) f ( [ x ] d ) , then the decision information system S K = ( U , A K d ) is locally consistent for x i . Based on this idea, a local optimal scale combination selection algorithm of a multi-scale decision system can be described as Algorithm 2.
The computation process of Algorithm 2 is similar to Algorithm 1, except the value of s i g ( K ) is the value of s i g ( x , K ) relevant to the given object x.
Example 5.
Look for the local optimal scale combination for x 1 in Table 3.
We apply scale combination relation diagram 1 to illustrate the calculation steps of optimal local scale selection for x 1 , by Definition 13, s i g ( x i , K ) = f ( [ x i ] A K ) ( f ( [ x i ] d ) ) .
(1) For the smallest scale combination, K 0 = ( 1 , 1 , 1 ) , by calculation, s i g ( x 1 , K ) = 0 , the system S K 0 is local consistent for x 1 ;
(2) For the coarsest scale combination K = ( 4 , 3 , 2 ) , by calculation, s i g ( x 1 , K ) = 1 , Hence, S K = ( U , A K d ) is local inconsistent for x 1 ;
(3) For each scale combination that belongs to (4,3,2) for the second layer, by calculation, s i g ( x 1 , 3 , 3 , 2 ) = 1 ; s i g ( x 1 , 4 , 2 , 2 ) = 0 ; s i g ( x 1 , 4 , 3 , 1 ) = 1 , Hence the scale combination ( 4 , 2 , 2 ) is the local optimal scale combination for x 1 .
Algorithm 2: Selecting the local optimal scale combination of a multi-scale decision system
Input: A consistent multi-scale decision systems S = ( U , A d ) and an object x U .
Output: A local optimal scale combination for x.
 Calculates f ( [ x ] d ) ;
 Let Queue=NULL
 Let o p t K = ( I 1 , I 2 , , I m )               //set the coarsest scale combination initially
 Let o p t S i g = Integer. MAX ̲ VALUE               //set maximum value to o p t S i g initially
 Queue. put( o p t K )
 While (true)
      While (not Queue.empty)
          Let K = Queue.get()                     //pick a scale combination K from Queue
          Let s i g ( K ) = f ( [ x ] A K ) ( f ( [ x ] d )        //calculate s i g ( x , K )
          If ( s i g ( K ) o p t S i g )              //choose the minimum s i g ( x , K ) and its K
               Let o p t S i g = s i g ( K )
               Let o p t K = K
          EndIf
          If ( o p t S i g ==0)       //if the optimal scale combination is found, return.
               return o p t K
          EndIf
     EndWhile              //loop until Queue. empty
     Queue.clear()       //empty the Queue, and construct next finer scale combination
     Let ( l 1 , l 2 , , l m ) = o p t K
     For each j ( 1 , 2 , , m )               //set finer scale to each attribute, respectively
          If ( l j > 1 )
               Let K = ( l 1 , l 2 , , l j 1 , l j 1 , l j + 1 , , l m )
               Queue. put(K)
          EndIf
     EndFor                                    //next to search finer scale combinations
 EndWhile

5. Local Optimal Scale Selection for Inconsistent Generalized Decision Information Systems

Let S = ( U , A d ) be a decision information system which defined by Definition 3, where U = { x 1 , x 2 , , x n } is the universe, A = { a 1 , a 2 , , a m } is the set of conditional attributes, d A is the decision attribute. If the decision information system S is locally consistent for x U , then d ( [ x ] A ) is unique. If the decision information system is globally consistent, d ( [ x ] A ) is unique for any x U . For an inconsistent decision information system, x U , such that d ( [ x ] A ) is not unique. For this reason, Wu and Leung put forward a definition of generalized decision in their paper [21].
Definition 14
([21]). Let S = ( U , A d ) be a decision information system. For B A , the generalized decisions of object x U is defined as
B ( x ) = d ( [ x ] B ) = { d ( y ) | y [ x ] B } .
Definition 15.
let A = ( a i j ) m × n , B = ( b i j ) m × n are m × n matrices. Define A B = ( a i j × b i j ) m × n .
The following theorem is easily obtained from Definitions 14 and 15.
Theorem 9.
Let S = ( U , A d ) be a decision information system, Let d j = d ( x j ) , x j U , K A ( i ) = f ( [ x i ] A ) ( d 1 , d 2 , , d n ) , H A ( i ) represents the set of all component elements in vector K A ( i ) , for convenience, we denote H A ( i ) = { d | d K A ( i ) } . If d j 0 , ( j = 1 , 2 , , n ) , then
A ( x i ) = H A ( i ) .
It should be pointed out that if there exist a x U and d ( x ) = 0 , the above theorem is not suitable. For this case, we may assign another value to the attribute d for x, which does not affect the classification and decision making.
Example 6.
Find the generalized decision of each object in the following inconsistent decision information system Table 4.
By calculation d = ( d ( x 1 ) , d ( x 2 ) , , d ( x n ) ) = ( 4 , 2 , 1 , 3 , 2 , 2 ) , f ( [ x 1 ] A ) = ( 1 , 1 , 0 , 0 , 0 , 0 ) . According to Theorem 9, we have K A ( 1 ) = f ( [ x 1 ] A ) d = ( 4 , 2 , 0 , 0 , 0 , 0 ) , therefore A ( x 1 ) = H A ( 1 ) = { 2 , 4 } . A ( x 1 ) = { 2 , 4 } , A ( x 2 ) = { 2 , 4 } , A ( x 3 ) = { 1 , 2 } , A ( x 4 ) = { 3 } , A ( x 5 ) = { 2 } , A ( x 6 ) = { 1 , 2 } can be obtained by the same method.
Definition 16.
Let S = ( U , A d ) be a multi-scale inconsistent decision information system, and ∑ is the collection of all scale combinations, the decision information systems corresponding to the scale combination K is S K = ( U , A K d ) . If A K ( x ) = A K 0 ( x ) holds, we call S K to be local generalized consistent for x. If S K is local generalized consistent for x, but for any L = ( l 1 , l 2 , , l m ) with K < L , S L = ( U , A L d ) is local generalized inconsistent for x, we call K the optimal local generalized consistence for x.
Theorem 10.
S K is a local generalized consistence for x i if and only if H A K ( i ) = H A K 0 ( i ) holds.
Proof. 
By Theorem 9, S K is local consistent for x i A K ( x i ) = A 0 K ( x i ) H A K ( i ) = H A K 0 ( i ) .
Theorem 11.
Let S = ( U , A d ) be an inconsistent multi-scale decision system, and ∑ is the collection of all scale combinations, K . Then K is the optimal scale combination for x i if and only if H A K ( i ) = H A K 0 ( i ) holds. Additionally, for any K < L , H A L ( i ) H A K 0 ( i ) .
Definition 17.
Let S = ( U , A d ) be an inconsistent multi-scale decision system, and ∑ is the collection of all scale combinations, K = ( l 1 , , l j , , l m ) . Then the significance of scale combination K for x i is defined as
S I G ( x i , K ) = H A K ( i ) H A K 0 ( i )
where H A K ( i ) H A K 0 ( i ) represents the cardinality of the set H A K ( i ) H A K 0 ( i ) .
If S I G ( x i , K ) = 0 , then H A K ( i ) H A K 0 ( i ) . Therefore, the decision information system S = ( U , A d ) is locally generalized consistent for x i . The computation process of Algorithm 3 is similar to Algorithm 2.
Algorithm 3: Selecting the global optimal scale combination of a multi-scale decision system
Input: A inconsistent multi-scale decision system S = ( U , A d ) and an object x U .
Output: A local generalized optimal scale combination for x.
 Let K 0 = ( 1 , 1 , , 1 ) ;                                     //set the finest scale combination
 Calculates H A K 0 ( x ) ;       // H A K 0 ( x ) correspondents to H A ( i ) with K 0 and x i in Theorem 9
 Let Queue=NULL
 Let o p t K = ( I 1 , I 2 , , I m )               //set the coarsest scale combination initially
 Let o p t S i g = Integer. MAX ̲ VALUE               //set maximum value to o p t S i g initially
 Queue. put( o p t K )
 While (true)
      While (not Queue.empty)
          Let K = Queue. get()              //pick a scale combination K from Queue
          Calculates H A K ( x ) ;    // H A K ( x ) correspondents to H A ( i ) with K and x i in Theorem 9
          Let s i g ( x , K ) = H A K ( x ) H A K 0 ( x )               //calculate s i g ( x , K )
          If ( s i g ( x , K ) o p t S i g )              //choose the minimum s i g ( x , K ) and its K
               Let o p t S i g = s i g ( x , K )
               Let o p t K = K
          EndIf
          If ( o p t S i g ==0)       //if the optimal scale combination is found, return.
               return o p t K
          EndIf    
     EndWhile              //loop until Queue. empty
     Queue.clear()       //empty the Queue, and construct next finer scale combination
     Let ( l 1 , l 2 , , l m ) = o p t K
     For each j 1 , 2 , , m        //set finer scale to each attribute, respectively
          If ( l j > 1 )
               Let K = ( l 1 , l 2 , , l j 1 , l j 1 , l j + 1 , , l m )
               Queue. put(K)
          EndIf
     EndFor              //next to search finer scale combination
 EndWhile
Example 7.
Search for the generalized local optimal scale combination of the following system Table 5 for x i .
(1) d = ( d ( x 1 ) , d ( x 2 ) , d ( x 3 ) , d ( x 4 ) , d ( x 5 ) , d ( x 6 ) ) = ( 1 , 2 , 4 , 4 , 4 , 3 ) , for scale combination K 0 = ( 1 , 1 , 1 ) , f ( [ x 1 ] A K 0 ) d = ( 1 , 2 , 0 , 0 , 0 , 0 ) , H A K 0 ( 1 ) = ( 1 , 2 ) ,
(2) For the coarsest scale K = ( 3 , 3 , 2 ) , f ( [ x 1 ] A K ) d = ( 1 , 2 , 4 , 0 , 0 , 3 ) , H A K ( 1 ) = ( 1 , 2 , 3 , 4 ) and S I G ( x 1 , K ) = 2 therefore, K = ( 3 , 3 , 2 ) is not the generalized local optimal scale for x 1 ;
(3) For each scale combination that belongs to (3,3,2) for the second layer, by calculation, S I G ( x 1 , 2 , 3 , 2 ) = 1 ; S I G ( x 1 , 3 , 2 , 2 ) = 1 ; S I G ( x 1 , 3 , 3 , 1 ) = 0 , and hence the ( 3 , 3 , 1 ) is the generalized local optimal scale combination for x 1 .

6. Conclusions

This paper introduces relational matrices and matrix calculations into multi-scale decision information systems. Some properties of multi-scale decision information systems are discussed based on matrix methods. Using the relation matrix to introduce the significance of scale combination, the method of optimal scale combination selection is studied, and the related algorithm is designed. The effectiveness of the method is illustrated by experiments. In the future, we will use the matrix method to study the classification and decision-making methods of multi-scale information systems.

7. Experiments and Analysis

In order to verify whether the Algorithms 1 and 2 are able to practically apply to choose the optimal combined-scale in multi-scale decision information system, some University of California Irvine (UCI) datasets are applied in the experiments. Table 6 shows the description of these datasets.
These data sets in Table 6 are single-scale decision datasets, and to extend these datasets from single-scale to multi-scale, some methods in [27] are adopted, which are described below.
Firstly, the finest level of scale of each attribute, that is, a 1 ( x ) is the value of x U at attribute a in the original dataset.
Secondly, for each attribute a, the standard deviation and minimum are denoted as s t d ( a ) and m i n ( a ) , then the second level of scale of x U at the attribute a in multi-scale dataset is denoted:
a 2 ( x ) = a ( x ) m i n ( a ) s t d ( a ) ,
where y means the largest integer v satisfied v y .
Thirdly, based on the previous level of scale, the next scale of attribute a can be made by merging some equivalence classes in previous level of scale. For example, suppose the equivalence classes of previous level of scale denoted S k = { a k ( x ) | x U } = { { 5 } , { 6 } , { 8 } , { 12 } , { 20 } } , then in the next level of scale, the class { 5 } and class { 6 } are merged into { 5 , 6 } , which denoted by { 6 } . That is, if a k ( x ) is belong to { 5 } or { 6 } , then a k + 1 ( x ) is { 6 } , and the consequence equivalence classes is S k + 1 = { { 6 } , { 8 } , { 12 } , { 20 } } .
Table 7 shows the result of experiments using multi-scale optimal combined dataset according to the Algorithm 1. In Table 7 it presents the Support Vector Machine (SVM) classification accuracy both under the raw datasets and the refined datasets of optimal combined multi-scale, and it demonstrates the Algorithm 1 works well.
The Table 7 also presents the rates of optimal, that is the percent of the attributes with level of scale ≥2 in the given refined dataset. For example, there are seven attributes in Auto_MPG raw dataset, the optimal combined scale of the refined dataset outcomes in the experiment is [ 1 , 1 , 1 , 5 , 5 , 3 , 1 ] , which indicates three of seven attributes are optimal, and the rate of optimal is 3 7 = 42.9 % . Similarly, the average of level of scale (LoS) in the refined Auto_MPG dataset is 17 7 = 2.43 , and relatively the LoS of raw Auto_MPG dataset is 1.
To compare the local multi-scale optimal method of Algorithm 2 with the global multi-scale optimal method of Algorithm 1, each datasets randomly select ≤5% instances of all and the final optimal combined-scale is the intersection of every optimal combined-scales obtained by the every single instance. For example, two instances are randomly selected in Auto_MPG, and their optimal combined-scales are [1,3,4,5,5,7,4] and [3,2,4,5,5,6,1] respectively, thus their final intersected combined-scale is [1,3,4,5,5,7,4]∧[3,2,4,5,5,6,1]=[1,2,4,5,5,6,1].
Table 8 shows the SVM accuracy, Rate of optimal, Average of LoS and time cost in Algorithm 2 relatively compare to Algorithm 2. It demonstrates that, the same as optimal performances of global multi-scale optimal algorithm can be achieved by choosing a small amount of instances in local multi-scale optimal algorithm.

Author Contributions

All authors contributed equally to this work.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant (No:11701258,11871259), Program for Innovative Research Team in Science and Technology in Fujian Province University, and Quanzhou High-Level Talents Support Plan under Grant 2017ZT012.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zadeh, L.A. Fuzzy sets and information granularity. Adv. Fuzzy Set Theory Appl. 1979, 11, 3–18. [Google Scholar]
  2. Pedrycz, W. Granular Computing: An Introduction; Physica-Verlag: Heidelberg, Germany, 2000; pp. 309–328. [Google Scholar] [CrossRef]
  3. Yao, Y.Y.; Miao, D.Q.; Xu, F.F. Granular Structures and Approximations in Rough Sets and Knowledge Spaces. In Rough Set Theory: A True Landmark in Data Analysis; Springer: Berlin/Heidelberg, Germany, 2009; pp. 71–84. [Google Scholar] [CrossRef]
  4. Li, J.; Ren, Y.; Mei, C.; Qian, Y.; Yang, X. A comparative study of multigranulation rough sets and concept lattices via rule acquisition. Knowl.-Based Syst. 2016, 91, 152–164. [Google Scholar] [CrossRef]
  5. Pal, S.K.; Ray, S.S.; Ganivada, A. Introduction to Granular Computing, Pattern Recognition and Data Mining. In Granular Neural Networks, Pattern Recognition and Bioinformatics; Springer International Publishing: Cham, Switzerland, 2017; pp. 1–37. [Google Scholar] [CrossRef]
  6. Han, S.E. Roughness measures of locally finite covering rough sets. Int. J. Approx. Reason. 2019, 105, 368–385. [Google Scholar] [CrossRef]
  7. Liu, H.; Cocea, M. Granular computing-based approach for classification towards reduction of bias in ensemble learning. Granul. Comput. 2017, 2, 131–139. [Google Scholar] [CrossRef]
  8. Wang, G.; Yang, J.; Xu, J. Granular computing: From granularity optimization to multi-granularity joint problem solving. Granul. Comput. 2017, 2, 105–120. [Google Scholar] [CrossRef]
  9. Sun, B.Z.; Ma, W.M.; Qian, Y.H. Multigranulation fuzzy rough set over two universes and its application to decision making. Knowl.-Based Syst. 2017, 123, 61–74. [Google Scholar] [CrossRef]
  10. Yao, Y. Three-way decision and granular computing. Int. J. Approx. Reason. 2018, 103, 107–123. [Google Scholar] [CrossRef]
  11. Mo, J.; Huang, H.L. (T, S)-Based Single-Valued Neutrosophic Number Equivalence Matrix and Clustering Method. Mathematics 2019, 7, 36. [Google Scholar] [CrossRef]
  12. Petrosino, A.; Ferone, A. Feature Discovery through Hierarchies of Rough Fuzzy Sets. In Granular Computing and Intelligent Systems: Design with Information Granules of Higher Order and Higher Type; Pedrycz, W., Chen, S.M., Eds.; Springer: Berlin/Heidelberg, Germany, 2011; pp. 57–73. [Google Scholar] [CrossRef]
  13. Ferone, A. Feature selection based on composition of rough sets induced by feature granulation. Int. J. Approx. Reason. 2018, 101, 276–292. [Google Scholar] [CrossRef]
  14. Pawlak, Z. Imprecise Categories, Approximations and Rough Sets. In Rough Sets: Theoretical Aspects of Reasoning about Data; Springer: Dordrecht, The Netherlands, 1991; pp. 9–32. [Google Scholar] [CrossRef]
  15. Zhu, P.F.; Hu, Q.H.; Zuo, W.M.; Yang, M. Multi-granularity distance metric learning via neighborhood granule margin maximization. Inf. Sci. 2014, 282, 321–331. [Google Scholar] [CrossRef]
  16. Zhu, P.F.; Hu, Q.H. Adaptive neighborhood granularity selection and combination based on margin distribution optimization. Inf. Sci. 2013, 249, 1–12. [Google Scholar] [CrossRef]
  17. Tan, A.H.; Wu, W.Z.; Li, J.J.; Lin, G.P. Evidence-theory-based numerical characterization of multigranulation rough sets in incomplete information systems. Fuzzy Sets Syst. 2016, 294, 18–35. [Google Scholar] [CrossRef]
  18. Yao, Y.; She, Y. Rough set models in multigranulation spaces. Inf. Sci. 2016, 327, 40–56. [Google Scholar] [CrossRef]
  19. Lin, G.P.; Liang, J.Y.; Qian, Y.H. An information fusion approach by combining multigranulation rough sets and evidence theory. Inf. Sci. 2015, 314, 184–199. [Google Scholar] [CrossRef]
  20. Yang, X.B.; Song, X.N.; Chen, Z.H.; Yang, J.Y. On multigranulation rough sets in incomplete information system. Int. J. Mach. Learn. Cybern. 2012, 3, 223–232. [Google Scholar] [CrossRef]
  21. Wu, W.Z.; Leung, Y. Theory and applications of granular labelled partitions in multi-scale decision tables. Inf. Sci. 2011, 181, 3878–3897. [Google Scholar] [CrossRef]
  22. Wu, W.Z.; Leung, Y. Optimal scale selection for multi-scale decision tables. Int. J. Approx. Reason. 2013, 54, 1107–1129. [Google Scholar] [CrossRef]
  23. Gu, S.M.; Wu, W.Z. On knowledge acquisition in multi-scale decision systems. Int. J. Mach. Learn. Cybern. 2013, 4, 477–486. [Google Scholar] [CrossRef]
  24. Wu, W.Z.; Qian, Y.; Li, T.J.; Gu, S.M. On rule acquisition in incomplete multi-scale decision tables. Inf. Sci. 2017, 378, 282–302. [Google Scholar] [CrossRef]
  25. She, Y.H.; Li, J.H.; Yang, H.L. A local approach to rule induction in multi-scale decision tables. Knowl.-Based Syst. 2015, 89, 398–410. [Google Scholar] [CrossRef]
  26. Li, F.; Hu, B.Q.; Wang, J. Stepwise optimal scale selection for multi-scale decision tables via attribute significance. Knowl.-Based Syst. 2017, 129, 4–16. [Google Scholar] [CrossRef]
  27. Li, F.; Hu, B.Q. A new approach of optimal scale selection to multi-scale decision tables. Inf. Sci. 2017, 381, 193–208. [Google Scholar] [CrossRef]
  28. Shafer, G. A Mathematical Theory of Evidence; Princeton University Press: Princeton, NJ, USA, 1976. [Google Scholar]
  29. Guan, J.W.; Bell, D.A.; Guan, Z. Matrix computation for information systems. Inf. Sci. 2001, 131, 129–156. [Google Scholar] [CrossRef]
  30. Liu, G.L. Rough Sets over the Boolean Algebras. In Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing; Ślęzak, D., Wang, G., Szczuka, M., Düntsch, I., Yao, Y., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; pp. 124–131. [Google Scholar]
  31. Wang, L.; Li, T.R. Matrix-Based Computational Method for Upper and Lower Approximations of Rough Sets. Pattern Recognit. Artif. Intell. 2011, 24, 756–762. [Google Scholar] [CrossRef]
  32. Huang, Y.Y.; Li, T.R.; Luo, C.; Horng, S.J. Matrix-Based Rough Set Approach for Dynamic Probabilistic Set-Valued Information Systems. In Rough Sets; Springer International Publishing: Cham, Switzerland, 2016; pp. 197–206. [Google Scholar]
  33. Tan, A.H.; Li, J.J.; Lin, Y.J.; Lin, G.P. Matrix-based set approximations and reductions in covering decision information systems. Int. J. Approx. Reason. 2015, 59, 68–80. [Google Scholar] [CrossRef]
  34. Tan, A.H.; Li, J.J.; Lin, G.P.; Lin, Y.J. Fast approach to knowledge acquisition in covering information systems using matrix operations. Knowl.-Based Syst. 2015, 79, 90–98. [Google Scholar] [CrossRef]
  35. Luo, C.; Li, T.R.; Yi, Z.; Fujita, H. Matrix approach to decision-theoretic rough sets for evolving data. Knowl.-Based Syst. 2016, 99, 123–134. [Google Scholar] [CrossRef]
  36. Hu, C.X.; Liu, S.X.; Liu, G.X. Matrix-based approaches for dynamic updating approximations in multigranulation rough sets. Knowl.-Based Syst. 2017, 122, 51–63. [Google Scholar] [CrossRef]
  37. Karczmarek, P.; Kiersztyn, A.; Pedrycz, W. An Application of Graphic Tools and Analytic Hierarchy Process to the Description of Biometric Features. In Artificial Intelligence and Soft Computing; Rutkowski, L., Scherer, R., Korytkowski, M., Pedrycz, W., Tadeusiewicz, R., Zurada, J.M., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 137–147. [Google Scholar]
  38. Xu, Y.H.; Wu, W.Z.; Tan, A.H. Optimal Scale Selections in Consistent Generalized Multi-scale Decision Tables. Rough Sets 2017, 10313, 185–198. [Google Scholar] [CrossRef]
  39. Gu, S.; Gu, J.; Wu, W.; Li, T.; Chen, C. Local optimal granularity selections in incomplete multi-granular decision systems. J. Comput. Res. Dev. 2017, 54, 1500–1509. [Google Scholar] [CrossRef]
  40. Wu, W.Z.; Chen, C.J.; Li, T.J.; Xu, Y.H. Comparative Study on Optimal Granularities in Inconsistent Multi-granular Labeled Decision Systems. Pattern Recognit. Artif. Intell. 2016, 29, 1095–1103. [Google Scholar] [CrossRef]
  41. Wu, W.Z.; Yang, L.; Tan, A.H.; Xu, Y.H. Granularity Selections in Generalized Incomplete Multi-Granular Labeled Decision Systems. J. Comput. Res. Dev. 2018, 55, 1263–1272. [Google Scholar] [CrossRef]
Figure 1. Partial ordered lattices of scale combinations.
Figure 1. Partial ordered lattices of scale combinations.
Mathematics 07 00290 g001
Table 1. A decision syetem.
Table 1. A decision syetem.
U a 1 a 2 a 3 a 4 d
x 1 21111
x 2 32222
x 3 21111
x 4 22113
x 5 11425
x 6 11321
x 7 32222
Table 2. Evaluation criteria.
Table 2. Evaluation criteria.
Score Evaluation Criteria 1 Evaluation Criteria 2 Evaluation Criteria 3 Evaluation Criteria 4
( 90 , 100 ] 1 e x c e l l e n t s u p e r Y
[ 80 , 89 ) 2 g o o d s u p e r Y
[ 70 , 79 ) 3 f a i r m i d d l e Y
[ 60 , 69 ) 4 p a s s m i d d l e Y
[ 50 , 59 ) 5 b a d l o w N
[ 0 , 49 ) 6 b a d l o w N
Table 3. A generalized multi-scale information system.
Table 3. A generalized multi-scale information system.
U Subject 1 a 1 1 a 1 2 a 1 3 a 1 4 Subject 2 a 2 1 a 2 2 a 2 3 Subject 3 a 3 1 a 3 2 d
x 1 783FMY941EY753Y1
x 2 614PMY624PY604Y0
x 3 694PMY783FY684Y1
x 4 981ESY585BN971Y1
x 5 366BLN723FY921Y0
x 6 634PMY763FY475N0
x 7 585BLN951EY921Y1
x 8 356BLN901EY733Y0
Table 4. An inconsistent decision system.
Table 4. An inconsistent decision system.
U a 1 a 2 a 3 d
x 1 0ES4
x 2 0ES2
x 3 1FL1
x 4 2BM3
x 5 0FS2
x 6 1FL2
Table 5. A generalized inconsistent multi-scale information system.
Table 5. A generalized inconsistent multi-scale information system.
U a 1 1 a 1 2 a 1 3 a 2 1 a 2 2 a 2 3 a 3 1 a 3 2 d
x 1 GSY2SY1S1
x 2 GSY2SY1S2
x 3 FMY4MY2S4
x 4 BLN5MY8L4
x 5 BLN7LN6M4
x 6 ESY3SY2S3
Table 6. The UCI dataset description.
Table 6. The UCI dataset description.
DatasetsInstancesAttributesClasses
Auto_MPG39273
CarEvaluation172863
Wine178133
German1000202
WDBC569302
Ionosphere351342
Table 7. Experiments results of Algorithm 1.
Table 7. Experiments results of Algorithm 1.
DatasetsOrignal-Dataset (SVM)Optimal MSC-Dataset (SVM)Rates of OptimalAverage of LoS
Auto_MPG0.66110.694942.9%2.43
CarEvaluation0.92670.949950.0%1.50
Wine0.35180.870392.3%5.15
German0.72710.723380.0%2.65
WDBC0.61410.959293.3%4.43
Ionosphere0.91520.933991.2%3.76
Table 8. The experiments results of Algorithm 2 compares to Algorithm 1.
Table 8. The experiments results of Algorithm 2 compares to Algorithm 1.
DatasetsPercent ofRelative AccuracyRelative RateRelative AverageRelative Time
Instances(SVM)of Optimalof LoSCost
Auto_MPG0.5%108.5%200.0%158.8%40.2%
CarEvaluation0.8%98.2%133.3%111.1%203.1%
Wine2.2%104.3%108.3%100.0%78.3%
German1.0%95.4%125.0%128.3%14.8%
WDBC0.4%87.8%103.6%151.1%70.6%
Ionosphere4.8%99.0%100.0%97.7%60.9%

Share and Cite

MDPI and ACS Style

Chen, Y.S.; Li, J.J.; Huang, J.X. Matrix Method for the Optimal Scale Selection of Multi-Scale Information Decision Systems. Mathematics 2019, 7, 290. https://doi.org/10.3390/math7030290

AMA Style

Chen YS, Li JJ, Huang JX. Matrix Method for the Optimal Scale Selection of Multi-Scale Information Decision Systems. Mathematics. 2019; 7(3):290. https://doi.org/10.3390/math7030290

Chicago/Turabian Style

Chen, Ying Sheng, Jin Jin Li, and Jian Xin Huang. 2019. "Matrix Method for the Optimal Scale Selection of Multi-Scale Information Decision Systems" Mathematics 7, no. 3: 290. https://doi.org/10.3390/math7030290

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop