Next Article in Journal
A Class of Quadratic Polynomial Chaotic Maps and Their Fixed Points Analysis
Next Article in Special Issue
Estimating Topic Modeling Performance with Sharma–Mittal Entropy
Previous Article in Journal
An Image Encryption Scheme Based on Block Scrambling, Modified Zigzag Transformation and Key Generation Using Enhanced Logistic—Tent Map
Previous Article in Special Issue
Model Selection for Non-Negative Tensor Factorization with Minimum Description Length
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Double-Granule Conditional-Entropies Based on Three-Level Granular Structures

1
School of Mathematical Sciences, Sichuan Normal University, Chengdu 610066, China
2
Institute of Intelligent Information and Quantum Information, Sichuan Normal University, Chengdu 610066, China
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(7), 657; https://doi.org/10.3390/e21070657
Submission received: 30 April 2019 / Revised: 26 June 2019 / Accepted: 27 June 2019 / Published: 3 July 2019
(This article belongs to the Special Issue Information-Theoretical Methods in Data Mining)

Abstract

:
Rough set theory is an important approach for data mining, and it refers to Shannon’s information measures for uncertainty measurements. The existing local conditional-entropies have both the second-order feature and application limitation. By improvements of hierarchical granulation, this paper establishes double-granule conditional-entropies based on three-level granular structures (i.e., micro-bottom, meso-middle, macro-top), and then investigates the relevant properties. In terms of the decision table and its decision classification, double-granule conditional-entropies are proposed at micro-bottom by the dual condition-granule system. By virtue of successive granular summation integrations, they hierarchically evolve to meso-middle and macro-top, to respectively have part and complete condition-granulations. Then, the new measures acquire their number distribution, calculation algorithm, three bounds, and granulation non-monotonicity at three corresponding levels. Finally, the hierarchical constructions and achieved properties are effectively verified by decision table examples and data set experiments. Double-granule conditional-entropies carry the second-order characteristic and hierarchical granulation to deepen both the classical entropy system and local conditional-entropies, and thus they become novel uncertainty measures for information processing and knowledge reasoning.

1. Introduction

Rough set theory can effectively implement data mining for the imprecise, inconsistent, and incomplete information [1], and it has been extensively applied in artificial intelligence and machine learning [2,3,4,5,6,7,8]. In rough set theory, attribute reduction based on decision tables is a main topic for approximate reasoning and knowledge discovery, and there are three main construction strategies: from the positive region, information measure, and a discernibility matrix  [9,10,11,12,13,14,15]. By virtue of the discernibility matrix, Wei et al. [16] proposed an incremental reduction algorithm for dynamic data; Ma et al. [17] utilized the compressed binary discernibility matrix to construct an incremental reduction algorithm for group dynamic data; moreover, Nie and Zhou [18] proposed a new discernibility matrix defined by local conditional-entropies to compute the reduction core.
Information theory originated from Shannon’s entropy system [19], and it provides an effective method for uncertainty measurement, such as in attribute reduction. Currently, information theory has been introduced into rough set theory for uncertainty analyses and information processing [20,21,22,23,24,25]. As far as attribute reduction is concerned, Miao [26] offered the informational representation of knowledge reduction and decision reduction, where entropy and mutual-information are highlighted; Wang et al. [27] conducted a comparative study on attribute reduction from the algebra and information viewpoints, where the conditional-entropy acts as a main tool; Jiang et al. [28] presented the relative decision entropy to propose a feature selection algorithm; Slezak [29] used the conditional-entropy to define approximate reducts; moreover, Qian and Shu [30] provided the mutual information criterion to evaluate candidate features in incomplete data. In general, the entropy, conditional-entropy, and mutual-information together constitute the classical information system with integrality and comprehensiveness, and they can function on rough set applications (such as attribute reduction) but may exhibit different emphases in different application scenarios. In addition, information-theoretic measures have multiple variational forms [31,32,33,34,35]. As far as conditional-entropies are concerned, they are extensively applied in rough set theory from multiple pointcuts [26,27,29,31,34,36,37,38,39], while uncertainty measurement and reduction construction still serve as two basic issues. Aiming at probabilistic rough sets, Deng and Yao [40,41] used Shannon’s entropy and conditional-entropy to interpret and determine probabilistic thresholds by an information-theoretic approach, and Ma et al. [42] considered variants of conditional-entropies to construct heuristic reduction algorithms for the probabilistic model. In particular, local conditional-entropies are put forward by adopting double condition-granules and their union locality [18], and they can distinctively determine a new discernibility matrix for reduction core computation; moreover, the information measures exhibit a novel feature of second-order expressions, especially when compared to the traditional entropy system with only single-granule descriptions [19,26,27].
Granular computing is a structural methodology of hierarchical computing and information processing [43,44], and its technology of multi-granularity and multiple levels is useful for uncertainty analyses and knowledge acquisition regarding data. In rough set theory, the information granulation is of extensive concern [45,46,47,48,49], and the granulation monotonicity plays an important role in attribute reduction [12,50,51,52]. In particular, a decision table acts as a formal background of data mining [12,53,54,55], and it involves condition/decision granules and classifications from granular structures. According to granular computing, Zhang and Miao [56] introduced three-layer granular structures of decision tables, and they further hierarchically constructed three-way informational measures based on weighted-entropies; moreover, Wang et al. [57] utilized three-layer granular structures to research three-way weighted combination-entropies. These studies adhere to three-level analyses, and the latter are directly related to granular computing [43] and three-way decisions [58], as well as their interplay. Recently, Yao [59] discussed three-way granular computing by making use of two particular types of three granules and three levels, where thinking in three levels results in an important model. Additionally, three-level analyses were extensively utilized in the location allocation and programming/optimization modeling [60,61,62].
According to [18], the new discernibility matrix is used for reduction core calculations, and its creative implementation mainly depends on local conditional-entropies. Therefore, local conditional-entropies focus on the granule-union locality rather than their underlying double-granule interaction, and the latter more essentially adheres to the second-order characteristic; moreover, they lack the condition granulation to restrict their uncertainty measurement function and information procession prospect based on knowledge. Motivated by the two issues, this paper utilizes the two-granular essence and three-hierarchical evolution to propose double-granule conditional-entropies based on three-level granular structures. Regarding the contribution, this novel type of information measures improves local conditional-entropies from both the granular interaction and hierarchical/conditional granulation, and they will achieve multiple important properties (including the integration hierarchy, number distribution, calculation algorithm, three bounds, and granulation non-monotonicity) to offer both robust measurement functions and knowledge-application prospects. Moreover, three-level granular structures here (including micro-bottom, meso-middle, macro-top) adopt only the condition part of decision table, and thus they differ from and push forward the previous ones, which include both the condition and decision parts [56].
The remainder of this paper is organized as follows. Section 2 reviews the decision table and local conditional-entropies; Section 3 proposes and studies double-granule conditional-entropies from three-level granular structures; Section 4 provides a decision table example for mechanism illustration; Section 5 makes data experiments for effectiveness verification; finally, Section 6 concludes this paper.

2. Decision Table and Its Existing Entropy Measures

Rough set theory [1] focuses on the data that are represented in an information table
( U , A T , { V a : a A T } , { I a : a A T } ) ;
U is the universe with finite objects, A T is the finite attribute set, V a is the value domain for a A T , and I a : U V a is an information function to endow each object x with a value I a ( x ) = a ( x ) on attribute a. The decision table is a special type of information table with A T = C D and C D = , where C and D denote the sets of condition attribute and decision attribute, respectively, and it is simply denoted by ( U , C D ) in this paper. Furthermore, the granulation construction usually considers two parts.
(1)
The condition attribute subset A C induces an equivalence relation
I N D ( A ) = { ( x , y ) U × U : a A , a ( x ) = a ( y ) } ,
and the latter provides the condition granulation or partition U / I N D ( A ) = { A i : i = 1 , , n } , where A i = [ x ] A i represents the equivalence granule to exhibit number | U / I N D ( A ) | = n .
(2)
Similarly, the decision attribute set D induces the equivalence relation I N D ( D ) and further decision classification U / I N D ( D ) = { D j : j = 1 , , m } , which consists of | U / I N D ( D ) | = m decision classes.
The decision table ( U , C D ) and its granulation from A C and D constitute the basic background for information measure construction. The probability space ( U , 2 U , P ) establishes the usual probability framework, where
P : 2 U Q , P ( X ) = | X | | U | , X U ,
and thus two usual probabilities are
P ( A i ) = | A i | | U | , P ( D j / A i ) = | A i D j | | A i | .
Definition 1
([26,27,56]). The entropy on condition A, conditional-entropy on D given A, and mutual-information between A and D are respectively defined by
H ( A ) = i = 1 n P ( A i ) log 2 P ( A i ) , H ( D / A ) = i = 1 n P ( A i ) j = 1 m P ( D j / A i ) log 2 P ( D j / A i ) , I ( A ; D ) = H ( D ) H ( D / A ) ,
where
H ( D ) = j = 1 m P ( D j ) log 2 P ( D j ) .
Theorem 1
([26,27,56]). The entropy, conditional-entropy, and mutual-information have granulation monotonicity. Concretely,
U / I N D ( A ) U / I N D ( B ) H ( B ) H ( A ) , H ( D / B ) H ( D / A ) , I ( B ; D ) I ( A ; D ) .
In terms of the decision table ( U , C D ) , the classical system of Shannon entropies has been introduced into rough set theory, as shown by Definition 1 and Theorem 1. As three basic information measures, the entropy, conditional-entropy, and mutual-information have uncertainty semantics and granulation monotonicity, so they are extensively used in attribute reduction and heuristic algorithms [26,27,42]. The granulation relation U / I N D ( A ) U / I N D ( B ) is equivalent to I N D ( A ) I N D ( B ) , that is,
B i * U / I N D ( B ) , A i U / I N D ( A ) , s . t . , B i * A i ,
and it is usually induced by A B C ; furthermore, relevant granulation monotonicity/non-monotonicity becomes an important index to assess and apply uncertainty measures.
According to the decision table and its formal structure, Zhang and Miao [56] recently introduced three-level granular structures, i.e.,
micro-bottom   ( A i , D j ) ,   meso-middle   ( U / I N D ( A ) , D j ) ,   macro-top   ( U / I N D ( A ) , U / I N D ( D ) ) ,
and further investigated weighted-entropy constructions. As a result, the previous entropy system (Equation (3)) is actually located at macro-top and has an equivalent construction from the weighted-entropy system; at meso-middle, Zhang et al. [10] established three-way informational class-specific reducts to be compared with the algebraic class-specific reducts [9].
In particular, Nie and Zhou [18] proposed a new discernibility matrix for computing the reduction core, and they tactfully utilized a kind of novel information of so-called local conditional-entropy. As our preliminary, the relevant entropy and matrix are reviewed as follows, where let U / I N D ( C ) = { C k : k = 1 , . . , r } and the cardinality form is mainly adopted.
Definition 2
([18]). The local conditional-entropy on decision table ( U , C D ) is defined by:
C p , C q U / I N D ( C ) ( 1 p , q r ) , H C p C q ( D / C ) = | C p | | C p C q | j = 1 m | C p D j | | C p | log 2 | C p D j | | C p | | C q | | C p C q | j = 1 m | C q D j | | C q | log 2 | C q D j | | C q | .
Definition 3
([18]). The discernibility matrix D M = ( r i j ) | U | × | U | on decision table ( U , C D ) is defined by:
r i j = c C , i f m i n ( | d x i | , | d x j | ) = 1 , c ( x i ) c ( x j ) , D ( x i ) D ( x j ) , c C , i f m i n ( | d x i | , | d x j | ) > 1 , c ( x i ) c ( x j ) , D ( x i ) D ( x j ) , H [ x i ] C [ x j ] C ( D / ( C { c } ) ) > H [ x i ] C [ x j ] C ( D / C ) , , o t h e r w i s e
where d x = { D ( y ) : y [ x ] C } ( x , y U ) represents the set of decision values induced by conditional class [ x ] C while | d x | means the corresponding cardinality [63]. In Equation (6), let [ x i ] C = C p , [ x j ] C = C q , and then
H [ x i ] C [ x j ] C ( D / ( C { c } ) ) = H C p C q ( D / ( C { c } ) ) = j = 1 m | ( C p C q ) D j | | C p C q | log 2 | ( C p C q ) D j | | C p C q |
is determined to represent the conditional-entropy of local decision table when accompanied by new universe C p C q after deleting attribute c; moreover, H [ x i ] C [ x j ] C ( D / C ) = H C p C q ( D / C ) is clear according to Equation (5).

3. Double-Granule Conditional-Entropies Based on Three-Level Granular Structures

The local conditional-entropy in Equation (5) implements effective uncertainty descriptions to guide the in-depth discernibility matrix and core calculation [18], thus exhibiting fundamental significance. However, this basic measure has three flawed aspects, and corresponding improvements for general applications.
(1)
According to Equation (5), the locality mainly refers to less range C p C q in universe U. More essentially, we can stand on the dual granules C p and C q to propose a novel notion of double-granule conditional-entropies, and it differs from the usual entropy system with only the single-granule representation which implies a kind of first-order style. Moreover, the measure properties are lacking in [18], and we will provide in-depth properties such as restriction bounds and granulation non-monotonicity.
(2)
Regarding granular structures, all decision classes D j ( j = 1 , , m ) (or decision classification U / I N D ( D ) ) are considered, but condition granules involve only two factors C p and C q . A condition partition U / I N D ( C ) ) needs considering in practice to provide a system description of knowledge granulation, so we also focus on granulation U / I N D ( C ) to introduce three-level granular structures for hierarchical constructions of double-granule conditional-entropies.
(3)
Finally, the initial concept is limited to only C for expressing the discernibility matrix and reduction core, and a general subset A C has better theoretical and practical prospects, especially for the knowledge-based applications (such as attribute reduction or feature selection).
Along the above thoughts, this section mainly establishes double-granule conditional-entropies based on a universal attribute-subset A C and investigates relevant algorithms and properties, and we particularly use a kind of three-level granular structures.
From a viewpoint of only condition granulation, basic descriptions of three-level granular structures are provided in Table 1, and relevant concepts are usually intuitionistic and descriptive according to a supporting figure with granular structures: Figure 1. Micro-bottom ( A p , A q ) focuses on only two granules, meso-middle
( A p , U / I N D ( A ) = { A q : q = 1 , , n } )
consists of one granule and a partition, while macro-top
( U / I N D ( A ) = { A p : p = 1 , , n } , U / I N D ( A ) = { A q : q = 1 , , n } )
considers the same partition with different construction origins. The three-level granular structures carry a kind of hierarchical integration (or decomposition) relationship, and they provide n × n , n, and one parallel patterns, respectively; they will be presented in a table form with the n × n mainbody data as well as the edge statistics. Moreover, they differ from the existing three-level granular structures for decision tables, which consider not only the condition granulation (with A i and U / I N D ( A ) ) but also decision granulation (with D j and U / I N D ( D ) ) [56].

3.1. Double-Granule Conditional-Entropy at Micro-Bottom

The local conditional-entropies are actually at only micro-bottom, i.e., ( C p , C q ) regarding C. As a basis of hierarchical development, this subsection improves local conditional-entropies to construct double-granule conditional-entropies at micro-bottom ( A p , A q ) ( p , q { 1 , , n } ), which comes from an arbitrary condition-attribute subset A C . We first suppose weight coefficients
ω p = | A p | | A p | + | A q | , ω q = | A q | | A p | + | A q | ,
where
ω p + ω q = 1 .
Definition 4.
At micro-bottom ( A p , A q ) , the double-granule conditional-entropy is defined by
H ( A p , A q ) ( D / A ) = ω p j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) ω q j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) = | A p | | A p | + | A q | j = 1 m | A p D j | | A p | log 2 | A p D j | | A p | | A q | | A p | + | A q | j = 1 m | A q D j | | A q | log 2 | A q D j | | A q | .
Proposition 1.
The double-granule conditional-entropy based on A p becomes
H ( A p , A p ) ( D / A ) = j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) = j = 1 m | A p D j | | A p | log 2 | A p D j | | A p | .
By using probabilistic and cardinal forms, Definition 4 proposes the double-granule conditional-entropy at micro-bottom. In contrast to the local conditional-entropy in [18], our measure generally adopts the same essence but a different viewpoint. In other words, Equation (9) with forms ( A p , A p ) and | A q | + | A p | is equivalent to Equation (5) with styles A p A p and | A q A p | when
A q A p | A q | + | A p | = | A q A p | ,
but the former becomes different and coherent when
A q = A p | A q | + | A p | = 2 | A q A p | > | A q A p | ;
moreover, it more tends to the double-granule description rather than the granule-union locality. In Equation (9), conditional-information measures
j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) , j = 1 m P ( D j / A q ) log 2 P ( D j / A q )
represent the uncertainty of decision classification U / I N D ( D ) regarding condition granules A p and A q , respectively, and they are integrated into H ( A p , A q ) ( D / A ) by two complementary weight coefficients ω p and ω q . As a result, H ( A p , A q ) ( D / A ) embodies a kind of information fusion of double-granule A p , A q to describe decision classification U / I N D ( D ) and its uncertainty, from the perspective of conditional information. Therefore, H ( A p , A q ) ( D / A ) is naturally called the double-granule conditional-entropy, and it is actually located at micro-bottom ( A p , A q ) . In particular, the double-granule measures utilize the double-granule fusion to capture a new feature of second-order, because main entropy systems (such as those in Equation (3)) utilize only the single-granule description which correspondingly refers to the so-called first-order information. Proposition 1 focuses on a specific case of A q = A p , and the concrete result H ( A p , A p ) ( D / A ) degenerates into a one-order measure regarding conditional-entropy.
Proposition 2.
At micro-bottom, double-granule conditional-entropies offer n × n values, i.e.,
H ( A p , A q ) ( D / A ) ( p , q = 1 , , n ) .
Since both A p and A q have n granules based on p = 1 , , n and q = 1 , , n , H ( A p , A q ) ( D / A ) offers number n × n (Proposition 2) to correspond to n × n micro-bottoms. The n × n kinds of double-granule conditional-entropies are arranged in Table 2, and the mainbody refers to an n × n square symmetric matrix where
H ( A p , A q ) ( D / A ) = H ( A q , A p ) ( D / A ) .
Based on Equation (9), Algorithm 1 resorts to a “for” loop to effectively offer a double-granule conditional-entropy H ( A p , A q ) ( D / A ) for two arbitrary granules A p , A q U / I N D ( A ) . Furthermore, we can achieve all n × n entropies values by adding two “for” loops regarding p = 1 , , n and q = 1 , , n .
Algorithm 1: Calculation of double-granule conditional-entropy at micro-bottom
Input: Decision table ( U , C D ) , target subset A C , and two special indexes
   p , q { 1 , , n } ;
Output: Double-granule conditional-entropy H ( A p , A q ) ( D / A ) at micro-bottom ( A p , A q ) .
1:
Compute U / I N D ( A ) to obtain two concrete granules A p , A q U / I N D ( A ) , and determine ω p , ω q .
2:
Compute U / I N D ( D ) to obtain all decision classes D j ( j = 1 , , m ).
3:
Let H p = 0 , H q = 0 .
4:
for j { 1 , . . , m } do
5:
   H p H p P ( D j / A p ) log 2 P ( D j / A p ) ,
   H q H q P ( D j / A q ) log 2 P ( D j / A q ) .
6:
end for
7:
Obtain H ( A p , A q ) ( D / A ) = ω p H p + ω q H q .
8:
return H ( A p , A q ) ( D / A ) .
Theorem 2.
At micro-bottom, the double-granule conditional-entropy has lower and upper bounds. Concretely,
H ̲ ( A p , A q ) ( D / A ) H ( A p , A q ) ( D / A ) H ¯ ( A p , A q ) ( D / A ) ,
where
H ̲ ( A p , A q ) ( D / A ) = | A p | 2 | U | j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) | A q | 2 | U | j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) , H ¯ ( A p , A q ) ( D / A ) = j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) .
Proof. 
| A p | , | A q | [ 1 , | U | ) implies
ω p = | A p | | A p | + | A q | | A p | 2 | U | , 1 , ω q = | A q | | A p | + | A q | | A q | 2 | U | , 1 ,
so H ( A p , A q ) ( D / A ) [ H ̲ ( A p , A q ) ( D / A ) , H ¯ ( A p , A q ) ( D / A ) ] . □
In Theorem 2, the double bounds of H ( A p , A q ) ( D / A ) are acquired by the enlarging and reducing of weight coefficients. Regarding Equation (12),
A q A p A q = A p | A q | = | A p | | U | 2 ω p | A p | | U | > | A p | 2 | U | , ω q | A q | | U | > | A q | 2 | U | ;
on the other hand,
A q = A p | A q | = | A p | > | U | 2 ω p = 1 2 | A p | 2 | U | , | A p | | U | , ω q = 1 2 | A q | 2 | U | , | A q | | U | .
In other words, ω p and ω q have theoretical lower bounds | A p | 2 | U | and | A q | 2 | U | , respectively, but they usually have closer lower bounds | A p | | U | and | A q | | U | , respectively. Therefore, H ( A p , A q ) ( D / A ) can theoretically achieve H ̲ ( A p , A q ) ( D / A ) , such as in the case
A p = A q = U ω p = 1 2 = | A p | 2 | U | ω q = 1 2 = | A q | 2 | U | ;
usually, it may be practically restricted by a better measure:
H ̲ ( A p , A q ) ( D / A ) = | A p | | U | j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) | A q | | U | j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) ,
which offers
H ̲ ( A p , A q ) ( D / A ) H ̲ ( A p , A q ) ( D / A ) .
We below provide another upper bound of H ( A p , A q ) ( D / A ) , which may be better than H ¯ ( A p , A q ) ( D / A ) in some cases.
Theorem 3.
At micro-bottom, the double-granule conditional-entropy has an upper bound. Concretely,
H ( A p , A q ) ( D / A ) H ( A p , A q ) * ( D / A ) = j = 1 m | A p D j | + | A q D j | | A p | + | A q | log 2 | A p D j | + | A q D j | | A p | + | A q | .
Proof. 
As shown in Figure 2, function f ( P ) = P log 2 P ( P [ 0 , 1 ] ) is convex, where  f ȃ ( P ) = 1 P ln 2 < 0 . Thus, let
P p = P ( D j / A p ) = | A p D j | | A p | , P q = P ( D j / A q ) = | A q D j | | A q | ,
and then the famous “Jensen’s inequality” in mathematics could induce
ω p + ω q = 1 ω p P p log 2 P p ω q P q log 2 P q ( ω p P q + ω p P q ) log 2 ( ω p P q + ω p P q ) ,
where
ω p P q + ω p P q = | A p | | A p | + | A q | | A p D j | | A p | + | A q | | A p | + | A q | | A q D j | | A q | = | A p D j | + | A q D j | | A p | + | A q | .
In other words, we can get
H ( A p , A q ) ( D / A ) = j = 1 m [ ω p P p log 2 P p ω q P q log 2 P q ] j = 1 m ( ω p P q + ω p P q ) log 2 ( ω p P q + ω p P q ) = j = 1 m | A p D j | + | A q D j | | A p | + | A q | log 2 | A p D j | + | A q D j | | A p | + | A q | = H ( A p , A q ) * ( D / A ) .
In Theorem 3, the convex property of information function f ( P ) = P log 2 P is utilized to provide a new upper bound H ( A p , A q ) * ( D / A ) of central measure H ( A p , A q ) ( D / A ) . When comparing Equations (7) and (17), we can surprisingly discover that H ( A p , A q ) * ( D / A ) highly adheres to
H A p A q ( D / ( A { a } ) ) = j = 1 m | ( A p A q ) D j | | A p A q | log 2 | ( A p A q ) D j | | A p A q | ,
which naturally comes from H C p C q ( D / ( C { c } ) ) (Equation (7)). In fact,
H ( A p , A q ) * ( D / A ) = H A p A q ( D / ( A { a } ) )
when A p A q ; when
A q = A p | A p D j | + | A q D j | | A p | + | A q | = | A p D j | | A p | | A p D j | 2 | A p | = | ( A p A q ) D j | | A p | + | A q | ,
where A p A q = A p , there is a difference between two measures, and we obtain
H ( A p , A p ) * ( D / A ) = j = 1 m | A p D j | | A p | log 2 | A p D j | | A p | j = 1 m | A p D j | 2 | A p | log 2 | A p D j | 2 | A p | = H A p A q ( D / ( A { a } ) ) .
Thus far, H ( A p , A q ) ( D / A ) has one lower bound H ̲ ( A p , A q ) ( D / A ) and two upper bounds H ¯ ( A p , A q ) ( D / A ) , H ( A p , A q ) * ( D / A ) . An interesting question naturally emerges, i.e., can we necessarily determine the size relationship between H ¯ ( A p , A q ) ( D / A ) and H ( A p , A q ) * ( D / A ) to provide an exact bound? Unfortunately, the answer is negative, and the later example and experiment will reveal the size uncertainty. We simply provide a mechanism analysis. Let
P p q = | A p D j | + | A q D j | | A p | + | A q | ,
and its numerator/denominator be the corresponding sum of numerators/denominators of P p and P q . According to [64], we can obtain
P p q [ m i n ( P p , P q ) , m a x ( P p , P q ) ]
but P p q produces an uncertainty location between P p and P q . In view of the information function f ( P ) = P log 2 P and its maximum point ( 1 e , 1 e ln 2 ) (Figure 2),
f ( P p ) + f ( P q ) , f ( P p q )
never having the necessary size relationships, so
H ¯ ( A p , A q ) ( D / A ) = j = 1 m ( f ( P p ) + f ( P q ) ) , H ( A p , A q ) * ( D / A ) = j = 1 m f ( P p q )
also never have the necessary size relationships. In summary, H ¯ ( A p , A q ) ( D / A ) and H ( A p , A q ) * ( D / A ) adopt different views to become irrelevant and interactive, and they together restrict H ( A p , A q ) ( D / A ) . With the addition of lower bound of H ̲ ( A p , A q ) ( D / A ) , there are in total three bounds to systematically emerge. Similar to H ( A p , A q ) ( D / A ) and its distributional Table 2, they can also be arranged in a table with an n × n square symmetric matrix, i.e., Table 3, and thus Table 3 correspondingly restricts Table 2.
Finally, consider relevant granulation monotonicity/non-monotonicity. In fact, micro-bottom and its double-granule conditional-entropies focus on only two condition granules and thus never consider the condition granulation and further monotonicity/non-monotonicity. Moreover, U / I N D ( A ) U / I N D ( B ) implies the granulation refining and granule decomposition from A to B; thus A p , A q U / I N D ( A ) and B p * , B q * U / I N D ( B ) exhibit complex correspondence and uncertainty change, so we cannot mine fine relationships between H ( A p , A q ) ( D / A ) and H ( B p * , B q * ) ( D / B ) .

3.2. Double-Granule Conditional-Entropy at Meso-Middle

As analyzed above, double-granule conditional-entropies at micro-bottom never consider the condition granulation to lack robust functions of uncertainty descriptions. In terms of fixed decision granulation U / I N D ( D ) , H ( A p , A q ) ( D / A ) at micro-bottom ( A p , A q ) involves only two condition granules A p , A q and their interactive uncertainty information. For the function promotion, the condition granulation U / I N D ( A ) with systematic granules is worth introducing based on double-granule conditional-entropy H ( A p , A q ) ( D / A ) . Thus, we will gradually strengthen the knowledge granulation U / I N D ( A ) to establish better double-granule conditional-entropies, by virtue of three-level granular structures (Table 1). This subsection discusses double-granule conditional-entropies at meso-middle
( A p , U / I N D ( A ) = { A 1 , , A n } ) ( p { 1 , , n } ) .
Definition 5.
At meso-middle ( A p , U / I N D ( A ) ) , the double-granule conditional-entropy is defined by
H ( A p ) ( D / A ) = q = 1 n ω p j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) q = 1 n ω q j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) = q = 1 n | A p | | A p | + | A q | j = 1 m | A p D j | | A p | log 2 | A p D j | | A p | q = 1 n | A q | | A p | + | A q | j = 1 m | A q D j | | A q | log 2 | A q D j | | A q | .
Corollary 1.
At meso-middle, the double-granule conditional-entropy has an analytic expression:
H ( A p ) ( D / A ) = q = 1 n | A p | | A p | + | A q | j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) q = 1 n | A q | | A p | + | A q | j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) .
Theorem 4.
Double-granule conditional-entropies have a hierarchical integration from micro-bottom to meso-middle, i.e.,
H ( A p ) ( D / A ) = q = 1 n H ( A p , A q ) ( D / A ) = H ( A p , A 1 ) ( D / A ) + + H ( A p , A q ) ( D / A ) + + H ( A p , A n ) ( D / A ) .
By Definition 5 (Corollary 1) and Theorem 4, meso-middle’s measure H ( A p ) ( D / A ) (which can also be noted by H ( A p , U / I N D ( A ) ) ( D / A ) ) hierarchically integrates double-granule conditional-entropies H ( A p , A q ) ( D / A ) by condition-granular summation on q = 1 , , n . Thus, H ( A p ) ( D / A ) inherits the features of double-granule and conditional-entropy, it considers a granule A p and condition granulation U / I N D ( A ) to be at meso-middle ( A p , U / I N D ( A ) ) , so it is called the double-granule conditional-entropy at meso-middle. As a transition, H ( A p ) ( D / A ) combines granule A p and partition U / I N D ( A ) to describe decision classification U / I N D ( D ) and its uncertainty, from the perspective of conditional information.
Similar to and based on previous discussions on H ( A p , A q ) ( D / A ) (Section 3.1), we will provide corresponding properties of H ( A p ) ( D / A ) , including the number distribution, calculation algorithm, three bounds, and granulation monotonicity/non-monotonicity.
Proposition 3.
At meso-middle, double-granule conditional-entropies offer n values, i.e.,
H ( A p ) ( D / A ) ( p = 1 , , n ) .
In Proposition 3, double-granule conditional-entropies naturally exhibit number n to correspond to n meso-middles. The n values can be stored in an n-dimension vector to be related to the previous distributional Table 2. By enlarging Table 2, they are represented by the marginal vector of the bottom or right in Table 4, and they exactly correspond to the relevant row/column sum of micro-bottom’s information values. According to Equations (21) and (23), Algorithm 2 resorts to two “for” loops to effectively offer a double-granule conditional-entropy H ( A p ) ( D / A ) for an arbitrary granule A p U / I N D ( A ) . In fact, the inner loop invokes Algorithm 1 to calculate an arbitrary double-granule conditional-entropy at micro-bottom, while the outer loop integrates n related bottomed measures to produce H ( A p ) ( D / A ) . Furthermore, we can achieve all n middle entropies values by adding a “for” loop regarding p = 1 , , n .
Algorithm 2: Calculation of double-granule conditional-entropy at meso-middle
Input: Decision table ( U , C D ) , target subset A C , and a special index p { 1 , , n } ;
Output: Double-granule conditional-entropy H ( A p ) ( D / A ) at meso-middle ( A p , U / I N D ( A ) ) .
  1:
Compute U / I N D ( A ) to obtain all condition classes A i ( i = 1 , , n ) and a fixed granule A p U / I N D ( A ) .
  2:
Compute U / I N D ( D ) to obtain all decision classes D j ( j = 1 , , m ).
  3:
Let H ( A p ) ( D / A ) = 0 .
  4:
for q { 1 , . . , n } do
  5:
  Compute ω p , ω q .
  6:
  Let H p = 0 , H q = 0 .
  7:
  for j { 1 , . . , m } do
  8:
    H p H p P ( D j / A p ) log 2 P ( D j / A p ) ,
    H q H q P ( D j / A q ) log 2 P ( D j / A q ) .
  9:
  end for
10:
  Obtain H ( A p , A q ) ( D / A ) = ω p H p + ω q H q .
11:
   H ( A p ) ( D / A ) H ( A p ) ( D / A ) + H ( A p , A q ) ( D / A ) .
12:
end for
13:
return H ( A p ) ( D / A ) .
Theorem 5.
At meso-middle, the double-granule conditional-entropy has a lower bound and two upper bounds. Concretely,
H ( A p ) ( D / A ) [ H ̲ ( A p ) ( D / A ) , H ¯ ( A p ) ( D / A ) ] , H ( A p ) ( D / A ) H ( A p ) * ( D / A ) ,
where
H ̲ ( A p ) ( D / A ) = q = 1 n H ̲ ( A p , A q ) ( D / A ) = n | A p | 2 | U | j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) q = 1 n | A q | 2 | U | j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) , H ¯ ( A p ) ( D / A ) = q = 1 n H ¯ ( A p , A q ) ( D / A ) = n j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) q = 1 n j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) , H ( A p ) * ( D / A ) = q = 1 n H ( A p , A q ) * ( D / A ) = q = 1 n j = 1 m | A p D j | + | A q D j | | A p | + | A q | log 2 | A p D j | + | A q D j | | A p | + | A q | .
Theorem 5 naturally comes from Theorems 2–4. The three bounds in Equation (25) hierarchically integrate previous three bounds at micro-bottom (Equations (11) and (17)) to correspondingly restrict H ( A p ) ( D / A ) . They can be supplemented into distributional Table 4, and following Table 5 provides the relevant part.
At meso-middle, H ( A p ) ( D / A ) introduces the condition granulation U / I N D ( A ) , but it still needs condition granule A p . Thus, we cannot make a positive assertion regarding granulation monotonicity/non-monotonicity. In fact, U / I N D ( A ) U / I N D ( B ) also implies chaos between H ( A p ) ( D / A ) and H ( B p * ) ( D / B ) .

3.3. Double-Granule Conditional-Entropy at Macro-Top

As analyzed above, double-granule conditional-entropies at meso-middle consider the condition granulation, but in an insufficient way, and H ( A p ) ( D / A ) also depends on a single condition granule A p . For the thorough granulation and robust description, systematic measures H ( A p ) ( D / A ) ( p = 1 , , n ) can be further integrated to generate double-granule conditional-entropies at macro-top. Based on the previous thought and result in Section 3.1 and Section 3.2, this subsection further discusses double-granule conditional-entropies at macro-top
( U / I N D ( A ) = { A p : p = 1 , , n } , U / I N D ( A ) = { A q : q = 1 , , n } ) ,
which is given in Table 1. We will directly provide the relevant integration definition, number distribution, calculation algorithm, three bounds, and we finally uncover an important conclusion of granulation non-monotonicity.
Definition 6.
At macro-top ( U / I N D ( A ) , U / I N D ( A ) ) , the double-granule conditional-entropy is defined by
H ( D / A ) = p = 1 n q = 1 n ω p j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) p = 1 n q = 1 n ω q j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) = p = 1 n q = 1 n | A p | | A p | + | A q | j = 1 m | A p D j | | A p | log 2 | A p D j | | A p | p = 1 n q = 1 n | A q | | A p | + | A q | j = 1 m | A q D j | | A q | log 2 | A q D j | | A q | .
Corollary 2.
At macro-top, the double-granule conditional-entropy has an analytic expression:
H ( D / A ) = p = 1 n q = 1 n | A p | | A p | + | A q | j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) p = 1 n q = 1 n | A q | | A p | + | A q | j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) .
Theorem 6.
Double-granule conditional-entropies have a hierarchical integration from micro-bottom and meso-middle to macro-top, i.e.,
H ( D / A ) = p = 1 n H ( A p ) ( D / A ) = H ( A 1 ) ( D / A ) + + H ( A n ) ( D / A ) = p = 1 n q = 1 n H ( A p , A q ) ( D / A ) = H ( A 1 , A 1 ) ( D / A ) + + H ( A n , A n ) ( D / A ) .
By Definition 6 (Corollary 2) and Theorem 6, macro-top’s measure H ( D / A ) hierarchically integrates meso-middle’s entropies H ( A p ) ( D / A ) by a single summation on p = 1 , , n , and thus it further hierarchically integrates micro-bottom’s entropies H ( A p , A q ) ( D / A ) by double summations on p , q = 1 , , n . As a result, H ( D / A ) inherits the features of double-granule and conditional-entropy. It considers only conditional granulation U / I N D ( A ) to be at macro-top ( U / I N D ( A ) , U / I N D ( A ) ) , so it is called the double-granule conditional-entropy at macro-top. As an ultimate measure, H ( D / A ) completely utilizes the U / I N D ( A ) granulation information to effectively describe decision classification U / I N D ( D ) and its uncertainty, thus holding robust measurement functions for knowledge granulation. Moreover, H ( D / A ) can be noted by H ( U / I N D ( A ) , U / I N D ( A ) ) ( D / A ) ).
Proposition 4.
At macro-top, the double-granule conditional-entropy offers only one value, i.e.,
H ( D / A )   at   macro-top   ( U / I N D ( A ) , U / I N D ( A ) ) .
In Proposition 4, the double-granule conditional-entropy naturally exhibits number 1 to correspond to the sole macro-top. In fact, the first top entropy comes from the fusion of either n middle entropies or n × n bottom entropies; thus, three-level entropies accord with three-level granular structures (Table 1) from the quantitative and structural perspective, and they embody the hierarchical integration. In particular, the sole conditional-entropy H ( D / A ) is put into the lower-right corner of Table 4, thus corresponding to the summations of central n × n micro values and marginal n meso values. According to Equations (26) and (28), Algorithm 3 resorts to three “for” loops to effectively offer the double-granule conditional-entropy H ( D / A ) . The two inner loops invoke Algorithm 2 to calculate an arbitrary double-granule conditional-entropy at meso-middle (where the central loop invokes Algorithm 1 to construct micro-bottom’s entropies), while the outer loop integrates n related meso-middle’s information values to produce H ( D / A ) . In other words, Algorithms 1–3 exhibit a kind of hierarchical evolution based on circulation development, and thus they constitute a novel kind of three-level algorithms.
Algorithm 3: Calculation of double-granule conditional-entropy at macro-top
Input: Decision table ( U , C D ) , target subset A C ;
Output: Double-granule conditional-entropy H ( D / A ) at Macro-Top ( U / I N D ( A ) , U / I N D ( A ) ) .
  1:
Compute U / I N D ( A ) to obtain all condition classes A i ( i = 1 , , n ).
  2:
Compute U / I N D ( D ) to obtain all decision classes D j ( j = 1 , , m ).
  3:
Let H ( D / A ) = 0 .
  4:
for p { 1 , . . , n } do
  5:
  Let H ( A p ) ( D / A ) = 0 .
  6:
  for q { 1 , . . , n } do
  7:
   Compute ω p , ω q .
  8:
   Let H p = 0 , H q = 0 .
  9:
   for j { 1 , . . , m } do
10:
     H p H p P ( D j / A p ) log 2 P ( D j / A p ) ,
     H q H q P ( D j / A q ) log 2 P ( D j / A q ) .
11:
   end for
12:
   Obtain H ( A p , A q ) ( D / A ) = ω p H p + ω q H q .
13:
    H ( A p ) ( D / A ) H ( A p ) ( D / A ) + H ( A p , A q ) ( D / A ) .
14:
  end for
15:
   H ( D / A ) H ( D / A ) + H ( A p ) ( D / A ) .
16:
end for
17:
return H ( D / A ) .
Theorem 7.
At macro-top, the double-granule conditional-entropy has a lower bound and two upper bounds. Concretely,
H ( D / A ) [ H ̲ ( D / A ) , H ¯ ( D / A ) ] , H ( D / A ) H * ( D / A ) ,
where
H ̲ ( D / A ) = p = 1 n H ̲ ( A p ) ( D / A ) = p = 1 n q = 1 n H ̲ ( A p , A q ) ( D / A ) = n p = 1 n | A p | 2 | U | j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) n q = 1 n | A q | 2 | U | j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) = n p = 1 n | A p | | U | j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) ,
H ¯ ( D / A ) = p = 1 n H ¯ ( A p ) ( D / A ) = p = 1 n q = 1 n H ¯ ( A p , A q ) ( D / A ) = n p = 1 n j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) n q = 1 n j = 1 m P ( D j / A q ) log 2 P ( D j / A q ) = 2 n p = 1 n j = 1 m P ( D j / A p ) log 2 P ( D j / A p ) ,
H * ( D / A ) = p = 1 n H ( A p ) * ( D / A ) = p = 1 n q = 1 n H ( A p , A q ) * ( D / A ) = p = 1 n q = 1 n j = 1 m | A p D j | + | A q D j | | A p | + | A q | log 2 | A p D j | + | A q D j | | A p | + | A q | .
Theorem 7 naturally comes from Theorems 2–6. The three bounds in Equations (30)–(32) hierarchically integrate previous three bounds at meso-middle and micro-bottom, and thus they become three new uncertainty measures at macro-top ( U / I N D ( A ) , U / I N D ( A ) ) to correspondingly restrict H ( D / A ) . They are supplemented into the bottom in the previous bound table: Table 5.
Theorem 8.
At macro-top, the double-granule conditional-entropy has granulation non-monotonicity. That is, U / I N D ( A ) U / I N D ( B ) cannot necessarily achieve
either   H ( D / A ) H ( D / B )   or   H ( D / A ) H ( D / B ) ,
and both cases can practically exist. In addition, the matched double bounds H ̲ ( D / A ) and H ¯ ( D / A ) (Equations (30) and (31)) also have the granulation non-monotonicity, and they cannot theoretically acquire
either   H ̲ ( D / A ) H ̲ ( D / B )   or   H ̲ ( D / A ) H ̲ ( D / B ) ,
either   H ¯ ( D / A ) H ¯ ( D / B )   or   H ¯ ( D / A ) H ¯ ( D / B ) .
At macro-top, the double-granule conditional-entropy completely breaks away from the condition granule dependence to establish the condition granulation description, so it becomes a powerful type of information measure for knowledge-based uncertainty representation. In terms of condition granulation, its non-monotonicity is finally revealed in Theorem 8, and the relevant evidence will be provided in the later example and experiment. Moreover, this fundamental non-monotonicity conclusion embodies information uncertainty, and it can be induced or explained by the previous complexity mechanism at micro-bottom and meso-middle. Based on macro-top and its granulation mechanism, the related three bounds (Equations (30)–(32)) and their monotonicity/non-monotonicity can be practically observed, and thus we also obtain the granulation non-monotonicity for H ̲ ( D / A ) and H ¯ ( D / A ) ; however, the case of upper bound H * ( D / A ) becomes a remaining problem.

4. Decision Table Example

In this section, the above theoretical constructions and properties are illustrated by a decision table example. By extracting a part of VOTING data set (which comes from UCI database [65]), we provide a practical decision table ( U , C D ) in Table 6 with
| U | = 8 , | C | = 11 , | D | = 1 .
According to this decision table,
U / I N D ( D ) = { D 1 = { x 1 , x 2 , x 8 } , D 2 = { x 3 , x 4 , x 5 , x 6 , x 7 } }
provides m = | U / I N D ( D ) | = 2 . As an example, A = { c 1 , c 2 , c 3 , c 4 , c 5 } is chosen to generate condition granulation
U / I N D ( A ) = { A 1 = { x 1 } , A 2 = { x 2 , x 7 , x 8 } , A 3 = { x 3 } , A 4 = { x 4 } , A 5 = { x 5 } , A 6 = { x 6 } } ,
where n = | U / I N D ( A ) | = 6 . By virtue of three-level granular structures (Table 1), double-granule conditional-entropies and their three bounds are calculated by relevant algorithms and definitions, and they are compactly listed in Table 7 and Table 8, respectively. The measures at micro-bottom, meso-middle, macro-top have numbers 36, 6, 1, respectively, and they correspond to the central 6 × 6 matrix, marginal 6-dimensional vector, lower-right-corner 1 digit, respectively. In part, we provide some processes of entropy calculation as follows.
P ( D 1 / A 1 ) log 2 P ( D 1 / A 1 ) = 0 , P ( D 2 / A 1 ) log 2 P ( D 2 / A 1 ) = 0 , P ( D 1 / A 2 ) log 2 P ( D 1 / A 2 ) = 0.3900 , P ( D 2 / A 2 ) log 2 P ( D 2 / A 2 ) = 0.5283 , H ( A 1 , A 1 ) ( D / A ) = 1 1 + 1 ( 0 + 0 ) + 1 1 + 1 ( 0 + 0 ) = 0 , H ( A 1 , A 2 ) ( D / A ) = 1 1 + 3 ( 0 + 0 ) + 3 1 + 3 ( 0.3900 + 0.5283 ) = 0.6887 ; H ( A 1 ) ( D / A ) = H ( A 1 , A 1 ) ( D / A ) + H ( A 1 , A 2 ) ( D / A ) + + H ( A 1 , A 6 ) ( D / A ) = 0 + 0.6887 + 0 + 0 + 0 + 0 = 0.6887 ; H ( D / A ) = H ( A 1 ) ( D / A ) + H ( A 2 ) ( D / A ) + + H ( A 6 ) ( D / A ) = 0.6887 + 4.3619 + 0.6887 + 0.6887 + 0.6887 + 0.6887 = 7.8055 .
By Table 7 and Table 8, we can make relevant verification analyses. First, entropies and bounds naturally present hierarchical integration relationships from micro-bottom to meso-middle to macro-top. Indeed, conditional-entropies are correspondingly restricted by three bounds. Moreover, the two types of upper bounds exactly have no necessary size relationships, and a part but powerful proof is provided as follows:
H ¯ ( A 1 , A 2 ) ( D / A ) = 0 . 9183 > 0 . 8113 = H ( A 1 , A 2 ) * ( D / A ) , H ¯ ( A 1 , A 3 ) ( D / A ) = 0 < 1 = H ( A 1 , A 3 ) * ( D / A ) ;
H ¯ ( A 1 ) ( D / A ) = 0.9183 < 4.8113 = H ( A 1 ) * ( D / A ) , H ¯ ( A 2 ) ( D / A ) = 6.4281 > 5.7296 = H ( A 2 ) * ( D / A ) .
Finally, the granulation non-monotonicity at macro-top (Theorem 8) is verified. For this, we chose a natural attribute-addition chain:
{ c 1 } { c 1 , c 2 } { c 1 , c 2 , , c 11 } .
C A k ( k { 1 , 2 , , 11 } ) denotes the attribute subset in the chain, and its granulation is represented by
U / I N D ( C A k ) = { C A k , 1 , , C A k , p , , C A k , | U / I N D ( C A k ) | } .
In other words, C A k , p corresponds to the kth chain element C A k to represent the pth condition granule in partition U / I N D ( C A k ) . According to the subset chain, Table 9 provides double-granule conditional-entropies, including both part values at micro-bottom ( C A k , p , C A k , q ) , meso-middle ( C A k , p , U / I N D ( C A k ) ) and all values (as well as the three bounds) at macro-top ( U / I N D ( C A k ) , U / I N D ( C A k ) ) . As a supporting detail, previous Table 7 and Table 8 actually embrace the chain element C A 5 and its partition U / I N D ( C A 5 ) = { { x 1 } , { x 2 , x 7 , x 8 } , { x 3 } , { x 4 } , { x 5 } , { x 6 } } , while double-granule conditional-entropies regarding attribute subset C A 2 = { c 1 , c 2 } and corresponding condition granulation
U / I N D ( C A 2 ) = { C A 2 , 1 = { x 1 , x 2 , x 7 , x 8 } , C A 2 , 2 = { x 3 } , C A 2 , 3 = { x 4 , x 6 } , C A 2 , 4 = { x 5 } }
are supplemented in Table 10 for better observation and illustration.
(1)
Since different chain subsets may have different equivalence partitions and granule numbers, the measures at micro-bottom and meso-middle consider condition granules to have a distinctive number and difficult correspondence. Table 9 focuses on the small and the same granule number, but relevant granules have different connotations. For example, the granules of the first one — C A k , 1 ( k = 1 , 2 , , 11 )—may be different. Thus, we cannot acquire the so-called granulation non-monotonicity assertion because of granulation incompletion, although the values at micro-bottom and meso-middle actually exhibit a kind of non-monotonic change in Table 9.
(2)
In contrast, macro-top offers the complete condition granulation, so we can effectively focus on value monotonicity/non-monotonicity for both double-granule conditional-entropies and their three bounds. Observing the bottom part of Table 9 in the enlargement chain direction, we can discover that the three types of information measures are all non-monotonic, i.e.,
H ( D / C A k ) , H ̲ ( D / C A k ) , H ¯ ( D / C A k ) ( except   H * ( D / C A k ) ) .
More vividly, the entropy and its three bounds regarding the chain are depicted in Figure 3, so the related granulation non-monotonicity becomes clearer. For example, the macro entropy value H ( D / C A k ) first increases and then decreases in the addition chain direction. Moreover, Table 9 and Figure 3 reflect the restriction properties of three bounds.

5. Data Experiments

In this section, the above theoretical results and their effectiveness are verified by data experiments. The new measures are mainly suitable for categorical (or nominal) data, which are usually used in the traditional rough set theory, and thus we adopt three relevant data sets from the UCI Machine Learning Repository [65], whose concrete descriptions on decision table ( U , C D ) are given in Table 11.
Similar to the above example, we also adopt the attribute-addition chain
C A 1 = { c 1 } C A | C | = { c 1 , c 2 , c | C | }
and its relevant symbol such as
U / I N D ( C A k ) = { C A k , 1 , , C A k , p , , C A k , | U / I N D ( C A k ) | } .
Note that this attribute-subset sequence (Equation (34)) can deeply and typically probe the hierarchical knowledge-granulation within a framework of the complete lattice ( 2 C , ) . As a representative manifestation, we provide two typical results regarding the first chain element C A 1 = { c 1 } and the last one C A | C | = C .
(1)
Regarding VOTING, { c 1 } and C induce three and 342 granules, respectively, and relevant double-granule conditional-entropies and three bounds are provided in Table 12 and Table 13, respectively.
(2)
Regarding SPECT, { c 1 } and C produce two and 169 granules, respectively, and relevant three-level measures and three bounds are provided in Table 14 and Table 15, respectively.
(3)
Regarding Tic-Tac-Toe, { c 1 } and C determine three and 958 granules, respectively, and relevant entropies and bounds are provided in Table 16 and Table 17, respectively.
From the perspective of macro-top, double-granule conditional-entropies and their three information bounds based on the attribute-enlargement chain are finally summarized in Figure 4. These tables and figures can be utilized to effectively verify all previous conclusions, including the hierarchy, algorithm, restriction, and non-monotonicity. In particular, double-granule conditional-entropies are confined by three bounds, thus supporting the boundedness (Theorems 2, 3, 5 and 7); moreover, the entropies and their matched double-bounds fluctuate up and down, thus proving relevant granulation non-monotonicity (Theorem 8).

6. Conclusions

The information measures implement fundamental uncertainty measurement in rough set theory and granular computing. The local conditional-entropies have the second-order feature, but they are limited to micro-bottom for describing discernibility matrix and reduction core [18]. In this paper, double-granule conditional-entropies achieve corresponding improvements of hierarchical/conditional granulation, and thus they become broader measures with uncertainty representation and information processing. They focus more on the double-granule interaction rather than granule-union locality, which is used in local conditional-entropies [18]. This strategy directly utilizes the second-order mechanism to implement more systematic and robust uncertainty measurements, especially when compared to the current mainstream of first-order information measures. In our studies, double-granule conditional-entropies and their hierarchies, granulation, algorithms, bounds, and non-monotonicity are acquired and verified at three-level granular structures (i.e., micro-bottom, meso-middle, macro-top), and these results underlie both the efficiency in information processing and effectiveness in knowledge-based data analyses. Furthermore, their future developments and in-depth applications can be explored as follows.
(1)
In contrast to the relevant technology in [56], the hierarchical granulation of three-level granular structures focuses on the conditional granulation and relevant number, and it can be generalized for granular computing.
(2)
The double-granule conditional-entropies and their three bounds become new types of information measures with the second-order feature. In contrast to the traditional first-order entropy system, their description power and application advantage need further practical verification.
(3)
The double-granule conditional-entropies have three-restrictive bounds and granulation non-monotonicity, which have been experimentally verified by a granulation-hierarchical sequence (i.e., Equation (34)). These results are worth deeply utilizing in uncertainty measurement and data mining.
(4)
The double-granule conditional-entropies originate from the local conditional-entropies to carry a potential and distinctive advantage of discernibility matrix representation, and they also have the complete conditional granulation to have application prospects in knowledge reasoning or acquisition. Both their relationships with the discernibility matrix and their functions on attribute reduction need be deeply researched by promoting the previous studies in [18].

Author Contributions

T.M. conceived the algorithms and implemented the experiments; X.Z. mined the hierarchies and properties; Z.M. analyzed the results. All authors read and revised the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grants 61673285 and 11671284), the Sichuan Science and Technology Project of China (Grants 19YYJC2845 and 2017JY0197), and the Sichuan Youth Science and Technology Foundation of China (Grant 2017JQ0046).

Acknowledgments

The authors thank all of the editors and reviewers for their valuable suggestions, which have substantially improved this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pawlak, Z. Rough set. Int. J. Comput. Inf. Sci. 1982, 11, 38–39. [Google Scholar] [CrossRef]
  2. Raza, M.S.; Qamar, U. Redefining core preliminary concepts of classic rough set theory for feature selection. Eng. Appl. Artif. Intell. 2017, 65, 375–387. [Google Scholar] [CrossRef]
  3. Saha, I.; Sarkar, J.P.; Maulik, U. Integrated rough fuzzy clustering for categorical data analysis. Fuzzy Sets Syst. 2019, 361, 1–32. [Google Scholar] [CrossRef]
  4. Qian, Y.H.; Liang, X.Y.; Wang, Q.; Liang, J.Y.; Liu, B.; Andrzej, S.; Yao, Y.Y.; Ma, J.M.; Dang, C.Y. Local rough set: A solution to rough data analysis in big data. Int. J. Approx. Reason. 2018, 97, 38–63. [Google Scholar] [CrossRef]
  5. Hu, M.J.; Yao, Y.Y. Structured approximations as a basis for three-way decisions in rough set theory. Knowl.-Based Syst. 2019, 165, 92–109. [Google Scholar] [CrossRef]
  6. Yang, X.B.; Liang, S.C.; Yu, H.L.; Gao, S.; Qian, Y.H. Pseudo-label neighborhood rough set: Measures and attribute reductions. Int. J. Approx. Reason. 2019, 105, 112–129. [Google Scholar] [CrossRef]
  7. Wang, Z.H.; Feng, Q.R.; Wang, H. The lattice and matroid representations of definable sets in generalized rough sets based on relations. Inf. Sci. 2019, 485, 505–520. [Google Scholar] [CrossRef]
  8. Luo, C.; Li, T.R.; Chen, H.M.; Fujita, H.; Zhang, Y. Incremental rough set approach for hierarchical multicriteria classification. Inf. Sci. 2018, 429, 72–87. [Google Scholar] [CrossRef] [Green Version]
  9. Yao, Y.Y.; Zhang, X.Y. Class-specific attribute reducts in rough set theory. Inf. Sci. 2017, 418–419, 601–618. [Google Scholar] [CrossRef]
  10. Zhang, X.Y.; Yang, J.L.; Tang, L.Y. Three-way class-specific attribute reducts from the information viewpoint. Inf. Sci. 2018. [Google Scholar] [CrossRef]
  11. Ma, X.A.; Yao, Y.Y. Three-way decision perspectives on class-specific attribute reducts. Inf. Sci. 2018, 450, 227–245. [Google Scholar] [CrossRef]
  12. Miao, D.Q.; Zhao, Y.; Yao, Y.Y.; Li, H.X.; Xu, F.F. Relative reducts in consistent and inconsistent decision tables of the Pawlak rough set model. Inf. Sci. 2009, 179, 4140–4150. [Google Scholar] [CrossRef]
  13. Lang, G.M.; Cai, M.J.; Fujita, H.; Xiao, Q.M. Related families-based attribute reduction of dynamic covering decision information systems. Knowl.-Based Syst. 2018, 162, 161–173. [Google Scholar] [CrossRef]
  14. Gao, C.; Lai, Z.H.; Zhou, J.; Wen, J.J.; Wong, W.K. Granular maximum decision entropy-based monotonic uncertainty measure for attribute reduction. Int. J. Approx. Reason. 2019, 104, 9–24. [Google Scholar] [CrossRef]
  15. Wang, C.Z.; Shi, Y.P.; Fan, X.D.; Shao, M.W. Attribute reduction based on k-nearest neighborhood rough sets. Int. J. Approx. Reason. 2019, 106, 18–31. [Google Scholar] [CrossRef]
  16. Wei, W.; Wu, X.Y.; Liang, J.Y.; Cui, J.B.; Sun, Y.J. Discernibility matrix based incremental attribute reduction for dynamic data. Knowl.-Based Syst. 2018, 140, 142–157. [Google Scholar] [CrossRef]
  17. Ma, F.M.; Ding, M.W.; Zhang, T.F.; Cao, J. Compressed binary discernibility matrix based incremental attribute reduction algorithm for group dynamic data. Neurocomputing 2019, 344, 20–27. [Google Scholar]
  18. Nie, H.M.; Zhou, J.Q. A new discernibility matrix and the computation of a core. J. Sichuan Univ. (Nat. Sci. Ed.) 2007, 44, 277–283. (In Chinese) [Google Scholar]
  19. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  20. Shiraz, R.K.; Fukuyama, H.; Tavana, M.; Caprio, D.D. An integrated data envelopment analysis and free disposal hull framework for cost-efficiency measurement using rough sets. Appl. Soft Comput. 2016, 46, 204–219. [Google Scholar] [CrossRef]
  21. Liang, J.Y.; Shi, Z.Z.; Li, D.Y.; Wierman, M.J. Information entropy, rough entropy and knowledge granularity in incomplete information systems. Int. J. Gen. Syst. 2006, 35, 641–654. [Google Scholar] [CrossRef]
  22. Wei, W.; Liang, J.Y. Information fusion in rough set theory: An overview. Inf. Fusion 2019, 48, 107–118. [Google Scholar] [CrossRef]
  23. Hu, Q.H.; Yu, D.R.; Xie, Z.X.; Liu, J.F. Fuzzy probabilistic approximation spaces and their information measures. IEEE Trans. Fuzzy Syst. 2006, 14, 191–201. [Google Scholar]
  24. Dai, J.H.; Wei, B.J.; Zhang, X.H. Uncertainty measurement for incomplete interval-valued information systems based on α-weak similarity. Knowl.-Based Syst. 2017, 136, 159–171. [Google Scholar] [CrossRef]
  25. Chen, Y.M.; Xue, Y.; Ma, Y.; Xu, F.F. Measures of uncertainty for neighborhood rough sets. Knowl.-Based Syst. 2017, 120, 226–235. [Google Scholar] [CrossRef]
  26. Miao, D.Q. Rough Set Theory and Its Application in Machine Learing. Ph.D. Thesis, Institute of Automation, The Chinese Academy of Sciences, Beijing, China, 1997. (In Chinese). [Google Scholar]
  27. Wang, G.Y.; Zhao, J.; An, J.J.; Wu, Y. A comparative study of algebra viewpoint and information viewpoint in attribute reduction. Fundam. Inf. 2005, 68, 289–301. [Google Scholar]
  28. Jiang, F.; Sui, Y.F.; Zhou, L. A relative decision entropy-based feature selection approach. Pattern Recognit. 2015, 48, 2151–2163. [Google Scholar] [CrossRef]
  29. Slezak, D. Approximate entropy reducts. Fundam. Inf. 2002, 53, 365–390. [Google Scholar]
  30. Qian, W.B.; Shu, W.H. Mutual information criterion for feature selection from incomplete data. Neurocomputing 2015, 168, 210–220. [Google Scholar] [CrossRef]
  31. Liang, J.Y.; Chin, K.S.; Dang, C.Y.; Yam, R.C.M. A new method for measuring uncertainty and fuzziness in rough set theory. Int. J. Gen. Syst. 2002, 31, 331–342. [Google Scholar] [CrossRef]
  32. Qian, Y.H.; Liang, J.Y. Combination entropy and combination granulation in rough set theory. Int. J. Uncertain. Fuzziness Knowl.-Based Syst. 2008, 16, 179–193. [Google Scholar]
  33. Hu, Q.H.; Che, X.J.; Zhang, L. Rank entropy-based decision trees for monotonic classifcation. IEEE Trans. Knowl. Data Eng. 2012, 24, 2052–2064. [Google Scholar] [CrossRef]
  34. Dai, J.H.; Xu, Q.; Wang, W.T.; Tian, H.W. Conditional entropy for incomplete decision systems and its application in data mining. Int. J. Gen. Syst. 2012, 41, 713–728. [Google Scholar] [CrossRef]
  35. Sun, L.; Zhang, X.Y.; Xu, J.C.; Zhang, S.G. An attribute reduction method using neighborhood entropy measures in neighborhood rough sets. Entropy 2019, 21, 155. [Google Scholar] [CrossRef]
  36. Chen, D.G.; Yang, W.X.; Li, F.C. Measures of general fuzzy rough sets on a probabilistic space. Inf. Sci. 2008, 178, 3177–3187. [Google Scholar] [CrossRef]
  37. Mi, J.S.; Leung, Y.; Wu, W.Z. An uncertainty measure in partition-based fuzzy rough sets. Int. J. Gen. Syst. 2005, 34, 77–90. [Google Scholar] [CrossRef]
  38. Hu, Q.H.; Zhang, L.; Zhang, D.; Pan, W.; An, S.; Pedrycz, W. Measuring relevance between discrete and continuous features based on neighborhood mutual information. Expert Syst. Appl. 2011, 38, 10737–10750. [Google Scholar] [CrossRef]
  39. Zhao, J.Y.; Zhang, Z.L.; Han, C.Z.; Zhou, Z.F. Complement information entropy for uncertainty measure in fuzzy rough set and its applications. Soft Comput. 2015, 19, 1997–2010. [Google Scholar] [CrossRef]
  40. Deng, X.F.; Yao, Y.Y. A multifaceted analysis of probabilistic three-way decisions. Fundam. Inf. 2014, 132, 291–313. [Google Scholar]
  41. Deng, X.F.; Yao, Y.Y. An information-theoretic interpretation of thresholds in probabilistic rough sets. Lect. Notes Comput. Sci. 2012, 7414, 369–378. [Google Scholar]
  42. Ma, X.A.; Wang, G.Y.; Yu, H.; Li, T.R. Decision region distribution preservation reduction in decision-theoretic rough set model. Inf. Sci. 2014, 278, 614–640. [Google Scholar] [CrossRef]
  43. Zadeh, L.A. Towards a theory of fuzzy information granulation and its centrality in human reasoning and fuzzy logic. Fuzzy Sets Syst. 1997, 90, 111–127. [Google Scholar] [CrossRef]
  44. Yao, Y.Y. A triarchic theory of granular computing. Granul. Comput. 2016, 1, 145–157. [Google Scholar] [CrossRef]
  45. Skowron, A.; Stepaniuk, J.; Swiniarski, R. Modeling rough granular computing based on approximation spaces. Inf. Sci. 2012, 184, 20–43. [Google Scholar] [CrossRef]
  46. Chiaselotti, G.; Gentile, T.; Infusino, F. Granular computing on information tables: Families of subsets and operators. Inf. Sci. 2018, 442–443, 72–102. [Google Scholar]
  47. Eissa, M.M.; Elmogy, M.; Hashem, M. Rough-granular computing knowledge discovery models for medical classification. Egypt. Inf. J. 2016, 17, 265–272. [Google Scholar] [CrossRef]
  48. Qian, Y.H.; Zhang, H.; Sang, Y.L.; Liang, J.Y. Multigranulation decision-theoretic rough sets. Int. J. Approx. Reason. 2014, 55, 225–237. [Google Scholar] [CrossRef]
  49. Li, J.H.; Mei, C.L.; Xu, W.H.; Qian, Y.H. Concept learning via granular computing: A cognitive viewpoint. Inf. Sci. 2015, 298, 447–467. [Google Scholar] [CrossRef]
  50. Wang, G.Y.; Ma, X.A.; Yu, H. Monotonic uncertainty measures for attribute reduction in probabilistic rough set model. Int. J. Approx. Reason. 2015, 59, 41–67. [Google Scholar] [CrossRef]
  51. Jia, X.Y.; Shang, L.; Zhou, B.; Yao, Y.Y. Generalized attribute reduct in rough set theory. Knowl.-Based Syst. 2016, 91, 204–218. [Google Scholar]
  52. Zhang, X.Y.; Miao, D.Q. Double-quantitative fusion of accuracy and importance: Systematic measure mining, benign integration construction, hierarchical attribute reduction. Knowl.-Based Syst. 2016, 91, 219–240. [Google Scholar]
  53. Calvanese, D.; Dumas, M.; Laurson, U.; Maggi, F.M.; Montali, M.; Teinemaa, I. Semantics analysis and simplification of DMN decision tables. Inf. Syst. 2018, 78, 112–125. [Google Scholar] [CrossRef]
  54. Liu, G.L.; Hua, Z.; Zou, J.Y. Local attribute reductions for decision tables. Inf. Sci. 2018, 422, 204–217. [Google Scholar] [CrossRef]
  55. Ge, H.; Li, L.S.; Xu, Y.; Yang, C.J. Quick general reduction algorithms for inconsistent decision tables. Int. J. Approx. Reason. 2017, 82, 56–80. [Google Scholar] [CrossRef]
  56. Zhang, X.Y.; Miao, D.Q. Three-layer granular structures and three-way informational measures of a decision table. Inf. Sci. 2017, 412–413, 67–86. [Google Scholar]
  57. Wang, J.; Tang, L.Y.; Zhang, X.Y.; Luo, Y.Y. Three-way weighted combination-entropies based on three-layer granular structures. Appl. Math. Nonlinear Sci. 2017, 2, 329–340. [Google Scholar] [Green Version]
  58. Yao, Y.Y. An outline of a theory of three-way decisions. In Rough Sets and Current Trends in Computing, Proceedings of the International Conference on Rough Sets and Current Trends in Computing, Chengdu, China, 17–20 August 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 1–17. [Google Scholar]
  59. Yao, Y.Y. Three-way decision and granular computing. Int. J. Approx. Reason. 2018, 103, 107–123. [Google Scholar] [CrossRef]
  60. Fard, A.M.F.; Hajaghaei-Keshteli, M. A tri-level location-allocation model for forward/reverse supply chain. Appl. Soft Comput. 2018, 62, 328–346. [Google Scholar] [CrossRef]
  61. Fathollahi-Fard, A.M.; Hajiaghaei-Keshteli, M.; Mirjalili, S. Hybrid optimizers to solve a tri-level programming model for a tire closed-loop supply chain network design problem. Appl. Soft Comput. 2018, 70, 701–722. [Google Scholar] [CrossRef]
  62. Gu, Y.; Cai, X.J.; Han, D.R.; Wang, D.Z.W. A tri-level optimization model for a private road competition problem with traffic equilibrium constraints. Eur. J. Operat. Res. 2019, 273, 190–197. [Google Scholar] [CrossRef]
  63. Ye, D.Y.; Chen, Z.J. A new discernibility matrix and the computation of a core. Acta Electr. Sin. 2002, 30, 1086–1088. (In Chinese) [Google Scholar]
  64. Zhang, X.Y.; Miao, D.Q. Quantitative/qualitative region-change uncertainty/certainty in attribute reduction: Comparative region-change analyses based on granular computing. Inf. Sci. 2016, 334–335, 174–204. [Google Scholar]
  65. Dua, D.; Graff, C. UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2019; Available online: http://archive.ics.uci.edu/ml (accessed on 3 July 2019).
Figure 1. Schematic diagram of three-level granular structures.
Figure 1. Schematic diagram of three-level granular structures.
Entropy 21 00657 g001
Figure 2. Convex figure of information function f ( P ) = P log 2 P .
Figure 2. Convex figure of information function f ( P ) = P log 2 P .
Entropy 21 00657 g002
Figure 3. Macro-top’s double-granule conditional-entropies and their three bounds based on an attribute-enlargement chain in the example.
Figure 3. Macro-top’s double-granule conditional-entropies and their three bounds based on an attribute-enlargement chain in the example.
Entropy 21 00657 g003
Figure 4. Macro-top’s double-granule conditional-entropies and their three information bounds based on an attribute-enlargement chain in data experiments.
Figure 4. Macro-top’s double-granule conditional-entropies and their three information bounds based on an attribute-enlargement chain in data experiments.
Entropy 21 00657 g004
Table 1. Three-level granular structures based on condition granulation of the decision table.
Table 1. Three-level granular structures based on condition granulation of the decision table.
Structure NamingComposition SystemGranular ScaleGranular LevelNumber of Parallel Patterns
Micro-Bottom ( A p , A q ) MicroBottom n × n
Meso-Middle ( A p , U / I N D ( A ) )
= ( A p , { A q : q = 1 , , n } )
MesoMiddlen
Macro-Top ( U / I N D ( A ) , U / I N D ( A ) )
= ( { A p : p = 1 , , n } , { A q : q = 1 , , n } )
MacroTop1
Table 2. Matrix distribution of double-granule conditional-entropies at micro-bottom.
Table 2. Matrix distribution of double-granule conditional-entropies at micro-bottom.
U / I N D ( A ) A 1 A q A n
A 1 H ( A 1 , A 1 ) ( D / A ) H ( A 1 , A q ) ( D / A ) H ( A 1 , A n ) ( D / A )
A p H ( A p , A 1 ) ( D / A ) H ( A p , A q ) ( D / A ) H ( A p , A n ) ( D / A )
A n H ( A n , A 1 ) ( D / A ) H ( A n , A q ) ( D / A ) H ( A n , A n ) ( D / A )
Table 3. Three bounds of double-granule conditional-entropies at micro-bottom.
Table 3. Three bounds of double-granule conditional-entropies at micro-bottom.
U / I N D ( A ) A 1 A q A n
A 1 [ H ̲ ( A 1 , A 1 ) ( D / A ) ( D / A ) , H ¯ ( A 1 , A 1 ) ( D / A ) ]
H ( A 1 , A 1 ) * ( D / A )
[ H ̲ ( A 1 , A q ) ( D / A ) , H ¯ ( A 1 , A q ) ( D / A ) ]
H ( A 1 , A q ) * ( D / A )
[ H ̲ ( A 1 , A n ) ( D / A ) , H ¯ ( A 1 , A n ) ( D / A ) ]
H ( A 1 , A n ) * ( D / A )
A p [ H ̲ ( A p , A 1 ) ( D / A ) , H ¯ ( A p , A 1 ) ( D / A ) ]
H ( A p , A 1 ) * ( D / A )
[ H ̲ ( A p , A q ) ( D / A ) , H ¯ ( A p , A q ) ( D / A ) ]
H ( A p , A q ) * ( D / A )
[ H ̲ ( A p , A n ) ( D / A ) , H ¯ ( A p , A n ) ( D / A ) ]
H ( A p , A n ) * ( D / A )
A n [ H ̲ ( A n , A 1 ) ( D / A ) , H ¯ ( A n , A 1 ) ( D / A ) ]
H ( A n , A 1 ) * ( D / A )
[ H ̲ ( A n , A q ) ( D / A ) , H ¯ ( A n , A q ) ( D / A ) ]
H ( A n , A q ) * ( D / A )
[ H ̲ ( A n , A n ) ( D / A ) , H ¯ ( A n , A n ) ( D / A ) ]
H ( A n , A n ) * ( D / A )
Table 4. Marginal distribution of double-granule conditional-entropies at meso-middle and macro-top.
Table 4. Marginal distribution of double-granule conditional-entropies at meso-middle and macro-top.
U / I N D ( A ) A 1 A q A n Meso-Middle
A 1 H ( A 1 , A 1 ) ( D / A ) H ( A 1 , A q ) ( D / A ) H ( A 1 , A n ) ( D / A ) H ( A 1 ) ( D / A )
A p H ( A p , A 1 ) ( D / A ) H ( A p , A q ) ( D / A ) H ( A p , A n ) ( D / A ) H ( A p ) ( D / A )
A n H ( A n , A 1 ) ( D / A ) H ( A n , A q ) ( D / A ) H ( A n , A n ) ( D / A ) H ( A n ) ( D / A )
Meso-Middle H ( A 1 ) ( D / A ) H ( A q ) ( D / A ) H ( A n ) ( D / A ) Macro-Top: H ( D / A )
Table 5. Three bounds of double-granule conditional-entropies at meso-middle and macro-top.
Table 5. Three bounds of double-granule conditional-entropies at meso-middle and macro-top.
U / I N D ( A ) H ( A p ) ( D / A ) H ̲ ( A p ) ( D / A ) H ¯ ( A p ) ( D / A ) H ( A p ) * ( D / A )
A 1 H ( A 1 ) ( D / A ) H ̲ ( A 1 ) ( D / A ) H ¯ ( A 1 ) ( D / A ) H ( A 1 ) * ( D / A )
A p H ( A p ) ( D / A ) H ̲ ( A p ) ( D / A ) H ¯ ( A p ) ( D / A ) H ( A p ) * ( D / A )
A n H ( A n ) ( D / A ) H ̲ ( A n ) ( D / A ) H ¯ ( A n ) ( D / A ) H ( A n ) * ( D / A )
Macro-Top H ( D / A ) H ̲ ( D / A ) H ¯ ( D / A ) H * ( D / A )
Table 6. A decision table.
Table 6. A decision table.
U c 1 c 2 c 3 c 4 c 5 c 6 c 7 c 8 c 9 c 10 c 11 D
x 1 224443444241
x 2 224422444231
x 3 343424244220
x 4 242324242240
x 5 442424344430
x 6 242422244440
x 7 224422234420
x 8 224422244341
Table 7. Information values of double-granule conditional-entropies in the example.
Table 7. Information values of double-granule conditional-entropies in the example.
U A 1 A 2 A 3 A 4 A 5 A 6 Meso-Middle
A 1 0 0.6887 0000 0.6887
A 2 0.6887 0.9183 0.6887 0.6887 0.6887 0.6887 4.3619
A 3 0 0.6887 0000 0.6887
A 4 0 0.6887 0000 0.6887
A 5 0 0.6887 0000 0.6887
A 6 0 0.6887 0000 0.6887
Meso-Middle 0.6887 4.3619 0.6887 0.6887 0.6887 0.6887 Macro-Top: 7.8055
Table 8. Three bounds of double-granule conditional-entropies in the example.
Table 8. Three bounds of double-granule conditional-entropies in the example.
U A 1 A 2 A 3 A 4 A 5 A 6 Meso-Middle
A 1 [ 0 , 0 ]
0
[ 0.1722 , 0.9183 ]
0.8113
[ 0 , 0 ]
1
[ 0 , 0 ]
1
[ 0 , 0 ]
1
[ 0 , 0 ]
1
[ 0.1722 , 0.9183 ]
4.8113
A 2 [ 0.1722 , 0.9183 ]
0.8113
[ 0.3444 , 1.8366 ]
0.9183
[ 0.1722 , 0.9183 ]
1
[ 0.1722 , 0.9183 ]
1
[ 0.1722 , 0.9183 ]
1
[ 0.1722 , 0.9183 ]
1
[ 1.2053 , 6.4281 ]
5.7296
A 3 [ 0 , 0 ]
1
[ 0.1722 , 0.9183 ]
1
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0.1722 , 0.9183 ]
2.0000
A 4 [ 0 , 0 ]
1
[ 0.1722 , 0.9183 ]
1
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0.1722 , 0.9183 ]
2.0000
A 5 [ 0 , 0 ]
1
[ 0.1722 , 0.9183 ]
1
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0.1722 , 0.9183 ]
2.0000
A 6 [ 0 , 0 ]
1
[ 0.1722 , 0.9183 ]
1
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0.1722 , 0.9183 ]
2.0000
Meso-Middle [ 0.1722 , 0.9183 ]
4.8113
[ 1.2053 , 6.4281 ]
5.7296
[ 0.1722 , 0.9183 ]
2.0000
[ 0.1722 , 0.9183 ]
2.0000
[ 0.1722 , 0.9183 ]
2.0000
[ 0.1722 , 0.9183 ]
2.0000
Macro-Top:
[ 2.0662 , 11.0196 ]
18.5049
Table 9. Double-granule conditional-entropies based on an attribute-enlargement chain in the example.
Table 9. Double-granule conditional-entropies based on an attribute-enlargement chain in the example.
LevelMeasure C A 1 C A 2 C A 3 C A 4 C A 5 C A 6 C A 7 C A 8 C A 9 C A 10 C A 11
Micro-Bottom H ( C A k , 1 , C A k , 1 ) ( D / C A k , )
H ( C A k , 1 , C A k , 2 ) ( D / C A k )
H ( C A k , 1 , C A k , 3 ) ( D / C A k )
H ( C A k , 2 , C A k , 2 ) ( D / C A k )
H ( C A k , 2 , C A k , 3 ) ( D / C A k )
1.0000
0.8571
0.8571
0
0
0.8113
0.6490
0.5409
0
0
0.8113
0.6490
0.5409
0
0
0.8113
0.6490
0.6490
0
0
0
0.6887
0
0.9183
0.6887
0
0.6887
0
0.9183
0.6887
0
1
0
0
0
0
1
0
0
0
0
1
0
0
0
0
1
0
0
0
0
1
0
0
0
Meso-Middle H ( C A k , 1 ) ( D / C A k )
H ( C A k , 2 ) ( D / C A k )
H ( C A k , 3 ) ( D / C A k )
2.7143
0.8571
0.8571
2.6502
0.6490
0.5409
2.6502
0.6490
0.5409
3.4074
0.6490
0.6490
0.6887
4.3619
0.6887
0.6887
4.3619
0.6887
0.6667
0.6667
0.6667
0
0
0
0
0
0
0
0
0
0
0
0
Macro-Top H ( D / C A k )
H ̲ ( D / C A k )
H ¯ ( D / C A k )
H * ( D / C A k )
4.4286
2.2500
6.0000
4.9409
4.4891
1.6226
6.4902
6.6951
4.4891
1.6226
6.4902
6.6951
6.0035
2.0282
8.1128
8.5789
7.8055
2.0282
11.0196
18.5409
7.8055
2.0282
11.0196
18.5409
9.0000
1.7500
14.0000
28.0196
0
0
0
30
0
0
0
30
0
0
0
30
0
0
0
30
Table 10. Double-granule conditional-entropies regarding C A 2 = { c 1 , c 2 } in the example.
Table 10. Double-granule conditional-entropies regarding C A 2 = { c 1 , c 2 } in the example.
U C A 2 , 1 C A 2 , 2 C A 2 , 3 C A 2 , 4 Meso-Middle
C A 2 , 1 0.8113 0.6490 0.5409 0.6490 2.6502
C A 2 , 2 0.6490 000 0.6490
C A 2 , 3 0.5409 000 0.5409
C A 2 , 4 0.6490 000 0.6490
Meso-Middle 2.6502 0.6490 0.5409 0.6490 Macro-Top: 4.4891
Table 11. Three UCI data sets.
Table 11. Three UCI data sets.
LabelName | U | | C | | U / IND ( C ) | | D | | U / IND ( D ) |
(1)VOTING4351634212
(2)SPECT1872216912
(3)Tic-Tac-Toe958995812
Table 12. Double-granule conditional-entropies in the VOTING data set.
Table 12. Double-granule conditional-entropies in the VOTING data set.
U / I N D ( C A 1 ) C A 1 , 1 C A 1 , 2 C A 1 , 3 Meso-Middle U / I N D ( C A 16 ) C A 16 , 1 C A 16 , 342 Meso-Middle
C A 1 , 1 0.98670.97820.83692.8018 C A 16 , 1 000
C A 1 , 2 0.97820.81130.65782.4473
C A 1 , 3 0.83690.65780.64792.1427 C A 16 , 342 000
Meso-Middle2.80182.44732.1427Macro-Top:
7.3918
Meso-Middle00Macro-Top:
0
Table 13. Three information bounds in the VOTING data set.
Table 13. Three information bounds in the VOTING data set.
U / I N D ( C A 1 ) C A 1 , 1 C A 1 , 2 C A 1 , 3 Meso-Middle U / I N D ( C A 16 ) C A 16 , 1 C A 16 , 342 Meso-Middle
C A 1 , 1 [ 1.0706 ,
1.9734 ]
0.9867
[ 0.5577 ,
1.7980 ]
0.9921
[ 0.8139 ,
1.6346 ]
0.9649
[ 2.4422 ,
5.4060 ]
2.9436
C A 16 , 1 [ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
221.3143
C A 1 , 2 [ 0.5577 ,
1.7980 ]
0.9921
[ 0.0448 ,
1.6226 ]
0.8113
[ 0.3009 ,
1.4592 ]
0.6596
[ 0.9034 ,
4.8798 ]
2.4630
C A 1 , 3 [ 0.8139 ,
1.6346 ]
0.9649
[ 0.3009 ,
1.4592 ]
0.6596
[ 0.5571 ,
1.2959 ]
0.6479
[ 1.6719 ,
4.3898 ]
2.2724
C A 16 , 342 [ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
221.3143
Meso-Middle [ 2.4422 ,
5.4060 ]
2.9436
[ 0.9034 ,
4.8798 ]
2.4630
[ 1.6719 ,
4.3898 ]
2.2724
Macro-Top:
[ 5.0174 ,
14.6755 ]
7.6790
Meso-Middle [ 0 , 0 ]
221.3143
[ 0 , 0 ]
221.3143
Macro-Top:
[ 0 , 0 ]
50132
Table 14. Double-granule conditional-entropies in the SPECT data set.
Table 14. Double-granule conditional-entropies in the SPECT data set.
U / I N D ( C A 1 ) C A 1 , 1 C A 1 , 2 Meso-Middle U / I N D ( C A 22 ) C A 22 , 1 C A 22 , 169 Meso-Middle
C A 1 , 1 0.21080.38150.5924 C A 22 , 1 001.5335
C A 1 , 2 0.38150.53990.9215 C A 22 , 169 001.5335
Meso-Middle0.59240.9215Macro-Top:
1.5139
Meso-Middle1.53351.5335Macro-Top:
513.0879
Table 15. Three information bounds in the SPECT data set.
Table 15. Three information bounds in the SPECT data set.
U / I N D ( C A 1 ) C A 1 , 1 C A 1 , 2 Meso-Middle U / I N D ( C A 22 ) C A 22 , 1 C A 22 , 169 Meso-Middle
C A 1 , 2 [ 0.1015 ,
0.4217 ]
0.2109
[ 0.1908 ,
0.7508 ]
0.4030
[ 0.2922 ,
1.1725 ]
0.6138
C A 22 , 1 [ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0.0332 ,
1.9457 ]
8.8982
C A 1 , 2 [ 0.1908 ,
0.7508 ]
0.4030
[ 0.2801 ,
1.0799 ]
0.5400
[ 0.4708 ,
1.8306 ]
0.9429
C A 22 , 169 [ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0.0332 ,
1.9457 ]
161.2140
Meso-Middle [ 0.2922 ,
1.1725 ]
0.6138
[ 0.4708 ,
1.8306 ]
0.9429
Macro-Top:
[ 0.7631 ,
14.6755 ]
1.5566
Meso-Middle [ 0.0332 ,
1.9457 ]
8.8982
[ 0.0332 ,
1.9457 ]
161.2140
Macro-Top:
[ 11.2085 ,
657.6332 ]
2867
Table 16. Double-granule conditional-entropies in the Tic-Tac-Toe data set.
Table 16. Double-granule conditional-entropies in the Tic-Tac-Toe data set.
U / I N D ( C A 1 ) C A 1 , 1 C A 1 , 2 C A 1 , 3 Meso-Middle U / I N D ( C A 9 ) C A 9 , 1 C A 9 , 958 Meso-Middle
C A 1 , 1 0.87420.92480.87942.6784 C A 9 , 1 000
C A 1 , 2 0.92480.98810.95092.8638
C A 1 , 3 0.87940.95090.89012.7203 C A 9 , 958 000
Meso-Middle2.67842.86382.7203Macro-Top:
8.2625
Meso-Middle00Macro-Top:
0
Table 17. Three information bounds in the Tic-Tac-Toe data set.
Table 17. Three information bounds in the Tic-Tac-Toe data set.
U / I N D ( C A 1 ) C A 1 , 1 C A 1 , 2 C A 1 , 3 Meso-Middle U / I N D ( C A 9 ) C A 9 , 1 C A 9 , 958 Meso-Middle
C A 1 , 1 [ 0.3814 ,
1.7483 ]
0.8741
[ 0.3635 ,
1.8622 ]
0.9404
[ 0.2859 ,
1.7642 ]
0.8796
[ 1.0308 ,
5.3748 ]
2.6940
C A 9 , 1 [ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
332
C A 1 , 2 [ 0.3635 ,
1.8622 ]
0.9404
[ 0.3455 ,
1.9762 ]
0.9881
[ 0.2680 ,
1.8781 ]
0.9628
[ 0.9770 ,
5.7165 ]
2.8913
C A 1 , 3 [ 0.2859 ,
1.7642 ]
0.8796
[ 0.2680 ,
1.8781 ]
0.9628
[ 0.1905 ,
1.7801 ]
0.8900
[ 0.7444 ,
5.4224 ]
2.7324
C A 9 , 958 [ 0 , 0 ]
0
[ 0 , 0 ]
0
[ 0 , 0 ]
626
Meso-Middle [ 1.0308 ,
5.3748 ]
2.6940
[ 0.9770 ,
5.7165 ]
2.8913
[ 0.7444 ,
5.4224 ]
2.7342
Macro-Top:
[ 2.7522 ,
16.5138 ]
8.3178
Meso-Middle [ 0 , 0 ]
332
[ 0 , 0 ]
626
Macro-Top:
[ 0 , 0 ]
415664

Share and Cite

MDPI and ACS Style

Mu, T.; Zhang, X.; Mo, Z. Double-Granule Conditional-Entropies Based on Three-Level Granular Structures. Entropy 2019, 21, 657. https://doi.org/10.3390/e21070657

AMA Style

Mu T, Zhang X, Mo Z. Double-Granule Conditional-Entropies Based on Three-Level Granular Structures. Entropy. 2019; 21(7):657. https://doi.org/10.3390/e21070657

Chicago/Turabian Style

Mu, Taopin, Xianyong Zhang, and Zhiwen Mo. 2019. "Double-Granule Conditional-Entropies Based on Three-Level Granular Structures" Entropy 21, no. 7: 657. https://doi.org/10.3390/e21070657

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop