Next Article in Journal
Research on a Sowing Depth Detection System Based on an Improved Adaptive Kalman Filtering Method
Next Article in Special Issue
A Context Awareness Hierarchical Attention Network for Next POI Recommendation in IoT Environment
Previous Article in Journal
MM-LMF: A Low-Rank Multimodal Fusion Dangerous Driving Behavior Recognition Method Based on FMCW Signals
Previous Article in Special Issue
Relative Knowledge Distance Measure of Intuitionistic Fuzzy Concept
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cost-Sensitive Multigranulation Approximation in Decision-Making Applications

1
School of Physics and Electronic Science, Zunyi Normal University, Zunyi 563002, China
2
Chongqing Key Laboratory of Computational Intelligence, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(22), 3801; https://doi.org/10.3390/electronics11223801
Submission received: 25 October 2022 / Revised: 13 November 2022 / Accepted: 16 November 2022 / Published: 18 November 2022
(This article belongs to the Special Issue Advances in Intelligent Data Analysis and Its Applications)

Abstract

:
A multigranulation rough set (MGRS) model is an expansion of the Pawlak rough set, in which the uncertain concept is characterized by optimistic and pessimistic upper/lower approximate boundaries, respectively. However, there is a lack of approximate descriptions of uncertain concepts by existing information granules in MGRS. The approximation sets of rough sets presented by Zhang provide a way to approximately describe knowledge by using existing information granules. Based on the approximation set theory, this paper proposes the cost-sensitive multigranulation approximation of rough sets, i.e., optimistic approximation and pessimistic approximation. Their related properties were further analyzed. Furthermore, a cost-sensitive selection algorithm to optimize the multigranulation approximation was performed. The experimental results show that when multigranulation approximation sets and upper/lower approximation sets are applied to decision-making environments, multigranulation approximation produces the least misclassification costs on each dataset. In particular, misclassification costs are reduced by more than 50% at each granularity on some datasets.

1. Introduction

As a human-inspired paradigm, granular computing (GrC) solves complex problems by utilizing multiple granular layers [1,2,3,4]. Zadeh [1] noted that information in granules refer to pieces, classes, and groups, into which complex information is divided in accordance with the characteristics and processes of understanding and decision-making. From the different views, GrC models mainly cover four types: fuzzy sets [5], rough sets [6], quotient spaces [7], and cloud models [8]. As representative models of GrC, rough sets describe uncertain concepts by upper and lower approximation boundaries, which have been applied to data mining [9,10], medical systems [11], attribute reductions [12,13], decision systems [14,15], and machine learning [16].
Regarding similarity, Zhang [17,18,19,20] presented the approximation set of rough sets, vague sets, rough fuzzy sets, rough vague sets, etc. These approximation models were developed by utilizing the existing equivalence classes to describe uncertain concepts. The approximation model has a higher similarity with the target concept than the upper/lower approximations. Furthermore, the approximation model has been applied in attribute reduction [21], image segmentation [22], the optimization algorithm [23], etc. Based on the approximation set theory, Yang [24,25] developed the approximation model of rough sets based on misclassification costs. In the process of cost-sensitive learning, the smaller misclassification costs will help to improve the decision-making qualities in real applications. Recently, from the perspective of three-way decisions [26,27,28,29], Yao [30] constructed a symbol–meaning–value (SMV) model for data analysis. In the three-way decision model, the equivalence classes in a boundary region will produce misclassification costs when they are used as approximation sets. Hence, the approximation model that is constructed from the perspective of similarity is no longer applicable to cost-sensitive scenarios. To minimize the misclassification costs of constructing the approximation set, we proposed the multigranulation approximation, i.e., the optimistic approximation model and pessimistic approximation model. Moreover, to search the optimal approximation layer for multigranulation rough sets [31] under the constraints, the algorithm of the cost-sensitive multigranulation approximation selection is further proposed to be applied to decision-making environments.
The following sections are arranged as follows: Section 2 presents the related works. Section 3 introduces the relevant definitions of the multigranulation rough set and approximation set. Section 4 introduces an approximate representation of the rough sets. Section 5 presents the cost-sensitive multigranulation approximations of rough sets and further introduces the optimal multigranulation approximation algorithm. To verify the availability of the proposed model, the related experiments and discussion are presented in Section 6. Ultimately, in Section 7, the conclusions are presented.

2. Literature Review

Rough sets are typically constructed based on a single binary relation. However, in many cases, they may be described in multiple granularity structures. In order to extend single granularity to multi-granularity in rough approximations, Qian [31] proposed the multigranulation rough set model (MGRS), where the upper/lower approximations were defined by multi-equivalence relations (multiple granulations) in the universe [32,33]. For the lower approximation of optimistic MGRS, at least one granular space was obtained, such that objects completely belonged to the target concept. For the lower approximation of pessimistic MGRS, objects completely belong to target concepts in each granular space. MGRS has two advantages: (1) In the process of decision-making applications, the decision of each decision maker may be independent of the same project (or an element) in the universe [34]. In this situation, the intersection operations between any two granularity structures will be redundant for decision-making [35]. (2) Extract decision rules from distributive information systems and groups of intelligent agents by using rough set approaches [34,36].
There are many works [33,34,35,37,38,39,40,41,42] on multigranulation rough sets. To extend the MGRS to the neighborhood information system, Hu [43,44] presented matrix-based incremental approaches to update knowledge about neighborhood information systems by changing the granular structures. From the perspective of uncertainty measure, Sun [39] proposed a feature selection based on fuzzy neighborhood multigranulation rough sets. Xu [38] proposed a dynamic approximation update mechanism of a multigranulation neighborhood rough set from a local viewpoint. Liu [35] introduced a parameter-free multi-granularity attribute reduction scheme, which is more effective for microarray data than other well-established attribute reductions. Based on the three-way decision theory, She [40] presented a five-valued logic approach for the multigranulation rough set model. From the above, however, the method of approximately describing the target concept with existing information granules is not given, which limits the application of the multigranulation rough set theory. Li [41] presented two kinds of local multigranulation rough set models in the ordered decision system by extending the single granulation environment to a multigranulation case. Zhang [42] constructed hesitant fuzzy multigranulation rough sets to handle the hesitant fuzzy information and group decision-making for person–job fit.

3. Preliminaries

In this section, some necessary definitions related to the multigranulation rough set and approximation set are reviewed to facilitate the framework of this paper. Let S = ( U , C D , V , f ) be a decision information table, where U is a non-empty finite set of objects, C is a non-empty finite set of condition attributes, D is a set of decision attributes, V is the set of all attribute values, and f : U × C is an information function.
Definition 1 
((Rough Sets) [6]). Let S = ( U , C D , V , f ) be a decision information table, A C and X U , the lower and upper approximation sets of X are given as follows:
A ̲ ( X ) = { x U [ x ] A X } , A ¯ ( X ) = { x U [ x ] A X ϕ } .
where [ x ] A denotes the equivalence class induced by U U A A , namely, U U A A = { [ x ] R } = { [ x ] 1 , [ x ] 2 , , [ x ] l } .
Based on the lower and upper approximations, the universe U can be divided into three disjoint regions, which are expressed as follows:
P O S A ( X ) = A ̲ ( X ) , B N D A ( X ) = A ¯ ( X ) A ̲ ( X ) , N E G A ( X ) = U A ¯ ( X ) .
Definition 2 
((Optimistic multigranulation rough sets) [36]). Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C and X U , then the lower and upper approximation sets of X related to A 1 , A 2 , , A m are given as follows:
i = 1 m A i O ̲ ( X ) = { x [ x ] A 1 X [ x ] A 2 X [ x ] A m X , x U } ,
i = 1 m A i O ¯ ( X ) = i = 1 m A i O ̲ ( X ) .
Then, ( i = 1 m A i O ̲ ( X ) , i = 1 m A i O ¯ ( X ) ) is called optimistic multigranulation rough sets. The lower and upper approximation sets of X in optimistic multigranulation rough sets are presented by multiple independent approximation spaces. The boundary regions are defined as follows:
B N D i = 1 m A i O ( X ) = i = 1 m A i O ¯ ( X ) i = 1 m A i O ̲ ( X ) .
Definition 3 
((Pessimistic multigranulation rough sets) [36]). Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C , and X U . The lower and upper approximation sets of X related to A 1 , A 2 , , A m are given as follows:
i = 1 m A i P ̲ ( X ) = { x [ x ] A 1 X [ x ] A 2 X [ x ] A m X , x U } ,
i = 1 m A i P ¯ ( X ) = i = 1 m A i P ̲ ( X ) .
Then, ( i = 1 m A i P ̲ ( X ) , i = 1 m A i P ¯ ( X ) ) is called pessimistic multigranulation rough sets. The lower and upper approximation sets of X in pessimistic multigranulation rough sets are presented by multiple independent approximation spaces. However, the strategy is different from optimistic multigranulation rough sets. The boundary region is defined as follows:
B N D i = 1 m A i P ( X ) = i = 1 m A i P ¯ ( X ) i = 1 m A i P ̲ ( X ) .
Definition 4 
((Approximation of rough sets) [17]). Let S = ( U , C D , V , f ) be a decision information table, A C and X U . U U A A = { [ x ] 1 , [ x ] 2 , , [ x ] l } is a granularity layer on U, then the α-approximation of X on U U R A is defined as follows:
A α ( X ) = [ x ] i μ ¯ ( [ x ] i ) α , [ x ] i U .
where 0 α 1 , μ ¯ ( [ x ] i ) = [ x ] i X [ x ] i denotes the membership degree of the equivalence class [ x ] i belongs to X.
Example 1. 
Let Table 1 be a decision information table, A 1 , A 2 , A 3 C and X U . For the element x 4 , the equivalence classes [ x 4 ] i ( i = 1 , 2 , 3 ) belonging to X in the multigranulation approximation space are as follows:
[ x 4 ] 1 = { x 1 , x 2 , x 3 , x 4 } ; [ x 4 ] 2 = { x 3 , x 4 , x 7 , x 8 } ; [ x 4 ] 3 = { x 2 , x 4 , x 6 , x 8 } .
Accordingly, the membership degrees are computed:
μ ¯ ( [ x 4 ] 1 ) = 0 + 0 + 0 + 1 4 = 0.25 ; μ ¯ ( [ x 4 ] 2 ) = 0 + 1 + 1 + 1 4 = 0.75 ; μ ¯ ( [ x 4 ] 3 ) = 0 + 1 + 1 + 1 4 = 0.75 .
If α is set to 0.5, considering the optimistic approximation, element x 4 will be classified into the optimistic lower approximation sets of X due to one of its membership degrees being greater than 0.5. However, if considering the pessimistic approximation, element x 4 will only be classified into the pessimistic upper approximation sets of X.
Based on the given conditions, we have:
X = 0.25 + 0.25 + 0.25 x 1 + 0.25 + 0.25 + 0.75 x 2 + 0.25 + 0.75 + 0.05 x 3 + 0.25 + 0.75 + 0.75 x 4 + 0.75 + 0.25 + 0.25 x 5 + 0.75 + 0.25 + 0.75 x 6 + 0.75 + 0.75 + 0.25 x 7 + 0.75 + 0.75 + 0.75 x 8 .
Then, the results of the optimistic approximations are shown as follows:
i = 1 m A i O ̲ ( X ) = { x 2 , x 3 , x 4 , x 5 , x 6 , x 7 , x 8 } , i = 1 m A i O ¯ ( X ) = { x 1 , x 2 , x 3 , x 4 , x 5 , x 6 , x 7 , x 8 } , B N D i = 1 m A i O ( X ) = { x 1 } .
Moreover, the results of the pessimistic approximations, in this case, are changed as follows:
i = 1 m A i P ̲ ( X ) = { x 8 } , i = 1 m A i P ¯ ( X ) = { x 1 , x 2 , x 3 , x 4 , x 5 , x 6 , x 7 , x 8 } , B N D i = 1 m A i P ( X ) = { x 1 , x 2 , x 3 , x 4 , x 5 , x 6 , x 7 } .

4. Cost-Sensitive Approximation Methods of the Rough Sets

Let S = ( U , C D , V , f ) be a decision information table, A C and X U . U U A A = { [ x ] 1 , [ x ] 2 , , [ x ] l } is a granularity layer on U. λ 12 represents the cost generated by taking an element belonging to X as the approximation, λ 21 represents the cost generated by taking an element that does not belong to X as the approximation. Furthermore, misclassification costs incurred by the equivalence classes in characterizing X on U U A A are given in the following:
λ Y = λ 12 ( 1 μ ¯ ( [ x ] i ) ) [ x ] i .
Misclassification costs incurred by the equivalence classes when not characterizing X on U U A A are given in the following:
λ N = λ 21 μ ¯ ( [ x ] i ) [ x ] i .
Herein, μ ¯ ( [ x ] i ) ( i = 1 , 2 , , l ) denotes the membership degree of [ x ] i belonging to X. Based on the Bayesian decision procedure, the minimum cost decision rules are expressed as follows:
(P)
If λ Y λ N , [ x ] i A ( X ) ;
(N)
If λ Y > λ N , [ x ] i A ( X ) .
It is clear that the above rules are only relevant to the loss function μ ¯ ( [ x ] i ) . From Formulas (8) and (9), the decision rules are re-expressed in the following:
(P1)
If μ ¯ ( [ x ] i ) λ 12 λ 12 + λ 21 , [ x ] i A ( X ) ;
(N1)
If μ ¯ ( [ x ] i ) < λ 12 λ 12 + λ 21 , [ x ] i A ( X ) .
Supposing γ = λ 12 λ 12 + λ 21 , then we have the following decision rules:
(P2)
If μ ¯ ( [ x ] i ) γ , [ x ] i A ( X ) ;
(N2)
If μ ¯ ( [ x ] i ) < γ , [ x ] i A ( X ) .
Definition 5. 
Let S = ( U , C D , V , f ) be a decision information table, A C and X U . U U A A = { [ x ] 1 , [ x ] 2 , , [ x ] l } is a granularity layer on U, then the cost-sensitive approximation model (CSA) of rough sets is defined as follows:
A ( X ) = [ x ] i μ ¯ ( [ x ] i ) λ 12 λ 12 + λ 21 , [ x ] i U .
Suppose 0 λ 12 , λ 21 1 , boundary region I and boundary region II are denoted by B N 1 ( X ) = { [ x ] i λ 12 λ 12 + λ 21 μ ¯ ( [ x ] i ) < 1 } and B N 2 ( X ) = { [ x ] i 0 < μ ¯ ( [ x ] i ) < λ 12 λ 12 + λ 21 } , respectively, then B N ( X ) = B N 1 ( X ) B N 2 ( X ) , and A ( X ) = B N 1 ( X ) P O S ( X ) . Figure 1 shows the CSA of rough sets, where B N 1 ( X ) is the dark blue region, which denotes the region in the boundary region used as the approximation. B N 2 ( X ) is the light blue region, which denotes the region in the boundary region not used as the approximation. Therefore, the region surrounded by the green broken line in Figure 1 constructs the approximations of rough sets, and the misclassification costs come from two uncertain regions, defined as follows:
D C ( A ( X ) ) = [ x ] i B N 1 ( X ) λ Y + [ x ] i B N 2 ( X ) λ N .
Theorem 1. 
Let S = ( U , A D , V , f ) be a decision information table, A 1 , A 2 , , A m C , then D C ( A 1 ( X ) ) D C ( A 2 ( X ) ) .
Proof of Theorem 1. 
Let U be a non-empty finite domain, U U A A 1 = { E 1 , E 2 , , E l } and U U A A 2 = { F 1 , F 2 , , F m } . Because A 1 A 2 , U U A 2 A 2 U U R 1 A 1 . According to the condition, for simplicity, supposing only one granule E 1 can be subdivided into two finer sub-granules by Δ A = A 2 A 1 (the more complicated cases can be transformed into this case, so we will not repeat them here). Without loss of generality, let E 1 = F 1 F 2 , E 2 = F 3 , E 3 = F 4 , , E l = F m ( m = l + 1 ), namely, U / A 2 = { F 1 , F 2 , E 2 , E 3 , , E l } . There are two cases to prove the theorem as follows:
(1)
Suppose μ ¯ ( E 1 ) γ , obviously, E 1 A 1 ( X ) .
Case 1. μ ¯ ( F 1 ) γ and μ ¯ ( F 2 ) γ . Namely, F 1 A ( X ) and F 2 A ( X ) . Case 1, in which the granules are subdivided in B N 1 ( X ) , can be shown in Figure 2a, then
Δ D C A 1 A 2 ( X ) = D C ( A 1 ( X ) ) D C ( A 2 ( X ) ) = λ 12 ( 1 μ ¯ ( E 1 ) ) E 1 λ 12 ( 1 μ ¯ ( F 1 ) ) F 1 λ 12 ( 1 μ ¯ ( F 2 ) ) F 2 . = ( E 1 F 1 F 2 + x i F 1 μ ( x i ) + x i F 2 μ ( x i ) x i E 1 μ ( x i ) ) λ 12 .
Because x i E 1 μ ( x i ) = x i F 1 μ ( x i ) + x i F 2 μ ( x i ) and E 1 = F 1 + F 2 , then Δ D C A 1 A 2 = 0 . Therefore, D C ( A 1 ( X ) ) = D C ( A 2 ( X ) ) .
Case 2. μ ¯ ( F 1 ) γ and μ ¯ ( F 2 ) < γ . Namely, F 1 A ( X ) and F 2 A ( X ) . Case 2, in which the granules are subdivided in B N 1 ( X ) , can be shown in Figure 2b, then
Δ D C A 1 A 2 = D C ( A 1 ( X ) ) D C ( A 2 ( X ) ) = λ 12 ( 1 μ ¯ ( E 1 ) ) E 1 λ 12 ( 1 μ ¯ ( F 1 ) ) F 1 λ 21 μ ¯ ( F 2 ) F 2 = F 2 ( λ 12 μ ¯ ( F 2 ) ( λ 21 + λ 12 ) )
Because μ ¯ ( F 2 ) < γ = λ 12 λ 12 + λ 21 and E 1 = F 1 + F 2 , then Δ D C A 1 A 2 ( X ) 0 . Therefore, D C ( A 1 ( X ) ) > D C ( A 2 ( X ) ) .
(2)
Suppose μ ¯ ( E 1 ) < γ , obviously, E 1 A 1 ( X ) .
Case 1. μ ¯ ( F 1 ) γ and μ ¯ ( F 2 ) < γ . Namely, F 1 A ( X ) and F 2 A ( X ) . Case 1, in which the granules are subdivided in B N 2 ( X ) , can be shown in Figure 2c, then
Δ D C A 1 A 2 = D C ( A 1 ( X ) ) D C ( A 2 ( X ) ) = μ ¯ ( E 1 ) E 1 λ 21 μ ¯ ( F 2 ) F 2 λ 21 ( 1 μ ¯ ( F 1 ) ) F 1 λ 12 = F 1 μ ¯ ( F 1 ) ( ( λ 21 + λ 12 ) λ 12 ) .
Because μ ¯ ( F 1 ) γ = λ 12 λ 12 + λ 21 , then Δ D C A 1 A 2 0 . Therefore, D C ( A 1 ( X ) ) D C ( A 2 ( X ) ) .
Case 2. μ ¯ ( F 1 ) < γ and μ ¯ ( F 2 ) < γ . Namely, F 1 R ( X ) and F 2 R ( X ) . Case 2, in which the granules are subdivided in B N 2 ( X ) , can be shown in Figure 2d, then
Δ D C A 1 A 2 = D C ( A 1 ( X ) ) D C ( A 2 ( X ) ) = μ ¯ ( E 1 ) E 1 λ 21 μ ¯ ( F 1 ) F 1 λ 21 μ ¯ ( F 2 ) F 2 λ 21 = ( x i E 1 μ ( x i ) x i F 1 μ ( x i ) x i F 2 μ ( x i ) ) λ 21 .
Because x i E 1 μ ( x i ) = x i F 1 μ ( x i ) + x i F 2 μ ( x i ) , then Δ D C A 1 A 2 = 0 . Therefore, D C ( A 1 ( X ) ) = D C ( A 2 ( X ) ) . □
Theorem 1 shows that the misclassification costs in the approximation model will monotonously decrease with the changing approximation space, in accordance with human cognitive mechanisms.

5. Cost-Sensitive Multigranulation Approximations and Optimal Granularity Selection Method

The multigranulation rough set model (MGRS) [43] extends single granularity to multi-granularity in rough approximations to describe an uncertain concept. MGRS is an expansion of the classical rough set, and the target concept is characterized by optimistic and pessimistic upper/lower approximation boundaries in MGRS, respectively. However, there is a lack of an approximate description of an uncertain concept by utilizing the equivalence classes in MGRS. In this section, based on the model proposed in Section 3, we further construct the approximations of MGRS.

5.1. Cost-Sensitive Multigranulation Approximations of Rough Sets

Definition 6. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C and X U . The optimistic membership degree of x U related to A 1 , A 2 , , A m is given as follows:
μ ¯ i = 1 m A i O ( x ) = max { μ ¯ ( [ x ] A i ) i = 1 , 2 , , m } .
Definition 7. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C and X U . The approximation model of the optimistic MGRS of X related to A 1 , A 2 , , A m is given as follows:
i = 1 m A i O ( X ) = { x μ ¯ ( [ x ] A 1 ) γ μ ¯ ( [ x ] A 2 ) γ μ ¯ ( [ x ] A m ) γ , x U } .
From the perspective of the optimistic membership degree, i = 1 m A i O ( X ) can be =expressed as follows:
i = 1 m A i O ( X ) = { x μ ¯ i = 1 m A i O ( x ) γ , x U } .
The corresponding decision regions are defined as follows:
B N 1 i = 1 m A i O ( X ) = { x 1 > μ ¯ i = 1 m A i O ( x ) γ , x U } , B N 2 i = 1 m A i O ( X ) = { x 0 < μ ¯ i = 1 m A i O ( x ) < γ , x U } , P O S i = 1 m A i O ( X ) = { x μ ¯ i = 1 m A i O ( x ) = 1 , x U } , N E G i = 1 m A i O ( X ) = { x μ ¯ i = 1 m A i O ( x ) = 0 , x U } .
We have
i = 1 m A i O ( X ) = B N 1 i = 1 m A i O ( X ) P O S i = 1 m A i O ( X ) .
The misclassification costs of approximations of optimistic MGRS come from two uncertain regions B N 1 i = 1 m A i O ( X ) and B N 2 i = 1 m A i O ( X ) , which are defined in the following:
D C ( i = 1 m A i O ( X ) ) = x B N 1 i = 1 m A i O ( X ) λ Y + x B N 2 i = 1 m A i O ( X ) λ N .
Theorem 2. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C , X U and X 1 , X 2 , , X n U . The following properties hold:
(1)
i = 1 m A i O ( X ) = i = 1 m A i ( X ) ;
(2)
i = 1 m A i O ( j = 1 n X j ) = i = 1 m ( j = 1 n A i O ( X j ) ) ;
(3)
i = 1 m A i O ( j = 1 n X j ) j = 1 n ( i = 1 m A i O ( X j ) ) ;
(4)
i = 1 m A i O ( j = 1 n X j ) j = 1 n ( i = 1 m A i O ( X j ) ) ;
(5)
i = 1 m A i O ̲ ( X ) i = 1 m A i O ( X ) i = 1 m A i O ¯ ( X ) .
Proof of Theorem 2. 
(1)
From Formula (14), i = 1 m A i O ( X ) = i = 1 m A i ( X ) obviously holds.
(2)
i = 1 m A i O ( j = 1 n X j ) = i = 1 m A i O ( j = 1 n X j ) = i = 1 m ( j = 1 n A i O ( X j ) ) .
(3)
i = 1 m A i O ( j = 1 n X j ) = i = 1 m ( j = 1 n A i ( X j ) ) = j = 1 n ( i = 1 m A i ( X j ) = j = 1 n ( i = 1 m A i O ( X j ) ) = j = 1 n ( i = 1 m A i O ( X j ) ) .
(4)
From X j j = 1 n X j , we have i = 1 m A i O ( X j ) i = 1 m A i O ( j = 1 n X j ) . Therefore, i = 1 m A i O ( j = 1 n X j ) j = 1 n ( i = 1 m A i O ( X j ) ) .
(5)
It is easy to prove by Formulas (1), (2), and (14).
 □
Definition 8. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C and X U . The pessimistic membership degree of x U related to A 1 , A 2 , , A m is given as follows:
μ ¯ i = 1 m A i P ( x ) = min { μ ¯ ( [ x ] A i ) i = 1 , 2 , , m } .
Definition 9. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C and X U . The approximation model of pessimistic MGRS of X related to A 1 , A 2 , , A m is given as follows:
i = 1 m A i P ( X ) = { x μ ¯ ( [ x ] A 1 ) γ μ ¯ ( [ x ] A 2 ) γ μ ¯ ( [ x ] A m ) γ , x U } .
From the perspective of the pessimistic membership degree, i = 1 m A i P ( X ) can be expressed as follows:
i = 1 m A i P ( X ) = { x μ ¯ i = 1 m A i P ( x ) γ , x U } .
The corresponding decision regions are expressed in the following:
B N 1 i = 1 m A i P ( X ) = { x 1 > μ ¯ i = 1 m A i P ( x ) γ , x U } , B N 2 i = 1 m A i P ( X ) = { x 0 < μ ¯ i = 1 m A i P ( x ) < γ , x U } , P O S i = 1 m A i P ( X ) = { x μ ¯ i = 1 m A i P ( x ) = 1 , x U } , N E G i = 1 m A i P ( X ) = { x μ ¯ i = 1 m A i P ( x ) = 0 , x U } .
We have
i = 1 m A i P ( X ) = B N 1 i = 1 m A i P ( X ) P O S i = 1 m A i P ( X ) .
The misclassification costs of approximations of optimistic MGRS come from two uncertain regions B N 1 i = 1 m A i P ( X ) and B N 2 i = 1 m A i P ( X ) , which are defined in the following:
D C ( i = 1 m A i P ( X ) ) = x B N 1 i = 1 m A i P ( X ) λ Y + x B N 1 i = 1 m A i P ( X ) λ N .
Theorem 3. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C , X U and X 1 , X 2 , , X n U . The following properties hold:
(1)
i = 1 m A i P ( X ) = i = 1 m ( A i ( X ) ) ;
(2)
i = 1 m A i P ( j = 1 n X j ) = j = 1 n ( i = 1 m A i P ( X j ) ) ;
(3)
i = 1 m A i P ( j = 1 n X j ) j = 1 n ( i = 1 m A i P ( X j ) ) ;
(4)
i = 1 m A i P ̲ ( X ) i = 1 m A i P ( X ) i = 1 m A i P ¯ ( X ) .
Proof of Theorem 3. 
(1)
x i = 1 m A i P ( X ) , according to Definition 9, μ ¯ ( [ x ] A i ) γ holds, i = 1 , 2 , , m . According to Definition 5, x A i ( X ) , i = 1 , 2 , , m and x i = 1 m ( A i ( X ) ) holds, i.e., i = 1 m A i P ( X ) i = 1 m ( A i ( X ) ) . x i = 1 m ( A i ( X ) ) , μ ¯ ( [ x ] A i ) γ , i = 1 , 2 , , m . According to Definition 7, x i = 1 m A i P ( X ) holds, i.e., i = 1 m ( A i ( X ) ) i = 1 m A i P ( X ) . Therefore, we have i = 1 m A i P ( X ) = i = 1 m ( A i ( X ) ) .
(2)
From the proof of (1), i = 1 m A i P ( j = 1 n X j ) = i = 1 m A i ( j = 1 n X j ) . According to Definition 7, i = 1 m A i ( j = 1 n X j ) = i = 1 m j = 1 n A i ( X j ) . Because i = 1 m A i ( X j ) = i = 1 m A i P ( X j ) , i = 1 m A i P ( j = 1 n X j ) = i = 1 m A i ( j = 1 n X j ) = j = 1 n i = 1 m A i ( X j ) = j = 1 n ( i = 1 m A i P ( X j ) ) .
(3)
x j = 1 n ( i = 1 m A i P ( X j ) ) , X k ( k { 1 , 2 , , n } ) , x i = 1 m A i P ( X j ) . According to Definition 7, X j , j = 1 , 2 , , n , μ ¯ ( [ x ] A i ) γ , , i = 1 , 2 , , m , x i = 1 m A i P ( j = 1 n ( X j ) ) holds. Therefore, i = 1 m A i P ( j = 1 n X j ) j = 1 n ( i = 1 m A i P ( X j ) ) .
(4)
It is easy to prove by Formulas (4), (5), and (19).
 □
Theorem 4. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C and X U , A = A 1 A 2 A m and X 1 , X 2 , , X n U . The following properties hold:
(1)
i = 1 m A i P ( X ) i = 1 m A i O ( X ) A ( X ) ;
(2)
i = 1 m 1 A i O ( X ) i = 1 m A i O ( X ) and i = 1 m 1 A i P ( X ) i = 1 m A i P ( X ) .
Proof of Theorem 4. 
(1)
According to Definition 6, we only need to prove i = 1 m A i P ( X ) i = 1 m A i O ( X ) . x i = 1 m A i P ( X ) , according to Definition 7, μ ¯ ( [ x ] A i ) γ . From Definition 5, we have x i = 1 m A i O ( X ) .
(2)
It is easy to prove according to Definitions 5 and 7.
 □
Lemma 1. 
Let S = ( U , C D , V , f ) be a decision information table, A C and X U , U U A A = { E 1 , E 2 , . . . , E l } is a granularity layer on U. The following properties hold:
(1)
E i B N 1 ( X ) λ E i Y E i B N 1 ( X ) λ E i N ;
(2)
E i B N 2 ( X ) λ E i N E i B N 2 ( X ) λ E i Y . Here, E i U U A A ( i = 1 , 2 , , l ) .
Proof of Lemma 1. 
(1)
λ E i Y λ E i N = λ 12 ( 1 μ ¯ ( E i ) ) E i λ 21 μ ¯ ( E i ) E i = E i ( λ 12 ( λ 12 + λ 21 ) ) μ ¯ ( E i ) , because E i B N 1 ( X ) , we have λ 12 λ 12 + λ 21 μ ¯ ( E i ) < 1 , then λ E i Y λ E i N , therefore E i B N 1 ( X ) λ E i Y E i B N 1 ( X ) λ E i N .
(2)
λ E i Y λ E i N = λ 12 ( 1 μ ¯ ( E i ) ) E i λ 21 μ ¯ ( E i ) E i = E i ( λ 12 ( λ 12 + λ 21 ) ) μ ¯ ( E i ) . Because E i B N 2 ( X ) , we have 0 < μ ¯ ( E i ) < λ 12 λ 12 + λ 21 , then λ E i Y λ E i N . Therefore, E i B N 2 ( X ) λ E i N E i B N 2 ( X ) λ E i Y .
 □
Lemma 1 shows that the misclassification costs incurred by the equivalence classes in characterizing X are not more than the misclassification costs incurred by the equivalence classes when they do not characterize X in B N 1 ( X ) . Moreover, misclassification costs incurred by the equivalence classes when they do not characterize X are not more than the misclassification costs incurred by the equivalence classes in characterizing X in B N 2 ( X ) .
Theorem 5. 
Let S = ( U , C D , V , f ) be a decision information table, A C and X U , the following properties hold:
(1)
x B N 1 i = 1 m A i O ( X ) λ Y x B N 1 i = 1 m A i O ( X ) λ N and x B N 1 i = 1 m A i P ( X ) λ Y x B N 1 i = 1 m A i P ( X ) λ N ;
(2)
x B N 2 i = 1 m A i O ( X ) λ N x B N 2 i = 1 m A i O ( X ) λ Y and x B N 1 i = 1 m A i P ( X ) λ N x B N 1 i = 1 m A i P ( X ) λ Y .
Proof of Theorem 5. 
From Lemma 1, Theorem 5 holds. □
Theorem 5 shows that misclassification costs incurred by the equivalence classes in characterizing X are not more than misclassification costs incurred by the equivalence classes when they do not characterize X in B N 1 i = 1 m A i O ( X ) and B N 1 i = 1 m R i P ( X ) . Moreover, misclassification costs incurred by the equivalence classes when they do not characterize X are not more than misclassification costs incurred by the equivalence classes in characterizing X in B N 2 i = 1 m A i O ( X ) and B N 2 i = 1 m A i P ( X ) .
D C ( i = 1 m A i O ( X ) ) , D C ( i = 1 m A i O ̲ ( X ) ) and D C ( i = 1 m A i O ¯ ( X ) ) denote the misclassification costs generated when i = 1 m A i O ( X ) , i = 1 m A i O ̲ ( X ) and i = 1 m A i O ¯ ( X ) are approximated to X, respectively. Then, the following theorem holds:
Theorem 6. 
Let S = ( U , C D , V , f ) be a decision information table, A C and X U , Then, D C ( i = 1 m A i O ( X ) ) D C ( i = 1 m A i O ̲ ( X ) ) and D C ( i = 1 m A i O ( X ) ) D C ( i = 1 m A i O ¯ ( X ) ) .
Proof of Theorem 6. 
When i = 1 m A i O ̲ ( X ) is taken as the approximation of X, D C ( i = 1 m A i O ̲ ( X ) ) = x B N ( X ) λ N ; when i = 1 m A i O ¯ ( X ) is considered the approximation of X, D C ( i = 1 m A i O ¯ ( X ) ) = x B N ( X ) λ Y ; when D C ( i = 1 m A i O ( X ) ) is considered the approximation of X, D C ( i = 1 m A i O ( X ) ) = x B N 1 ( X ) λ Y + x B N 2 ( X ) λ N .
Because B N ( X ) = B N 1 ( X ) B N 2 ( X ) , we have:
D C ( i = 1 m A i O ̲ ( X ) ) = x B N 1 ( X ) λ N + x B N 2 ( X ) λ N , D C ( i = 1 m A i O ¯ ( X ) ) = x B N 1 ( X ) λ Y + x B N 2 ( X ) λ Y .
Therefore, according to Theorem 5,
D C ( i = 1 m A i O ( X ) ) D C ( i = 1 m A i O ̲ ( X ) ) , D C ( i = 1 m A i O ( X ) ) D C ( i = 1 m A i O ¯ ( X ) ) .
Theorem 6 indicates that when D C ( i = 1 m A i O ( X ) ) , D C ( i = 1 m A i O ̲ ( X ) ) and D C ( i = 1 m A i O ¯ ( X ) ) are used as approximations of X, respectively, D C ( i = 1 m A i O ( X ) ) generates the least misclassification costs. □
D C ( i = 1 m A i P ( X ) ) , D C ( i = 1 m A i P ̲ ( X ) ) and D C ( i = 1 m A i P ¯ ( X ) ) denote the misclassification costs generated when i = 1 m A i P ( X ) , i = 1 m A i P ̲ ( X ) and i = 1 m A i P ¯ ( X ) are approximated to X, respectively.
Theorem 7. 
Let S = ( U , C D , V , f ) be a decision information table, A C and X U . Then, D C ( i = 1 m A i P ( X ) ) D C ( i = 1 m A i P ̲ ( X ) ) and D C ( i = 1 m A i P ( X ) ) D C ( i = 1 m A i P ¯ ( X ) ) .
Proof of Theorem 7. 
Similar to Theorem 6, Theorem 7 is easy to prove. □
From Theorem 7, when D C ( i = 1 m A i P ( X ) ) , D C ( i = 1 m A i P ̲ ( X ) ) and D C ( i = 1 m A i P ¯ ( X ) ) are used as approximations of X, respectively, D C ( i = 1 m A i P ( X ) ) generates the least misclassification costs. Theorems 6 and 7 reflect the advantages of the multigranulation approximation sets that are used for approximating the target concept.

5.2. The Optimal Multigranulation Approximation Selection Method

The objects in the boundary region may be reclassified under different granularities. As a result, the equivalence classes used to represent the approximation set will be changed in the boundary region. In practical applications, the optimal approximation selection should consider both the misclassification and test costs. In MGRS, uncertain concepts characterized in a finer approximation layer result in lower misclassification costs, and test costs increase with the added attributes. Therefore, it is essential to find a balance between misclassification and test costs.
Lemma 2. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C , A 1 A 2 , , A m , and X U , then x U , μ ¯ i = 1 m 1 A i O ( x ) μ ¯ i = 1 m A i O ( x ) .
Lemma 3. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C , A 1 A 2 , , A m , and X U , then x U , μ ¯ i = 1 m 1 A i P ( x ) μ ¯ i = 1 m A i P ( x ) .
Theorem 8. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C , A 1 A 2 , , A m and X U , only when μ ¯ ( x ) ( x B N 1 i = 1 m 1 A i O ( X ) ) changed with the attribute increased, then D C ( i = 1 m 1 A i O ( X ) ) D C ( i = 1 m A i O ( X ) ) .
Proof of Theorem 8. 
D C ( i = 1 m A i O ( X ) ) = x B N 1 i = 1 m A i O ( X ) λ Y + x B N 2 i = 1 m A i O ( X ) λ N = x B N 1 i = 1 m A i O ( X ) λ 12 ( 1 μ ¯ ( x ) ) + x B N 2 i = 1 m A i O ( X ) λ 21 μ ¯ ( x ) .
According to Lemma 2, we have
μ ¯ i = 1 m 1 A i O ( x ) μ ¯ i = 1 m A i O ( x ) .
Obviously, D C ( i = 1 m 1 A i O ( X ) ) D C ( i = 1 m A i O ( X ) ) . □
From Theorem 7, for optimistic MGRS, to reduce the misclassification costs of the approximation, we can add the attribute that only changes the membership of objects in x B N 1 i = 1 m 1 A i O ( X ) .
Theorem 9. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C and A 1 A 2 , , A m , X U , only when μ ¯ ( x ) ( x B N 2 i = 1 m 1 A i P ( X ) ) changed with the attributes increased, then D C ( i = 1 m 1 A i P ( X ) ) D C ( i = 1 m A i P ( X ) ) .
Proof of Theorem 9. 
D C ( i = 1 m A i P ( X ) ) = x B N 1 i = 1 m A i P ( X ) λ Y + x B N 2 i = 1 m A i P ( X ) λ N = x B N 1 i = 1 m A i P ( X ) λ 12 ( 1 μ ¯ ( x ) ) + x B N 2 i = 1 m A i P ( X ) λ 21 μ ¯ ( x ) .
According to Lemma 3, we have
μ ¯ i = 1 m 1 A i P ( x ) μ ¯ i = 1 m A i P ( x ) .
Obviously, D C ( i = 1 m 1 A i P ( X ) ) D C ( i = 1 m A i P ( X ) ) . □
From Theorem 9, for pessimistic MGRS, to reduce the misclassification costs of the approximation, we can add the attribute that only changes the membership of objects in x B N 2 i = 1 m 1 R i O ( X ) .
In practical applications, on the one hand, the factors included in the test cost, such as money, time, environment, etc., are hard to evaluate objectively. On the other hand, these factors are hard to be integrated because of their different dimensions. In this section, we will evaluate test costs in an attribute-driven form, which are more objective.
Definition 10. 
Let S = ( U , C D , V , f ) be a decision information table, a C and X U , the significance of a is defined as follows:
S i g ( a , C , D ) = D C C { a } D C C .
Definition 11. 
Let S = ( U , C D , V , f ) be a decision information table, a C , A C and X U ; the test cost to construct A ( X ) is defined as follows:
T C A = a A S i g ( a , C , D ) .
In this paper, for simplicity, to present the optimal granularity selection of the multigranulation approximation, we only use the optimistic MGRS as an example.
Definition 12. 
Let S = ( U , C D , V , f ) be a decision information table, A 1 , A 2 , , A m C and X U ; the test cost to construct i = 1 m A i O ( X ) can be defined as follows:
T C i = 1 m A i O = i = 1 m T C A i .
In this paper, the misclassification and test costs for user requirements are represented as D C u and T C u , respectively. A multigranulation approximation i = 1 k A i O ( X ) is selected to meet the constraints D C i = 1 k A i O ( X ) D C u and T C i = 1 k A i O ( X ) T C u , then the related decision are made on i = 1 k A i O ( X ) . Figure 3 presents the optimal multigranulation approximation selection of optimistic MGRS. Herein, i = 1 3 A i O ( X ) complies with the requirements of the misclassification costs and fails to comply with the requirements of the test costs; A 1 O ( X ) complies with the requirements of the test cost and fails to comply with the requirements of misclassification costs; i = 1 2 A i O ( X ) complies with both requirements of misclassification costs and test costs, enabling effective calculations according to granularity optimization. Similarly, the optimal approximation selection of pessimistic MGRS is the same. We formalize the computation as an optimization problem:
arg min k C o s t i = 1 k A i O ( X ) ,
s . t . ξ D C i = 1 k A i O ( X ) D C u ; T C i = 1 k A i O ( X ) T C u .
where C o s t i = 1 k A i O ( X ) = ξ D C i = 1 k A i O + T C i = 1 k A i O , and C o s t i = 1 k A i O ( X ) denotes the total cost for constructing i = 1 k A i O ( X ) . ξ = U i = 1 m A i O ( X ) 1 D C ( A m ( X ) ) reflects the contribution degree of the multigranulation approximation layer for the misclassification costs of i = 1 k A i O ( X ) .

5.3. Case Study

Table 2 is an evaluation form of the company venture capital given by five experts. U = { x 1 , x 2 , , x 900 } is 900 investment plans, which are evaluated by 5 experts. The risk level is divided into three categories, including high and low. Suppose that D C u = 5 and T C u = 2.5 .
(1)
According to the above conditions, the attribute significance can be computed by Formula (25), which is shown in Table 3.
(2)
Each multigranulation approximation is obtained by adding attributes in ascending order of attribute significance, E 4 E 3 E 2 E 1 E 5 in each stage, respectively. We represent the attributes as follows: A 1 = E 4 , A 2 = E 3 , A 3 = E 2 , A 4 = E 1 , and A 5 = E 5 .
(3)
For each multigranulation approximation layer, D C i = 1 k A i O , T C i = 1 k A i O , and C o s t i = 1 k A i O ( X ) are computed by Formulas (16), (24), and (25), respectively, where k = 1 , 2 , , 5 and the results are displayed in Table 4.
C o s t i = 1 k A i O ( X ) changes with the increased attributes and only C o s t i = 1 3 A i O ( X ) and C o s t i = 1 4 A i O ( X ) satisfy D C i = 1 k A i O ( X ) D C u and T C i = 1 k A i O ( X ) T C u at the same time. According to the Formula (25), we choose the multigranulation approximation layer with the lowest total cost from the above layers; its corresponding approximation layer is i = 1 3 A i O ( X ) . Therefore, i = 1 3 A i O ( X ) is the optimal multigranulation approximation used for deciding investment plans, because it possesses lower misclassification costs, i.e., from the perspective of optimistic MGRS, E 4 , E 3 , and E 2 are reasonable expert sets. The analysis of the case study shows that the proposed method can search for a reasonable approximation under the constraint conditions.

6. Simulation Experiment and Result Analysis

6.1. Simulation Experiment

In this section, the effectiveness and rationality of our model are demonstrated as shown by illustrative experiments. The computer used in the experiments was a WIN 10 operating system with 3.10-GHz CPU and 16.0 GB RAM, and the programming software was MATLAB R2022a. The capability of the proposed model was evaluated on six UCI datasets, which are shown in Table 5. In our experiments, we randomly took away some known attribute values from datasets 10–12 to create incomplete decision systems. The missing values are randomly distributed on all conditional attributes.
From Figure 4, for classical rough sets, the misclassification costs of the approximation model monotonously decrease with the granularity being finer, which complies with human cognitive habits.
In Figure 5, O D C 1 , O D C 2 , O D C 3 and O D C 4 represent x B N 1 i = 1 m A i O ( X ) λ Y , x B N 1 i = 1 m A i O ( X ) λ N , x B N 2 i = 1 m A i O ( X ) λ N and x B N 2 i = 1 m A i O ( X ) λ Y , respectively. P D C 1 , P D C 2 , P D C 3 and P D C 4 represent x B N 1 i = 1 m A i P ( X ) λ Y , x B N 1 i = 1 m A i P ( X ) λ N , x B N 2 i = 1 m A i P ( X ) λ N and x B N 1 i = 1 m A i P ( X ) λ Y , respectively. From Figure 5, under different granular layers, misclassification costs incurred by the equivalence classes in approximating X are always less than or equal to misclassification costs incurred by the equivalence classes when they do not characterize X in B N 1 i = 1 m A i O ( X ) and B N 1 i = 1 m A i P ( X ) . Moreover, misclassification costs incurred by equivalence classes when they do not characterize X are less than or equal to the misclassification costs incurred by the equivalence classes in approximating X in B N 2 i = 1 m A i O ( X ) and B N 2 i = 1 m A i P ( X ) . This is consistent with Theorem 4.
In Figure 6, the horizontal and vertical axes denote the granularity and misclassification costs, respectively. O D C _ l o w e r , O D C _ u p p e r and O D C represent D C ( i = 1 m A i O ̲ ( X ) ) , D C ( i = 1 m A i O ¯ ( X ) ) and D C ( i = 1 m A i O ( X ) ) . P D C _ l o w e r , P D C _ u p p e r and P D C represent D C ( i = 1 m A i P ̲ ( X ) ) , D C ( i = 1 m A i P ¯ ( X ) ) and D C ( i = 1 m A i P ( X ) ) , respectively, namely, the misclassification costs generated when i = 1 m A i O ̲ ( X ) , i = 1 m R i O ¯ ( X ) and i = 1 m A i O ( X ) are approximated to X and the misclassification costs generated when i = 1 m A i P ̲ ( X ) , i = 1 m A i P ¯ ( X ) and i = 1 m A i P ( X ) are approximated to X. Obviously, compared with i = 1 m A i O ̲ ( X ) and i = 1 m A i O ¯ ( X ) , the misclassification costs of i = 1 m A i O ( X ) are the least on each granularity. Similarly, compared with i = 1 m A i P ̲ ( X ) , i = 1 m A i P ¯ ( X ) , the misclassification costs of i = 1 m A i P ( X ) are the least on each granularity. This is consistent with Theorems 5 and 6.

6.2. Results and Discussions

According to the above experiments, compared with the upper/lower approximation sets, we can conclude that the multigranulation approximations have the following advantages when applied to decision-making environments:
(1)
The misclassification costs of the approximation model monotonously decrease with the granularity being finer;
(2)
In multigranulation approximations, under different granular layers, the misclassification costs incurred by the equivalence classes in approximating X are less than or equal to the misclassification costs incurred by the equivalence classes when they do not characterize X in boundary region I of optimistic and pessimistic rough sets. Moreover, the misclassification costs incurred by equivalence classes when they do not characterize X are less than or equal to the misclassification costs incurred by the equivalence classes in approximating X in boundary region II of optimistic and pessimistic rough sets;
(3)
Compared with the upper/lower approximation sets, the misclassification costs of the multigranulation approximations are the least on each granularity.

7. Conclusions

In MGRS, optimistic and pessimistic upper/lower approximation boundaries are utilized to characterize uncertain concepts. They still cannot take advantage of the known equivalence classes to establish the approximation of an uncertain concept. To handle the problem, cost-sensitive multigranulation approximations of rough sets were constructed. Furthermore, an optimization mechanism of the multigranulation approximations is proposed, which selects the optimal approximation to obtain the minimum misclassification costs under the conditions. The case study shows that the proposed algorithm is capable of searching for a rational approximation under restraints. Finally, the experiments demonstrate that the multigranulation approximations possess the least misclassification costs. In particular, our models apply to the decision-making environment where each decision-maker is independent. Moreover, our models are useful for extracting decision rules from distributive information systems and groups of intelligent agents through rough set approaches [34,36]. Figure 7 presents a diagram that summarizes the works conducted in this paper. Herein, we present the process of the cost-sensitive multigranulation approximations of rough sets; according to different granulation mechanisms, our approach can be extended to uncertainty models, i.e., vague sets, shadow sets, and neighborhood rough sets. These results will be important to contribute to the progress of the GrC theory.
Our future work will focus on the following two aspects: (1) We hope to build a more reasonable three-way decision model based on our model from optimistic and pessimistic perspectives; (2) we wish to combine the model with the cloud model theory to extend our model to construct a multigranulation approximation model with bidirectional cognitive computing. This will offer more cognitive advantages and benefits in application fields with uncertainty from multiple perspectives, i.e., image segmentation, clustering, and recommendation systems.

Author Contributions

Conceptualization, J.Y.; methodology, J.Y., J.K. and Q.L.; writing—original draft, J.Y. and J.K.; writing—review and editing, J.Y., J.K., Q.L. and Y.L.; data curation, J.Y., Q.L. and Y.L.; supervision, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Science Foundation of China (No. 6206049), Excellent Young Scientific and Technological Talents Foundation of Guizhou Province (QKH-platform talent (2021) No. 5627), the Key Cooperation Project of Chongqing Municipal Education Commission (HZ2021008), Guizhou Provincial Science and Technology Project (QKH-ZK (2021) General 332), Science and Technology Top Talent Project of Guizhou Education Department (QJJ2022(088)), Key Laboratory of Evolutionary Artificial Intelligence in Guizhou (QJJ[2022] No. 059) and the Key Talens Program in digital economy of Guizhou Province, Electronic Manufacturing Industry University Research Base of Ordinary Colleges and Universities in Guizhou Province (QJH-KY Zi (2014) No. 230-2).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This study was mainly completed at the Chongqing Key Laboratory of Computational Intelligence, Chongqing University of Posts and Telecommunications, and the authors would like to thank the laboratory for its assistance.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zadeh, L.A. Toward a theory of fuzzy information granulation and its centrality in human reasoning and fuzzy logic. Fuzzy Sets Syst. 1997, 90, 111–127. [Google Scholar] [CrossRef]
  2. Bello, M.; Nápoles, G.; Vanhoof, K.; Bello, R. Data quality measures based on granular computing for multi-label classification. Inf. Sci. 2021, 560, 51–67. [Google Scholar] [CrossRef]
  3. Pedrycz, W.; Chen, S. Interpretable Artificial Intelligence: A Perspective of Granular Computing; Springer Nature: Berlin/Heidelberg, Germany, 2021; Volume 937. [Google Scholar]
  4. Li, J.; Mei, C.; Xu, W.; Qian, Y. Concept learning via granular computing: A cognitive viewpoint. IEEE Trans. Fuzzy Syst. 2015, 298, 447–467. [Google Scholar] [CrossRef] [PubMed]
  5. Zadeh, L. Fuzzy sets. Inf. Control. 1965, 8, 338–353. [Google Scholar] [CrossRef] [Green Version]
  6. Pawlak, Z. Rough sets. Int. J. Comput. Inf. Sci. 1982, 11, 341–356. [Google Scholar] [CrossRef]
  7. Zhang, L.; Zhang, B. The quotient space theory of problem solving. In Proceedings of the International Workshop on Rough Sets, Fuzzy Sets, Data Mining, and Granular-Soft Computing, Chongqing, China, 26–29 May 2003; pp. 11–15. [Google Scholar]
  8. Li, D.Y.; Meng, H.J.; Shi, X.M. Membership clouds and membership cloud generators. J. Comput. Res. Dev. 1995, 32, 15–20. [Google Scholar]
  9. Colas-Marquez, R.; Mahfouf, M. Data Mining and Modelling of Charpy Impact Energy for Alloy Steels Using Fuzzy Rough Sets. IFAC-Pap. 2017, 50, 14970–14975. [Google Scholar] [CrossRef]
  10. Hasegawa, K.; Koyama, M.; Arakawa, M.; Funatsu, K. Application of data mining to quantitative structure-activity relationship using rough set theory. Chemom. Intell. Lab. Syst. 2009, 99, 66–70. [Google Scholar] [CrossRef]
  11. Santra, D.; Basu, S.K.; Mandal, J.K.; Goswami, S. Rough set based lattice structure for knowledge representation in medical expert systems: Low back pain management case study. Expert Syst. Appl. 2020, 145, 113084. [Google Scholar] [CrossRef] [Green Version]
  12. Chebrolu, S.; Sanjeevi, S.G. Attribute Reduction in Decision-Theoretic Rough Set Model using Particle Swarm Optimization with the Threshold Parameters Determined using LMS Training Rule. Procedia Comput. Sci. 2015, 57, 527–536. [Google Scholar] [CrossRef] [Green Version]
  13. Abdolrazzagh-Nezhad, M.; Radgohar, H.; Salimian, S.N. Enhanced cultural algorithm to solve multi-objective attribute reduction based on rough set theory. Math. Comput. Simul. 2020, 170, 332–350. [Google Scholar] [CrossRef]
  14. Beaubier, S.; Defaix, C.; Albe-Slabi, S.; Aymes, A.; Galet, O.; Fournier, F.; Kapel, R. Multiobjective decision making strategy for selective albumin extraction from a rapeseed cold-pressed meal based on Rough Set approach. Food Bioprod. Process. 2022, 133, 34–44. [Google Scholar] [CrossRef]
  15. Landowski, M.; Landowska, A. Usage of the rough set theory for generating decision rules of number of traffic vehicles. Transp. Res. Procedia 2019, 39, 260–269. [Google Scholar] [CrossRef]
  16. Tawhid, M.; Ibrahim, A. Feature selection based on rough set approach, wrapper approach, and binary whale optimization algorithm. Int. J. Mach. Learn. Cybern. 2020, 11, 573–602. [Google Scholar] [CrossRef]
  17. Zhang, Q.H.; Wang, G.Y.; Yu, X. Approximation sets of rough sets. J. Softw. 2012, 23, 1745–1759. [Google Scholar] [CrossRef]
  18. Zhang, Q.H.; Wang, J.; Wang, G.Y. The approximate representation of rough-fuzzy sets. Chin. J. Comput. Jisuanji Xuebao 2015, 38, 1484–1496. [Google Scholar]
  19. Zhang, Q.; Wang, J.; Wang, G.; Yu, H. The approximation set of a vague set in rough approximation space. Inf. Sci. 2015, 300, 1–19. [Google Scholar] [CrossRef]
  20. Zhang, Q.H.; Zhang, P.; Wang, G.Y. Research on approximation set of rough set based on fuzzy similarity. J. Intell. Fuzzy Syst. 2017, 32, 2549–2562. [Google Scholar] [CrossRef]
  21. Zhang, Q.H.; Yang, J.J.; Yao, L.Y. Attribute reduction based on rough approximation set in algebra and information views. IEEE Access 2016, 4, 5399–5407. [Google Scholar] [CrossRef]
  22. Yao, L.Y.; Zhang, Q.H.; Hu, S.P.; Zhang, Q. Rough entropy for image segmentation based on approximation sets and particle swarm optimization. J. Front. Comput. Sci. Technol. 2016, 10, 699–708. [Google Scholar]
  23. Zhang, Q.H.; Liu, K.X.; Gao, M. Approximation sets of rough sets and granularity optimization algorithm based on cost-sensitive. J. Control. Decis. 2020, 35, 2070–2080. [Google Scholar]
  24. Yang, J.; Yuan, L.; Luo, T. Approximation set of rough fuzzy set based on misclassification cost. J. Chongqing Univ. Posts Telecommun. (Nat. Sci. Ed.) 2021, 33, 780–791. [Google Scholar]
  25. Yang, J.; Luo, T.; Zeng, L.J.; Jin, X. The cost-sensitive approximation of neighborhood rough sets and granular layer selection. J. Intell. Fuzzy Syst. 2022, 42, 3993–4003. [Google Scholar] [CrossRef]
  26. Siminski, K. 3WDNFS—Three-way decision neuro-fuzzy system for classification. Fuzzy Sets Syst. 2022, in press. [Google Scholar] [CrossRef]
  27. Subhashini, L.; Li, Y.; Zhang, J.; Atukorale, A.S. Assessing the effectiveness of a three-way decision-making framework with multiple features in simulating human judgement of opinion classification. Inf. Process. Manag. 2022, 59, 102823. [Google Scholar] [CrossRef]
  28. Subhashini, L.; Li, Y.; Zhang, J.; Atukorale, A.S. Integration of semantic patterns and fuzzy concepts to reduce the boundary region in three-way decision-making. Inf. Sci. 2022, 595, 257–277. [Google Scholar] [CrossRef]
  29. Mondal, A.; Roy, S.K.; Pamucar, D. Regret-based three-way decision making with possibility dominance and SPA theory in incomplete information system. Expert Syst. Appl. 2023, 211, 118688. [Google Scholar] [CrossRef]
  30. Yao, Y.Y. Symbols-Meaning-Value (SMV) space as a basis for a conceptual model of data science. Int. J. Approx. Reason. 2022, 144, 113–128. [Google Scholar] [CrossRef]
  31. Qian, Y.H.; Liang, J.Y.; Dang, C.Y. Incomplete multigranulation rough set. IEEE Trans. Syst. Man-Cybern.-Part Syst. Humans 2009, 40, 420–431. [Google Scholar] [CrossRef]
  32. Huang, B.; Guo, C.X.; Zhuang, Y.L.; Li, H.X.; Zhou, X.Z. Intuitionistic fuzzy multigranulation rough sets. Inf. Sci. 2014, 277, 299–320. [Google Scholar] [CrossRef]
  33. Li, F.J.; Qian, Y.H.; Wang, J.T.; Liang, J. Multigranulation information fusion: A Dempster-Shafer evidence theory-based clustering ensemble method. Inf. Sci. 2017, 378, 389–409. [Google Scholar] [CrossRef]
  34. Liu, X.; Qian, Y.H.; Liang, J.Y. A rule-extraction framework under multigranulation rough sets. Int. J. Mach. Learn. Cybern. 2014, 5, 319–326. [Google Scholar] [CrossRef]
  35. Liu, K.; Li, T.; Yang, X.; Ju, H.; Yang, X.; Liu, D. Hierarchical neighborhood entropy based multi-granularity attribute reduction with application to gene prioritization. Int. J. Approx. Reason. 2022, 148, 57–67. [Google Scholar] [CrossRef]
  36. Qian, Y.H.; Liang, X.Y.; Lin, G.P.; Guo, Q.; Liang, J. Local multigranulation decision-theoretic rough sets. Int. J. Approx. Reason. 2017, 82, 119–137. [Google Scholar] [CrossRef]
  37. Qian, Y.H.; Zhang, H.; Sang, Y.L.; Liang, J. Multigranulation decision-theoretic rough sets. Int. J. Approx. Reason. 2014, 55, 225–237. [Google Scholar] [CrossRef]
  38. Xu, W.; Yuan, K.; Li, W. Dynamic updating approximations of local generalized multigranulation neighborhood rough set. Appl. Intell. 2022, 52, 9148–9173. [Google Scholar] [CrossRef]
  39. Sun, L.; Wang, L.; Ding, W.; Qian, Y.; Xu, J. Feature selection using fuzzy neighborhood entropy-based uncertainty measures for fuzzy neighborhood multigranulation rough sets. IEEE Trans. Fuzzy Syst. 2020, 29, 19–33. [Google Scholar] [CrossRef]
  40. She, Y.H.; He, X.L.; Shi, H.X.; Qian, Y. A multiple-valued logic approach for multigranulation rough set model. Int. J. Approx. Reason. 2017, 82, 270–284. [Google Scholar] [CrossRef]
  41. Li, W.; Xu, W.; Zhang, X.Y.; Zhang, J. Updating approximations with dynamic objects based on local multigranulation rough sets in ordered information systems. Artif. Intell. Rev. 2021, 55, 1821–1855. [Google Scholar] [CrossRef]
  42. Zhang, C.; Li, D.; Zhai, Y.; Yang, Y. Multigranulation rough set model in hesitant fuzzy information systems and its application in person-job fit. Int. J. Mach. Learn. Cybern. 2019, 10, 717–729. [Google Scholar] [CrossRef]
  43. Hu, C.; Zhang, L.; Wang, B.; Zhang, Z.; Li, F. Incremental updating knowledge in neighborhood multigranulation rough sets under dynamic granular structures. Knowl.-Based Syst. 2019, 163, 811–829. [Google Scholar] [CrossRef]
  44. Hu, C.; Zhang, L. Dynamic dominance-based multigranulation rough sets approaches with evolving ordered data. Int. J. Mach. Learn. Cybern. 2021, 12, 17–38. [Google Scholar] [CrossRef]
Figure 1. The approximation of rough sets (surrounded by the broken line).
Figure 1. The approximation of rough sets (surrounded by the broken line).
Electronics 11 03801 g001
Figure 2. The granules subdivided in B N 1 ( X ) and B N 2 ( X ) of the cost-sensitive approximation model of rough sets. And all the red circles in the figure represent the set X. In addition, (a) shows the case 1 in which the granules are subdivided in B N 1 ( X ) ; (b) shows the case 2 in which the granules are subdivided in B N 1 ( X ) ; (c) shows the case 1 in which the granules are subdivided in B N 2 ( X ) ; (d) shows the case 2 in which the granules are subdivided in B N 2 ( X ) .
Figure 2. The granules subdivided in B N 1 ( X ) and B N 2 ( X ) of the cost-sensitive approximation model of rough sets. And all the red circles in the figure represent the set X. In addition, (a) shows the case 1 in which the granules are subdivided in B N 1 ( X ) ; (b) shows the case 2 in which the granules are subdivided in B N 1 ( X ) ; (c) shows the case 1 in which the granules are subdivided in B N 2 ( X ) ; (d) shows the case 2 in which the granules are subdivided in B N 2 ( X ) .
Electronics 11 03801 g002
Figure 3. The optimal granularity selection of the optimistic multigranulation approximation. And the red circles in the figure represent the set X.
Figure 3. The optimal granularity selection of the optimistic multigranulation approximation. And the red circles in the figure represent the set X.
Electronics 11 03801 g003
Figure 4. The misclassification cost with the changing granularity on each dataset.
Figure 4. The misclassification cost with the changing granularity on each dataset.
Electronics 11 03801 g004
Figure 5. The misclassification cost of two boundary regions under different granularities.
Figure 5. The misclassification cost of two boundary regions under different granularities.
Electronics 11 03801 g005
Figure 6. The misclassification costs of i = 1 m A i O ̲ ( X ) , i = 1 m A i O ¯ ( X ) , i = 1 m A i O ( X ) , i = 1 m A i P ̲ ( X ) , i = 1 m A i P ¯ ( X ) , i = 1 m A i P ( X ) with the changing granularity of each dataset.
Figure 6. The misclassification costs of i = 1 m A i O ̲ ( X ) , i = 1 m A i O ¯ ( X ) , i = 1 m A i O ( X ) , i = 1 m A i P ̲ ( X ) , i = 1 m A i P ¯ ( X ) , i = 1 m A i P ( X ) with the changing granularity of each dataset.
Electronics 11 03801 g006
Figure 7. Diagram of works conducted in this paper.
Figure 7. Diagram of works conducted in this paper.
Electronics 11 03801 g007
Table 1. A decision information table.
Table 1. A decision information table.
A 1 A 2 A 3 X
x 1 0000
x 2 0010
x 3 0100
x 4 0111
x 5 1000
x 6 1011
x 7 1101
x 8 1111
Table 2. Evaluation form of the company’s venture capital.
Table 2. Evaluation form of the company’s venture capital.
Firm E 1 E 2 E 3 E 4 E 5 D
x 1 33331High
x 2 21232High
x 3 21212High
·······
·······
·······
x 898 33223Low
x 899 31331Low
x 900 11331Low
Table 3. Description of attribute significance.
Table 3. Description of attribute significance.
Attribute E 1 E 2 E 3 E 4 E 5
S i g ( a , C , D ) 0.740.560.540.360.77
Table 4. The description of cost in each multigranulation approximation.
Table 4. The description of cost in each multigranulation approximation.
A 1 O ( X ) i = 1 2 A i O ( X ) i = 1 3 A i O ( X ) i = 1 4 A i O ( X ) i = 1 5 A i O ( X )
ξ D C 8.35.94.73.63.5
T C 0.360.91.462.22.97
C o s t 8.666.86.165.86.47
Table 5. The description of datasets.
Table 5. The description of datasets.
IDDatasetAttribute CharacteristicsInstancesCondition Attributes
1BankInteger3912
2Breast-CancerInteger6999
3CarInteger17286
4ENB2012dataReal7688
5MushroomInteger812422
6TicInteger9589
7Air QualityReal935812
8ConcreteReal10308
9HcvReal56910
10WisconsinReal6999
11ZooInteger10116
12BalanceInteger6254
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, J.; Kuang, J.; Liu, Q.; Liu, Y. Cost-Sensitive Multigranulation Approximation in Decision-Making Applications. Electronics 2022, 11, 3801. https://doi.org/10.3390/electronics11223801

AMA Style

Yang J, Kuang J, Liu Q, Liu Y. Cost-Sensitive Multigranulation Approximation in Decision-Making Applications. Electronics. 2022; 11(22):3801. https://doi.org/10.3390/electronics11223801

Chicago/Turabian Style

Yang, Jie, Juncheng Kuang, Qun Liu, and Yanmin Liu. 2022. "Cost-Sensitive Multigranulation Approximation in Decision-Making Applications" Electronics 11, no. 22: 3801. https://doi.org/10.3390/electronics11223801

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop