1. Introduction
Rough set is a powerful tool for data analysis and knowledge discovery. Initially introduced by the Polish scholar Pawlak [
1] in 1982. Since its inception, rough set theory has undergone substantial advancements in both theoretical extensions and practical applications. For instance, Stepaniuk and Skowron [
2] explored a three-way rough set-based approach for decision granule approximation and developed a novel approximation space to optimize the characterization of compound decision granules. Janusz and Ślęzak [
3] proposed rough set methods for attribute clustering and selection in which greedy heuristics are incorporated to reduce computation times and generate multiple decision reducts. Inspired by the concepts of lower and upper approximations in rough set theory, Siminski and Wakulicz-Deja [
4] proposed an extended forward inference algorithm. This method enhances classical forward inference by enabling the process to continue after an inference failure through the interactive supplementation of missing facts. Rough set is established upon the indiscernibility relations that exist within the universe of discourse. However, the indiscernibility relations may sometimes restrict the practical applications of rough set [
5]. To overcome this limitation, Zakowsk [
6] proposed covering-based rough sets by utilizing cover relations instead of indiscernibility relations. To address issues of the minimal and maximal descriptions of covering-based rough sets, Wang et al. [
7] devised matrix approaches that facilitate these descriptions and introduced two types of covering information reductions under these descriptions. Zhu [
8] further enriched the field by identifying four distinct types of covering-based rough sets and investigated their topological properties, which has been instrumental in understanding the connections among different covering-based rough sets. Moreover, Zhang et al. [
9] built the equivalences between the four types of covering-based rough sets and traditional rough set, and also presented sufficient and necessary conditions for these equivalences. Yao et al. [
10] presented a framework for approximating covering-based rough sets by constructing four distinct neighborhoods, six novel coverings, and two sub-systems. They also proposed approximation operators based on elements, granules, and subsystems relevant to covering-based rough sets and discussed the relationships between these operators and existing ones. Janicki et al. [
11] defined optimal approximation in the context of similarity measures and provided an algorithm that employs the Jaccard Index as an optimal approximation operator.
Formal Concept Analysis (FCA), introduced by Wille in 1982, has evolved into a valuable method utilized in data mining, information retrieval, knowledge discovery, and various other fields [
12,
13,
14,
15,
16,
17]. Some scholars have expanded FCA by introducing multiple types of formal concepts, which significantly enriches the original FCA. In 2014, Qi et al. [
18] established three-way operators and their inverses and introduced two distinct types of three-way concepts: the object-induced concept and the attribute-induced concept. Furthermore, they integrated these concepts within the framework of three-way decision to construct three-way concept lattices. Zhang et al. [
19] proposed a novel algorithm, termed 3WOC, which can directly and quickly generate all OEO concepts with lower time complexity. The introduction of the three-way concept lattice has since prompted widespread research interests across multiple scholarly fields including lattice construction [
20,
21,
22,
23,
24], lattice reduction [
25,
26,
27], rule acquisition derived from lattice analysis [
28,
29,
30], and conceptual knowledge [
31,
32,
33].
Some researchers have incorporated rough sets into formal concept analysis to investigate the uncertainty present in FCA [
34,
35,
36]. Yao et al. [
37] utilized approximation operators to represent and elucidate classical concept lattices and studied three distinct types of concept lattices: object-oriented, attribute-oriented, and complement-oriented concept lattices. Furthermore, Yao [
34] developed a set of operators drawing from both lattice theory and set theory and investigated the interconnections between these operators. Building on these foundations, Monhanty [
38] generalized approximation operators on the formal concepts and represented undefinable sets by employing two definable sets in a concept context. Mao et al. [
39] proposed the variable precision rough set approximations within concept lattices informed by FCA, thereby facilitating the approximation of undefinable sets and analyzing trends in the changes of the
-upper and lower approximations. Recently, some researchers have traditionally constructed a complete concept lattice before employing approximation operators to approximately characterize conceptual knowledge. However, this type of approach overlooks the fact that constructing a concept lattice is an NP-hard problem. Existing approximation methods require the computation of all concepts, which results in significant computational time, especially when dealing with massive datasets. If approximate characterization can be achieved without building the entire concept lattice, the computational cost of mining conceptual knowledge would be significantly reduced. Therefore, lowering the time complexity of the computational process is a meaningful research task. This study aims to optimize the computational process of approximation methods by improving knowledge representation, thereby decreasing computational time. This improvement lays the foundation for a more efficient and concise knowledge representation in subsequent three-way concept lattices.
Building on the foundational work reference [
39], this paper introduces an innovative approach for developing approximation operators within the framework of three-way concept lattices. The primary contribution of this work is to extend the framework of the object-induced three-way concept. The organization of the rest of this paper is outlined below.
Section 2 provides an overview of the foundational concepts of formal concept analysis.
Section 3 and
Section 4 elaborate on the extended framework of the OE concept:
Section 3 defines the concept of maximal description and introduces the lower and upper approximation OE concepts;
Section 4 proposes an optimization algorithm for approximation operators and analyzes its time complexity.
Section 5 provides experimental validation of the proposed algorithm. Finally, conclusions are drawn in
Section 6.
2. Preliminaries
This section will provide an overview of essential definitions and properties involved in this paper.
Definition 1 ([
40]
). Let L be a lattice. For , if , , and implies that or , then is meet-irreducible. Definition 2 ([
41]
). A formal context is composed of two sets, P and Q, along with a relation R that links elements of P to those of Q. The elements of P are referred to as the objects, while the elements of Q are known as the attributes. To indicate that an object is related to an attribute , we use the notation or .
For an object , we denote the object intent simply as instead of . Accordingly, we define to represent the attribute extent of the attribute q.
Let and . The following defines the pair of operators used in formal concept analysis: A formal concept is defined as a pair where , and it satisfies and . The notation represents the set of all concepts associated with the context K.
We denote the object concept as and the attribute concept as .
Definition 3 ([
18]
). Let be a formal context with and . The pair of negative operators is defined as follows:Here, . Definition 4 ([
18]
). Let be a formal context and . The three-way operators are defined as follows: Dually, for and , we define their inverses by Proposition 1 ([
18]
). Let be a formal context. For , and , the following properties hold: - (1)
- (2)
,
- (3)
,
- (4)
,
- (5)
,
- (6)
,
- (7)
,
- (8)
,
- (9)
,
- (10)
,
- (11)
.
Definition 5 ([
18]
). Let be a formal context. The pair is referred to as an object-induced three-way concept (abbreviated as an concept) if and only if and . The set Z is termed the extension of the concept (shortened to ), and the set of all concepts is known as the object-induced three-way concept lattice of , shortened to . We denote the meet-irreducible elements in by and for the extension of the meet-irreducible elements .
We denote the attribute OE concept as or . If , then or . That is, the meet-irreducible element in must be the attribute OE concept.
Definition 6 ([
39]
). Let be a concept lattice. For and , is referred to as the lower approximation concept, while is designated as the upper approximation concept, where the lower approximation of Y is defined asand the upper approximation of Y is defined asHere, . 3. Three-Way Approximations
Definition 7. Let be a formal context. For any , the definition of the maximal description of y is given by is referred to as the minimal neighborhood of the maximal description of y, denoted as . To clarify Definition 7, we illustrate it through Example 1.
Example 1. Table 1 presents a formal context K and Figure 1 depicts the details of its . By Definition 1 and Definition 5, we can identify the following meet-irreducible elements in : The extension of the concept in FCA can be considered analogous to the definable set in a rough set, but they are not precisely equivalent [
38]. When a given set of objects serves as the extension of an OE concept, it is deemed definable. Conversely, set that does not correspond to any OE concept’s extension being classified as undefinable. In order to describe an undefinable set of objects
by using the extensions in the OE concept lattice, we introduce formal definitions of the lower and upper approximation OE concepts, which are elaborated upon below:
Definition 8. Let be a formal context. For a set of objects and , the lower approximation of Y is defined byand the upper approximation of Y is defined bywhere . We call and the lower approximation concept and upper approximation concept, respectively.
Example 2 (building upon Example 1). Let and .
The maximal description is as follows: The minimal neighborhood of the maximal description is given as follows:and Example 2 illustrates a specific scenario in which the statement “If , then ” holds. However, the general applicability of this statement may be questionable. In order to clarify this, Example 3 shows that the aforementioned statement is not universally valid.
Example 3 (building upon Example 2). Let .
The lower and upper approximation concepts are presented below:andThus, Theorem 1. For 0.5 1 and , the following properties are valid:
- (1)
,
- (2)
,
- (3)
,
- (4)
,
- (5)
,
- (6)
If ,
- (7)
If ,
- (8)
,
- (9)
,
- (10)
,
- (11)
.
Proof. (1) Since
and
, then
Hence, is proven.
(2) Similarly to (1), can be proven.
(3) Since
and
, then
Hence, is proven.
(4) Similarly to (3), we can prove that .
(5) By Definition 8, holds.
(6) Since , then By Definition 8, is proven.
(7) Similarly to (6), can be proven.
(8) Since
and
, then
We then obtain Therefore, can be proven.
(9) Similarly to (8), can be proven.
(10) According to (6), since ), then and . Therefore, can be proven.
(11) The same as (10), can be proven. □
Subsequently, we discuss further properties of varying thresholds for the lower and upper approximations.
Theorem 2. Let be a formal context. For and , the following properties are true:
- (1)
,
- (2)
.
Proof. (1) According to Definition 8, since , then and . Therefore, is proven.
(2) Since , we then obtain ; then, and . Therefore, is proven. □
Theorem 3. Let be a formal context. For and , thenwhere denotes the complement of Y. Proof. For
, it holds that
- (1)
If , then , which is equivalent to . Hence, (1) can be proven.
- (2)
Similarly, if , then , which is equivalent to . Thus (2) can be proved. □
According to Definition 8, the establishment of the concept model introduced in this paper depends on the parameter . As a special case, we will examine the unique characteristics of the proposed model when .
Theorem 4. Let be an concept lattice. For a subset of objects , if , then
- (1)
,
- (2)
.
Proof. (1) For
, since
and
, we get
; thus,
. By Definition 8, then
(2) Similarly, can be proven. □
Assume that ; then, and hold, which is consistent with Theorem 2.
Theorem 5. Let be an concept lattice. For , then .
Proof. Let
,
. By Theorem 4, then
Then,
. By Proposition 1(3) and Proposition 1(4),
.
According to Propositon 1(1), it can easily seen that ; then, ,. Furthermore, by Proposition 1(3) and Proposition 1(5), , . Therefore, . □
Theorem 6. Let be an concept lattice. For , then .
Proof. Let
. By Theorem 4, we obtain
Then,
. By Proposition 1(3) and Proposition 1(4), we have
. Therefore,
. □
4. Approximating Undefinable Sets in a Three-Way Concept
In this section, we present an optimization algorithm (three-way approximation optimization algorithm, TAO) to approximate undefinable sets.
4.1. Proposed Algorithm
The purpose of the TAO algorithm is to identify the lower and upper approximations for an undefinable set of objects. Initially, a maximal description is established on the basis of meet-irreducible elements to create its minimal neighborhood set. Afterwards, the covering-based rough set is employed to derive the the lower and upper approximations.
The TAO algorithm is made up of three parts: Algorithm 1, which is tasked to find the minimal neighborhood of the maximal description; Algorithm 2, which computes the lower approximation; and Algorithm 3, which calculates the upper approximation.
Reducing redundant information is a key objective of the TAO algorithm. To gauge the effectiveness of the TAO algorithm in handling redundant information, we employ the redundancy ratio and reduction ratio for quantitative measurement.
Definition 9 (redundancy ratio)
. where, the upper approximation redundancy ratio, measures the diversity between the upper approximation and the undefinable set Y. , the lower approximation redundancy ratio, measures the difference between the lower approximation and the undefinable set Y. Moreover, a higher indicates stronger data redundancy, which may degrade the algorithm’s performance. Additionally, when , depends entirely on , and when , depends entirely on .
Definition 10 The reduction ratio () serves to evaluate the comparative efficacy of concept reduction between two algorithms. Here, denotes the count of concepts generated by Algorithm 1 and represents the count of concepts generated by Algorithm 2. Mathematically, the value of lies within the interval . A higher reduction ratio implies that Algorithm 1 outperforms Algorithm 2 in generating fewer concepts. Specifically, when , this indicates that Algorithm 1 exhibits no reduction effect compared to Algorithm 2, resulting in . Conversely, as approaches 1, this signifies that Algorithm 1 achieves a substantial reduction in the number of concepts, highlighting its superior concept reduction capability.
Algorithm 1 Determining the minimal neighborhood of the maximal description |
Input: the formal context Output: - 1:
while do - 2:
according to Definition 5, find out the set of extension of meet-irreducible elements , denote it as . - 3:
end while - 4:
while do - 5:
if , then - 6:
for do - 7:
for do - 8:
Compute the description by Equation ( 1). - 9:
end for - 10:
end for - 11:
end if - 12:
end while - 13:
Generate the minimal neighborhood of the maximal description . - 14:
return
|
Algorithm 2 Computing the lower approximation |
Input: , threshold , Output: .
- 1:
. - 2:
if then - 3:
if then - 4:
- 5:
else - 6:
- 7:
end if - 8:
else - 9:
- 10:
end if - 11:
return
|
Algorithm 3 Calculating the upper approximation |
Input: , ,threshold , Output: - 1:
. - 2:
if then - 3:
if then - 4:
- 5:
else - 6:
- 7:
end if - 8:
else - 9:
- 10:
end if - 11:
return
|
4.2. Algorithm Complexity Analysis
Let
be a formal context. While meet-irreducible elements are classified as attribute concepts, not all attribute concepts are necessarily meet-irreducible elements [
40]; therefore,
. Since the statement that object OE concepts are indeed OE concepts holds true,
.
The primary operations of Procedure 1 are divided into two components: finding and generating . The time complexity for the first component is and the time complexity for the second component is . Therefore, the time complexity of the whole of Procedure 1 is . The primary operations of Procedure 2 are dedicated to describing the lower approximation, with the overall time complexity of Procedure 2 being . Likewise, the complexity analysis of Procedure 3 parallels that of Procedure 2 and is not detailed here; thus, the time complexity of Procedure 3 also remains as . As a result, the overall time complexity of the entire algorithm is .
The method in [
39] demonstrates effective results in computing lower and upper approximations in classical concepts and can also be expanded to the three-way concept. The time complexity of the method [
39] is composed of two parts: constructing all OE concepts and computing lower and upper approximations. This method involves creating
and subsequently performing a one-by-one comparison for all OE concepts. The set of all OE concepts, denoted as
, is calculated by reference [
22], with the time complexity of constructing
represented as
. The second part could require, at most,
of time. Consequently, the overall time complexity of the method in [
39] is
.
The TAO algorithm offers the advantage of efficiently calculating the lower and upper approximations for large-scale formal contexts. Therefore, in such contexts, it is usually observed that
. As a result, the TAO algorithm exhibits lower complexity compared to the method in [
39].