SFCA: A Scalable Formal Concepts Driven Architecture for Multi-Field Knowledge Graph Completion

: With the proliferation of Knowledge Graphs (KGs), knowledge graph completion (KGC) has attracted much attention. Previous KGC methods focus on extracting shallow structural information from KGs or in combination with external knowledge, especially in commonsense concepts (generally, commonsense concepts refer to the basic concepts in related ﬁelds that are required for various tasks and academic research, for example, in the general domain, “Country” can be considered as a commonsense concept owned by “China”), to predict missing links. However, the technology of extracting commonsense concepts from the limited database is immature, and the scarce commonsense database is also bound to speciﬁc verticals (commonsense concepts vary greatly across verticals, verticals refer to a small ﬁeld subdivided vertically under a large ﬁeld). Furthermore, most existing KGC models reﬁne performance on public KGs, leading to inapplicability to actual KGs. To address these limitations, we proposed a novel Scalable Formal Concept-driven Architecture (SFCA) to automatically encode factual triples into formal concepts as a superior structural feature, to support rich information to KGE. Speciﬁcally, we generate dense formal concepts ﬁrst, then yield a handful of entity-related formal concepts by sampling and delimiting the appropriate candidate entity range via the ﬁltered formal concepts to improve the inference of KGC. Compared with commonsense concepts, KGC beneﬁts from more valuable information from the formal concepts, and our self-supervision extraction method can be applied to any KGs. Comprehensive experiments on ﬁve public datasets demonstrate the effectiveness and scalability of SFCA. Besides, the proposed architecture also achieves the SOTA performance on the industry dataset. This method provides a new idea in the promotion and application of knowledge graphs in AI downstream tasks in general and industrial ﬁelds.


Introduction
Knowledge Graphs(KGs), as a structured representation of interlinked descriptions of concepts, entities, relations and events, provide effective support for question answering [1], recommendation systems [2,3], information retrieval [4,5], and natural language processing [6]. By analyzing the public KGs (Freebase [7], YAGO [8], and DBPedia [9]), the incompleteness of KGs is the inevitable problem limited by existing KGs construction technology, requiring KGC to infer new facts. Among all available Knowledge Graph Completion (KGC) research, the knowledge graph embedding (KGE) model shows efficiency and significant performance, which embeds KGs components (entities and relations) into a latent space to learn topological structure information.

Limitations
Judging by the results of recent methods, combining external information performed better than the structure information-driven method. Among various external information, commonsense concepts are recognized as more appropriate and effective by many researchers in enhancing the KGE model. However, there are still several challenges to catching commonsense concepts from data or a knowledge base.
(1) For the DBpedia-related KGs, only some famous KGs have their commonsense knowledge base, and the number of including commonsense concepts is not large. In general, the commonsense concepts from the DBpedia KGs are hard to share with other verticals KGs because commonsense concepts are mainly appropriate for the corresponding KGs (The greater the difference between common sense concepts in the more specialized fields, the more difficult it is to reuse). (2) For the specific KGs, such as industrial KGs, do not even have a commonsense knowledge base. Meanwhile, the commonsense concepts of specific KGs are hard to collect since they are commonly defined by corresponding researchers or experts. For the same reason, the previous models or algorithms [23,24] performed poorly on automatic commonsense extraction (The more specialized domains rely more on human-defined commonsense concepts).

Motivation
After analyzing the dilemmas involved in fetching external information, we meet three puzzles. What kind of concept can be defined as commonsense concepts? Which part of the commonsense concept provides valuable information in the KGC model? Is it possible to refine a concept from data to instead commonsense concept? The essence of the commonsense concept is summarized in many instances by human cognition. Thus, we try to explore new metaphysical 'concepts' from the latent space of KGs. Inspired by data mining theories, we argue that formal concepts with lattice structures are similar to ontologies with tree structures and can guide instance knowledge in KGC tasks. Thus, we use formal concepts to represent a concept, including entities and relations subset of KG. Figure 1a shows a schematic diagram of the ontology and concept lattice structure.
In light of the definition of formal concept analysis in data mining, a formal concept is an idea or category defined by a concrete or specific set of rules, guidelines, or properties. Extending to KGs, formal concepts can be treated as an ensemble of two sets. One is its extension to denote entity sets. The other is its intention to represent latent relations of entities. For the first set, our identification of instance membership in formal concepts relies on the formal concept's instance set. By comparing the source instances and instances in the formal concept instances collection to judge whether this instance belongs to the formal concept. As shown in Figure 1b, In the lower part of the figure, diamonds represent formal concepts, and the squares and the rectangles to the right of the diamond, respectively, represent the objects and properties that make up the formal concept. The source of formal concepts is shown in the upper part of the figure. Specifically, huskies can be recognized as an instance in the formal concept of the "Dog" and an instance in the formal concept of the "Sled Dog" since it is contained in the instances collection of "Dog" and "Sled Dog" in formal concepts (The three names of "Dog", "Herding Dog" and "Sled Dog" are matched by ourselves according to the instance set an attribute set of formal concept, formal concepts themselves have no names, but the individual elements of instance sets and property sets have their names). For the second set, formal concepts involve potential information of instances concepts, including possible properties (According to the different attributes of the instance, we can find the different formal concepts corresponding to the instance that are more focused on a certain attribute in meaning). For example, Border Collies, in this instance, hide the attribute of "Hunting" and the attribute of "Herding", which are respectively contained in the formal concept of "Dog" and the formal concept of "Herding Dog". possible properties (According to the different attributes of the instance, we can find the different formal concepts corresponding to the instance that are more focused on a certain attribute in meaning). For example, Border Collies, in this instance, hide the attribute of "Hunting" and the attribute of "Herding", which are respectively contained in the formal concept of "Dog" and the formal concept of "Herding Dog". This sparked our interest in exploring the role of formal concepts in KGC research. Formal concepts can be found from the binary relations between known instances and attributes and exactly corresponds to the entity and relation in KG. For example, in a triplet SPO, P can be regarded as an attribute of S, and S can be regarded as an instance with the P property, from which we can naturally mine the formal concept in KG. In this work, we apply formal concepts, such as metaphysical structural information, to enhance the KGE model for the KGC task. Compared with commonsense concepts, formal concepts have the following advantages: (1) Formal concepts can be efficient and automatically generated, while commonsense concepts require expensive manual annotation. The formal concept is derived from the KG itself and belongs to the information of the KGs. On the contrary, the commonsense concept must be manually annotated for information outside the KG. (2) Formal concepts are not subject to KGs, while commonsense concepts are limited to the corresponding KGs. To be detailed, the formal concept can be applied to any KG, including a commonsense KG, while the commonsense concept is only applicable to the KG that has corresponding commonsense in theory and has artificially annotated commonsense concepts.

Architecture for KGC
Based on this, we proposed a Scalable Formal Concepts driven Architecture (SFCA) to extract formal concepts from the KGs to improve the performance of KGE. The SFCA conidia of three modules: (1) Formal concept extraction module (FCE) extracts the formal concepts in triples and links the formal concepts with the entities. This framework applies to all knowledge graphs and fully benefits from the selfsupervision of the data itself. The contributions of the proposed method are summarized as follows: (1) We propose a scalable KGC architecture based on formal concept analysis to generate formal concepts from KG. In comparison, existing methods either perform poorly or require commonsense concepts of manual extraction. To our knowledge, we are the first to apply formal concepts to KGC. (2) We design a coarser-to-fine formal concept extraction strategy to choose streamlined formal concepts of entities in the KG that can be used to improve the computational efficiency of the model.

KGE Models
According to KGE models' input data, the KGE models currently in use can be broadly split into two main streams: (1) The KG structural information-based methods include translation-based and semantic matching models. Translation-based models [10,12,15] are models that utilize entity and relation embeddings to compute translation scores, where relations represent the translation operations between entities. Translating embeddings for modeling multi-relational data (TransE) [10] is the pioneer of the translation-based model, which embeds entities and relations into the space with the same dimension, and regards relations as translation operations between entity vectors. The advantage of TransE is that it is simple and efficient, while it cannot model various relation patterns. Knowledge graph embedding by relational rotation in complex space (RotatE) [15] embeds entities and relations in a complex vector space, treats relations as rotation operations between entity vectors, effectively models and infers various relational patterns, and refreshes the best results of the KGC task. Learning hierarchy-aware knowledge graph embeddings for link prediction (HAKE) [12] effectively embeds the semantic hierarchy by mapping the entity to the polar coordinate system, achieving the STOA result of the KGC task. Semantic matching models [11,13,14,16] compute semantic matching scores for entity and relation embeddings in the latent space. A three-way model for collective learning on multi-relational data (RESCAL) [16] treats entities as vectors and relations as matrices and calculates scores with bilinear functions. Embedding entities and relations for learning and inference in knowledge bases (DistMult) [13] simplifies RESCAL by restricting the relation matrix to be a diagonal matrix. Complex embeddings for simple link prediction (ComplEx) [14] extend DistMult to embed entities and relations into complex space. Quaternion knowledge graph embeddings (QuatE) [11] embed a hypercomplex value with three imaginary components to represent entities and model the relation as a rotation on a 4-dimensional space (hypercomplex space), thus unifying ComplEx [14] and RotatE [15]. (2) External information-based methods focus on adding extra information to enrich the KGE models. Most models that add external information [18,20,21] utilize the logic rules mined from the knowledge graph to improve the link prediction results. Fast rule mining in ontological knowledge bases with AMIE + + (AMIE+) [20], End-to-end differentiable rule mining on knowledge graphs (DRUM) [21], and Knowledge graph embedding with iterative guidance from soft rules (RUGE) [18] automatically mine the logic rules in KG and apply the logic rules to KGC tasks. A considerable part of the models that add internal information [17,19,22] use the known concept information corresponding to the entity to improve the link prediction outcomes. Representation learning of knowledge graphs with hierarchical types (TKRL) [17] uses the type information of the entity to design a scoring function with a hierarchical projection matrix for the entity, which improves the performance of the KGC task, Type-based multiple embedding representations for knowledge graph completion (TransT) [22] adopts entity types to construct relation types and takes the similarity between relative entities and relations as prior knowledge, and utilizes prior knowledge to improve the KGC task results. A scalable commonsense-aware framework for multi-view knowledge graph completion (CAKE) [19] leverages commonsense concepts of entities to improve the quality of negative sampling and the accuracy of link prediction candidate entities. Ontology-guided Entity Alignment via Joint Knowledge Graph Embedding (OntoEA) [25] embeds ontology and knowledge graphs together to get better entity embedding.

KG Embedding with Ontology
At present, many scholars are studying KG embedding with ontology. Semantically smooth embedding for knowledge graphs (SSE) [26] models the intrinsic geometry of KG based on the assumption that entities belonging to the same semantic category are close to each other in the embedding space (semantically smooth). SSE [26] uses two manifold learning algorithms, Laplacian Eigenmap and Local Linear Embedding, as regularization terms to model the smoothness hypothesis. Differentiating concepts and instances for knowledge graph embedding (TransC) [27] models concept embeddings as a sphere and assumes that the embedding vectors corresponding to instances of categories belonging to concepts should lie in this sphere. TKRL [17] focuses on the Type hierarchy of KG and believes that entities have different representations under different categories. And use the layered type as a mapping matrix and use type encoders to design. A knowledge-driven representation learning method with ontology information constraints (TransO) [28] model considers the limitations of the three types of ontology information, type information, relation constraint information, and hierarchical structure information, and maps entities and relationships to the ontology information-limited space. Based on the TransE model, loss functions are defined in the basic space and ontology information-limited space and combine the two for representation learning. Knowledge graph embedding with hierarchical relation structure (HRS) [29] learns representations of Relation clusters, Relations, and Sub-relations separately. It sums the three as the embedding vector of the relationship, thus modeling the hierarchical structure of the relationship. A representation learning method for knowledge graphs with relation hierarchical structure (TransRHS) [30] builds on HRS [29]. It encodes each relation as a vector and relation-specific spheres in the same space. TransRHS [30] uses the relative positions between vectors and spheres to model subrelationships, which embodies the inherent generalization relationship between relations. The Universal representation learning of knowledge bases by jointly embedding instances and ontological concepts (JOIE) [31] model proposes a method that considers that KG and ontology use different embedding spaces and enables cross-space interaction between the two embeddings. A Modified Joint Knowledge Graph Embedding Model for Concepts and Instances (JECI++) [32] simplifies hierarchical concepts and links instances to them, making identifying cases based on neighboring instances and simplified concepts easier. It uses circular convolution to locate instances in the embedding space and employs CBOW and Skip-Gram strategies to embed simplified concepts and instances jointly.

Formal Concept Analysis
Formal Concept Analysis (FCA) [33] is a powerful and widely used method in information science that enables the creation of a concept hierarchy or formal ontology from a given set of objects and their properties. This approach is based on the mathematical theory of lattices and ordered sets, which allows for the identification of shared properties and relationships between objects in a structured and systematic way.
The resulting hierarchy of concepts represents a logical and intuitive organization of the objects and their properties, with each concept capturing a group of objects that share a set of common attributes. Moreover, the sub-concepts in the hierarchy represent a more specific grouping of the objects and a superset of the attributes from the abovementioned concepts.
Introduced by Rudolf Wille in 1981, Formal Concept Analysis (FCA) [33] has become a fundamental tool in various fields, including data mining, text mining, machine learning, knowledge management, semantic web, software development, chemistry, and biology. Its practical applications are diverse and numerous, ranging from discovering hidden patterns and relationships in data sets to developing more effective search algorithms and enhancing the quality of knowledge representation in various domains.

Methodology
We will introduce the SFCA framework in this section. As shown in Figure 2, the whole architecture includes three modules. The FCE module realizes the conversion of KG to formal context and formal context to formal concepts. The FCS module realizes the conversion of the whole formal concepts to the partial formal concepts. Finally, the FCLP module implements partial formal concepts supervision for link prediction.
As shown in Figure 2, In the Vertical field KG, points of different colors correspond to different types of real-world things (type information is not included in this KG). In the formal context of the FCE module, the various entities and relations are denoted by name abbreviations and the color of the type to which they belong. The FCE module's concept lattice's formal concept is depicted as a diamond, and the colored rectangle on the right side of the diamond identifies the different kinds of entities and relations that make up the formal concept's entity set and relation set, respectively. As shown in Figure 2, In the Vertical field KG, points of different colors correspond to different types of real-world things (type information is not included in this KG). In the formal context of the FCE module, the various entities and relations are denoted by name abbreviations and the color of the type to which they belong. The FCE module's concept lattice's formal concept is depicted as a diamond, and the colored rectangle on the right side of the diamond identifies the different kinds of entities and relations that make up the formal concept's entity set and relation set, respectively.
As shown in Figure 2, In the formal concept sampling module in the lower right corner, the initial input is the natural mapping relationship between the formal concept and the entity. First, the natural mapping between the formal concept and the relationship is considered to find the relationship between each triple "head entity-relationship" and the formal concept. And then consider the natural mapping of formal concepts and entities again to find the correspondence between each triplet and the "formal concept triplet". Finally, in the link prediction module driven by the formal concept in the lower left corner, whether the new triplet combined with the triplet of the missing head entity or the missing tail entity and the candidate entity can correspond to the known "formal concept triplet" is initially carried out a filter, and then enter the ranking stage to output scores.  As shown in Figure 2, In the formal concept sampling module in the lower right corner, the initial input is the natural mapping relationship between the formal concept and the entity. First, the natural mapping between the formal concept and the relationship is considered to find the relationship between each triple "head entity-relationship" and the formal concept. And then consider the natural mapping of formal concepts and entities again to find the correspondence between each triplet and the "formal concept triplet". Finally, in the link prediction module driven by the formal concept in the lower left corner, whether the new triplet combined with the triplet of the missing head entity or the missing tail entity and the candidate entity can correspond to the known "formal concept triplet" is initially carried out a filter, and then enter the ranking stage to output scores.

Preliminary Knowledge of FCA
For better understanding, we first provide a brief introduction to Formal Concept Analysis [19] (FCA). FCA is a method for knowledge representation, information management and data analysis. Generally, we regarded FCA as a conceptual clustering method used to determine implicit associations between objects and attributes. Formal context, formal concept, and concept lattice are three central notions in FCA. The following are the key definitions: Definition 1. Let G be the set of objects, M be the set of attributes, and I be the binary relationship between object set G and attribute set M. Then, the triple (G, M, I) is a formal context. Definition 2. Given a formal context K = (G, M, I), for A⊆ G, B⊆ M, the following operations are defined: A is the set of attributes shared by all objects in A; B is the set of objects corresponding to all attributes in B.
If A = B, B = A , then the binary (A, B) is a concept in the formal context, where A is the intent of the concept (A, B) and B is the extent of the concept (A, B).    The concept lattice can be visualized by a Hasse diagram. For example, the corresponding concept lattice in Example 1 is shown in Figure 3.

KGE Score Function
Inspired by the CAKE [12], we adopt a scalable architecture design that takes any KGE model as a plugin module for SFCA direct usage. The KGE model plays the role of extracting the entity and relation embeddings in architecture. Here, we give a uniform symbol E (h, r, t) to clearly describe the score function of any KGE model for assessing the plausibility of a triple (h, r, t). Table 2 shows the definitions of various KGE models. Table 2. Details of several knowledge graph embedding models.

Score Function Parameters
KGC is commonly divided into three subtasks: triple classification, link prediction and relation prediction. In this work, we only concern about link prediction, not involving other KGC tasks. The link prediction task refers to finding the missing entity when the head or tail entity in the triple is missing. Specifically, we treat link prediction as an entity prediction task that searches for the reasonable entity when the triple's a head entity or tail entity is missing. Every entity in KGs will be considered a candidate when encountering a missing entity or in a triple query. We choose the top n (n = 1, 3, 10) hits of correct entities as predicted results by ranking the scores of the candidate entities.

Formal Concept Extraction Module
According to the formal concepts definition (see in Section 3.1.1), SFCA automatically generate formal concepts from arbitrary KGs without external annotated knowledge. To get high-quality formal concepts, we developed a FCE module to generate massive corresponding formal concepts from KGs in mining valuable information. All generated formal concepts contain corresponding entity set and relation set.
Each fact triple set will first be encoded into a two-dimensional tabular as the formal context. Then, we produce formal concepts by the entity and relation binary relations in these formal contexts. The formal representation of KG: where E is the set of entities, R is the set of relations from E to E. In this paper, entity e i is regarded as the object and relation r j as the attribute of e i . The formal context can be obtained by a KG and then the concept lattice K can be induced by the formal context.
For every e i ∈ E, let C The mapping from the formal concepts to the entity is formulated as:

Formal Concept Sampling Module
An obvious problem with dense formal concepts produced by the FCE module: Not all formal concepts are requisite. Similar semantic information and negative gain information consist of dense formal concepts (Not all formal concepts can have corresponding real concepts in the real world, so formal concepts need to be screened before using them). Thus, we proposed an FCS module to reduce complexity and improve the quality of formal concepts.
An entity can be mapped into many formal concepts in a sizable KG. Formal concepts, including the same entity, can be considered a hierarchy. Among all formal concepts, we argue that those located in the top-most node formal concepts involve the most valuable information, and relation in factual triples is sufficient to filter out them (Here, the topmost formal concept refers to the formal concept that has a non-zero and minimum number of elements in the relationship set, and the number of elements in the entity set owned by this formal concept is closest to all entities). As shown in Figure 4, our FCS module is designed by a rough-to-fine sampling strategy in the usage of instance relations and partial order relations. Specifically, the formal concepts mapped by Entity 2 are marked by the blue dots in the concept lattice on the left top. The formal concepts mapped by Entity 2 and relation e after the first sampling are marked by the blue dots in the left middle concept lattice. The formal concept mapped by entity 2 and relation e after the second sampling is marked by the blue dots in the concept lattice on the lower left. The blue dots in the green parallelogram on top of the figure are the sampled formal concepts.
(1) Sampling with instance relations: In the first stage, we only use the instance relations for sampling by inferring that formal concepts mapped by the same entity under different relations should differ. Given an instance triplet, if a formal concept in the set is mapped to the head entity of the instance triplet and the relation of the instance triplet belongs to the relation set of the formal concept, the formal concept will be mapped to the entity-relation pair composed of the head entity and the relation of the instance triple: For every (e i , r j , e k ) ∈ KG, let C temp e i r j = (A l , B l ) (A l , B l ) ∈ C temp e i , r j ∈ B l . Where (A l , B l ) is the concept in the concept lattice K.
We can get the following mapping: for every (A l , B l ) ∈ C temp e i r j , g (A l , B l )) =(e i , r j ).
(2) Sampling with partial order relations: After the first sampling stage, the selected formal concept set mapped to the head entity-relation pair still contains some formal concepts, and the formal concepts in this set can form a partial order relation between formal concepts. Thus, to simplify the mapping, we select the formal concept with the same link position for all formal concept links. Here, we select the most valuable formal concept in the formal concept set. Substitute formal concept set as a mapping formal concept of head entity-relation pair.
Given an instance triplet, if a formal concept in the set mapped to the head entityrelation pair of the instance triplet and the relation set of the formal concept is included in the relation set of all formal concepts in the formal concept set, then the formal concept is selected to be mapped to the entity-relation pair composed of the instance triple head entity and relation, according to Definition 3, there exists a maximum concept in the C temp e i r j . For every (e i , r j , e k ) ∈ KG, let C

Formal Concept-Driven Link Prediction Module
To find better candidate entities and improve prediction outcomes, we propose a novel two-stage formal concept supervised link prediction mechanism. In the first stage, candidate entities are selected from the perspective of formal concepts; specifically, accepting a query ( , ,?) h r to filter plausible formal concepts of tail entities using the set of After sampling twice, the mapping from formal concepts to instance entities and the formal concept triples can be obtained-the mapping of entity-relation pairs from formal concepts to instance triples after sampling twice.
For every e i ∈ E, let C We can get the following mapping: A collection of formal concept triples is denoted as FC, where each triple consists of a head entity's formal concept set C f inal hr and a tail entity's formal concept set C f inal t associated with their instance-level relation r, defined as:

Formal Concept-Driven Link Prediction Module
To find better candidate entities and improve prediction outcomes, we propose a novel two-stage formal concept supervised link prediction mechanism. In the first stage, candidate entities are selected from the perspective of formal concepts; specifically, accepting a query (h, r, ?) to filter plausible formal concepts of tail entities using the set of formal concept triples FC, the set of candidate formal concepts of tail entity t is C f inal t , and then determine the entity belonging to the formal concept set as the candidate entity.
In the second stage, for each candidate entity e i that has been screened, the score is calculated by the scoring function. The candidate triplet that finally calculates the score is: Among them, E(h, r, e i ) is the scoring function used to train the KGE model, and then the prediction results will arrange the scores of the candidate entities in ascending order and output the top n hits of the correct entities.

Experiments
For a comprehensive comparison, we evaluate our SFCA on five real-world datasets and one industry dataset. In this section, the detail of the setting of the experiment will be introduced. First, the performance of SFCA on four public datasets will be shown. Second, the comparison with common sense concepts will be discussed third, and the effectiveness of real data will be proved finally.

Datasets
Our evaluation is based on five public datasets (FB15K237 [34], YAGO3-10 [35], WN18RR [36], NELL-995 [37], DBpedia-242 [19]) and an industrial KG datasets collecting from the workshop of actual factory. Table 3 shows the statistics of public and industry datasets. FB15K237 is a link prediction dataset created from FB15k. FB15k-237 was created by Toutanova and Chen to ensure that the test and evaluation datasets do not have an inverse relation with test leakage. YAGO3-10 is a benchmark dataset for knowledge base completion. It is a subset of YAGO3 (YAGO3 is an extension of YAGO) and contains entities associated with at least ten different relations. WN18RR is a link prediction dataset created by WN18, a subset of WordNet. However, many text triples are obtained by inversely finding triples from the training set. The WN18RR dataset was therefore created to ensure that the evaluation dataset does not have an inverse relation with test leaks. NELL-995 is a subset of NELL suitable for multi-hop inference proposed from the 995th iteration of the NELL system. Useless triples are first removed using relations that occur more than 2 M times in the NELL dataset. After this step, the triples with Top-200 relations are selected, and the dataset is obtained after adding the inverse triples. DBpedia-242 is extracted from DBpedia [6], which contains 242 concepts. It is worth mentioning that the entities in FB15K237, YAGO3-10 and NELL-995 have a corresponding ontology, while in WN18RR, The entity does not have a corresponding ontology.

Baselines
We compare our SFCA model with five baseline models, including TransE [10], Dist-Mult [13], ComplEx [14], RotatE [15], and HAKE [12], and we also integrate these baseline models into our framework. All baseline models are KG-structured models. As a result, our framework does not require the input of external data, and we show through experiments that our model can be applied to most models without the input of external expert data.

Implementation Details
We use the Adam optimizer for training, and all models adopt the self-adversarial negative sampling method. In terms of the parameters, under the same dataset, we use the same parameters of different baseline models, including embedding size, batch size, negative sampling size, learning rate, margin, and sampling temperature. All experiments are performed on Pytorch and NVIDIA Quadro RTX 5000 GPU.

Evaluation Protocol
We choose three recognized evaluation metrics for comparison, including mean rank (MR), mean reciprocal rank (MRR), and the proportion of top-N rankings of correct entities (Hits@N). Notably, we filtered out all candidate triples in the datasets. The detailed computing formulas and notation definition as shown in Table 4.  Table 5 shows the link prediction performance of SFCA on the four public datasets. The formal concepts-driven KGE module improved significantly, which has an average increase of 11.83% (3.78 points), 16.53% (4.12 points), and 19.13% (5.82 points) in the MRR indicators of different baselines on the three datasets of FB15K237, YAGO3-10 and NELL-995, respectively. On WN18RR, the Hit@10 indicator of different baselines improved by an average of 2.63% (1.34 points). These results prove formal concepts, as metaphysical features of KG-structured information, are more splendid and effective for link prediction.

Common Sense Concepts vs. Formal Concepts
We also compare our SFCA with an external information-based method: Commonsense-Aware Knowledge Embedding (CAKE) [19] framework. The comparison results are obtained by combining uniform sampling [10] and self-adversarial sampling [15] with the KGE model TransE [10] and RotatE [15]. Table 6 presents the link prediction evaluation results on the three datasets. By comparison, our SFCA has an average increase of 147.54% (45.13 points) and 9.34% (3.33 points) higher than CAKE in the MRR indicators of different baselines on the two datasets of FB15K237 and NELL-995, respectively. On DBpedia-242, in the best effect on each indicator of different baselines, the highest is 14.47% (2.3 points) higher than CAKE, and the lowest is 3.36% (1.5 points) lower than CAKE. These results prove formal concepts are more effective in most cases than common sense concepts for the KGC task. Table 7 shows the link prediction evaluation results on the industrial KG dataset. By comparison, we can see that our SFCA framework has improved by more than 13.11% (7.3 points) in the MRR indicators of different baselines on the fault diagnosis industrial dataset. Our SFCA achieves remarkable performance in industrial filed KGs. With this result, we believe our SFCA can play well in multi-field KGs.

Conclusions
Motivated by the formal concept analysis theory, we propose a novel scalable formal concept-driven knowledge graph completion framework (SFCA) applying to multiple verticals. SFCA can automatically generate formal concepts from KG with a coarse-tofine extraction strategy and a formal concept-supervised link prediction module to filter candidate entities from the perspective of formal concepts. Experiments on five public datasets demonstrate the effectiveness and scalability of SFCA. In addition, our model is also experimentally performed on real industrial datasets to demonstrate that the model has high performance in both general and industrial domains.
Despite the performance of SFCA, it is still some areas for improvement. First, the application of our method on the KGC task is based on the closed-world hypothesis, which considers any triplet not explicitly present in the graph as a negative triplet. Second, our method only applies to the KGC task and does not extend to other knowledge-related tasks. Thus, our further work is to study the application of formal concept analysis on KGC tasks under the open-world hypothesis and explore the application of formal concept analysis on various concept-related KG tasks, such as life-long learning. Third, this paper's knowledge graph embedding-related tasks only involve knowledge graph completion. Future work can consider other knowledge map-related tasks such as named entity recognition and relationship extraction, that is, using formal concept analysis to assist tasks such as named entity recognition and relationship extraction.