Next Article in Journal
Electrical Bioimpedance Analysis for Evaluating the Effect of Pelotherapy on the Human Skin: Methodology and Experiments
Previous Article in Journal
MEMS Shielded Capacitive Pressure and Force Sensors with Excellent Thermal Stability and High Operating Temperature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning

1
College of Information Science and Engineering, Xinjiang University, Urumqi 830046, China
2
Xinjiang Multilingual Information Technology Laboratory, College of Information Science and Engineering, Xinjiang University, Urumqi 830046, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(9), 4250; https://doi.org/10.3390/s23094250
Submission received: 12 March 2023 / Revised: 15 April 2023 / Accepted: 22 April 2023 / Published: 25 April 2023
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Transforming the task of information extraction into a machine reading comprehension (MRC) framework has shown promising results. The MRC model takes the context and query as the inputs to the encoder, and the decoder extracts one or more text spans as answers (entities and relationships) from the text. Existing approaches typically use multi-layer encoders, such as Transformers, to generate hidden features of the source sequence. However, increasing the number of encoder layers can lead to the granularity of the representation becoming coarser and the hidden features of different words becoming more similar, potentially leading to the model’s misjudgment. To address this issue, a new method called the multi-granularity attention multi-scale self-learning network (MAML-NET) is proposed, which enhances the model’s understanding ability by utilizing different granularity representations of the source sequence. Additionally, MAML-NET can independently learn task-related information from both global and local dimensions based on the learned multi-granularity features through the proposed multi-scale self-learning attention mechanism. The experimental results on two information extraction tasks, named entity recognition and entity relationship extraction, demonstrated that the method was superior to the method based on machine reading comprehension and achieved the best performance on the five benchmark tests.

1. Introduction

The information extraction task is a critical component of natural language processing, aimed at extracting structured knowledge from unstructured text [1,2,3,4,5]. Recently, pre-trained language models (PLMs) [6,7,8,9,10,11,12,13,14] have led to significant advancements in two classic information extraction tasks: named entity recognition (NER) and relation extraction (RE). In their work, Li et al. [15,16] described these tasks as requiring an understanding of context information and the ability to answer questions (queries). First, the corresponding queries are constructed based on entity or relationship types. Next, queries and context are connected and representations are learned using PLMs. Finally, the labels of each token are predicted, resulting in good performance on NER and RE tasks. Figure 1 illustrates the process of the model based on machine reading comprehension in identifying named entities.
Although joint encoding of query and context can capture inter-dependencies across the entire sequence, in a deep learning setting, the hidden features of words become increasingly similar with the increase of the encoder layers. This means that the hidden features no longer represent individual word information and become a fuzzy concept [17], as illustrated in Figure 2. Additionally, generating queries through crowdsourcing incurs significant human and time costs and has poor portability. Therefore, many researchers have transformed query generation into a template-filling problem. Examples of generating queries for different entity types are shown in Table 1, and examples of generating queries for different relationship types and entity types are shown in Table 2. When the query is long, connecting the query and context will cause the proportion of the context to relatively decrease. Therefore, as the query length increases, the information of individual words in the context will be diluted by the information of other words, resulting in the loss of some original information.
To address the issues discussed earlier, we propose a multi-granularity attention network. Our approach involves first independently encoding the query and context using dual encoders. Next, we employed a bidirectional attention network to capture word-level interactions between the query words and context words from two dimensions. We then used long-term-memory-gated attention to filter the original text information, preserving the source text representation of words relevant to the task. Even if the same phenomenon occurs due to the excessive number of encoding layers and the loss of original text information due to the query length, our model can still learn fine-grained features of individual words and retain the original text representation through long-term-memory-gated attention, thus enhancing its ability to read and understand text information. Finally, we combined the aforementioned components with the bidirectional attention vectors to construct a multi-granularity attention network. To the best of our knowledge, there is currently no work that uses dual encoders to solve the NER and RE tasks.
In order to better help the model understand text and extract information for downstream tasks, a representation method that can describe and highlight the characteristics of a sequence can be relatively more effective. To this end, inspired by the human reader’s reasoning process of revisiting specific paragraphs or questions after single-round comprehension, we propose a self-learning mechanism that, with the guidance of the learning mechanism, focuses on different parts of the document each time from the perspectives of overall meaning and finer details and effectively collects evidence from the entire article to match the answer. As shown in Figure 3, MAML-NET can extract relationship triplets (its, ART, parks), (firm, GEN–AFF, New York), and (empire, PART–WHOLE, its) through the learning mechanism, by simulating the reasoning process of human readers. Specifically, we established a fusion module for interactive features and deep contextual semantic features, combining multi-head self-attention with depthwise separable convolution. At the same time, from both global and local perspectives, we used interactive information and deep contextual information to dynamically refine feature representations that match themselves. For global self-learning, we used self-attention to capture it; for local self-learning, we used deep separable multi-scale convolution, which is computed in parallel with self-attention. Overall, our proposed self-learning mechanism can effectively highlight key features in the text and help the model better understand and extract information for downstream tasks.
In summary, our contribution is as follows:
• A new information extraction method based on a reading comprehension model is proposed, which explores and utilizes the multi-granularity representation of the source sequence, encourages the model to fully utilize the interaction information and original information between the text and query, and enhances the expression ability of reading comprehension, thereby improving the information extraction capability.
• In order to dynamically collect evidence for candidate answers from multi-granularity text representations, a multi-scale self-learning network is proposed to refine semantic representations from global and local perspectives.

2. Related Work

2.1. Entity and Relation Extraction

Entity extraction, also known as named entity recognition, was initially approached using rule-based methods, which were expensive to build [18]. In recent years, deep learning has gained attention in various fields, and various neural network models have been proposed for this task. For example, Straková et al. [19] treated the nested NER as a sequence-to-sequence problem. Luan et al. [20] proposed a general information extraction method using dynamically constructed span graphs that share span representations. Li et al. [15] transformed the NER task into a machine reading comprehension task, where each entity type’s feature is a natural language query and entities are extracted by answering these queries (extracting two different categories of overlapping entities requires answering two separate questions).
Traditional entity relation extraction systems use a pipeline approach [21]. First, an entity extraction model identifies entities in the text. Then, a relation extraction model predicts the relations between the identified entities. Although this separate learning method makes processing the two sub-tasks more flexible, it ignores their inherent interaction and is severely affected by error propagation. To mitigate this problem, later work proposed joint learning methods. Early joint learning methods were based on crowdsourcing to build features, which was expensive. In subsequent research, several neural-network-based methods were proposed.
For example, Miwa et al. [22] proposed an end-to-end method that uses a tree-shaped LSTM model with parameter sharing to extract entities and relations. To better model the semantic relationship between entities, Katiyar et al. [23] improved upon this method using an attention mechanism to address the error propagation problem caused by the use of dependency trees. Zhang et al. [24] developed an end-to-end relation extraction model based on global optimization and used new LSTM features to better represent context. They also proposed a new method to integrate syntax information to facilitate global information learning. Sun et al. [25] proposed a graph convolutional network that operates on the entity relation bipartite graph. By introducing binary relation classification tasks, the bipartite graph structure of entity relations can be more interpretable and effectively utilized. At the same time, they pointed out that existing joint learning methods can predict entity spans correctly, but perform poorly when predicting entity types. Fu et al. [26] proposed a GraphRel end-to-end relation extraction model that uses graph convolutional networks (GCNs) to jointly learn named entities and entity relations. Zhang et al. [24] proposed a soft pruning model that can automatically learn how to select relevant substructures to focus on for relation extraction tasks. Shen et al. [27] proposed a triggered memory flow framework to enhance the interactive capabilities of entity recognition and relation extraction. Li et al. [16] viewed the relation extraction task as a multi-turn QA task. Zhao et al. [28] addressed the issue of a single query not being able to fully describe the meaning of entities and relations due to semantic diversity in multi-turn QA tasks. They designed a model that introduces diversity QA mechanisms and two answer selection strategies to integrate different answers.

2.2. NLP Tasks Based on the MRC Approach

Recently, there has been a trend of converting non-reading comprehension tasks in NLP into machine reading comprehension tasks. For example, Levy et al. [29] first converted the relation extraction task into the form of reading comprehension, which could adapt to zero-shot scenario problems. Mccann et al. [30] proposed a universal framework based on MRC, formalizing ten tasks such as question answering, sentiment analysis, and machine translation into question-answering tasks. Li et al. [15] proposed a unified MRC framework and applied it to the named entity recognition task. Yang et al. [31] proposed a label-aware enhanced language representation based on MRC and applied it to the named entity recognition task and time detection task. Du et al. [32] converted the event extraction task into a question task. Recent work has proposed a method of converting entity and relation extraction tasks into question-answering tasks. Although inspired by Li et al. [16], our work has the following characteristics: firstly, unlike the former using joint encoding to encode the query and context, MAML-NET independently encodes them through two pre-trained language model encoders. Secondly, MAML-NET uses bidirectional attention to establish the word-level interaction between the query and context and uses the original long-term memory gate attention to retain the important original information. Finally, through the original multi-scale self-learning network, it further understands deeper semantic information between the query and context. With the help of these features, MAML-NET’s performance was significantly improved.

3. Methodology

3.1. Overview

The multi-granularity attention and multi-scale self-learning network, as shown in Figure 4, consists of three main parts: (1) transforming the dataset of a specific task into a reading-comprehension-style dataset; (2) independently encoding the query and context using a Transformer-based encoder and establishing a multi-granularity representation of the source sequence; (3) passing the multi-granularity feature representation of the source sequence through the multi-scale self-learning network to further enhance the potential for reading comprehension at both the global and local levels.

3.2. Formalization of Tasks

Given a context sequence x i = ( w 1 , w 2 , , w n ) from the training set D, our task was to predict the relation between entities based on the information in each x i , where each entity is assigned a label y e Y e and each relation is assigned a label y r Y r . Y e is a set of entity types (e.g., ORG, GPE, etc.), and  Y r is a set of relation types (e.g., ORG–AFF, PER–SOC, etc.). For the RE task, there may be multiple relations for a subject/object pair or the same entity may be shared between the subject and object, and we constructed a relation–object query for each subject.

3.2.1. Named Entity Recognition

Given an input sequence x i , identify all candidate spans E i = ( w i , w j ) x i , and assign a label y e Y e to each candidate span. Our goal was to maximize the likelihood of the training set D:
i = 1 | D | w i , w j E i p w i , w j x i , q = i = 1 | D | w i E i p start w i x i , q s w j E i p end w j x i , q s
where p s t a r t ( w i | x i , q s ) and p e n d ( w j | x i , q s ) represent the probabilities of the start and end span positions for any y e Y e category corresponding to x i .

3.2.2. Entity Relation Extraction

The goal was to maximize the data likelihood of the training set D for a given input sequence x i and a set of potential relationship triplets T i = ( s , r , o ) :
i = 1 | D | ( s , r , o ) T i p ( s , r , o ) x i , q = i = 1 | D | s T i p s s x i , q s ( r , o ) T i s p o ( r , o ) s , x i , q o = i = 1 | D | s T i p s s x i , q s o T i | | s , r p o o s , r , x i , q o
where s, r, o, q s , and  q o , respectively, represent the subject, relation, object, and queries constructed based on the types of subjects and objects. T i | s denotes the set of relation triplets guided by subjects. T i | ( s , r ) denotes the object set guided by ( s u b j e c t , o b j e c t ) pairs. Firstly, a subject classifier p s ( s | x i , q s ) is learned, which is used to identify all head entities in x i . It is important to note that this head entity may not necessarily have a corresponding object and relation, i.e., the head entity may not be the subject. For each relation r, an object classifier p o ( o | ( s , r , x i , q o ) ) is learned, which is used to identify the object corresponding to a specific relation for a given subject. Thus, the relation extraction task is formalized as a classification problem of identifying the subject and the object corresponding to a specific relation.

3.3. Input Layer

Given a query and context, we used X q R m and X c R n to represent the tokenized input of the query and context, respectively. They are independently mapped to a standard pre-trained bidirectional Transformer [33] for learning, resulting in the preliminary feature representations of the query and context: O B E R T q R ( m + 2 ) × d and O B E R T c R ( n + 2 ) × d . The specific operations are shown in Formulas (3) and (4).
O B E R T q = B E R T q ( X q )
O B E R T c = B E R T c ( X c )
where m and n represent the lengths of the query and context sequences, respectively, and d represents the dimensionality of the BERT output.

3.4. Multi-Scale Attention

3.4.1. Interactive Attention

To connect and integrate information from the context and query words, we employed a bidirectional attention flow network (BiDAF) [34]. This mechanism was designed to capture word-level interactions, where the interaction attention estimates the impact relationship between each query and context word.
The attention is calculated in two dimensions, from context to query and from query to context. Both attentions originate from the shared correlation matrix U R ( n + 2 ) × ( m + 2 ) , which is between the context O B E R T c and query O B E R T q . U i j represents the correlation between the i-th context word and the j-th query word. The correlation matrix is calculated as follows:
U i j = W i j [ O B E R T i c ; O B E R T j q ; O B E R T i c · O B E R T j q ]
where W i j R 3 d is a trainable weight matrix, · represents matrix multiplication, and ; represents matrix concatenation. We used U to compute attention vectors on both dimensions.
Query-to-context attention: The query-to-context (C2Q) attention mechanism measures the relevance between each context word and each query word, which is crucial for identifying entities. This attention mechanism simulates the process in which humans search for answers in text based on questions, enabling the model to selectively read the text based on the query. The specific structure is shown on the left side of Figure 5.
We calculated the attention score s i q c for each context word associated with the query. The specific operation is shown in Formula (6).
s i q c = softmax ( m a x ( U i , : ) )
where m a x ( U i , : ) represents the maximum value across the i-th row of the shared relevance matrix U. We then calculated the query aware context representing r. The specific operation is shown in Formula (7).
O q 2 c = i = 1 n + 2 s i q c · O B E R T c
Context-to-query attention The context-to-query (C2Q) attention represents the correlation between each query word and all words in the context. This attention mechanism aims to simulate the process of human beings reviewing questions after reading the text, thereby enabling the model to better understand the query: what the question is about. The specific structural system is shown on the right of Figure 5.
The attention score s i c q R m + 2 represents the relevance of each context word to the query words, as shown in Formula (8).
s i c q = softmax U i , :
Then, we generated the context-aware query representation O c 2 q R ( n + 2 ) × d , as shown in Formula (9).
O c 2 q = i = 1 m + 2 s i c q · O B E R T q
Finally, we concatenated the attention vectors from both dimensions and used this as the final interaction attention representation O I A R ( n + 2 ) × d , as shown in Formula (10).
O I A = [ O c 2 q ; O q 2 c ]

3.4.2. Long-Term-Memory-Gated Attention

As deep neural networks are prone to homogeneity and the loss of the original information, we decided to incorporate the original information into the output results to reduce the information loss and achieve long-term memory. Therefore, we designed a long-term-memory-gated attention mechanism to allow the model to remember the original information of the context while controlling the flow of the context original information, capturing the source sequence features relevant to the task. The specific architecture is shown in Figure 6.
Long-term-memory-gated attention ( O L T M G A R ( n + 2 ) × d ) is calculated as Formulas (11)–(15).
O L T M G A = O S O C
O C = σ ( W s f [ O B E R T C ; S M H S A ] )
S M H S A = softmax ( Q · K T d k ) · V
Q , K , V = f ( O B E R T c )
f ( O B E R T c ) Q = W q · O B E R T c + b q K = W k · O B E R T c + b k V = W v · O B E R T c + b v
where Q, K, and V are abstract matrices projected from the initial embedding matrix O B E R T c of the context. s M H S A is based on multiple scale-dot attention with 12 attention heads. W s f R d × 2 d represents trainable parameters. σ is an elementwise sigmoid function used to control the flow of information. ⊙ denotes elementwise multiplication.
Finally, the long-term-memory-gated attention flow O L T G A is aggregated with the interaction attention representation O I A , and the result is normalized through layer normalization, resulting in the multi-granularity representation O M G A R ( n + 2 ) × 3 d . The process is shown in Formulas (16) and (17).
O I A l i n e a r = L i n e a r ( O I A )
O M G A = L N ( O I A l i n e a r O L T G A )
where L i n e a r denotes a linear transformation and ⊕ denotes vector addition. L N represents layer normalization.

3.5. Multi-Scale Self-Learning Layer

Although using the multi-granularity representation O M G A can enhance the representation of the source sequence by incorporating different levels of views, the enhanced source representation is still limited in understanding the information in the context, as a candidate answer typically contains information from multiple windows. In the constructed reading-comprehension-style relation extraction dataset, we enabled the model to learn from itself using the multi-granularity representation O M G A , based on the multi-granularity information contained in O M G A , to understand the overall idea of the context and important details related to the task. To achieve this, we designed two strategies to enable the model to self-learn at different scales.

3.5.1. Global-Information-Gated Attention

Although query-to-context can allow the model to selectively read text with a question in mind and context-to-query can allow the model to read text more broadly and deepen its understanding of the query, this is clearly not enough because just reading selectively and broadly will make it difficult to understand the true meaning of the text. Therefore, we hope to use this strategy to let the model review the entire text again and clarify the main idea of the text. For a given O M G A , it is first mapped to three representations, and then, the final representation G a t t R ( n + 2 ) × d is obtained. The specific operation is shown in Formulas (18) and (19).
L M G S ( O G M A ) = softmax ( Q · K T d k ) · V 1
V 1 = W v V 1
where Q , K , V 1 = L i n e a r 1 ( O G M A ) , L i n e a r 2 ( O G M A ) , L i n e a r 3 ( O G M A ) , and  W V is the projection parameter.

3.5.2. Local-Features-Focused Attention

In order to simulate human reading habits, after full-text scanning, question-based reading, and overall in-depth reading, we hope that the model can learn the key information related to the query in the text through its own understanding, rather than just relying on the query. To achieve this, we used convolutional operations to capture more refined local features, so that the model can grasp the details related to the task. We chose a deep convolution with pointwise projection and context transformation characteristics for the convolutional operations. We chose one variant of deep convolution, dynamic convolution, for the operation. Each convolutional submodule in the dynamic convolution contains convolutional kernels of different sizes, and it is called multi-scale dynamic convolution because of the different scales of the convolutional kernels. The convolutional kernel size k is calculated as shown in Formulas (20) and (21).
C o n v k ( O G M A ) = W o u t M S C o n v k ( V 2 )
V 2 = W v O G M A
where W V and W o u t are learnable hyperparameters and W V is a pointwise projecting transformation matrix. The projection operation performed on the input of multi-scale dynamic convolution V 2 is the same as V 1 , and MSConv represents the multi-scale dynamic convolution. The local self-learning sequence (LS) in the output of O G M A through multi-scale dynamic convolution is L S R ( n + 2 ) × d . The specific operation is shown in Formulas (22) and (23).
L S i , c = M S C o n v k ( O G M A )
M S C o n v k ( O G M A ) = j = 1 k ( softmax ( c = 1 d W j , c Q O G M A i , c ) · O G M A i + j k + 2 2 , c )
where d is the hidden layer size.
Shared projection: To represent the context sequence in the same hidden space, we shared the projections V 1 = W V V 1 and V 2 = W V O G M A . By sharing the projections, the global and local self-attention representations are mapped to the same hidden space. If two different projection matrices W 1 and W 2 are used, it is referred to as independent projection, i.e.,  V 1 = W V 1 V 1 .
Dynamically selected convolutional kernels: In order to dynamically and autonomously select the weights of the convolutional kernels of different scales, a gating mechanism was used. The specific operation is shown in Formula (24).
Conv O G M A = i = 1 n exp a i j = 1 n exp a j Conv k i O G M A
Then, we fused the global and local self-learned features to obtain the final label prediction vector O T R ( n + 2 ) × d , which was achieved through the following operation.
After the fusion of the global self-learning feature and the local self-learning representation, the final label prediction vector O T R ( n + 2 ) × d is obtained. The specific operation is shown in Formula (25):
O T = L S G S
where ⊕ represents matrix addition operation.

3.6. Decoding Layer

In the relation extraction task, we formalized the task as a query-based labeling problem. The softmax layer takes the label prediction vector O T through a multilayer perceptron and calculates the normalized probabilities of the BIOES entity labels, thus transforming the binary task of predicting the start and end indices into a five-classification task.
During training, we jointly trained the two stages of the objective function: the multiple answer task for subject extraction and the single-answer task for joint relation and object extraction. The two tasks are cascaded using a parameter-sharing strategy. The objective function is calculated as shown in Formulas (26)–(28).
L r e = λ L s u b j e c t + ( 1 λ ) L ( r e l a t i o n , o b j e c t )
L s u b j e c t = s T i log p s ( s | x i , q s )
L r e l a t i o n , o b j e c t = o T i | ( s , r ) log p o ( o | s , r , x i , q o )
where [ 0 , 1 ] is a hyperparameter that controls the weighting between the two objective functions. The experimental results showed that an α of 0.25 achieved a good balance and led to a good performance.
In the nested named entity recognition task, we used two n-class classifiers to predict the start and end indices. The objective function was calculated as shown in Formulas (29)–(32).
L n e r = L s t a r t + L e n d + α L m a t c h
L m a t c h = w i , w j E i P m a t c h ( p s t a r t , P e n d )
L s t a r t = w i E i p s t a r t ( w i | x i , q s )
L e n d = w j E i p e n d ( w j | x i , q s )
where L m a t c h , L s t a r t , and  L e n d represent the losses for start–end matching, start prediction, and end prediction, respectively. α [ 0 , 1 ] .

4. Experimental Section

4.1. Datasets

For the NER task, we conducted experiments on the ACE2004, ACE2005, and GENIA datasets. To ensure fairness, we followed the same data splitting as Yang et al. [35] and Chen et al. [36] for the ACE2004 and ACE2005 datasets, and followed Katiyar et al. [37] for the the GENIA dataset, which split the training, development, and test sets into a ratio of 8.1:0.9:1.0.
For the entity relation extraction task, we conducted experiments on the ACE2004 and ACE2005 relation datasets, which were collected from various domains such as news agencies and online forums by the Linguistic Data Consortium. To ensure fairness, we followed the same data splitting as Luan et al. [20], who split the ACE2005 dataset into training, development, and test sets and performed 5-fold cross-validation on the ACE2004 dataset.

4.2. Experimental Setups

We developed MAML-NET using Python and PyTorch. Two independent BERT-base-uncased [6] were used to encode the query and context. During training, AdamW [38] was used as the optimizer, with a learning rate of 2 × 10 5 , 20 epochs, and a batch size of 30. To regularize the model, different dropout rates were applied to different layers of MAML-NET, with the best setting ranging from 0.2 to 0.4. In addition, MAML-NET uses an early stopping strategy based on the validation loss. If the validation loss does not improve over the last 5 epochs, training will be terminated early. It should be noted that, on the GENIA dataset, biobert-base-cased-v1.2 [39] was used for encoding.

4.3. Named Entity Recognition

4.3.1. Baseline Models for Named Entity Recognition

(1) Straková et al. [23] viewed nested NER as a sequence-to-sequence problem, where each token in the input sequence is assigned to its corresponding entity and a recursive neural network is used to capture the nested entity structure. (2) Luan et al. [20] proposed a general method for information extraction using dynamically constructed span graphs to share span representations. By representing all possible entities as a dynamically constructed graph, entity representations can be shared and relationships between entities can be captured. (3) Li et al. [15] transformed the NER task into a QA task by treating named entities as questions and using the context as answers. By training the model with QA data during training, NER can be performed without labeled entities. (4) Yu et al. [40] used the idea of dependency tree parsing of graphs to score the beginning and ending tokens of entities in a sentence and used these scores to determine the boundaries of entities. (5) Hou et al. [41] injected semantic type word embeddings into entity embeddings to reduce differences in contextuality. This can help the model better distinguish entities and improve the performance of NER.

4.3.2. Experimental Results of Named Entity Recognition

The Table 3 shows the performance of MAML-NET on the ACE04, ACE05, and GENIA datasets relative to the previous state-of-the-art methods. Compared with the baseline model, MAML-NET improved the F1-score by 0.12%, 0.18%, and 0.32%, respectively. We speculate that this is because MAML-NET fully utilizes context and query information, not only to learn prior knowledge of named entities through the query, but also to exploit the interaction between the query and context, i.e., selectively reading the context based on the query and deepening the understanding of the query through reading the text. In addition, the model can alleviate the problem of information loss in the original sequence through the long-term-memory-gated attention mechanism. Finally, by repeatedly comprehensively grasping and studying the details of the text, the model achieved excellent results. However, Li et al. [15] only transformed the NER task into an MRC task and then processed the problem through a general deep neural network, which can only obtain prior knowledge of entities, and it is difficult to analyze the details and interaction between the text and questions. Yu et al. [40] and Straková et al. [23] did not use the query to transform the NER task into the MRC task, but learned the text information through RNNs, and then, the decoder recognized the entities based on this deep text information. We believe that, although their methods can avoid the problem of losing the original text information due to long queries, they also cut off the possibility of the model learning prior knowledge of entities in the query. However, MAML-NET can not only learn the deep text information, but also effectively alleviate the problem of similar word hidden features through the long-term-memory-gated network. Although Yu et al. [40] fused fine-grained semantic information into entity embedding to reduce uniqueness, we believe that this fusion strategy is not necessarily better than the text–query interaction strategy adopted in this paper. The results showed that our multi-scale attention mechanism and multi-scale self-learning mechanism were effective for named entity recognition tasks.

4.4. Baseline Models for Relation Extraction

We compared our approach with the following baselines: (1) Zhang et al. [24] transformed the relation extraction task into a table-filling problem. (2) Miwa et al. [22] proposed the first joint entity relation extraction method and used tree-based LSTMs to capture dependency information. (3) Straková et al. [23] replaced the tree-based structure with attentional LSTMs. (4) Sun et al. [25] used graph convolutional networks to address the problem of exploring entity relations. (5) Li et al. [16] proposed an MRC-based method. (6) Zhao et al. [28] extended Li et al.’s [16] method by constructing multiple queries for each context. (7) Shen et al. [27] memorized the entity and relationship category information learned in the task through the constructed memory module.

Experimental Results of Relation Extraction

Table 4 shows the performance of MAML-NET compared to state-of-the-art methods on the ACE2004 and ACE2005 datasets. It can be observed that MAML-NET outperformed all baseline models in terms of entity and relation extraction performance on both datasets. Specifically, MAML-NET improved the F1-scores of entity and relation extraction by 4.3% and 9.8%, respectively, on the ACE2004 dataset and by 2.0% and 0.9%, respectively, on the ACE2005 dataset. This indicated that MAML-NET performed better on the RE task compared to the NER task. We speculated that this is because MAML-NET can learn both the prior knowledge of named entities and entity relations in RE task. Furthermore, due to the presence of error propagation, the advantage obtained in the entity recognition stage was not simply transmitted to the relation extraction task, but continued to expand. Additionally, the MRC framework with multiple queries performed better than the single-query framework. We believe that this was because the multi-query framework allowed the model to more finely divide the RE task, rather than simply being a vague task such as “Find all relation”. However, neither the single-query framework nor the multi-query framework for RE models achieved the same level of performance as MAML-NET. We speculated that, on the one hand, although MAML-NET is based on a single-query framework; due to the different query design strategies, the performance of MAML-NET’s single-query framework was not inferior to the multi-query framework. On the other hand, unlike other models that only use BERT to learn the interaction information between query and context, MAML-NET has a dedicated interaction module, which can better utilize the implicit association information in the query and context.

4.5. Ablation Study

In order to investigate which aspects improved the performance of our model, we conducted an ablation study on the ACE2005 relation extraction dataset to understand the impact of each module on the model. We can observe the following results from Table 5.
Multi-scale attention (MSM): When the multi-scale self-learning network (MSM) is removed, compared to MAML-NET, the F1-score for entity and relation extraction decreased by 1.1% and 2%, respectively. We believe this is because the multi-scale self-learning layer can not only help the model understand the text as a whole, but also enable the model to discover important details relevant to the task, ultimately allowing the model to dynamically collect evidence from the entire paragraph to answer the question, resulting in improved performance in entity and relation extraction.
Long-term-memory-gated attention (LTMGA): When the long-term-memory-gated attention flow (LTMGA) was removed, compared to MAML-NET, the F1-score for entity and relation extraction decreased by 0.6% and 1.7%, respectively. This is reasonable because, although the interaction between the query and context can provide complementary information to each other, it also causes some loss of the original context information. The introduction of our long-term-memory-gated attention flow can alleviate this phenomenon, resulting in improved performance in both entity and relation extraction.
Interactive attention (IA): When using interactive attention, compared to MAML-NET, the F1-score for entity and relation extraction decreased by 1.6% and 2.9%, respectively. Compared to the joint encoding method (-IA-MSM-LTMGA), the F1-score for entity and relation extraction decreased by 4.8% and 3.5%, respectively. We believe this is because interactive attention can selectively collect relevant representations of other words in the source sequence during the word-encoding process based on the similarity between words in the text and the question. This not only alleviates the problem of encoding information similarity between different words, but also increases the distinguishability of encoding information between dissimilar words.

4.6. Case Study

To demonstrate the effectiveness of the MAML-NET model in improving entity and relation extraction by enhancing the reading comprehension ability, we present the examples shown in Table 6.
Table 6 compares the output of our MAML-NET with the previous state-of-the-art reading-comprehension-based relation extraction model (Multi-QA). In the first example, Multi-QA was unable to determine the triplets (units, ORG–AFF, republican guard) and (units, PHYS, capital). In the second example, Multi-QA was unable to determine the triplets (outlet, GEN–AFF, Milton Keynes) and (outlet, PART–WHOLE, easy). In the third example, Multi-QA was unable to determine the triplets (forces, PHYS, home) and (microbiologist, ART, home). From these examples, we observed that MAML-NET performed much better than Multi-QA when the subject or object participates in multiple relationship triplets at the same time. We speculated that, on the one hand, in NER tasks, the existence of nested named entities may cause some named entities to be unidentifiable if the model cannot understand the text from both the global and local perspectives. On the other hand, for RE tasks, if the model ignores certain key details, some relationships between entities will not be identified. This also demonstrates that MAML-NET has a good ability to process global information and local features.

5. Conclusions

This article proposed a multi-scale self-learning network based on multi-granularity attention to address the issue of homogeneity that arises with the increasing number of network layers in RE tasks, as well as the problem of the loss of the text’s original information in MRC tasks. Firstly, to effectively utilize the unique query information in MRC tasks, we proposed a multi-granularity attention mechanism, which allowed the model to learn the interaction information between the context and query in the source sequence through bidirectional attention and retained the task-related original context information through long-term-memory-gated attention flow. Then, to enable the model to consider both the global information and local features of the text, we proposed a self-learning mechanism to further improve the model’s understanding of the context. Finally, we optimized the model through joint learning. The experiment showed that our model was effective in entity relation extraction tasks.
Future research can be conducted in the following directions. First, although the introduced query can provide prior knowledge helpful for the task, the knowledge provided was still limited. Therefore, knowledge graphs or other knowledge can be used as prior knowledge for data augmentation. Second, the model still lacks entity recognition and relation extraction in long sentences. Therefore, multi-hop reading comprehension methods can be used to address entity recognition and relation extraction tasks in long sentences.

Author Contributions

Conceptualization, S.L.; methodology, W.S.; investigation, Z.J.; visualization, L.K.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Xinjiang Uygur Autonomous Region (No. 2019D01C060) and the National Natural Science Foundation of China (No. 61966034).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Guo, X.; Zhou, H.; Su, J.; Hao, X.; Tang, Z.; Diao, L.; Li, L. Chinese agricultural diseases and pests named entity recognition with multi-scale local context features and self-attention mechanism. Comput. Electron. Agric. 2020, 179, 105830. [Google Scholar] [CrossRef]
  2. Chen, X.; Zhang, N.; Li, L.; Yao, Y.; Deng, S.; Tan, C.; Huang, F.; Si, L.; Chen, H. Good visual guidance makes A better extractor: Hierarchical visual prefix for multimodal entity and relation extraction. arXiv 2022, arXiv:2205.03521. [Google Scholar]
  3. Yuan, L.; Cai, Y.; Wang, J.; Li, Q. Joint multimodal entity-relation extraction based on edge-enhanced graph alignment network and word-pair relation tagging. arXiv 2022, arXiv:2211.15028. [Google Scholar]
  4. Yang, S.; Tu, K. Bottom-up constituency parsing and nested named entity recognition with pointer networks. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, 22–27 May 2022; Muresan, S., Nakov, P., Villavicencio, A., Eds.; Association for Computational Linguistics: Toronto, ON, Canada, 2022; pp. 2403–2416. [Google Scholar]
  5. Lou, C.; Yang, S.; Tu, K. Nested named entity recognition as latent lexicalized constituency parsing. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, 22–27 May 2022; Muresan, S., Nakov, P., Villavicencio, A., Eds.; Association for Computational Linguistics: Toronto, ON, Canada, 2022; pp. 6183–6198. [Google Scholar]
  6. Devlin, J.; Chang, M.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional Transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  7. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. Roberta: A robustly optimized BERT pretraining approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  8. Sun, Y.; Wang, S.; Li, Y.; Feng, S.; Tian, H.; Wu, H.; Wang, H. Ernie 2.0: A continual pre-training framework for language understanding. arXiv 2019, arXiv:1907.12412. [Google Scholar] [CrossRef]
  9. Xu, Y.; Li, M.; Cui, L.; Huang, S.; Wei, F.; Zhou, M. Layoutlm: Pre-training of text and layout for document image understanding. In Proceedings of the KDD ’20: The 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual Event, CA, USA, 23–27 August 2020; Gupta, R., Liu, Y., Tang, J., Aditya Prakash, B., Eds.; ACM: New York, NY, USA, 2020; pp. 1192–1200. [Google Scholar]
  10. Xu, Y.; Xu, Y.; Lv, T.; Cui, L.; Wei, F.; Wang, G.; Lu, Y.; Florêncio, D.A.F.; Zhang, C.; Che, W.; et al. Layoutlmv2: Multi-modal pre-training for visually-rich document understanding. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, 1–6 August 2021; Association for Computational Linguistics: Toronto, ON, Canada, 2021; pp. 2579–2591. [Google Scholar]
  11. Sun, Y.; Wang, S.; Feng, S.; Ding, S.; Pang, C.; Shang, J.; Liu, J.; Chen, X.; Zhao, Y.; Lu, Y.; et al. Ernie 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation. arXiv 2021, arXiv:2107.02137. [Google Scholar]
  12. Wang, Q.; Dai, S.; Xu, B.; Lyu, Y.; Zhu, Y.; Wu, H.; Wang, H. Building chinese biomedical language models via multi-level text discrimination. arXiv 2021, arXiv:2110.07244. [Google Scholar]
  13. Ouyang, L.; Wu, J.; Jiang, X.; Almeida, D.; Wainwright, C.L.; Mishkin, P.; Zhang, C.; Agarwal, S.; Slama, K.; Ray, A.; et al. Training language models to follow instructions with human feedback. arXiv 2022, arXiv:2203.02155. [Google Scholar]
  14. Chen, Z.; Balan, M.M.; Brown, K. Language models are few-shot learners for prognostic prediction. arXiv 2023, arXiv:2302.12692. [Google Scholar]
  15. Li, X.; Feng, J.; Meng, Y.; Han, Q.; Wu, F.; Li, J. A unified mrc framework for named entity recognition. arXiv 2019, arXiv:1910.11476. [Google Scholar]
  16. Li, X.; Yin, F.; Sun, Z.; Li, X.; Yuan, A.; Chai, D.; Zhou, M.; Li, J. Entity-relation extraction as multi-turn question answering. In Proceedings of the Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019. [Google Scholar]
  17. Chen, N.; Liu, F.; You, C.; Zhou, P.; Zou, Y. Adaptive bidirectional attention: Exploring multi-granularity representations for machine reading comprehension. In Proceedings of the ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2020. [Google Scholar]
  18. Manning, C.D.; Surdeanu, M.; Bauer, J.; Finkel, J.R.; Bethard, S.; McClosky, D. The Stanford CoreNLP natural language processing toolkit. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, Baltimore, MD, USA, 23–24 June 2014. [Google Scholar]
  19. Straková, J.; Straka, M.; Hajič, J. Neural architectures for nested ner through linearization. arXiv 2019, arXiv:1908.06926. [Google Scholar]
  20. Luan, Y.; Wadden, D.; He, L.; Shah, A.; Osten-dorf, M.; Hajishirzi, H. A general framework for information extraction using dynamic span graphs. arXiv 2019, arXiv:1904.03296. [Google Scholar]
  21. Zelenko, D.; Aone, C.; Richardella, A. Kernel methods for relation extraction. J. Mach. Learn. Res. 2003, 3, 1083–1106. [Google Scholar]
  22. Miwa, M.; Bansal, M. End-to-end relation extraction using lstms on sequences and tree structures. arXiv 2016, arXiv:1601.00770. [Google Scholar]
  23. Katiyar, A.; Cardie, C. Going out on a limb: Joint extraction of entity mentions and relations without dependency trees. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, BC, Canada, 30 July–4 August 2017. [Google Scholar]
  24. Zhang, M.; Yue, Z.; Fu, G. End-to-end neural relation extraction with global optimization. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 9–11 September 2017. [Google Scholar]
  25. Sun, C.; Gong, Y.; Wu, Y.; Gong, M.; Duan, N. Joint type inference on entities and relations via graph convolutional networks. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019. [Google Scholar]
  26. Fu, T.J.; Li, P.H.; Ma, W.Y. Graphrel: Modeling text as relational graphs for joint entity and re-lation extraction. In Proceedings of the Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019. [Google Scholar]
  27. Shen, Y.; Ma, X.; Tang, Y.; Lu, W. A trigger-sense memory flow framework for joint entity and relation extraction. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 1704–1715. [Google Scholar]
  28. Zhao, T.; Yan, Z.; Cao, Y.; Li, Z. Asking effective and diverse questions: A machine reading comprehen-sion based framework for joint entity-relation extraction. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, Yokohama, Japan, 7–15 January 2021; pp. 3948–3954. [Google Scholar]
  29. Levy, O.; Seo, M.; Choi, E.; Zettlemoyer, L. Zero-shot relation extraction via reading comprehension. arXiv 2017, arXiv:1706.04115. [Google Scholar]
  30. McCann, B.; Keskar, N.S.; Xiong, C.; Socher, R. The natural language decathlon: Multitask learning as question answering. arXiv 2018, arXiv:1806.08730. [Google Scholar]
  31. Yang, P.; Cong, X.; Sun, Z.; Liu, X. Enhanced language representation with label knowledge for span extrac-tion. arXiv 2021, arXiv:2111.00884. [Google Scholar]
  32. Du, X.; Cardie, C. Event extraction by answering (almost) natural questions. arXiv 2020, arXiv:2004.13625. [Google Scholar]
  33. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30, Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  34. Seo, M.; Kembhavi, A.; Farhadi, A.; Hajishirzi, H. Bidirectional attention flow for machine comprehension. arXiv 2016, arXiv:1611.01603. [Google Scholar]
  35. Yang, B.; Mitchell, T.M. Joint extraction of events and entities within a document context. arXiv 2016, arXiv:1609.03632. [Google Scholar]
  36. Chen, Y.; Xu, L.; Kang, L.; Zeng, D.; Zhao, J. Event extraction via dynamic multi-pooling convolutional neural networks. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics(ACL2015), Beijing, China, 26–31 July 2015. [Google Scholar]
  37. Katiyar, A.; Cardie, C. Nested named entity recognition revisited. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), New Orleans, LA, USA, 1–6 June 2018. [Google Scholar]
  38. Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  39. Lee, J.; Yoon, W.; Kim, S.; Kim, D.; Kim, S.; So, C.H.; Kang, J. BioBERT: A pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 2020, 36, 1234–1240. [Google Scholar]
  40. Yu, J.; Bohnet, B.; Poesio, M. Named entity recognition as dependency parsing. In Proceedings of the Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020. [Google Scholar]
  41. Hou, F.; Wang, R.; He, J.; Zhou, Y. Improving entity linking through semantic reinforced entity embeddings. arXiv 2021, arXiv:2106.08495. [Google Scholar]
Figure 1. The process of recognizing entities by the named entity recognition model based on machine reading comprehension framework.
Figure 1. The process of recognizing entities by the named entity recognition model based on machine reading comprehension framework.
Sensors 23 04250 g001
Figure 2. The information expressed by the hidden features of words has changed from fine-grained representation to coarse-grained representation.
Figure 2. The information expressed by the hidden features of words has changed from fine-grained representation to coarse-grained representation.
Sensors 23 04250 g002
Figure 3. Generation of a candidate relationship triple. Under the action of the multi-scale self-learning mechanism, MAML-NET can extract not only the triples connected by yellow lines, but also the triples connected by two green lines. Without the effect of the multi-scale self-learning mechanism, only triples of yellow lines can be extracted.
Figure 3. Generation of a candidate relationship triple. Under the action of the multi-scale self-learning mechanism, MAML-NET can extract not only the triples connected by yellow lines, but also the triples connected by two green lines. Without the effect of the multi-scale self-learning mechanism, only triples of yellow lines can be extracted.
Sensors 23 04250 g003
Figure 4. Overview of multi-granularity attention multi−scale self−learning network.
Figure 4. Overview of multi-granularity attention multi−scale self−learning network.
Sensors 23 04250 g004
Figure 5. Diagram of interactive attention.
Figure 5. Diagram of interactive attention.
Sensors 23 04250 g005
Figure 6. A description of long-term-memory-gated attention.
Figure 6. A description of long-term-memory-gated attention.
Sensors 23 04250 g006
Table 1. Examples of generating queries based on program templates for different entity types. The blue font is the entity type to be filled.
Table 1. Examples of generating queries based on program templates for different entity types. The blue font is the entity type to be filled.
Entity TypeQuery
Facility (FAC)find all facility entities in the context.
Geopolitical (GPE)find all geopolitical entities in the context.
Location (LOC)find all location entities in the context.
Organization (ORG)find all organization entities in the context.
Person (PER)find all person entities in the context.
Vehicle (VEH)find all vehicle entities in the context.
Weapon (WEA)find all weapon entities in the context.
Table 2. Examples of generating question queries based on program templates for different relationship types, where e h e a d represents the header entity obtained through the entity extraction phase. The header entity is not necessarily a real subject.
Table 2. Examples of generating question queries based on program templates for different relationship types, where e h e a d represents the header entity obtained through the entity extraction phase. The header entity is not necessarily a real subject.
Relation TypeHead TypeTail TypeQuery
ARTGPEFACfind all facility entities in the context that
have an artifact relationship with geopolitical
entity e h e a d .
ARTGPEWEHfind all vehicle entities in the context that
have an artifact relationship with geopolitical
entity e h e a d .
PHYSFACFACfind all facility entities in the context that
have a physical relationship with facility
entity e h e a d .
PART–WHOLEGPEGPEfind all geopolitical entities in the context that
have a part–whole relationship with geopolitical
entity e h e a d .
PART–WHOLEGPELOCfind all location entities in the context that
have a part–whole relationship with geopolitical
entity e h e a d .
Table 3. Comparison of the performance of ACE2004, ACE2005, and GENIA. † indicates the results we achieved through their code. Bold numbers indicate the best results.
Table 3. Comparison of the performance of ACE2004, ACE2005, and GENIA. † indicates the results we achieved through their code. Bold numbers indicate the best results.
ACE2004
ModelPRF1
Straková et al. [19]--84.40
Luan et al. [20]--84.70
Yu et al. [40] †85.4285.9285.67
Li et al. [15] †86.3885.0785.72
MAML-NET86.8284.8885.84
ACE2005
ModelPRF1
Straková et al. [19]--84.33
Li et al. [15] †85.4884.3684.92
Yu et al. [40] †84.5084.7284.61
Hou et al. [41]83.9585.3984.66
MAML-NET85.2684.9585.10
GENIA
ModelPRF1
Straková et al. [19]--76.44
Li et al. [15] †79.6276.878.19
Yu et al. [40] †79.4378.3278.87
Hou et al. [41]79.4578.9479.19
MAML-NET (ours)79.6579.3779.51
Table 4. Performance comparison of ACE2004 and ACE2005. † and †† represent baselines based on the MRC framework, and †† represent baselines with multiple queries for the context. Bold numbers indicate the best results.
Table 4. Performance comparison of ACE2004 and ACE2005. † and †† represent baselines based on the MRC framework, and †† represent baselines with multiple queries for the context. Bold numbers indicate the best results.
DatasetModelEntityRelation
PRFPRF
ACE2004Miwa et al. [22]80.882.981.848.748.148.4
Straková et al. [23]81.278.179.646.445.345.7
Li et al. [16] †84.482.983.650.148.749.4
MAML-NET87.988.887.957.960.259.2
ACE2005Miwa et al. [22]82.983.983.457.254.055.6
Straková et al. [23]84.081.382.655.551.853.6
Zhang et al. [24]--83.5--57.5
Sun et al. [25]83.983.283.664.955.159.6
Li et al. [16] †84.784.984.864.856.260.2
Zhao et al. [28] †85.184.284.657.861.959.8
Zhao et al. [28] ††85.985.285.562.062.262.1
Shen et al. [27]86.787.587.662.263.462.8
MAML-NET89.588.989.669.458.863.7
Table 5. Ablation trials on ACE2005. Bold numbers indicate the best results.
Table 5. Ablation trials on ACE2005. Bold numbers indicate the best results.
DatasetModelEntityRelation
PRFPRF
ACE2005MAML-NET89.589.889.669.458.863.7
-MSL89.088.188.566.257.761.7
-LTMGA88.989.189.066.058.662.0
-MSM&LTMGA87.488.688.064.957.260.8
-IA&MSM&LTMGA84.784.984.864.856.260.2
Table 6. Comparing the MAML-NET model with multi-QA [16]. The triples in red represent the results extracted by MAML-NET.
Table 6. Comparing the MAML-NET model with multi-QA [16]. The triples in red represent the results extracted by MAML-NET.
Sentence 1… US officials say some intelligence indicates a red line may have been
drawn around the capital with republican guard units ordered to use
chemical weapons once US and allied troops cross it. …
Multi-QA((PER, units), ART, (WEA, weapons));
((PER, troops), ORG–AFF, (GPE, US)).
MAML-NET((PER, units), ART, (WEA, weapons));
((PER, troops), ORG–AFF, (GPE, US));
((PER, units), ORG–AFF, (ORG, republican guard));
((PER, units), PHYS, (GPE, capital)).
Sentence 2… The deadlock, and subsequent lack of any films, has been threatening
to de-rail the debut of easy Cinema s first outlet in Milton Keynes, just
north of London, which is due to open its doors on May 23. …
Multi-QA((GPE, Milton Keynes), PHYS, (GPE, London)).
MAML-NET((GPE, Milton Keynes), PHYS, (GPE, London));
((ORG, outlet), GEN–AFF, (GPE, Milton Keynes);
((ORG, outlet), PART–WHOLE, (ORG, easy)).
Sentence 3…And as part of that effort, US special forces today raided the home
of the Iraqi microbiologist known as Dr. Germ. …
Multi-QA((PER, forces), ORG–AFF, (GPE, US));
((PER, microbiologist), GEN–AFF, (GPE, Iraqi)).
MAML-NET((PER, forces), ORG–AFF, (GPE, US));
((PER, microbiologist), GEN–AFF, (GPE, Iraqi));
((PER, forces), PHYS, (FAC, home));
((PER, microbiologist), ART, (FAC, home)).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, W.; Liu, S.; Liu, Y.; Kong, L.; Jian, Z. Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning. Sensors 2023, 23, 4250. https://doi.org/10.3390/s23094250

AMA Style

Sun W, Liu S, Liu Y, Kong L, Jian Z. Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning. Sensors. 2023; 23(9):4250. https://doi.org/10.3390/s23094250

Chicago/Turabian Style

Sun, Weiwei, Shengquan Liu, Yan Liu, Lingqi Kong, and Zhaorui Jian. 2023. "Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning" Sensors 23, no. 9: 4250. https://doi.org/10.3390/s23094250

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop