Next Article in Journal
Structural Design of Aerostatic Bearing Based on Multi-Objective Particle Swarm Optimization Algorithm
Next Article in Special Issue
Joint Syntax-Enhanced and Topic-Driven Graph Networks for Emotion Recognition in Multi-Speaker Conversations
Previous Article in Journal
Interference Mitigation for Metro Train Location Based on Karhunen–Loeve Decomposition
Previous Article in Special Issue
A Personalized Multi-Turn Generation-Based Chatbot with Various-Persona-Distribution Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Temporal Extraction of Complex Medicine by Combining Probabilistic Soft Logic and Textual Feature Feedback

1
College of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan 430065, China
2
Key Laboratory of Rich-Media Knowledge Organization and Service of Digital Publishing Content, National Press and Publication Administration, Beijing 100038, China
3
Institute of Big Data Science and Engineering, Wuhan University of Science and Technology, Wuhan 430065, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2023, 13(5), 3348; https://doi.org/10.3390/app13053348
Submission received: 10 January 2023 / Revised: 27 February 2023 / Accepted: 5 March 2023 / Published: 6 March 2023
(This article belongs to the Special Issue Natural Language Processing (NLP) and Applications)

Abstract

:
In medical texts, temporal information describes events and changes in status, such as medical visits and discharges. According to the semantic features, it is classified into simple time and complex time. The current research on time recognition usually focuses on coarse-grained simple time recognition while ignoring fine-grained complex time. To address this problem, based on the semantic concept of complex time in Clinical Time Ontology, we define seven basic features and eleven extraction rules and propose a complex medical time-extraction method. It combines probabilistic soft logic and textual feature feedback. The framework consists of two parts: (a) text feature recognition based on probabilistic soft logic, which is based on probabilistic soft logic for negative feedback adjustment; (b) complex medical time entity recognition based on text feature feedback, which is based on the text feature recognition model in (a) for positive feedback adjustment. Finally, the effectiveness of our approach is verified in text feature recognition and complex temporal entity recognition experimentally. In the text feature recognition task, our method shows the best F1 improvement of 18.09% on the Irregular Instant Collection type corresponding to utterance l 17 . In the complex medical temporal entity recognition task, the F1 metric improves the most significantly, by 10.42%, on the Irregular Instant Collection type.

1. Introduction

In the medical field, multiple types of temporal information describe the occurrence of and state change in medical events, and it plays an essential role in conducting disease risk prediction [1,2], early detecton of diseases [3], disease trajectory tracking, adverse reaction detection, etc. [4,5]. Most of these temporal information representations are complex times with characteristics such as fuzziness and irregularity [6]. To realize the representation and reasoning of complex time in clinical texts, Hu et al. [7] proposed Clinical Time Ontology (CTO) and defined the semantic concepts of each type of complex time.
Temporal entity recognition originated in the sixth and seventh Message Understanding Conferences (MUC-6 [8] and MUC-7 [9]), mainly focusing on formal texts such as news articles [10,11] and later evolving to other domains [12,13]. The survey shows the importance of medical time entity identification in constructing timelines and clinical decision support systems, etc. [14]. However, medical texts contain many proper names and different habits of medical recorders [15], which results in the inability to directly apply time recognition methods from other fields to the medical area. With the release of shared tasks in medical Natural Language Processing (NLP), researchers have begun to focus on the study of medical temporal recognition [16,17]. The most commonly used temporal recognition methods are rule-based methods [18,19], statistical learning-based methods [20], and deep learning-based methods [21,22]. Rule-based methods perform recognition by manually generalizing the expression rules of temporal entities. Although they have high accuracy, they require much time and effort and have poor portability. Statistical learning-based methods perform recognition by selecting feature templates and sequence annotation sets, but the recognition effect depends more on the annotated corpus. Deep-learning-based methods can automatically learn features from the corpus to achieve entity recognition, which makes it more advantageous than the above two methods. However, these methods mainly target coarse-grained simple moments or periods, such as Date (e.g., 6 April 1994), Duration (e.g., two weeks), and Time (e.g., Thursday evening), with less attention to fine-grained complex time. If they are applied directly to fine-grained complex medical time recognition, their accuracy is low, which limits the utilization of medical information, such as aiding medical decision-making systems.
After analyzing the temporal text in electronic medical record data, we found the reason for the lower effectiveness of the existing models in identifying fine-grained complex temporal. The main reason is that there are more overlapping text features between different temporal types, and the text descriptions of some temporal types face the problem of a large span. The above can make it more challenging to identify complex temporal entities correctly. Specific examples are shown in Figure 1 and Figure 2.
In Figure 1, “2007年至2008年 (from 2007 to 2008)” is a text description of Absolute Interval (AIST), and “2007年至2008年初 (from 2007 to early 2008)” is a time description of a Fuzzy Interval (FITL). Since there are only weak differences in text composition between the above two time types, and the overlapping features are relatively large, the existing models often face the problem of type recognition errors.
In Figure 2, “2000年8月,2005年1月19日,2009年12月23日 (in August 2000, 19 January 2005, and 23 December 2009)” is a time description of Irregular Instant Collection (IISTC), which has a large span. Therefore, the existing models often identify four entities of Absolute Instants (AISTs).
To address the above issues, we develop specific rules for complex medical time and introduce them into the deep learning model through probabilistic soft logic (PSL) and text feedback mechanisms to extract complex medical time. After analyzing time text in the electronic medical record and the concept of complex time semantics in the CTO, we define seven basic features and eleven extraction rules and propose a complex medical time extraction method combining probabilistic soft logic and textual feature feedback (PSLF-CTR). The extraction method consists of two parts: (a) PSL-based Text Feature Recognition (PSL-TFR), which is based on probabilistic soft logic for negative feedback adjustment; (b) entity recognition, which is based on the trained text features in (a) for positive feedback adjustment.
In summary, our main contributions in this paper are as follows:
  • We realize the atomization and regularization of complex medical time texts. We split the features of medical time and propose seven basic features after analyzing the text of medical time. Then, exploring the textual concept of complex medical time in the CTO, we combine the basic features and first-order predicates and propose eleven extraction rules.
  • We apply the PSL to text feature recognition. The PSL, a first-order logical predicate probabilistic inference framework, is combined with rules defining the correspondence of each time type to obtain the satisfaction distance corresponding to eleven types of time. The distance is used as a feedback factor in the model to improve the effectiveness of the text-feature-recognition model.
  • We add a text feedback mechanism to the entity recognition model and validate the results of the entity recognition model on the trained text feature model. A positive feedback factor is added based on the validation results, improving the complex time entity recognition model.

2. Related Work

2.1. CTO

The CTO defines eleven types of time in the medical domain, including simple time and complex time [7]. Among them, time with precision and absoluteness is a simple time, such as Absolute Instant, Absolute Interval, and time with characteristics such as ambiguity and repetition is a complex time. Of the eleven types of time in CTO, the remaining nine time types other than the two simple times are complex times, and the specific classification is shown in Table 1.

2.2. Time Entity Recognition

Temporal entity recognition is one of the essential tasks in medical named entity recognition [7], and scholars at home and abroad have conducted a series of studies on it. For English texts, Chang et al. [23] proposed SUTime, a time tagger for identifying and normalizing time expressions in English texts. It obtains time expressions using combining and filtering strings. Strotgen et al. [24] proposed HeidelTime, a processing system with multilingual features. This system introduces lexical annotation and manual setting rules for recognizing and normalizing temporal expressions. These approaches are temporal annotators implemented by designing rules using a fixed method that requires more labor, so subsequent researchers started to focus on ways to extend or automatically be able to generate rules. Zhong et al. [25] implement the lightweight time tagger SynTime. It is mainly based on boundary expansion and uses heuristics to design generic rules. Although it saves labor costs somewhat, it assumes that the words are all correctly labeled, which would otherwise significantly affect its results. And then, with the advent of machine learning, its models started to be applied in this field. Moharasar et al. [26] proposed a semi-supervised framework for detecting temporal expressions in clinical texts to enhance the accuracy of temporal entity recognition through a large number of unannotated texts. Ding et al. [27] propose a pattern-based approach PTime, which automatically generates and selects patterns for identifying temporal expressions. MacAvaney et al. [28] proposed conditional random field and decision tree integration, which uses lexical, syntactic, semantic, distributional, and rule-based feature approaches to implement a cross-domain clinical temporal-information-extraction framework. With the development of deep learning, researchers have started to apply it to the field. Hossain et al. [29] proposed a temporal information extraction system using long short-term memory (LSTM) recurrent neural networks (RNN) and word embedding, which achieved good results on the TempEval-2 dataset. Patra et al. [30] proposed a model that combines rules and neural models that identify relevant date-time entities and their associated negation constraints for a specific task. Li et al. [31] proposed a method for extracting clinical event expressions and corresponding temporal information based on the architecture of recurrent neural networks. These methods are relatively mature based on English text, but because of the grammatical differences between Chinese and English [32], they do not work well in Chinese text recognition.
For Chinese text, Lin et al. [33] proposed a recognition method based on regular expressions. Still, developing rules is time-consuming and labor-intensive and has poor domain adaptation. Later, Wu et al. [34] proposed combining conditional random fields with a temporal lexicon for temporal expression recognition. However, it relies on a temporal trigger lexicon and is less effective for time recognition without trigger words. With the emergence of deep learning methods, Liu et al. [35] used a BiLSTM-CRF approach to extract temporal information in geological texts. Kai et al. [36] used the BERT-BiLSTM-CRF model to extract temporal information from labeled media messages. These two extraction methods regarding geological time have significant strengths in their fields and strong domain specificity. Zhu et al. [37] proposed a Chinese temporal expression recognition method based on the BERT-FLAT-CRF model. It uses BERT to enhance word vector expressions, FLAT to fuse temporal lexical features, and CRF to extract optimal sequence annotation. However, these temporal entity recognition methods focus more on coarse-grained simple time and ignore fine-grained complex time. Fine-grained complex temporal descriptive text features are difficult to distinguish, so recognizing them is challenging. Therefore, we introduce a logical framework to improve the recognition of text features.
Probabilistic Soft Logic (PSL) [38] is a probabilistic reasoning framework that uses first-order logic to represent complex relationships succinctly and uses soft truth values between [0, 1] as operation values. Thus, it can capture the uncertain rows and incompletion inherent in real-world knowledge [39]. In recent years, it has been applied to various fields such as sentiment classification [39], collective classification [40], embedding uncertain knowledge graphs [41], and joint estimation of user and publisher feasibility in fake news detection [42].
Therefore, we introduce PSL, a first-order logic framework for probabilistic reasoning, into the deep learning approach. We embed the probability calculation results into a deep learning model through a text feedback mechanism, which is used to better discriminate text features of complex medical temporal and achieve complex temporal recognition.

3. Methodology

Based on the CTO, we define seven kinds of atomic statements and eleven kinds of extraction rules r 1 r 11 for complex medical time, the detailed definitions of which are shown in Section 3.1. By combining them with PSL, we realize the integration of logic rules in entity recognition tasks to further improve the accuracy of complex time recognition. Therefore, we design a complex medical time extraction with PSL and text feature feedback, whose overall framework is shown in Figure 3. The model mainly consists of two parts: text feature recognition and entity recognition. Text feature recognition (PSL-TFR) adjusts the model with negative feedback by using the calculation results of the PSL program as a negative feedback factor. The entity recognition (PSLF-CTR) adjusts itself with positive feedback by using the validation results of the text feature recognition model as a positive feedback factor. The specific process is as follows.
(1) Training of the PSL-TFR model: The processed electronic medical record data are input into the text feature recognition model. The obtained results are fed into a probabilistic soft logic program for calculation, in which the PSL approach is introduced to calculate the corresponding satisfaction distance d r according to the rules r 1 r 11 . This d r is used as a feedback factor to train the PSL-TFR model with negative feedback, thus improving the model’s ability to recognize temporal text features.
(2) Training of the PSLF-CTR model: The electronic medical record information is input into the entity recognition model. Then, the obtained text information is input into the trained PSL-TFR model to obtain the prediction results of the complex time text feature. The data with labeling results are consistent with the prediction results identified for positive feedback training, improving the entity recognition capability.
The following section describes the details of the calculation of d r and the feedback regulation.

3.1. Basic Extraction Rules for Complex Medical Time

Textual descriptions of time usually have multiple textual features, so there are still some challenges in determining the time type corresponding to time text. Based on the definitions of each type of time in terms of semantics, we extract seven basic semantic features and design eighteen atomic statements, as shown in (1)–(18). Among them, (1)–(7) are used for the basic text features, and (8)–(18) describe the conclusion.
For the “relativity” of time, if there are modifiers with relative meaning in the sentence, such as “之前(before)” or “之后(after)”, then the time expressed in the sentence is “relative”. Let “is relative ()” be the predicate and T be the time text; then, the atomic statement can be expressed as follows:
l 1 = i s r e l a t i v e T
For the “ fuzziness” of time, if there are modifiers with fuzzy meaning in the sentence, such as “大约(about)” or “大概(around)”, then the time expressed in the sentence is “fuzzy”. Let “is fuzzy ()” be the predicate and T be the time text; then, the atomic statement can be expressed as follows:
l 2 = i s f u z z y T
For the “periodicity” of time, if there are modifiers with periodic meaning in the sentence, such as “每天(every day)” or “每周(every week)”, then the time expressed in the sentence is “periodic”. Let “is periodic ()” be the predicate and T be the time text; then, the atomic statement can be expressed as follows:
l 3 = i s p e r i o d i c T
For the “incompleteness” of time, if there is a modifier with incomplete meaning in the sentence, such as “时间不详(time unknown)”, then the time expressed in the sentence is “incomplete”. Let “is incomplete ()” be the predicate and T be the time text; then, the atomic statement can be expressed as follows:
l 4 = i s i n c o m p l e t e T
The “continuity” of time often exists in the form of a period. If there is a modifier in the sentence that has the meaning of a period, such as “至(to)” or “持续(lasting)”, then the time expressed in the sentence is “interval”. Let “is an interval ()” be the predicate and T be the time text; then, the atomic statement can be expressed as follows:
l 5 = i s a n i n t e r v a l T
The “non-continuity” of time often exists in the form of moments. Since its textual features are not obvious and there is mutual exclusivity between moments and periods, it can be expressed as the negation of the predicate “is an interval ()”. Let “is instant()” be the predicate and T be the time text; then, the atomic statement can be expressed as follows:
l 6 = i s i n s t a n t T = ¬ l 5
For the “discrete” nature of time, if there are modifiers with multiple meanings in the sentence, such as “第一次(first time)” or “多次(many times)”, then the time expressed in the sentence is “multiple”. Let “is multiple ()” be the predicate and T be the time text; then, the atomic statement can be expressed as follows:
l 7 = i s m u l t i p l e T
Since logic rules often consist of h e a d and b o d y , the part h e a d is the conclusion that determines the type of time. It needs to be expressed using atomic statements. The atomic statements for describing the conclusion are defined as follows:
l 8 = A b s o l u t e I n s t a n t T
l 9 = R e l a t i v e I n s t a n t T
l 10 = F u z z y I n s t a n t T
l 11 = A b s o l u t e I n t e r v a l T
l 12 = R e l e I n t e r v a l T
l 13 = F u z z y I n t e r v a l T
l 14 = I n c o m p l e t e I n t e r v a l T
l 15 = P e r i o d i c I n s t a n t C o l l e c t i o n T
l 16 = P e r i o d i c I n t e r v a l C o l l e c t i o n T
l 17 = I r r e g u l a r I n s t a n t C o l l e c t i o n T
l 18 = I r r e g u l a r I n t e r v a l C o l l e c t i o n T
Let R l denote the set of original statements. According to the definition of time types, different time types can be defined logically using the atomic statement l i and the analytic (∨), combined (∧), and negation (¬) symbols. The definitions are shown below.
Regarding the judgment of AIST, the temporal text should satisfy “is instant(T)”, “is not relative(T)”, “is not fuzzy(T)”, and “is not multiple(T)”. Thus, the logical rule can be defined as follows:
r 1 : l 6 ¬ l 1 ¬ l 2 ¬ l 7 l 8
Regarding the judgment of RIST, the temporal text should satisfy “is instant(T)”, “is relative(T)”, “is not fuzzy(T)”, and “is not multiple(T)”. Thus, the logic rule can be defined as follows:
r 2 : l 6 l 1 ¬ l 2 ¬ l 7 l 9
Regarding the judgment of FIST, the temporal text should satisfy “is instant(T)”, “is fuzzy(T)”, and “is not multiple(T)”. Thus the logic rule can be defined as follows:
r 3 : l 6 l 2 ¬ l 7 l 10
Regarding the judgment of AITL, the temporal text should satisfy “is an interval(T)”, “is not relative(T)”, “is not fuzzy(T)”, “is not incomplete(T)”, and “is not multiple(T)”. Thus the logic rule can be defined as follows:
r 4 : l 5 ¬ l 1 ¬ l 2 ¬ l 4 ¬ l 7 l 11
Regarding the judgment of RITL, the temporal text should satisfy “is an interval(T)”, “is relative(T)”, “is not fuzzy(T)”, and “is not incomplete(T)” and “is not multiple(T)”. Thus the logic rule can be defined as follows:
r 5 : l 5 l 1 ¬ l 2 ¬ l 4 ¬ l 7 l 12
Regarding the judgment of FITL, the temporal text should satisfy “is an interval(T)”, “is relative(T)”, “is fuzzy(T)”, “is not incomplete(T)”, and “is not multiple(T)”. Thus the logic rule can be defined as follows:
r 6 : l 5 l 2 ¬ l 4 ¬ l 7 l 13
Regarding the judgment of IITL, the temporal text should satisfy “is an interval(T)”, “is incomplete(T)”, and “is not multiple(T)”. Therefore the logic rule can be defined as follows:
r 7 : l 5 l 4 ¬ l 7 l 14
Regarding the judgment of PISTC, the temporal text should satisfy “is instant(T)”, “is multiple(T)”, and “is periodicity(T)”. Therefore the logic rule can be defined as Equation (26), and by the same token, the logic rule of PITLC can be defined as shown in (27).
r 8 : l 7 l 6 l 3 l 15
r 9 : l 7 l 5 l 3 l 16
Regarding the judgment of IISTC, the temporal text should satisfy “is instant(T)”, “is multiple(T)”, and “is not periodicity(T)”. Thus, the logic rule is shown in Equation (28) and, by the same token, the IITLC are defined as shown in Equation (29).
r 10 : l 7 l 6 ¬ l 3 l 17
r 11 : l 7 l 5 ¬ l 3 l 18

3.2. Feedback Factor Calculation Based on Probabilistic Soft Logic

Probabilistic soft logic is a framework based on first-order logic rules for probabilistic reasoning. The difference is that it uses the [0, 1] interval of soft truth values of random variables, not only the extreme values 0 and 1 [39]. The following will introduce the relevant concepts of PSL:
Definition 1.
Given a set of atomic statements l ˜ , I ( l ) denotes the soft truth value that satisfies the atomic statement l.
Definition 2.
A rule r ˜ usually consists of a disjunctive clause for an atomic statement or a negative atomic statement, for example η r : B 1 B 2 B m H 1 H 2 H n , where B 1 B 2 B m is the body r b o d y of the rule, H 1 H 2 H n is the head r h e a d of the rule, and η r [ 0 , 1 ] is the weight of the rule r, indicating the prior confidence of the rule.
Definition 3.
A specific atomic statement l and a specific rule r are instantiations of l ˜ and r ˜ , such as l 10 , r 3 : l 6 l 2 ¬ l 7 l 10 .
Definition 4.
Define the computation of the base logic operator in PSL according to the L u k a s i e w i c z t equation:
I ( l 1 l 2 ) = min I l 1 + I l 2 , 1
I ( l 1 l 2 ) = max I l 1 + I l 2 1 , 0
I ¬ l 1 = 1 I l 1
I r b o d y r h e a d = I ¬ r b o d y r h e a d
Definition 5.
According to the interpretation I, d r ( I ) denotes the satisfaction distance of rule r. The calculation is as follows.
d r I = max I r b o d y I r h e a d , 0
Definitions 1–5 describe how PSL calculates the soft truth values and satisfiability distances of compound rules in detail. PSL provides theoretical support for calculating the satisfiability probabilities of compound logic rules.
Combine the probability that each atomic statement in l 1 l 18 is satisfied and its corresponding rule r. After the logical operation defined in PSL, the distance of satisfaction of this time atomic statement, the formula is shown in (35). If d r > 0, the model predictions are contradictory, and the model needs to be adjusted by negative feedback, using d r as a feedback factor to amplify its loss.
d r = max I r c b o d y I r c h e a d , 0 = max I r c b o d y I r c o r r e c t , 0

3.3. Text Feature Feedback Mechanism

3.3.1. Negative Feedback Mechanism

The PSLF-CTR model incorporates the logical rules related to temporal text representation into the machine learning model. Based on the BERT-TextCNN model, we designed the PSL-TFR model to improve the model’s accuracy in recognizing each type of temporal text feature. Figure 4 shows its structure.
The training process of PSL-TFR adds a negative feedback adjustment mechanism to the Bert-TextCNN model. The set of probabilities P s e t = P l j can be obtained after the Softmax layer that the text feature satisfies for the atomic utterance l 1 l 18 . The formula is shown in (36).
P l j = e l j i m e l j
where m is the number of atomic statements and takes the value of eighteen. The obtained P s e t and the corresponding r c to the actual type are input to the PSL calculation program, which in turn yields each atomic statement d r . If d r > 0, the model prediction results are contradictory, and the model needs to be adjusted by negative feedback, using d r as a feedback factor to expand its loss. Since the text feature recognition method incorporates logic rules for negative feedback adjustment, the loss function of the new model is based on the Bert-TextCNN model with the addition of the feedback factor d r . The calculation method is shown in (37):
L = y log y ^ + 1 y log 1 y ^ + λ d r
where λ is the feedback coefficient, taking values in the range (0, 5], indicating the weight of the negative feedback factor.

3.3.2. Positive Feedback Mechanism

Conventional named-entity-recognition models usually encode the text first and then tag the text using CRF methods. Due to the excellent encoding capability of the BERT model, the vectorized representation of text is achieved using the BERT model. In addition, considering the bidirectional semantic features of character text, the BiLSTM model is used to learn the semantic information embedded in the text. Finally, the CRF method is used to label all characters. The structure of PSLF-CTR, an entity-recognition model based on text feature feedback, is shown in Figure 5.
The model training process is based on the Bert-BiLSTM-CRF model, adding a positive-feedback adjustment mechanism. Let the model have the temporal type T y p e i based on the text and decode the label information of the input sentence after the CRF model. The probability is obtained that the input character x is labeled T y p e i , and its formula is shown in (38):
P T y p e | x = 1 Z x exp h T y p e 1 | x + k = 1 n 1 g T y p e k , T y p e k + 1 + h T y p e k + 1 | x
where Z ( x ) is the normalization factor, h is the emission matrix, and g T y p e k , T y p e k + 1 denotes the transfer matrix of adjacent positions. The text is input to the PSL-TFR model, and then the satisfaction of the text for complex temporal features is obtained. If the labeling result agrees with the feature recognition result, the model is rewarded with feedback, and the loss is reduced with α as the feedback factor. The loss function of the BERT-BiLSTM-CRF model is shown in (39). After fusing the temporal text features, the model’s loss function is shown in (40).
L o r i g i n a l = log P T y p e 1 , T y p e 2 , , T y p e n x
L = L o r i g i n a l α

4. Experimental Results and Discussion

4.1. Datasets

In this paper, the model is labeled and trained on a private dataset and validated on a public dataset. The private dataset is derived from the clinical desensitized data of the partner hospital. The processed sample is given in the link https://github.com/wangdaven/EMR.git (accessed on 27 October 2022), because the hospital requires it not to be made public. The public dataset is derived from The China Conference on Knowledge Graph and Semantic Computing (CCKS) 2020 data. Since both datasets do not contain complex temporal-type annotations, we invited three annotators to manually evaluate the model-identification results. For the annotation, we annotated 600 private EHRs and 300 open-sourced EHRs in CCKS using the BIO annotation method. On average, each private EHR contains 604 characters, and each public EHR includes 441 characters. The number of each time-type label in the data in both datasets is shown in Figure 6.
The BIO annotation method is to add a label to each word in the sentence. This label consists of two parts: (a) the location of the entity to which the word belongs, where B indicates that the word is the first word of the entity, I indicates that the word is the middle word of the entity, and O indicates that it is not an entity; (b) the entity type corresponding to the word, such as the entity types AITL and PITLC.

4.2. Experimental Settings and Metrics

In this paper, we use the large-scale Chinese word vector data open-sourced by Tencent AI-LAb Lab in 2018 with a dimension of 200, so we set the Embedding-Dim of all models to 200 and the Optimizer to Adam. All experiments were run on a machine equipped with a Tesla P-100 PCIE graphics card and an Intel(R) Xeon(R) CPU E5-2600 v4 @ 2.00GHz model CPU. Through several experiments, the results achieved by the following parameter settings were found to be better under the available computing resources, as shown in Table 2, where the effects of the values of the negative feedback coefficient λ in Equation (37) and the positive feedback factor α in Equation (40) on model F1 are shown in Figure 7.
Named entity identification includes both entity boundaries and entity types, so “Exact-match Evaluation” is used to evaluate the validity of the model [43]. This method recognizes an entity as correct only if both entity boundaries and types are true [44]. It involves three metrics: precision (P), recall (R), and F1 and is calculated as shown in (41)–(43).
P = T P / ( T P + F P )
R = T P / ( T P + F N )
F 1 = P × R × 2 / ( P + R )
True Positive (TP): The number of entities that are predicted to be positive samples and are actually positive samples.
True Negative (TN): The number of entities that are predicted to be negative samples and are actually negative samples.
False Positive (FP): The number of entities that are predicted to be positive samples but are actually negative samples.
False Negative (FN): The number of entities that are predicted to be negative samples but are actually positive samples.
Precision (P) represents the accuracy of the prediction in the positive sample results; a larger value indicates that more of the predicted samples are correct. Recall (R) represents the probability of predicting a positive sample among the actual positive samples; a larger value indicates that more of the actual positive samples are predicted. The F1 metric is the average of precision and recall, and a larger value indicates better model quality.

4.3. Results and Analyses

4.3.1. Comparison Experiments

PSLF-CTR performs positive feedback conditioning on itself, which is based on the recognition results of the PSL-TFR model, and uses the PSL calculation results as its feedback factor during the training of the PSL-TFR model. We designed comparison experiments between PSLF-CTR and BERT+FLAT+CRF and BERT+BiLSTM+CRF models in a private dataset to evaluate the competitiveness of the PSLF-CTR model in the temporal entity recognition task.
(1) BERT+BiLSTM+CRF model is the baseline model of this paper. It is a BiLSTM+CRF sequence annotation model using Google’s pre-trained BERT for word embedding.
(2) The BERT+FLAT+CRF model uses BERT to boost word vector expressions and FLAT to fuse temporal lexical features, and it finally extracts the optimal temporal expressions by CRF.
We explore the impact of complex temporal text features by comparing the three experiments on complex temporal entity recognition. Finally, we evaluate the ability that uses the PSLF-CTR model to recognize complex temporal entities on a publicly available dataset. The results are shown in Table 3 and Table 4.
Table 3 shows that overall, the PSLF-CTR model had better recognition than the baseline and BERT+FLAT+CRF models. It had the highest F1 in each type of time recognition except for the AIST. For the recognition of AIST, the BERT+FLAT+CRF model performed the best, and the PSLF-CTR model had less than 1% improvement in F1 over the baseline model. The analysis of the samples shows that the main reasons are the large sample size, the obvious text features, and the short text.
For the recognition of complex times such as FIST, FITL, PISTC, IISTC, PITLC, and IITLC, the recognition effect of the PSLF-CTR model was improved more obviously, especially for the recognition of IISTC, the F1 improvement reached 10.42%. The baseline model was less effective in recognizing these six types of time, with F1 below 75% and even a PITLC below 50%. After analyzing their texts, we found the reason is that these six types of time all involve more complex temporal text features and fewer occurrences of IISTC and PITLC entities. Although the recognition effect of the BERT-FLAT-CRF model was better than that of the baseline model, there was a slight improvement compared to the baseline model for only individual lexical differences (e.g., FITL). The reason for the better performance of the PSLF-CTR model in combination with Table 5 is that the PSL-TFR model is sufficient to obtain a high accuracy rate in predicting the atomic statement l 2 (“ i s f u z z y T ”) and the FIST, and FITL corresponds to the atomic statement l 10 , l 13 , whose logic rule is r 3 : l 6 l 2 ¬ l 7 l 10 , r 6 : l 5 l 2 ¬ l 4 ¬ l 7 l 13 , so the PSLF-CTR model can receive good feedback from PSL-TFR during the training process, thus improving the accuracy rate of identifying this type of entity. For PISTC, IISTC, PITLC, and IITLC, since the PSL-TFR model also has good accuracy in identifying atomic statements l 7 (“ i s m u l t i p l e T ”) and l 3 (“ i s p e r i o d i c T ”), the combination of the corresponding rules r 8 : l 7 l 6 l 3 l 15 , r 9 : l 7 l 5 l 3 l 16 , r 10 : l 7 l 6 ¬ l 3 l 17 and r 11 : l 7 l 5 ¬ l 3 l 18 , etc. makes the PSLF-CTR model perform better than the original model in identifying complex time-type entities.
After adequately training PSLF-CTR models, we randomly selected 100 medical record data from the 300 in the public dataset for evaluation in this paper. The recognition results are shown in Table 4. For the AIST and RIST, the identification results of all three models were better, reaching more than 90%. Analyzing the data shows that the reason is a large number of samples and more obvious features. For other types of time, although the BERT-FLAT-CRF model was better than the baseline model for most of the time types, it even had the highest F1 value for RITL and IITL. However, it was less effective than the PSLF-CTR model in recognizing complex temporal entities with longer text lengths and larger feature spans (e.g., IITLC and IISTC). Among them, the PITLC had a special recognition effect, marked as 50% in all three models. The analysis shows that the reason is that there are fewer temporal entities of this type, so the experimental results obtained are subject to chance.
From Table 3 and Table 4, PSLF-CTR significantly improves recognition of complex temporal entities compared to the baseline model BERT+BiLSTM+CRF and performs worse in recognition of some types of entities compared to the BERT-FLAT-CRF model. However, compared to the overall, PSLF-CTR performs better in recognizing all types of temporal entities, especially complex ones.

4.3.2. Ablation Study

To explore the effect of logical rules on the model recognition of complex temporal text features, we design the ablation experiments on a private dataset containing PSL-TFR and BERT-TextCNN models. The experimental results for l 1 l 18 recognition are shown in Table 5 and Table 6.
In Table 5, it can be seen that the difference between the two models is not significant since l 1 l 7 does not involve logic rules, and the negative feedback mechanism has limited influence on the model. The PSL-TFR scores the highest for the recognition of text features corresponding to l 1 atomic utterances and the lowest for the l 4 atomic utterances, in terms of precision, recall, and F1 metrics. Except for l 4 , PSL-TFR obtained high scores in precision (P). However, in recall (R), PSL-TFR performed differently for different text feature recognition and performed poorly in recognizing l 4 , l 7 . By analyzing the experimental results, we found that the number of texts satisfying the features of “ i s i n c o m p l e t e T ” and “ i s m u l t i p l e T ” is small in the dataset, which is not conducive to the model capturing the text features.
The results of l 8 l 18 atomic statements are shown in Table 6. The model without using any logical rules for constraints performs poorly overall in recognition of complex temporal text features and only performs well in recognizing the corresponding text features corresponding to l 8 and l 11 . The F1 score of BERT-TextCNN does not exceed 50% in recognizing text features corresponding to l 17 and l 18 . The PSL-TFR model improves in recognizing l 8 l 18 all the corresponding text features and achieves a larger improvement in l 10 , l 12 , l 14 l 18 feature recognition (complex time types such as RITL, FIST, and FITL). By analyzing the results of the two comparison experiments, we found that the BERT-TextCNN model often identifies FIST as AIST. At the same time, there are fewer such errors in the PSL-TFR model, which reduces the value of FP and thus improves the precision (P). Moreover, the precision (P) of l 16 is improved more, and the analysis of the results shows that the same value of FP is reduced, but the precision (P) index is improved more because it is less numerous. Overall, Table 6 verifies that using the text feature recognition model PSL-TFR with fused logic rules has better complex temporal text feature recognition capability.

5. Conclusions

In this paper, we propose a complex medical time recognition method combining PSL and text feedback mechanisms. First, we define seven basic features and eleven extraction rules based on the characteristics of complex medical time. Then these eleven rules are incorporated into the deep learning model through PSL and a text feedback mechanism to improve its ability to recognize complex medical time. Finally, the effectiveness of this method is verified by comparison experiments, and the PSL is proved to have an enhancing effect on text feature recognition using ablation experiments.
There are several limitations in this study, and in the future, we will continue our in-depth analysis to address the existing problems:
  • The data used are not a publicly available normalized dataset. Therefore, we consider how to construct a Chinese medical time dataset and achieve normalized annotation of temporal entities and temporal relationships while protecting privacy.
  • The model in this paper is only for a Chinese dataset. We will verify whether this method can be used for other languages and explore multilingual temporal recognition models in the future.
  • This paper only uses feedback adjustment to adjust the model after the model obtains each training result. Another promising research direction is to explore embedding the rules in the model during the model training.

Author Contributions

Conceptualization, J.G. and D.W.; methodology, J.G. and D.W.; formal analysis, J.G. and F.X.; investigation, F.X.; resources, J.G.; data curation, D.W. and D.H.; writing—original draft preparation, D.W.; writing—review and editing, J.G., D.W. and F.G.; supervision, J.G. and F.G.; funding acquisition F.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China grant number U1836118, Key Laboratory of Rich Media Digital Publishing, Content Organization and Knowledge Service grant number ZD2022-10/05, and National key research and development program grant number 2020AAA0108500.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ben-Assuli, O.; Jacobi, A.; Goldman, O.; Shenhar-Tsarfaty, S.; Rogowski, O.; Zeltser, D.; Shapira, I.; Berliner, S.; Zelber-Sagi, S. Stratifying individuals into non-alcoholic fatty liver disease risk levels using time series machine learning models. J. Biomed. Inform. 2022, 126, 103986. [Google Scholar] [CrossRef] [PubMed]
  2. Abbasimehr, H.; Paki, R.; Bahrini, A. A novel approach based on combining deep learning models with statistical methods for COVID-19 time series forecasting. Neural Comput. Appl. 2022, 34, 3135–3149. [Google Scholar] [CrossRef] [PubMed]
  3. Khanday, A.M.U.D.; Rabani, S.T.; Khan, Q.R.; Rouf, N.; Mohi Ud Din, M. Machine learning based approaches for detecting COVID-19 using clinical text data. Int. J. Inf. Technol. 2020, 12, 731–739. [Google Scholar] [CrossRef] [PubMed]
  4. Hettige, B.; Wang, W.; Li, Y.; Le, S.; Buntine, W.L. MedGraph: Structural and Temporal Representation Learning of Electronic Medical Records. In Proceedings of the ECAI 2020—24th European Conference on Artificial Intelligence, Santiago de Compostela, Spain, 29 August–8 September 2020; pp. 1810–1817. [Google Scholar]
  5. Lee, W.; Kim, G.; Yu, J.; Kim, Y. Model Interpretation Considering Both Time and Frequency Axes Given Time Series Data. Appl. Sci. 2022, 12, 12807. [Google Scholar] [CrossRef]
  6. Hauskrecht, M.; Liu, Z.; Wu, L. Modeling Clinical Time Series Using Gaussian Process Sequences. In Proceedings of the 13th SIAM International Conference on Data Mining, Austin, TX, USA, 2–4 May 2013; pp. 623–631. [Google Scholar]
  7. Hu, D.; Wang, M.; Gao, F.; Xu, F.; Gu, J. Knowledge Representation and Reasoning for Complex Time Expression in Clinical Text. Data Intell. 2022, 4, 573–598. [Google Scholar] [CrossRef]
  8. Grishman, R.; Sundheim, B. Message Understanding Conference-6: A Brief History. In Proceedings of the 16th International Conference on Computational Linguistics, COLING 1996, Center for Sprogteknologi, Copenhagen, Denmark, 5–9 August 1996; pp. 466–471. [Google Scholar]
  9. Chinchor, N. Appendix E: MUC-7 Named Entity Task Definition (Version 3.5). In Proceedings of the Seventh Message Understanding Conference, MUC 1998, Fairfax, VA, USA, 29 April–1 May 1998. [Google Scholar]
  10. Setzer, A.; Gaizauskas, R.J. Annotating Events and Temporal Information in Newswire Texts. In Proceedings of the Second International Conference on Language Resources and Evaluation, LREC 2000, Athens, Greece, 31 May–2 June 2000; pp. 1287–1294. [Google Scholar]
  11. Fu, Y.; Dhonnchadha, E.U. A Pattern-mining Driven Study on Differences of Newspapers in Expressing Temporal Information. arXiv 2020, arXiv:2011.12265. [Google Scholar]
  12. Kim, A.; Pethe, C.; Skiena, S. What time is it? Temporal Analysis of Novels. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, 16–20 November 2020; pp. 9076–9086. [Google Scholar]
  13. Zarcone, A.; Alam, T.; Kolagar, Z. PATE: A Corpus of Temporal Expressions for the In-car Voice Assistant Domain. In Proceedings of the 12th Language Resources and Evaluation Conference, LREC 2020, Marseille, France, 11–16 May 2020; pp. 523–530. [Google Scholar]
  14. Madkour, M.; Benhaddou, D.; Tao, C. Temporal data representation, normalization, extraction, and reasoning: A review from clinical domain. Comput. Methods Programs Biomed. 2016, 128, 52–68. [Google Scholar] [CrossRef] [Green Version]
  15. Hao, T.; Rusanov, A.; Weng, C. Extracting and Normalizing Temporal Expressions in Clinical Data Requests from Researchers. In Proceedings of the Smart Health—International Conference, ICSH 2013, Beijing, China, 3–4 August 2013; pp. 41–51. [Google Scholar]
  16. Sun, W.; Rumshisky, A.; Uzuner, Ö. Evaluating temporal relations in clinical text: 2012 i2b2 Challenge. J. Am. Med. Inform. Assoc. 2013, 20, 806–813. [Google Scholar] [CrossRef] [Green Version]
  17. Bethard, S.; Savova, G.; Chen, W.; Derczynski, L.; Pustejovsky, J.; Verhagen, M. SemEval-2016 Task 12: Clinical TempEval. In Proceedings of the 10th International Workshop on Semantic Evaluation, SemEval@NAACL-HLT 2016, San Diego, CA, USA, 16–17 June 2016; pp. 1052–1062. [Google Scholar]
  18. Tong, W.; Yaqian, Z.; Xuanjing, H.; Lide, W. Chinese Time Expression Recognition Based on Automatically Generated Basic-Time-Unit Rules. J. Chin. Inf. Process. 2010, 24, 3–11. [Google Scholar]
  19. Viani, N.; Kam, J.; Yin, L.; Bittar, A.; Dutta, R.; Patel, R.; Stewart, R.; Velupillai, S. Temporal information extraction from mental health records to identify duration of untreated psychosis. J. Biomed. Semant. 2020, 11, 2. [Google Scholar] [CrossRef] [Green Version]
  20. Jinfeng, Y.; Qiubin, Y.; Yi, G.; Zhipeng, J. An Overview of Research on Electronic Medical Record Oriented Named Entity. Acta Autom. Sin. 2014, 40, 1537–1562. [Google Scholar]
  21. Li, P.; Huang, H. UTA DLNLP at SemEval-2016 Task 12: Deep Learning Based Natural Language Processing System for Clinical Information Identification from Clinical Notes and Pathology Reports. In Proceedings of the 10th International Workshop on Semantic Evaluation, SemEval@NAACL-HLT 2016, San Diego, CA, USA, 16–17 June 2016; pp. 1268–1273. [Google Scholar]
  22. Chikka, V.R. CDE-IIITH at SemEval-2016 Task 12: Extraction of Temporal Information from Clinical documents using Machine Learning techniques. In Proceedings of the 10th International Workshop on Semantic Evaluation, SemEval@NAACL-HLT 2016, San Diego, CA, USA, 16–17 June 2016; pp. 1237–1240. [Google Scholar]
  23. Chang, A.X.; Manning, C.D. SUTime: A library for recognizing and normalizing time expressions. In Proceedings of the Eighth International Conference on Language Resources and Evaluation, LREC 2012, Istanbul, Turkey, 23–25 May 2012; pp. 3735–3740. [Google Scholar]
  24. Strötgen, J.; Gertz, M. A Baseline Temporal Tagger for all Languages. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, EMNLP 2015, Lisbon, Portugal, 17–21 September 2015; pp. 541–547. [Google Scholar]
  25. Zhong, X.; Sun, A.; Cambria, E. Time Expression Analysis and Recognition Using Syntactic Token Types and General Heuristic Rules. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, ACL 2017, Vancouver, BC, Canada, 30 July–4 August 2017; Volume 1, pp. 420–429. [Google Scholar]
  26. Moharasan, G.; Ho, T.B. A semi-supervised approach for temporal information extraction from clinical text. In Proceedings of the 2016 IEEE RIVF International Conference on Computing & Communication Technologies, Research, Innovation, and Vision for the Future, RIVF 2016, Hanoi, Vietnam, 7–9 November 2016; pp. 7–12. [Google Scholar]
  27. Ding, W.; Gao, G.; Shi, L.; Qu, Y. A Pattern-Based Approach to Recognizing Time Expressions. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, the Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, the Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, HI, USA, 27 January–1 February 2019; pp. 6335–6342. [Google Scholar]
  28. MacAvaney, S.; Cohan, A.; Goharian, N. GUIR at SemEval-2017 Task 12: A Framework for Cross-Domain Clinical Temporal Information Extraction. In Proceedings of the 11th International Workshop on Semantic Evaluation, SemEval@ACL 2017, Vancouver, BC, Canada, 3–4 August 2017; pp. 1024–1029. [Google Scholar]
  29. Hossain, T.; Rahman, M.M.; Islam, S.M. Temporal Information Extraction from Textual Data using Long Short-term Memory Recurrent Neural Network. J. Comput. Technol. Appl. 2018, 9, 1–8. [Google Scholar]
  30. Patra, B.; Fufa, C.; Bhattacharya, P.; Lee, C. To Schedule or not to Schedule: Extracting Task Specific Temporal Entities and Associated Negation Constraints. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, 16–20 November 2020; pp. 8445–8455. [Google Scholar]
  31. Li, Z.; Li, C.; Long, Y.; Wang, X. A system for automatically extracting clinical events with temporal information. BMC Med. Inform. Decis. Mak. 2020, 20, 198. [Google Scholar] [CrossRef]
  32. Guominl, S.; Sanqian, Z.; Fenlil, J.; Songyanl, J. Temporal Information Extraction and Normalization Method in Chinese Texts. J. Geomat. Sci. Technol. 2019, 36, 538–544. [Google Scholar]
  33. Jing, L.; Defang, C.; Chunfa, Y. Automatic TIMEX2 tagging of Chinese temporal information. J. Tsinghua Univ. Technol. 2008, 48, 117–120. [Google Scholar]
  34. Qiong, W.; Degen, H. Temporal Information Extraction Based on CRF and Time Thesaurus. J. Chin. Inf. Process. 2014, 28, 169–174. [Google Scholar]
  35. Wencong, L.; Chunju, Z.; Chen, W.; Xueying, Z.; Yueqin, Z.; Shoutao, J.; Yanxu, L. Geological Time Information Extraction from Chinese Text Based on BiLSTM-CRF. Adv. Earth Sci. 2021, 36, 211–220. [Google Scholar]
  36. Ma, K.; Tan, Y.; Tian, M.; Xie, X.; Qiu, Q.; Li, S.; Wang, X. Extraction of temporal information from social media messages using the BERT model. Earth Sci. Inform. 2022, 15, 573–584. [Google Scholar] [CrossRef]
  37. Lejun, Z.; Weimin, W. Chinese Time Expression Recognition Based on BERT-FLAT-CRF Model. Softw. Guide 2021, 20, 59–63. [Google Scholar]
  38. Kimmig, A.; Bach, S.; Broecheler, M.; Huang, B.; Getoor, L. A short introduction to probabilistic soft logic. In Proceedings of the NIPS Workshop on Probabilistic Programming: Foundations and Applications, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1–4. [Google Scholar]
  39. Gridach, M. A framework based on (probabilistic) soft logic and neural network for NLP. Appl. Soft Comput. 2020, 93, 106232. [Google Scholar] [CrossRef]
  40. Broecheler, M.; Getoor, L. Computing marginal distributions over continuous Markov networks for statistical relational learning. In Proceedings of the Advances in Neural Information Processing Systems 23: 24th Annual Conference on Neural 478 Information Processing Systems 2010, Vancouver, BC, Canada, 6–9 December 2010; pp. 316–324. [Google Scholar]
  41. Chen, X.; Chen, M.; Shi, W.; Sun, Y.; Zaniolo, C. Embedding Uncertain Knowledge Graphs. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, HI, USA, 27 January–1 February 2019; pp. 3363–3370. [Google Scholar]
  42. Chowdhury, R.; Srinivasan, S.; Getoor, L. Joint Estimation of User And Publisher Credibility for Fake News Detection. In Proceedings of the CIKM ’20: The 29th ACM International Conference on Information and Knowledge Management, Virtual Event, Ireland, 19–23 October 2020; pp. 1993–1996. [Google Scholar]
  43. Li, J.; Sun, A.; Han, J.; Li, C. A Survey on Deep Learning for Named Entity Recognition. IEEE Trans. Knowl. Data Eng. 2022, 34, 50–70. [Google Scholar] [CrossRef] [Green Version]
  44. Bethard, S.; Savova, G.; Palmer, M.; Pustejovsky, J. SemEval-2017 Task 12: Clinical TempEval. In Proceedings of the 11th International Workshop on Semantic Evaluation, Vancouver, BC, Canada, 3–4 August 2017; pp. 565–572. [Google Scholar]
Figure 1. A medical record fragment containing temporal text with overlapping text features. In the figure, both blue and red fonts are time entities. * is used as a substitute for Chinese characters involving private information.
Figure 1. A medical record fragment containing temporal text with overlapping text features. In the figure, both blue and red fonts are time entities. * is used as a substitute for Chinese characters involving private information.
Applsci 13 03348 g001
Figure 2. A medical record fragment containing temporal text with a large span of features described. In the figure, both blue and red fonts are time entities. * is used as a substitute for Chinese characters involving private information.
Figure 2. A medical record fragment containing temporal text with a large span of features described. In the figure, both blue and red fonts are time entities. * is used as a substitute for Chinese characters involving private information.
Applsci 13 03348 g002
Figure 3. The overall framework of PSLF-CTR. The blue part is the structure of PSL-TRF model.
Figure 3. The overall framework of PSLF-CTR. The blue part is the structure of PSL-TRF model.
Applsci 13 03348 g003
Figure 4. The PSL-TFR model structure.
Figure 4. The PSL-TFR model structure.
Applsci 13 03348 g004
Figure 5. The PSLF-CTR model structure.
Figure 5. The PSLF-CTR model structure.
Applsci 13 03348 g005
Figure 6. The number of various types of time in the datasets.
Figure 6. The number of various types of time in the datasets.
Applsci 13 03348 g006
Figure 7. The effect of λ an α taking value on model F1. (a) λ taking the value on the PSL-TFR model, (b) α taking the value on the PSLF-CTR model.
Figure 7. The effect of λ an α taking value on model F1. (a) λ taking the value on the PSL-TFR model, (b) α taking the value on the PSLF-CTR model.
Applsci 13 03348 g007
Table 1. Time types in CTO and their examples.
Table 1. Time types in CTO and their examples.
Syntactic FormsSemantic ConceptsAcronymsExamplesEglish Examples
InstantAbsoluteAIST2021年12月4日4 December 2021
RelativeRIST昨天Yesterday
FuzzyFIST6点左右Around 6:00 p.m.
IntervalAbsoluteAITL2021年11月至12月November to December 2021
RelativeRITL昨天早6点到9点Yesterday from 6:00 a.m. to 9:00 a.m.
FuzzyFITL大概下午5点开始8点结束About 5:00 p.m. start 8:00 p.m. end
IncompleteIITL昨天结束服药Finished taking medication yesterday
CollectionPeriodic InstantPISTC每天下午三点Daily at 3:00 p.m.
Irregular InstantIISTC10点、 11点、 20点分别有抽搐发作Seizures at 10:00, 11:00 and 20:00 o’clock
Periodic Interval CollectionPITLC每年3至5月March to May every year
Irregular IntervalIITLC在昨天下午9点至12点、今天早上5点至6点、下午4点至11点分别进行抢救Resuscitation took place yesterday from 9:00 pm to 12:00 pm, this morning from 5:00 a.m. to 6:00 a.m., and from 4:00 p.m. to 11:00 p.m.
Table 2. Experimental parameter setting.
Table 2. Experimental parameter setting.
ParameterValue
Max Length256
Learning Rate0.001
batch-size32
dropout0.1
λ 1
α 0.15
Table 3. The result of private dataset time recognition.
Table 3. The result of private dataset time recognition.
BERT+BiLSTM+CRFBERT+FLAT+CRFPSLF-CTR
TYPEPRF1PRF1PRF1
AIST92.8895.8394.3393.8196.5495.1593.84 (0.96)95.81 (−0.02)94.82 (0.49)
RIST85.1385.6785.4085.5386.9086.2189.14 (4.01)89.14 (3.47)89.14 (3.74)
FIST81.8677.6779.7180.6579.9180.2881.17 (−0.69)82.65 (4.98)81.90 (2.19)
AITL86.4085.7186.0685.8386.5186.1788.00 (1.60)87.30 (1.59)87.65 (1.59)
RITL75.8979.5577.6876.8380.1178.4378.75 (2.86)81.97 (2.42)80.33 (2.62)
FITL68.6081.3874.4569.1981.5174.8476.92 (8.32)82.19 (0.81)79.47 (5.02)
IITL78.5791.6784.6277.2791.8983.9579.76 (1.19)90.54 (−1.13)84.81 (0.19)
PISTC60.3368.4664.1462.1348.9848.9864.41 (4.08)70.95 (2.49)67.52 (3.38)
IISTC47.3769.2356.2552.9469.2360.0064.29 (16.92)69.23 (0.00)66.67 (10.42)
PITLC44.4452.1748.0051.8560.8756.0053.85 (9.41)60.87 (8.70)57.14 (9.14)
IITLC71.4371.4371.4377.7877.7877.7877.78 (6.35)77.78 (6.35)77.78 (6.35)
Table 4. The results of manual evaluation of the CCKS dataset.
Table 4. The results of manual evaluation of the CCKS dataset.
BERT+BiLSTM+CRFBERT+FLAT+CRFPSLF-CTR
TYPEPRF1PRF1PRF1
AIST93.9294.1394.0294.5994.8194.7094.52 (0.60)93.45 (−0.68)93.98 (−0.04)
RIST91.8992.7392.3192.1792.7392.4591.34 (−0.55)92.73 (0.00)92.03 (−0.28)
FIST85.3979.1782.1687.3679.1783.0687.64 (2.25)81.25 (2.08)84.32 (2.16)
AITL84.6255.0066.6780.0060.0068.5792.31 (7.69)60.00 (5.00)72.73 (6.06)
RITL68.3865.0366.6775.7669.9372.7376.15 (7.77)69.23 (4.20)72.53 (5.86)
FITL72.8879.6376.1172.4177.7877.1973.77 (0.89)83.33 (3.70)78.26 (2.15)
IITL85.0085.0085.0084.4495.0089.4186.05 (1.05)92.50 (7.5)89.16 (4.16)
PISTC45.7152.1748.7346.1552.1748.9846.60 (0.89)52.17 (0.00)49.23 (0.50)
IISTC79.1795.0086.3682.6195.0088.3784.44 (5.27)95.00 (0.00)89.41 (3.05)
PITLC50.0050.0050.0050.0050.0050.0050.00 (0.00)50.00 (0.00)50.00 (0.00)
IITLC36.3670.0047.8638.4671.4350.0041.67 (5.31)71.43 (1.43)52.63 (4.77)
Table 5. L-atom statements in private datasets predict the results of controlled experiments.
Table 5. L-atom statements in private datasets predict the results of controlled experiments.
BERT+textCNNPSL-CTR
PRF1PRF1
l 1 99.8196.8998.3399.34 (−0.47)94.23 (−2.66)96.72 (−1.61)
l 2 97.4286.6491.7297.18 (−0.24)87.22 (0.58)91.93 (0.21)
l 3 96.9791.1094.6597.75 (0.78)89.21 (−1.89)93.29 (−1.36)
l 4 78.7955.3265.0080.25 (1.46)61.41 (6.093)69.58 (4.58)
l 5 96.6493.9895.2998.73 (2.08)92.38 (−1.60)95.45 (0.16)
l 6 97.6092.6495.0697.59 (−0.01)92.87 (0.23)95.17 (0.11)
l 7 96.8464.0477.0992.97 (−3.87)65.58 (1.54)76.91 (−0.18)
Table 6. L-atom statements in private datasets predict the results of controlled experiments.
Table 6. L-atom statements in private datasets predict the results of controlled experiments.
BERT+textCNNPSL-CTR
PRF1PRF1
l 8 90.3993.9992.1597.93 (7.54)96.08 (2.09)97.00 (4.84)
l 9 81.6377.1679.3392.20 (10.57)91.58 (14.42)91.89 (12.56)
l 10 73.0877.9875.4585.93 (12.85)82.96 (4.98)84.42 (8.97)
l 11 90.5382.1386.1393.88 (3.35)94.64 (12.51)94.26 (8.13)
l 12 68.5671.6070.0582.47 (13.91)79.74 (8.14)81.08 (11.03)
l 13 83.2470.8676.5592.35 (9.11)80.10 (9.24)85.79 (9.24)
l 14 66.6776.0071.0386.68 (20.01)80.91 (4.91)83.70 (12.67)
l 15 82.1383.4082.7693.11 (10.98)86.77 (3.37)89.83 (7.07)
l 16 60.8658.2359.5283.93 (23.07)67.14 (8.91)74.60 (15.09)
l 17 47.2347.8447.5363.64 (16.41)67.74 (19.90)65.63 (18.09)
l 18 45.0044.0044.4958.62 (13.62)56.67 (12.67)57.63 (13.13)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gu, J.; Wang, D.; Hu, D.; Gao, F.; Xu, F. Temporal Extraction of Complex Medicine by Combining Probabilistic Soft Logic and Textual Feature Feedback. Appl. Sci. 2023, 13, 3348. https://doi.org/10.3390/app13053348

AMA Style

Gu J, Wang D, Hu D, Gao F, Xu F. Temporal Extraction of Complex Medicine by Combining Probabilistic Soft Logic and Textual Feature Feedback. Applied Sciences. 2023; 13(5):3348. https://doi.org/10.3390/app13053348

Chicago/Turabian Style

Gu, Jinguang, Daiwen Wang, Danyang Hu, Feng Gao, and Fangfang Xu. 2023. "Temporal Extraction of Complex Medicine by Combining Probabilistic Soft Logic and Textual Feature Feedback" Applied Sciences 13, no. 5: 3348. https://doi.org/10.3390/app13053348

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop