Next Article in Journal
On the Need for Healthcare Informatics Training among Medical Doctors in Jordan: A Pilot Study
Previous Article in Journal
Development and Internal Validation of an Interpretable Machine Learning Model to Predict Readmissions in a United States Healthcare System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Generating Paraphrase Using Simulated Annealing for Citation Sentences

School of Electrical Engineering and Informatics, Bandung Institute of Technology, Bandung 40312, Indonesia
*
Author to whom correspondence should be addressed.
Informatics 2023, 10(2), 34; https://doi.org/10.3390/informatics10020034
Submission received: 29 January 2023 / Revised: 7 March 2023 / Accepted: 8 March 2023 / Published: 30 March 2023
(This article belongs to the Section Machine Learning)

Abstract

:
The paraphrase generator for citation sentences is used to produce several sentence alternatives to avoid plagiarism. Furthermore, the generation results need to pay attention to semantic similarity and lexical divergence standards. This study proposed the StoPGEN model as an algorithm for generating citation paraphrase sentences with stochastic output. The generation process is guided by an objective function using a simulated annealing algorithm to maintain the properties of semantic similarity and lexical divergence. The objective function is created by combining the two factors that maintain these properties. This study combined METEOR and PINC Scores in a linear weighting function that can be adjusted for its value tendency in one of the matrix functions. The dataset of citation sentences that had been labeled with paraphrases was used to test StoPGEN and other models for comparison. The StoPGEN model, with the citation sentences dataset, produced a BLEU score of 55.37, outperforming the bidirectional LSTM method with a value of 28.93. StoPGEN was also tested using Quora data by changing the language source in the architecture section resulting in a BLEU score of 22.37, outperforming UPSA 18.21. In addition, the qualitative evaluation results of the citation sentence generation based on respondents obtained an acceptance value of 50.80.

1. Introduction

Paraphrase generation produces new text from the input with different wording but the same information [1]. The generation machine aims to create sentences with high lexical divergence and maintain semantic similarity. Furthermore, the generator is often equated with a machine translation, but the input and output are sentences in the same language [2]. The generation of a paraphrase needs to consider several criteria. Moreover, sentences in scientific papers are usually argumentative [3], where one statement is bound in context with another, either in a causal paragraph or vice versa. The resulting new sentence should not have plagiarism characteristics [4]. Scientific papers contain many equivalent or multilevel compound sentences; hence, the output form is more complex [5]. In the domain of scientific papers, paraphrasing can be found in several events [6], such as:
  • The abstract is a paraphrase of the sentence in the body of the paper
  • The introductory part has a paraphrase equivalent to the methodology section
  • The conclusion has a paraphrase equivalent to the experimental section
  • Definition sentences have paraphrase equivalents with others that define the same construct
  • The citation sentence that quotes the same paper is a paraphrase.
This study used the potential of the citation sentence as the paraphrase collection, which can be obtained from various papers. The citation sentence was selected because it is often considered a part that increases a paper’s plagiarism value.
It also has many potentials for paraphrasing when collected. Therefore, the dataset was collected from open-source Computing Language Papers (ACL Anthology). Citation sentences have several functions, including citing weaknesses, contrasts, methods, and data similarities, as well as problem bases or neutral ones [7]. The citation sentence used in this study is limited to only one of the citation targets. This was carried out to limit the context of the purpose of each delivery in the scientific argument on sentences.
Inspired by the Unsupervised Paraphrasing of Simulated Annealing [8], a generate and test model architecture was developed with the same algorithm but different objective functions and strategies. Furthermore, this study combined two matrix functions, namely METEOR [9] and PINC Score [10], to capture semantic similarities and lexical differences. The two matrix functions were combined in a linear weighted function [11], which can be adjusted to the tendency of its value. The language source that makes a substitute or addition successor to the input sentence is built with word embedding [12]. The sentence candidate selection strategy uses the n-gram language model [13].
Approaches for rule-based paraphrase generation are based on hand-crafted and automatically collected paraphrase rules. These rules were mostly hand-crafted in the early works [14]. Because of the enormous manual work required, some researchers have attempted to collect paraphrases automatically [15]. Unfortunately, because of the limitations of the extraction methods, long and complex additional patterns have been generated, affecting performances.
Thesaurus-based approaches start by extracting all synonyms for the words to be substituted from a thesaurus. The best choice is then selected according to the context of the source phrase [16]. Although simple and effective, the diversity of the generated paraphrases tends to limit this strategy.
Seq2Seq models were initially used for paraphrase generation with recent advances in neural networks, particularly the sequence-to-sequence architecture [17]. Convolutional neural networks (CNNs) have also been used to build seq2seq models since they have fewer parameters, and so train faster [18]. The Transformer [19] model has shown attempting to cut performance on a variety of text generation tasks. A transformer was developed for the seq2seq model because of the Transformer’s improved capacity to capture long-term dependencies in sentences [20].
The dataset of citation sentences was tested using other methods with a supervised approach, such as LSTM, a type of recurrent neural network that can be used for paraphrase generation by learning to capture long-term dependencies in input text sequences [21]. Bi-LSTM for paraphrase generator is a neural network model that utilizes bidirectional processing of input sequences to generate paraphrases with a focus on capturing contextual information [22]. Transformer is a neural network architecture for natural language processing that uses self-attention mechanisms to encode and generate paraphrased sentences. Furthermore, this study used the unsupervised method by modifying UPSA [8] to compare the ability to produce paraphrased sentences. Apart from the citation sentence data, the paraphrased data from the questions in Quora [23] and Twitter [24] showed that the architecture built could be used for other domains. The results of StoPGEN generation were compared with UPSA. Variant autoencoder is a neural network architecture that can be used for generating paraphrases by learning a compressed representation of the input text and then using it to generate a new text [25]. LagVAE is the improvement of the variant autoencoder [26]. CGHM (Concept-Phrase Hypergraph Model) is a model that generates paraphrases by leveraging semantic concepts and syntactic information to build a hypergraph representation of the original sentence [23].
The method of building the corpus and paraphrase generation has been widely developed for social media and news. However, the existing paraphrase generation method cannot be directly adapted to produce new sentences in scientific papers. Paraphrase generation is very dependent on the available language resources. The previously developed method was unable to provide alternative generation, so it requires a stochastic generation method.
This study contributed to producing paraphrase generation methods for citation sentences from scientific papers. This study focuses on developing the architecture of generating methods. Paraphrase detection formulas are used as objective functions. The method developed has a stochastic output and produces a different alternative output but the conference to the objective function. The proposed method has been tested in several corpora.

2. Related Work

2.1. Corpus Construction

The construction of the corpus paraphrase is known as paraphrase extraction. This is a task to generate the collection of paraphrased sentence pairs from large documents [27]. The extraction result can be a collection of words or phrases, such as PPDB [26], which uses two-language pivoting. It can also be the paraphrased sentence pair, such as MSRP [28], which is obtained from a news collection using a supervised approach. Other corpora, such as PIT [29], were compiled from tweets using the similarity object delivery approach. Each text unit and domain have unique characteristics because of its specific information purpose. State or the art of constructing a paraphrase corpus can be seen in Table 1.
It is necessary to observe the authors’ characteristics in conveying information when extracting paraphrases from scientific paper sources. Authors of scientific papers write information using three approaches, namely paraphrasing, summarizing, and translating [31]. Abstract sentences with body parts can be collected to build a paraphrase corpus [5]. However, citation sentences have the greatest potential to build a paraphrase corpus from these papers [32]. The construction of the citation paraphrase corpus in this study is a small contribution to paraphrase generation research.

2.2. Objective Function

The generation model built with the generate and test model requires an objective function to guide the generation process. In paraphrasing, the objective function is a formula to measure the paraphrase value of two pairs of sentences (usually a value between 0 to 1). Studies in this section are usually grouped in the task text similarity measurement.
Paraphrasing is a task that is very similar to machine translation; therefore, the evaluation approach of the translation can be used for paraphrasing. Furthermore, evaluation techniques, such as NIST [33], BLEU [34], and WMT [35], can be combined into a formula to assess the results of paraphrase generation evaluated based on the available data [36].
The Term Frequency Kullback–Leibler Divergence (TF-KLD) data representation is the best technique for measuring paraphrases in the MSRP dataset [37]. Prior to the classification, the matrix is converted into a latent representation with TF-KLD features, and the SVM algorithm is subsequently used for classification. The evaluation was carried out by comparing the standard TF-IDF resulting in an accuracy of 80.4% and an F1 Score of 85.9%.
Another approach for measuring the paraphrase output is the use of deep learning to build a sentence representation and simply compare it in vector form [32]. Apart from the neural network architecture, a Convolution Neural Network (CNN) model, which consists of composition layers, decomposition, and proximity estimation, can be used to measure the paraphrase generation results [38]. Text representation with word embedding is often used when a deep learning approach is applied.
A model is developed to measure paraphrase in the domain of scientific papers. The Siamese neural network architecture is used to study the similarity and dissimilarity based on corpus labeled true and false for sentences from scientific papers [39] with an accuracy rate of 64%. Furthermore, the SVM model can be developed by engineering word features, such as Euclidean distance, cosine similarity, and sentence length [40], with an accuracy rate of 61.9%. Both studies utilized a learning-based approach and were strongly influenced by the quality of the corpus used.
In this study, the objective function was built based on semantic similarity and lexical divergence. To combine the two, a formula that can configure the tendency to one aspect was built. The objective function formation is explained in the experiment section.

2.3. Paraphrase Generator

Paraphrase generation is a task to generate new sentences from the input. Furthermore, various language resources are needed in this process. The general approach of generating paraphrases uses a machine translation set to produce sentences in the same language [2].
Paraphrase generation can be found in various domains, such as news [41], where the generator can be used to package news content or form variations of headlines [42]. It can also be found in social media domains such as Twitter [29]. Paraphrase generation in these various domains aims to produce semantic similarity, compression, concatenation, and sentence simplification [43].
The sequence-to-sequence learning is a technique developed with a deep approach to paraphrase generation [44]. The main construction of this model is the Recurrent Neural Network (RNN) or Long-sort Term Memory (LSTM) units. The deep learning approach was developed with the Transformer and inspired the use of this technique in paraphrase generation [45].

2.4. Simulated Annealing

Simulated Annealing (SA) is an effective algorithm in the solution search on a very large dimension space [46]. The advantage of the algorithm is its ability to avoid the local maximum of the optimization function. Furthermore, the algorithm is inspired by heavy industrial processing that utilizes the lowering of an object’s temperature and manipulates it to the desired shape. The temperature drop factor determines the fault tolerance in the search process of the solution space. The error is acceptable when the temperature is still high and less likely to be accepted towards the end of the temperature drop.
In the sentence generation case, it can be stated that χ is a very large sentence dimension space, and f ( x ) is the objective function of generating new sentences. The main target of Simulated Annealing is to determine the sentence x with the maximum value f x . There is a generation step in every search, which can be called t , while the sentence generated can be referred to as x t . Simulated Annealing will select x t + 1 , which has undergone a change from x t as the current step when the f value is greater. Otherwise, tolerance will be calculated based on the probability value e f x t f x t + 1 T controlled by T (temperature), which decreases at each step, resulting in 0 in the last step. For example, in x t + 1 a smaller f value is obtained, and x t + 1 can be accepted as a new step when the result of generating a random number r < e f x t f x t + 1 T . Simulated Annealing is inspired by the metallurgical process of cooling materials. At the beginning of the search, the temperature T is usually very high and allows x t + 1 to be accepted even though the value of f is smaller. Theoretically, this can avoid the local maximum’s optimization function and guarantee the global maximum’s achievement [47].

3. Methods

3.1. Generator Model

The process of forming new sentences is developed following the generate and test model mechanism. The use of the simulated annealing algorithm requires the changing instances process. Furthermore, input sentences are considered instances that are processed into different sentence forms. The sentence changes are carried out at the lexical level. In addition, there are three factors to consider when forming new sentences, namely action (type of shape change), language model (change reference source), and sentence change strategy.
In one of the generation processes, an object is selected with a choice of substitution, insertion, and deletion actions. The replacement process is carried out by selecting equivalent words from the language model, while the addition of objects is conducted by inserting a new object based on the selection in the model. The substitute word is selected based on the neighbor context in the input sentence; therefore, paying attention to the word before and after is necessary. Meanwhile, the deletion process removes the selected object from the sentence. The architecture of the method for paraphrase generation using simulated annealing is illustrated in Figure 1.
Editing steps is a strategy for the sequence of action to be carried out. It can be selected in two ways, namely selecting, randomly and sequentially, from the beginning to the end of the sentence. Furthermore, the editing steps will determine how to stop the sentence generation process. When using a random scheme, the termination process entirely depends on the temperature setting of the simulated annealing algorithm. However, the process stops depending on the number of words in the sentence when conducted sequentially. With this strategy, the temperature is still used, and only the reduction process is adjusted to the number of lexical targets processed. An example of changing sentences in the generation process is shown in Figure 2.
A successor is a substitute or additional text unit that will be included in a sentence. The source used to obtain the successor was obtained from the language model with the word embedding approach. This language model was created using word2vec [48] with a corpus as a source for scientific papers collection. Word2vec has two models, namely CBOW and Skip-Gram. CBOW allows performing successor search processes for substitution and insertion actions. Meanwhile, the skip-gram was not used because it could not find the successor for the insertion action. The cosine similarity Formula (1) is used to determine the successor in the substitution action by comparing the target vector with the candidates in the word2vec dictionary. Meanwhile, the probability vector Formula (2) is used to determine the successor by comparing the neighbor word vectors from the target position with the candidates in the word2vec dictionary.
The formula for calculating similarity:
S i m A , B = i = 1 n A i B i i = 1 n A i 2 i = 1 n B i 2
The formula for calculating vector probability:
P r o b = n σ 2 + 1 n
w h e r e   σ 2 : the variance of the elements. n : number the elements
The substitution and insertion actions make it possible to produce several new sentences because the results of the candidate selection allow more than one successor. Therefore, the sentences formed from all possible successors can be very diverse. In the simulated annealing algorithm, a state can only be filled by a sentence; hence, a selection strategy is required. The n-gram language model [13] is used to measure the word object arrangement probability of all possible sentences that can be formed. Function (3) assesses the new sentence structure when the successor is included.
The formula for the n-gram probability language model:
p L M w n | w n N + 1 : n 1 = C w n N + 1 : n 1 w n C w n N + 1 : n 1
where
p L M w n | w n N + 1 : n 1
represents the probability of the word w n given the previous N 1 words w n ( N 1 ) , …, w n 1
C w n N + 1 : n 1 w n
represents the count of the N-gram w n ( N 1 ) , …, w n in the training corpus
C w n N + 1 : n 1
represents the count of the N 1 gram w n ( N 1 ) , …, w n 1 in the training corpus
The probability of choosing an action was analyzed in the experiment to see its influence on the generation outcome. Furthermore, multiple combinations of odds were allowed from three available actions; substitution, insertion, and deletion. For this reason, an action opportunity distribution scenario was prepared, as shown in Table 2.

3.2. Objective Function

Simulated annealing works by optimizing the objective function. In the case of paraphrase generation, an objective function that evaluates each state based on semantic similarity and lexical divergence is required. Therefore, this study utilized two calculations of these aspects in one formula. The formula used is weighted linear (4) inspired by Maximal Marginal Relevance (MMR), which selected the information based on two contradictory parameters, one strengthening and the other weakening [11].
W e i g h t e d   L i n e r = α P 1 + 1 α P 2
where
  • P 1 = p a r a m e t e r   v a l u e   1
  • P 2 = p a r a m e t e r   v a l u e   2
  • α = m u l t i p l i e r   c o n s t a n t a   ( 0 1 )
Two calculation matrices are needed to fill in the parameter values in the formula, each representing semantic similarities and lexical differences. The METEOR was selected to obtain the semantic similarity value [9], and PINC Score [10] to obtain the lexical divergence. The two calculated functions fill in the two parameters in function (2).
METEOR [34] is a basic parameter used to measure semantic similarity. Furthermore, it is commonly used to assess the translated text quality. Compared to other measuring tools, it has the ability to calculate the semantic similarity value from different lexical items because it is equipped with a lexical similarity dictionary. In machine translation, METEOR’s measurement results are closer to human judgment than other measuring instruments such as BLEU.
METEOR works by calculating the Precision (P) value, which is the ratio of the appropriate number of N-Grams in all translation results, and the Recall (R), which is the ratio value of the appropriate n-gram number in the translation results to the reference sentence. Furthermore, FMeans (5) is also calculated, namely the harmonic-mean value, which prioritizes the recall effect.
F M e a n s = 10 P R R + 9 P
Apart from the calculation of FMeans, METEOR considers the penalty value (6), which is the least number of Chunks in the phrase related to the reference, divided by the corresponding N-Gram number. The penalty reduces the FMean value to better consider the appropriate Chunk.
P e n a l t y = 0.5 # c h u n k # u n i g r a m s _ m a t c h
The final METEOR value is calculated by the Formula (7):
M S = F M e a n s 1 P e n a l t y
Pinc Score [10] was used because of the need for a formula to calculate the lexical divergence between the input and generated sentences. The score (8) was used to complete the parameters of the weighted linear function. Furthermore, it calculates the number of words that do not intersect from two text inputs with N-Grams and can be adjusted as needed. The score is obtained by subtracting the number 1 from the suitability value of the two sentence units.
P S O , S = 1 N n = 1 N 1 n g r a m o n g r a m s n g r a m s
where
  • O = s y s t e m   o u t p u t   s e n t e n c e
  • S = r e f e r e n c e   s e n t e n c e
  • n g r a m = n u m b e r   o f   m a t c h i n g   w o r d s
  • N = t h e   s e n s i t i v i t y   v a l u e
This study proposed the PScore, a weighted linear formula that uses METEOR and PINC Scores as two inverse parameters. A multiplier constant of 0.9 was obtained for METEOR. The PINC Score had a constant of 0.1. This value was selected based on the measurement of pre-experimental results. The weight used is unbalanced to emphasize one aspect. In the formula used, the meteor gets almost absolute value because the results obtained can prioritize semantic similarity and a little lexical divergence.
P s c o r e = 0.9 M S + 1 0.9 P S
StoPGEN was proposed as a method for generating paraphrase sentences based on the simulated annealing algorithm and developed with three types of architecture. The first uses a random target word selection scheme as done by UPSA. The second uses a sequential selection scheme from the beginning of the sentence to the end. Meanwhile, the third uses a sequential selection scheme with the probabilistic language model to select successors. The third architecture with complete functionality is described in Algorithm 1.
Algorithm 1: Simulated Annealing for Sentence Generator
FUNCTION StoPGEN(x [0…n]: Array of token) return S [0…n]: Array of token
  {function of sentence generator using simulated annealing}
DECLARATION
  T, initial temperature
  ∆T, temperature drop
  s, current state {state is array of token}
  Pscore, objective function
  pLM (), language model probability function {return score}
  most_similar (), get most similar token from word index
  Neighbors (), get most similar token by neighbors from word index
ALGORITHM
  T ← T0
  Sk ← x
  while T > 0 and n do
   action ← get_random_action ()
   if action = subtitution do
    sk + 1[n] = max (pLM (most_similar (sk + 1[n])))
   else if action = insertion do
    sk + 1[n] = max (pLM (neighbors ([sk + 1[n − 2], sk + 1[n − 1], sk + 1[n + 1], sk + 1[n + 2]]))
   else
    remove(sk + 1[n])
    ∆E ← Pscore (sk + 1)−Pscore (sk)
    if min (1, e−∆E/T) >= rand (0,1) then
     sk ← sk + 1
    end if
     T ← T − ∆T
    end while
   return sk {new array of token}

3.3. Dataset

Paraphrasing citation pairs were obtained from papers on the Association of Computational Linguistics (ACL) website (https://aclanthology.org/, accessed on 1 August 2019). The construction of the dataset runs in two stages, first, the candidate corpus development, and second, the labeling of the corpus sentences by the annotator. The construction of a candidate corpus is the extracting process from all collected papers to select citation sentences. Each sentence is processed by clustering technique based on the text features and the target citation similarity. The details of the parallel corpus candidate development process were published [6].
The following steps are taken in the process of getting a corpus:
  • Sentences from scientific papers will be collected based on their function as citation sentences, abstracts, and content. This process uses the Dr. Inventor Framework [49].
  • From the extraction results, the sentences selected white only have one citation target. The sentences that have two or more citation target was ignored.
  • After the selection is made, the clustering process is carried out. Clustering sentence citations is the process of grouping sentences to get candidate pairs of sentences that have the same meaning. The clustering of citation sentences uses the K-Means algorithm [50] with Jaccard similarity, bigram representation, and TF-IDF.
  • After the clustering process, each sentence in a cluster is paired with one another as a corpus candidate for the labeling stage.
Labeling is the last stage of producing a parallel corpus. Five annotators were used to label paraphrased or non-paraphrased. The number of data generated from the annotation process was 5720 pairs of sentences, with 4975 paraphrases and 745 non-paraphrases.
This study also utilized the Quora dataset in the question pair form, as used by UPSA [8], with CGHM as the setting of the dataset [23]. Furthermore, paraphrased sentence datasets from Twitter social media were also used [24]. Both of these datasets are used to test the performance of the architectures that are created and compared with other methods such as UPSA [8], Variant Auto Encoder [25], Lagging Variant Auto Encoder [51], and CGHM [23].

4. Experiment

4.1. Dataset Evaluation

The Fleiss kappa formula is commonly used to measure inter-annotator agreement when multiple annotators provide judgments on a categorical item. To evaluate the paraphrase dataset, multiple annotators can be assigned to rate each pair of sentences as a paraphrase or not. The label given by each annotator can be represented as a matrix, with each row and column representing a sentence pair and each cell containing the label given by a specific annotator. The Fleiss kappa formula can then be used to calculate the level of agreement among annotators, with higher scores indicating higher levels of agreement. The resulting score can provide insights into the quality of the dataset and the consistency of the annotations, which can inform decisions regarding the inclusion or exclusion of specific sentence pairs in the dataset. In Table 3 you can significance leve of dataset quality based on Fleiss Kappa.

4.2. Experiment Scenario

The paraphrase generation experiment was structured to examine the architectural abilities developed in the main domain of scientific paper sentences and the general domain. The model variations were used in all action factors in order to see the ability to produce paraphrased sentences in scientific paper sentences. As a result, 21 experimental scenarios can be produced (three StoPGEN architectures multiplied by seven action factor scenarios).

4.3. Method Comparison

4.3.1. UPSA

The main baseline of our study is the UPSA method [8]. This method is similar to the algorithm developed. However, they differ in the objective function, object selection scheme, and language domain used.

4.3.2. Variant Auto Encoder

In a previous study, the variant autoencoder method with the main LSTM architecture was used to generate the standard Twitter and Quora datasets [25]. This model uses the 300-dimensional LSTM trained with the non-parallel dataset. The mechanism is to maximize the loglikelihood in the inference process with the sentence variations choice obtained from the latent space feature.

4.3.3. Lagging Variant Auto Encoder

A simpler model from VAE was developed by increasing the sequence model learning ability, namely LagVAE [51], and reported to have better performance in the standard datasets used.

4.3.4. CGHM

CGHM [23] was developed using Metropolis–Hastings, which is a method for taking word space samples for sentence making. This method’s results outperformed VAE in the case of latent space samples. Moreover, CGHM is an unsupervised paraphrasing technique that excels on standard datasets.

4.3.5. Modified UPSA

The UPSA baseline model was developed by changing the language model used. The UPSA language model is the Twitter and Quora domains. This replacement was built from scientific paper collections. The model includes two parts, namely, the candidate token selection and the probabilistic model for the objective function.

4.3.6. LSTM Encoder-Decoder

LSTM encoder-decoder was widely used in machine translation [21]. The code was rewritten by including the scientific papers’ parallel sentences. This technique was included in the supervised method to examine the performance of the developed dataset.

4.3.7. Bi-Directional LSTM

Similar to the LSTM encoder-decoder, bi-directional is commonly used in generating interpreter machines [22]. This method adds an alternating learning scheme to sequentially arranged LSTM cells. It was only used in this study for the scientific paper’s domain.

4.3.8. Transformer

The Transformer-based generation model utilizes the alignment token to see how closely related two text units are and subsequently aligns them into a matrix. Therefore, it can be trained as a spatial data model [19,24]. The method was originally developed for machine translation, and the most basic transformer architecture was used for paraphrase generation.

4.4. Matrix Evaluation

4.4.1. BLEU Score

BLEU (bilingual evaluation understudy) is an evaluation method on the natural language generator that examines how close the machine output is to the sentence reference [34]. This matrix measures the proximity of the results to the customizable word segment. The supervised BLEU Score generation model is used to calculate one output based on the reference sentence. In generating unsupervised and stochastic models, the score can calculate all output results with reference sentences as well as the average and best results.

4.4.2. Rouge

Rouge (Recall-Oriented Understudy for Gisting Evaluation) is a software package for calculating. The generated sentences proximity with references that have variations includes ROUGE 1 (R1), ROUGE 2 (R2), and ROUGE L (RL) [52]. ROUGE 1 calculates the unigram overlap of the output sentences by reference, while ROUGE 2 calculates the overlapping bigrams. ROUGE L (longest common subsequence) calculates based on the cut of the longest segment in the compared sentence. This research uses rouge version 1.0.1.

5. Results

5.1. Dataset Agreement

We conducted an agreement analysis among five observers in evaluating a dataset consisting of 3476 paraphrase or non-paraphrase classification data. Each observer provided a classification of paraphrase or non-paraphrase for each data point. Out of the total of 5720 data, 4975 data were classified as paraphrase by all observers, 745 data were classified as non-paraphrase by all observers, and 269 data had different classifications among the observers.
Fleiss kappa was used to assess the amount of agreement among the observers. The study resulted in a kappa score of 0.67, showing that the observers were in good agreement. This kappa number shows a substantial level of agreement at the 0.05 alpha level, according to the Fleiss kappa significance table. Yet, according to the criteria in the Fleiss kappa significance table, this kappa number still falls in the “fair agreement” category.

5.2. Quantitative Evaluation

The StoPGEN model ability was tested with several architectural variants and all scenarios of action factors. Furthermore, its performance was compared to the public and the citation sentence datasets from scientific papers. The use of the compared model was grouped into two approaches, namely supervised and unsupervised. This study used the evaluation matrix BLEU, Rouge 1, Rouge 2, and Rouge L.

5.2.1. Generate the Quora and Twitter

The developed model was pre-tested with a standard dataset to examine its visibility. It can outperform both the supervised and unsupervised models. Furthermore, it significantly outperformed CGHM as the best-supervised model on the Quora dataset. Meanwhile, the Twitter dataset only exceeded by a few points. The unsupervised model and the basis for developing it, namely UPSA, can be outperformed in both datasets by numbers that are not too far apart. This is because UPSA and StoPGEN have the advantage of being quite adaptive in any domain. Table 4 showed the performances of the metodes on standard dataset.

5.2.2. Generate Citation Sentences

StoPGEN was developed with three variations, each with a different nature in generating sentences.
  • StoPGEN1: Generates a sentence with a random action by selecting a random word position.
  • StoPGEN2: Generates sentences with random actions by selecting words sequentially based on their position order in the sentence.
  • StoPGEN3: Generates sentences with random actions by selecting words sequentially based on their position order in the sentence and using language models to select candidate accessors.
The results showed that the 3rd variation, StoPGEN outperformed the previous two, indicating that the word selection order and language model filters can produce better sentences. The best scenario is normally distributing the probability for each action factor. Although this factor can be adjusted dynamically as needed, the qualitative results showed that dividing it evenly yields better performance. Table 5 shows the evaluation results of all StoPGEN variants with all action factor scenarios.
The citation dataset was tested using StopGEN3 and compared with the others (Table 6). The supervised model was selected based on its general use in the machine translation domain. Meanwhile, the selected unsupervised model, namely UPSA and the modified UPSA, is the main baseline of this study. The UPSA was modified by replacing the source language used with the scientific paper domain. It also uses a language model as successor words and objective functions.
We conduct more experiments to back up the findings of this study. The supervised model was trained with a pair of citation paraphrase sentences that were labeled true. All the models used a sequence-to-sequence approach. The experimental results showed that the best-supervised model to use is the Bidirectional LSTM [22], outperforming other machine translation models with a BLEU value of 28.93.
The original UPSA was directly used without changing the source. The results were worse than the UPSA, which was modified by replacing the source language with a corpus of scientific papers. However, StopPGEN can still outperform the modified UPSA. It also outperformed all models producing BLEU 55.37, Rouge 1 71.28, Rouge 2 47.46, and Rouge L 66.32.

5.3. Qualitative Evaluation

The qualitative evaluation was used to examine the StoPGEN generation results ability in the context of the resulting language. A survey was conducted on 30 readers with specifications in the computational linguistics field to see the acceptance. This study compared the acceptance of all variations of StoPGEN generation results and its comparison with other models. The results for each model are further detailed, showing direct examples.

5.3.1. Stochastic StoPGEN Results

The StoPGEN model developed can produce stochastic paraphrase sentences at any time of generation. As a result, the generated sentences can be customized in accordance with the tendency between lexical differences or semantic similarities. The results of generating sentences five times in this experiment are shown in Table 7.
Table 7 shows the generation results with an actual value of 0.9 for the Meteor Score and 0.1 for the Pinc Score. The architecture used was StoPGEN, which qualitatively and quantitatively produced the highest value. The 10 generated results showed that all examples were quite acceptable, although not all were grammatically perfect. This variation was the objective target of the model developed from this study.

5.3.2. Human Acceptability

A survey was conducted by showing the generation results, and the reader was subsequently asked to select the acceptance status for each sentence (True/False). In this survey, the reader was assisted with a coloring aid as a sign of the selected action.
Table 8 shows the generation results of all StoPGEN variants. The results of StoPGEN1 obtained sentences with a higher lexical divergence because the target was selected randomly. It also allowed the same unit position of the text to get repeated action. Meanwhile, StoPGEN2 produced sentences with better semantic similarity, and the number of sentence units was relatively the same as the input because the target sentences were selected sequentially. StoPGEN3 obtained new sentences that were more readable than the others because a language model filtered replacement/addition text units.
The survey results on the StoPGEN variant showed that the 3rd model had the highest value, with half the average acceptance value of all survey answers. This further supports the quantitative results that this model variation is the best. Meanwhile, the other two produced quite different values (see Table 9).
In the second survey, we compare the paraphrase result of the baseline method, UPSA, and UPSA-modified approaches. (see Table 10). As a result, StoPGEN3 had an acceptance value that outperformed the UPSA and UPSA baseline models, which was modified for the scientific paper domain. Furthermore, the basic UPSA had a low acceptance value because the source language used was not modified.
Table 10 compares the UPSA outputs [8], where the source language was modified for the scientific papers domain and the best variant of the developed StoPGEN. The UPSA generation results showed a decrease in sentences because it produced many Out of Vocab (OOV). Furthermore, the results did not appear as there were language model limitations. Even though the output of the modified UPSA was not much OOV (lost), the readability was low. This is due to the different configurations of simulated annealing. Therefore, the StoPGEN output results are more acceptable with lexical variations and semantic similarities that are still maintained.
Table 11 show paraphrase acceptability from all compared model, based on survey. StoPGEN3 obtained the highest acceptance value even though the results differed from the previous survey. The acceptance rate of 50.80 indicates that more than half of the displayed results had received acceptance from readers. Values that are less than the first survey can be considered insignificant.

6. Conclusions

This study succeeded in developing a paraphrase generation model for the scientific paper domain, specifically citation sentences. Furthermore, it developed StoPGEN as a sentence generation model with three variants. The best results were shown by the variants that used sequential word selection strategies, with equal opportunities for substitution, insertion, and deletion actions. A probability-based language model was used to select a replacement token or sentence filler at the end of the generation.
The method in variant three has shown the best results with the action factor G. That action factor is a strategy for selecting text change actions (substitution, insertion, and deletion) with equal probability. In other action factor strategies, changes in probability have been analyzed but did not show better results.
The StoPGEN generation results were compared with other models in the domain. Our model outperformed the supervised one which is commonly used for machine translation sentence generation. This research compared the result with deep learning models such as LSTM, Bi-directional LSTM, variant autoencoder, and Transformer. The StoPGEN can outperform the results of these methods. We see the dataset as the factor, as we know that it was not big enough for a deep learning neural network base method. We see that the method we have developed is suitable for low language resources. It also outperformed the unsupervised UPSA and modified UPSA, which had its source language adapted for scientific papers.
Furthermore, the developed method performed better on public datasets than in other studies. StoPGEN had a superior performance for the Twitter and Quora datasets, with BLEU scores of 6.26 and 22.37, outperforming the other models.
In qualitative measurement by survey, the best StoPGEN variant had an acceptance value of 50.96. Meanwhile, the true value was 50.80 when the model was compared with others. From the output observations, StoPGEN can produce stochastic outputs while maintaining semantic similarities but with lexical differences.

7. Future Works

There are various drawbacks to this study. This study method is dependent on the corpus domain. According to qualitative evaluation data, the level of reader revenue remains around 50%. There are still grammatical mistakes in selecting successors in substitution and insertion actions.
Further research can be developed by considering several aspects, including lexical, syntactic, and semantic dimensions. In this study, the paraphrasing aspect only considers semantic similarity and language divergence. Further research can pay attention to fluency, diversity, and coherence. All aspects mentioned are expected to be implemented to make a better model. With a better model, the value of reader acceptance will be increased.
Future research could take several different directions that would build on the results and contributions of this work. Objective functions can be built by combining several properties of paraphrasing, such as similarity, fluency, diversity, coherence, and linguistic correctness. Utilizing the latest transformer-based language resources such as GPT-3, Bert, Roberta, and T5.
The latest transformer-based language resources can be used to improve paraphrase generation by providing more accurate and diverse paraphrases. These models can be fine-tuned on large-scale datasets to learn the relationship between different sentences and generate high-quality paraphrases. Additionally, incorporating syntax information into these models can further improve their performance. The use of unaligned pre-trained models can also help generate domain-specific paraphrases.

Author Contributions

Conceptualization, R.I. and M.L.K.; Methodology, R.I.; Software, R.I; Formal analysis, R.I.; Data curation, M.L.K.; Writing—original draft, R.I.; Writing—review and editing, M.L.K.; Supervision, R.M. (Rinaldi Munir), M.L.K., R.M. (Rila Mandala) and D.H.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Educational Fund Management Institution—Ministry of Finance, Indonesia. The APC was funded by the Educational Fund Management Institution—Ministry of Finance and Bandung Institute of Technology.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We would like to express our appreciation to D.H.W., who contributed to this research before passing away. We are grateful for their dedication to leading this project and providing valuable insights during the initial stage of the research. Their presence is deeply missed, and their contribution will always be remembered and appreciated.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Androutsopoulos, I.; Malakasiotis, P. A Survey of Paraphrasing and Textual Entailment Methods. J. Artif. Intell. Res. 2010, 38, 135–187. [Google Scholar] [CrossRef] [Green Version]
  2. Quirk, C.; Brockett, C.; Dolan, W. Monolingual Machine Translation for Paraphrase Generation; Association for Computational Linguistics: Stroudsburg, PA, USA, 2003. [Google Scholar]
  3. Lisa, C.M.L. Merging Corpus Linguistics and Collaborative Knowledge. Ph.D. Thesis, University of Birmingham, Birmingham, UK, 2009. [Google Scholar]
  4. Barrom-Cedeno, A.; Vila, M.; Marti, A. Plagiarism Meets Paraphrasing: Insights for the Next Generation in Automatic Plagiarism Detection. Assoc. Comput. Linguist. 2013, 10, 1–309. [Google Scholar] [CrossRef] [Green Version]
  5. Kittredge, R. Paraphrasing for Condensation in Journal Abstracting. J. Biomed. Inform. 2002, 35, 265–277. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Ilyas, R.; Widiyantoro, D.H.; Khodra, M.L. Building Candidate Monolingual Parallel Corpus from Scientific Papers. In Proceedings of the 2018 International Conference on Asian Language Processing, IALP, Bandung, Indonesia, 15–17 November 2018; pp. 230–233. [Google Scholar] [CrossRef]
  7. Teufel, S.; Siddharthan, A.; Tidhar, D. An Annotation Scheme for Citation Function. In Proceedings of the COLING/ACL 2006–SIGdial06: 7th SIGdial Workshop on Discourse and Dialogue, Sydney, Australia, 15–16 July 2006; pp. 80–87. [Google Scholar] [CrossRef] [Green Version]
  8. Liu, X.; Mou, L.; Meng, F.; Zhou, H.; Zhou, J.; Song, S. Unsupervised Paraphrasing by Simulated Annealing. arXiv 2019, 302–312. [Google Scholar] [CrossRef]
  9. Banerjee, S.; Lavie, A. METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgments. In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization; Association for Computational Linguistics: Stroudsburg, PA, USA, 2005; pp. 65–72. [Google Scholar]
  10. Chen, D.L.; Dolan, W.B. Collecting Highly Parallel Data for Paraphrase Evaluation. In Proceedings of the ACL-HLT 2011, 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, Portland, OR, USA, 19–24 June 2011; Volume 1, pp. 190–200. [Google Scholar]
  11. Carbonell, J.; Goldstein, J. Use of MMR, Diversity-Based Reranking for Reordering Documents and Producing Summaries. In SIGIR Forum (ACM Special Interest Group on Information Retrieval); ACM: New York, NY, USA, 1998; pp. 335–336. [Google Scholar] [CrossRef] [Green Version]
  12. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient Estimation of Word Representations in Vector Space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  13. NADAS, A. Estimation of Probabilities in the Language Model of the IBM Speech Recognition System. In IEEE Transactions on Acoustics Speech and Signal Processing; IEEE: New York, NY, USA, 1984; p. 27. [Google Scholar] [CrossRef]
  14. McKeown, K.R. Paraphrasing Questions Using Given and New Information. Am. J. Comput. Linguist. 1983, 9, 1. [Google Scholar]
  15. Lin, D.; Pantel, P. Discovery of Inference Rules for Question-Answering. Nat. Lang. Eng. 2001, 7, 343–360. [Google Scholar] [CrossRef] [Green Version]
  16. Kauchak, D.; Barzilay, R. Paraphrasing for Automatic Evaluation. In Proceedings of the Main Conference on Human Language Technology Conference of the North American Chapter of the Association of Computational Linguistics; Association for Computational Linguistics: New York, NY, USA, 2006; pp. 455–462. [Google Scholar]
  17. Prakash, A.; Hasan, S.A.; Lee, K.; Datla, V.; Qadir, A.; Liu, J.; Farri, O. Neural Paraphrase Generation with Stacked Residual LSTM Networks. arXiv 2016, arXiv:1610.03098. [Google Scholar]
  18. Vizcarra, G.; Ochoa-Luna, J. Paraphrase Generation via Adversarial Penalizations. In Proceedings of the Sixth Workshop on Noisy User-generated Text (W-NUT 2020); Association for Computational Linguistics: New York, NY, USA, 2020; pp. 249–259. [Google Scholar]
  19. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. Adv. Neural Inf. Process. Syst. 2017, 2017, 5999–6009. [Google Scholar]
  20. Wang, S.; Gupta, R.; Chang, N.; Baldridge, J. A Task in a Suit and a Tie: Paraphrase Generation with Semantic Augmentation. Proc. Conf. AAAI Artif. Intell. 2019, 33, 7176–7183. [Google Scholar] [CrossRef] [Green Version]
  21. Luong, M.T.; Pham, H.; Manning, C.D. Effective Approaches to Attention-Based Neural Machine Translation. In Proceedings of the EMNLP 2015: Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 1412–1421. [Google Scholar] [CrossRef] [Green Version]
  22. Ma, X.; Hovy, E. End-to-End Sequence Labeling via Bi-Directional LSTM-CNNs-CRF; Association for Computational Linguistics: Berlin, Germany, 2016; Volume 2, pp. 1064–1074. [Google Scholar] [CrossRef] [Green Version]
  23. Miao, N.; Zhou, H.; Mou, L.; Yan, R.; Li, L. CGMH: Constrained Sentence Generation by Metropolis-Hastings Sampling. AAAI Conf. Artif. Intell. 2019, 33, 6834–6842. [Google Scholar] [CrossRef] [Green Version]
  24. Lan, W.; Qiu, S.; He, H.; Xu, W. A Continuously Growing Dataset of Sentential Paraphrases. In Proceedings of the EMNLP 2017—Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 1224–1234. [Google Scholar]
  25. Bowman, S.R.; Vilnis, L.; Vinyals, O.; Dai, A.M.; Jozefowicz, R.; Bengio, S. Generating Sentences from a Continuous Space. In Proceedings of the CoNLL 2016—20th SIGNLL Conference on Computational Natural Language Learning, Berlin, Germany, 11–12 August 2016; pp. 10–21. [Google Scholar]
  26. Bhagat, R.; Hovy, E. What Is Paraphrase; Association for Computational Linguistics: New York, NY, USA, 2013; Volume 39. [Google Scholar] [CrossRef]
  27. Ganitkevitch, J.; Van Durme, B.; Callison-Burch, C. PPDB: The Paraphrase Database. In Proceedings of the NAACL-HLT–Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Atlanta, GA, USA, 9–14 June 2013; pp. 758–764. [Google Scholar]
  28. Dolan, W.B.; Brockett, C. Automatically Constructing a Corpus of Sentential Paraphrases. In Proceedings of the Third International Workshop on Paraphrasing, Yamamoto, Japan, 14 October 2005; pp. 9–16. [Google Scholar]
  29. Xu, W.; Ritter, A.; Callison-burch, C.; Dolan, W.B.; Ji, Y. Extracting Lexically Divergent Paraphrases from Twitter. Trans. Assoc. Comput. Linguist. 2014, 2, 435–448. [Google Scholar] [CrossRef]
  30. Pavlick, E.; Rastogi, P.; Ganitkevitch, J.; Durme, B.V.; Callison-Burch, C. PPDB 2.0: Better Paraphrase Ranking, Fine-Grained Entailment Relations, Word Embeddings, and Style Classification. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Short Papers), Beijing, China, 26–31 July 2015; pp. 425–430. [Google Scholar]
  31. Shi, L. Rewriting and Paraphrasing Source Texts in Second Language Writing. J. Second Lang. Writ. 2012, 21, 134–148. [Google Scholar] [CrossRef]
  32. Teufel, S. Do “Future Work” Sections Have a Purpose? Citation Links and Entailment for Global Scientometric Questions. In Proceedings of the 2nd Joint Workshop on Bibliometric-enhanced Information Retrieval and Natural Language Processing for Digital Libraries, Tokyo, Japan, 7–11 August 2017. [Google Scholar]
  33. Takahashi, K.; Ishibashi, Y.; Sudoh, K.; Nakamura, S. Multilingual Machine Translation Evaluation Metrics Fine-Tuned on Pseudo-Negative Examples for WMT 2021 Metrics Task. In Proceedings of the WMT 2021–6th Conference on Machine Translation, Online, 10–11 November 2021; pp. 1049–1052. [Google Scholar]
  34. Papineni, K.; Roukos, S.; Ward, T.; Zhu, W. BLEU: A Method for Automatic Evaluation of Machine Translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL), Philadephia, PA, USA, 6–12 July 2002; pp. 311–318. [Google Scholar]
  35. Doddington, G. Automatic Evaluation of Machine Translation Quality Using N-Gram Co-Occurrence Statistics; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 2002; p. 138. [Google Scholar] [CrossRef] [Green Version]
  36. Madnani, N.; Tetreault, J.; Chodorow, M. Re-Examining Machine Translation Metrics for Paraphrase Identification. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Seattle, WA, USA, 10–15 July 2012; pp. 182–190. [Google Scholar]
  37. Ji, Y.; Eisenstein, J. Discriminative Improvements to Distributional Sentence Similarity. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing (EMNLP 2013), Seattle, WA, USA, 18–21 October 2013; pp. 891–896. [Google Scholar]
  38. Brad, F. Neural Paraphrase Generation Using Transfer Learning; Association for Computational Linguistics: New York, NY, USA, 2017. [Google Scholar]
  39. Aziz, A.A.; Djamal, E.C.; Ilyas, R. Siamese Similarity Between Two Sentences Using Manhattan’s Recurrent Neural Networks. In Proceedings of the 2019 International Conference of Advanced Informatics: Concepts, Theory and Applications (ICAICTA), Yogyakarta, Indonesia, 20–22 September 2019; pp. 1–6. [Google Scholar]
  40. Saputro, W.F.; Djamal, E.C.; Ilyas, R. Paraphrase Identification Between Two Sentence Using Support Vector Machine. In Proceedings of the 2019 International Conference on Electrical Engineering and Informatics (ICEEI), Nanjing, China, 8–10 November 2019; pp. 406–411. [Google Scholar]
  41. Wubben, S.; Bosch, A.V.D.; Krahmer, E. Paraphrase Generation as Monolingual Translation: Data and Evaluation. In Proceedings of the 6th International Natural Language Generation Conference, Meath, Ireland, 7–9 July 2010; pp. 203–207. [Google Scholar]
  42. Mallinson, J.; Sennrich, R.; Lapata, M. Paraphrasing Revisited with Neural Machine Translation. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, Valencia, Spain, 3–7 April 2017; pp. 881–893. [Google Scholar]
  43. Zhao, S.; Lan, X.; Liu, T.; Li, S. Application-Driven Statistical Paraphrase Generation. In Proceedings of the 47th Annual Meeting of the ACL and the 4th IJCNLP of the AFNLP, Suntec, Singapore, 2–7 August 2009; pp. 834–842. [Google Scholar]
  44. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to Sequence Learning with Neural Networks. Nips 2014, 27, 3104–3112. [Google Scholar] [CrossRef] [Green Version]
  45. Parikh, A.P.; Täckström, O.; Das, D.; Uszkoreit, J. A Decomposable Attention Model for Natural Language Inference. arXiv 2016, arXiv:1606.01933. [Google Scholar] [CrossRef] [Green Version]
  46. Kirkpatrick, S.; Gelatt, C.D.; Vecchi, M.P. Optimization by Simulated Annealing. Science 1983, 220, 671–680. [Google Scholar] [CrossRef] [PubMed]
  47. Granville, V.; Rasson, J.P.; Krivánek, M. Simulated Annealing: A Proof of Convergence. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 652–656. [Google Scholar] [CrossRef]
  48. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.; Dean, J. Distributed Representations Ofwords and Phrases and Their Compositionality. Adv. Neural Inf. Process. Syst. 2013, 26, 1–9. [Google Scholar]
  49. Donoghue, D.P.; Saggion, H.; Dong, F.; Hurley, D.; Abgaz, Y.; Zheng, X.; Corcho, O.; Careil, J.-M.; Mahdian, B.; Zhao, X. Towards Dr Inventor: A Tool for Promoting Scientific Creativity; Jozef Stefan Institute: Ljubljana, Slovenia, 2014; p. 5. [Google Scholar]
  50. Hartigan, J.A.; Wong, M.A. Algorithm AS 136: A K-Means Clustering Algorithm. Applied Statistics 1979, 28, 100. [Google Scholar] [CrossRef]
  51. He, J.; Spokoyny, D.; Neubig, G.; Berg-Kirkpatrick, T. Lagging Inference Networks and Posterior Collapse in Variational Autoencoders. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019; pp. 1–15. [Google Scholar]
  52. Lin, C.-Y. ROUGE: A Package for Automatic Evaluation of Summaries. In Proceedings of the Text Summarization Branches Out; Association for Computational Linguistics: Barcelona, Spain, 2004; pp. 74–81. [Google Scholar]
Figure 1. Method architecture for paraphrase generator.
Figure 1. Method architecture for paraphrase generator.
Informatics 10 00034 g001
Figure 2. Example of the changing process of the sentence form based on action.
Figure 2. Example of the changing process of the sentence form based on action.
Informatics 10 00034 g002
Table 1. Paraphrase corpus state of the art.
Table 1. Paraphrase corpus state of the art.
NoPaperYearNameDomainTechnique
1Ganitkevitch et al. [27]2013PPDBFreePivoting
2Pavlick et al. [30] 2013PPDB 2.0FreePivoting
3Dolan et al. [28]2005MSRPFreeSVM
4Xu et al. [29]2014PITTwitterMulti-instance
learning
Table 2. Action factor probability scenarios.
Table 2. Action factor probability scenarios.
Action FactorSubstitutionInsertionDeletion
A90%5%5%
B80%10%10%
C70%15%15%
D60%20%20%
E50%25%25%
F40%30%30%
GUniform
Table 3. Significance table for Fleiss kappa.
Table 3. Significance table for Fleiss kappa.
Kappa ScoreSignificance
<0No agreement
0.01–0.20Slight agreement
0.21–0.40Fair agreement
0.41–0.60Moderate agreement
0.61–0.80Substantial agreement
0.81–0.99Almost perfect agreement
1Perfect agreement
Table 4. Performances on a standard dataset.
Table 4. Performances on a standard dataset.
ModelTwitterQuora
BLEURouge 1Rouge 2BLEURouge 1Rouge 2
Supervised
VAE3.4615.133.4013.9644.5522.64
LagVAE3.7417.203.7915.5249.2026.12
CGHM5.3219.965.4415.7348.7326.12
Unsupervised
UPSA5.3019.965.4418.2159.5132.63
StoPGEN6.2628.608.7522.3761.0940.79
In Table 4 it can be seen that StoPGEN evaluation results, which are given bold, show the best results
Table 5. Performances of StoPGEN on the citation dataset.
Table 5. Performances of StoPGEN on the citation dataset.
ModelAction FactorBLEURouge 1Rouge 2Rouge L
StoPGEN1A44.7458.6439.7155.28
B46.2360.3440.6356.73
C47.4761.9241.4258.11
D48.7863.5642.2559.54
E50.2265.1343.3361.06
F51.5966.7144.2262.43
G52.4567.8545.0163.48
StoPGEN2A27.8044.2328.6941.65
B35.0750.6132.3347.48
C40.7255.8535.3152.16
D45.2060.0638.0256.19
E49.1663.7340.8659.71
F52.4766.9843.5962.70
G54.5868.9845.5464.62
StoPGEN3A27.6946.0428.5542.87
B35.1352.8332.8649.23
C40.8358.2636.4154.27
D45.7762.6839.6958.42
E49.8066.3142.8261.78
F53.3669.4445.6564.69
G55.3771.2847.4666.32
In Table 5 it can be seen that StoPGEN that use scenario G, which are given bold, show the best results
Table 6. Performance compared model on citation dataset.
Table 6. Performance compared model on citation dataset.
ModelBLEURouge 1Rouge 2Rouge L
Unsupervised
StoPGEN355.3771.2847.4666.32
UPSA21.2045.9315.4341.55
Modified UPSA33.8151.2526.6745.94
Supervised
LSTM encoder-decoder25.7722.607.6820.13
bidirectional LSTM28.9326.1011.7523.44
Transformer18.9120.707.8318.46
Table 7. Stochastic results example.
Table 7. Stochastic results example.
InputWe use pre-trained glove (pennington et. Al., 2014) embeddings for our purposes
TargetWe use glove (pennington et. Al., 2014) for our word embeddings
Output 1Used trained glove (pennington et. Al., 2014) embeddings for implement through
Output 2Use pre-300 dimensional glove embeddings (pennington et. Al., 2014) word glove embeddings for our purposes
Output 3We use glove trained pretrained embedding (pennington et. Al., 2014) matrix for our through
Output 4Use trained glove word vectors trained glove word (pennington et. Al., 2014) embeddings for our purposes
Output 5We use pre-trained (pennington et. Al., 2014) embeddings for our purposes glove
Table 8. Paraphrase generation result of the proposed methods.
Table 8. Paraphrase generation result of the proposed methods.
InputStoPGEN1StoPGEN2StoPGEN3
the model was implemented using tensorflow
(abadi et. al., 2015)
the model was based this
implemented transformer using tensorflow implemented
(abadi et. al.,2015)
the implemented transformer model transformer was transformer pytorch using tensorflow
(abadi et. al., 2015)
the our implemented our model based was pytorch
using tensorflow
(abadi et. al., 2015)
du et al. (2017) pioneered
nn-based qg by adopting the seq2seq architecture (sutskever et at, 2014)
du et al. (2017) pioneered
nn-based qg by adopting seq2seq architecture based (sutskever et at, 2014)
du al et al. (2017) pioneered nn-based qg by furthermore seq2seq implement (sutskever et at, 2014)du et al. (2017) pioneered
nn-based qg by using seq2seq models
(sutskever et at, 2014)
we use the spanish-english ner corpus introduced in the 2018 calcs competition
(aguilar et. al., 2018), which contains a total of 67,223 tweets with 808,663 tokens
details dimensional 200 the corpus introduced in the data calcs competition
(aguilar et. al., 2018) contains a of 67,223 tweets with 808,663 scratch tokens
details employ pre
spanish-english ner corpus introduced in the 2018 calcs competition (aguilar et. al., 2018) which contains was
descent total of 67,223 tweets with poor tokens
we use the spanish-english corpus in 2018 calcs
competition (aguilar et. al., 2018) been it contains by was work a total 67,223
preserving tweets with 808,663 tokens
Table 9. Acceptability generation by all variants of StoPGEN.
Table 9. Acceptability generation by all variants of StoPGEN.
ModelAcceptability (with Action Label)
StoPGEN126.45
StoPGEN227.09
StoPGEN350.96
Table 10. Result of paraphrase generation by unsupervised methods.
Table 10. Result of paraphrase generation by unsupervised methods.
InputUPSAModified UPSAStoPGEN3
Tokenizing the output sentence: all words except
special tokens are segmented by farasa (abdelali et al., 2016) and then tokenized with arabert tokenizer
The uk offers all words except special tokens of
segmented by
(abdelali et al., 2016)
The sentence all from initialize words from except user tokens are by farasa (abdelali et al., 2016) and then with tokenizerTokenizing the output
sentence: all words tokens
segmented farasa
(abdelali et al., 2016) and then tokenized with arabert
tokenizer
We use the spanish-english ner corpus introduced in the 2018 calcs competition
(aguilar et al., 2018), which contains a total of 67,223 tweets with 808,663 tokens
We use the huge ner
currency corpus
introduced in the 2016 competition in
(aguilar et al., 2018)
We use the spanishenglish
corpus in 2018 calcs
competition (aguilar et al., 2018) been it contains by was work a total 67,223 preserving tweets with 808,663 tokens
We the spanish-english ner corpus in the 2018 calcs competition (aguilar et al., 2018), which contains a total of 67,223 tweets with 808,663
tokens
The first one is heuristic rules such as treating identical words as the seed
(artetxe et al., 2017), but this kind of method is restricted to languages sharing the alphabet
The one is heuristic rules such as treating identical words regarding the seed (artetxe et al., 2017)First one transformer is
transformer formation such bleu treating from senses
identical from words as tialize the seed (artetxe et al., 2017) but partially work kind work of this method is this restricted to languages sharing the
alphabet
The one heuristic rules such as the seed (artetxe et al., 2017), but this kind of method is
restricted to languages sharing the alphabet
Table 11. Acceptability generation results by the compared model.
Table 11. Acceptability generation results by the compared model.
ModelAcceptability (without Action Label)
UPSA16.80
Modified UPSA26.40
StoPGEN350.80
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ilyas, R.; Khodra, M.L.; Munir, R.; Mandala, R.; Widyantoro, D.H. Generating Paraphrase Using Simulated Annealing for Citation Sentences. Informatics 2023, 10, 34. https://doi.org/10.3390/informatics10020034

AMA Style

Ilyas R, Khodra ML, Munir R, Mandala R, Widyantoro DH. Generating Paraphrase Using Simulated Annealing for Citation Sentences. Informatics. 2023; 10(2):34. https://doi.org/10.3390/informatics10020034

Chicago/Turabian Style

Ilyas, Ridwan, Masayu Leylia Khodra, Rinaldi Munir, Rila Mandala, and Dwi Hendratmo Widyantoro. 2023. "Generating Paraphrase Using Simulated Annealing for Citation Sentences" Informatics 10, no. 2: 34. https://doi.org/10.3390/informatics10020034

APA Style

Ilyas, R., Khodra, M. L., Munir, R., Mandala, R., & Widyantoro, D. H. (2023). Generating Paraphrase Using Simulated Annealing for Citation Sentences. Informatics, 10(2), 34. https://doi.org/10.3390/informatics10020034

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop