Next Article in Journal
Navigating the Maps: Euclidean vs. Road Network Distances in Spatial Queries
Next Article in Special Issue
Cross-Project Defect Prediction Based on Domain Adaptation and LSTM Optimization
Previous Article in Journal
Personalized Advertising in E-Commerce: Using Clickstream Data to Target High-Value Customers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Specification Mining Based on the Ordering Points to Identify the Clustering Structure Clustering Algorithm and Model Checking

Cyberspace Security and Computer College, Hebei University, Baoding 071000, China
*
Author to whom correspondence should be addressed.
Algorithms 2024, 17(1), 28; https://doi.org/10.3390/a17010028
Submission received: 19 December 2023 / Revised: 8 January 2024 / Accepted: 8 January 2024 / Published: 10 January 2024
(This article belongs to the Special Issue Algorithms in Software Engineering)

Abstract

:
Software specifications are of great importance to improve the quality of software. To automatically mine specifications from software systems, some specification mining approaches based on finite-state automatons have been proposed. However, these approaches are inaccurate when dealing with large-scale systems. In order to improve the accuracy of mined specifications, we propose a specification mining approach based on the ordering points to identify the clustering structure clustering algorithm and model checking. In the approach, the neural network model is first used to produce the feature values of states in the traces of the program. Then, according to the feature values, finite-state automatons are generated based on the ordering points to identify the clustering structure clustering algorithm. Further, the finite-state automaton with the highest F-measure is selected. To improve the quality of the finite-state automatons, we refine it based on model checking. The proposed approach was implemented in a tool named MCLSM and experiments, including 13 target classes, were conducted to evaluate its effectiveness. The experimental results show that the average F-measure of finite-state automatons generated by our method reaches 92.19%, which is higher than most related tools.

1. Introduction

With the increasing demand for new functionalities in software, developers often release multiple updates to meet the relevant needs. However, software is frequently released without standard documentation, which leads to the software specification becoming outdated [1]. Moreover, software specification mining requires the expertise of professional developers, since it is a task that requires highly specialized skills to analyze the log documents. Nevertheless, specification mining is necessary as it enables us to identify abnormal system behavior by analyzing system logs. In specification mining, the linear-time temporal logic (LTL) formula and finite-state automaton (FSA) model are generally used to represent software specifications.
Many methods have been proposed to extract system behavior specifications, which allows developers to make informed decisions about the behavior of the system [2,3,4]. Hila Peleg et al. proposed a static technique to mine software specifications directly from the system source code without executing the system [5], but it is not applicable to large-scale systems. Caroline Lemieux et al. proposed an approach to efficiently mine specifications expressed as temporal logic formulas from the execution traces of a system [6]. However, the quality of the mined specifications is not perfect. For instance, if the methods in the input execution traces frequently occur in a particular order or the amount of input traces is too small, the FSAs inferred by k-tails [7] and many other tools are often not generalized and overfitted to the input execution traces. To solve this problem, Tien-Duy et al. proposed deep specification mining (DSM) [8,9], which combines an long short-term memory (LSTM) neural network [10] with the K-means algorithm [11] to mine finite-state automatons (FSAs). LSTM is used in DSM because LSTM is better at learning long-term dependencies and is scalable for long sequences. DSM leverages deep learning techniques to effectively extract the specifications of large-scale systems. It also utilizes clustering algorithms to generate FSAs that more concisely capture the specifications of the systems. However, DSM still faces challenges such as the problem of incorrect merging among states in FSAs generated by clustering feature values and random parameter selection in the clustering process.
To address the above-mentioned issues, we improve DSM by generating FSAs with the ordering points to identify the clustering structure (OPTICS) clustering algorithm [12], and refining the FSAs based on model checking to improve the quality of the FSAs. The proposed specification mining approach can briefly be described as follows:
  • DSM is employed to generate the feature values of states in the traces of the program;
  • The OPTICS clustering algorithm first generates a reachability plot based on the feature values. Then, the range of the clustering radius is determined according to the reachability plot, and several suitable radii are selected within this range to cluster the feature values to obtain the FSAs;
  • A model selection algorithm is used to select the FSA with the highest F-measure from the generated FSAs. Specifically, the F-measure is defined in Section 3.2;
  • Texada [6] is employed to generate LTL properties specifying the desired behaviors of the system. Then, model checking of the FSA is performed to verify whether the selected FSA satisfies these properties. If not, we refine the FSA by adding and deleting the corresponding states and transitions in it according to the type of the property which the FSA violates, so that the traces that violate the property are removed from the FSA.
The three main contributions of this paper are as follows:
  • We propose a specification mining approach based on the OPTICS clustering algorithm and model checking. Specifically, during the generation of FSAs, we employ the OPTICS clustering algorithm to improve the clustering effect and solve the parameter setting problem so that the parameters can be intuitively set. In addition, we refine the generated FSAs based on model checking to alleviate the problem of incorrect merging among states in FSAs generated by clustering feature values and improve the quality of specifications;
  • We implemented our approach in a tool called MCLSM;
  • We conducted experiments on 13 target library classes from [8,9,13,14] to evaluate the performance of our tool, and the results show that the average F-measure of the FSAs generated by MCLSM reaches 92.19%, which is 175.85%, 197.67%, 66.95%, 177.17%, 233.41%, 367.49%, and 28.48% higher than that of FSAs generated by the most related tools, i.e., 1-tails [7], 2-tails [7], SEKT 1 [13], SEKT 2 [13], CONTRACTOR++ [13], TEMI [13], and DSM [8,9], respectively.
The remainder of the paper is structured as follows. Section 2 provides the background information. Section 3 introduces the specification mining approach based on the OPTICS clustering algorithm and model checking in detail. We evaluate our approach in Section 4. Section 5 presents the related work and Section 6 concludes the paper.

2. Preliminary

2.1. Model Checking

Model checking is a formal verification technique that determines whether certain properties are satisfied by a system model. In model checking, the system to be verified is modeled as a state transition diagram or finite state automaton. The desired specifications are usually expressed as a set of temporal logic formulas, such as LTL [15] and computing tree logic (CTL) [16]. The basic idea of model checking is to check whether a given specification is satisfied by automatically traversing all possible states of the system. If the system does not satisfy the specification, the model checker generates a counterexample that helps the designer to fix defects in the design.
Here, we focus on model checking of LTL properties. An LTL formula ∅ over a countable set P r of atomic propositions is inductively defined as follows:
ϕ : : = p | ¬ ϕ | ϕ 1 ϕ 2 | X ϕ | ϕ 1 ϕ 2
where p P r represents an atomic proposition, ϕ 1 and ϕ 2 are all well-formed LTL formulas. For the semantics of these formulas, please refer to [15].
The abbreviations ∨ and → are defined as usual. Some useful derived formulas are given as follows:
F ϕ = def t r u e ϕ G ϕ = def ¬ F ¬ ϕ

2.2. OPTICS Clustering Algorithm

In the past few decades, the research of clustering algorithms has undergone significant development. Early, K-means [11] clustering algorithms provided a simple and intuitive framework, but their applicability is limited by the data shape and cluster structure. Then, the hierarchical clustering algorithm introduces a hierarchical structure, which provides convenience for understanding the relationship between clusters at different levels. Next, Martin Ester et al. [17] proposed the DBSCAN algorithm, which focuses on density, making clustering more robust for data with an arbitrary shape and density distribution. In recent years, K-means clustering, the Gaussian mixture model (GMM), and mixtures of multivariate clustering algorithms have been applied to the field of chemistry, which has promoted the application of unsupervised learning in other disciplines [18]. With the development of technology, Mantas Lukauskas et al. [19] proposed a density clustering method based on improved inverse formula density estimation. The new method has a good effect when dealing with low dimensional data. In addition, Zhenzhou Wang [20] proposed a clustering method based on morphological operations that worked well on 2D and 3D data.
The OPTICS clustering algorithm creates an augmented ordering of the database representing its density-based clustering structure. Let the object set be O = { o 1 , o 2 , …, o m }. Some important definitions associated with the OPTICS clustering algorithm are defined as follows:
  • ϵ D o m a i n : For o j O, its ϵ D o m a i n is a subset of O containing objects whose distance from o j is not greater than ϵ . That is, N ϵ ( o j ) = { o i O| distance( o i , o j ) ≤ ϵ }. The number of objects in N ϵ ( o j ) is noted as N ϵ ( o j ) . Usually, e p s is used to represent the clustering radius ϵ ;
  • Core object: for any o j O, o j is a core object if N ϵ ( o j ) m i n _ s a m p l e s , where m i n _ s a m p l e s is an integer constant;
  • Core distance: the minimum radius that makes an object o a core object is called the core distance of o;
  • Reachability distance: the reachability distance of an object p with respect to a core object o denoted as r d ( p , o ) is the maximum value between the actual distance of o to p and the core distance of o.
Let the clustering radius be infinite, i.e.,  e p s = i n f . OPTICS generates the reachability plot as follows:
  • The core object queue C q , ordered queue O q , result queue R q , and reachability distance queue R d q are initialized to empty;
  • All core objects in O are chosen based on m i n _ s a m p l e s and e p s , and then added to C q ;
  • If there is an element in C q that has not been processed, jump to 4, and otherwise 8.
  • An unprocessed object o is randomly taken out from C q and placed in R q . Then, it is marked as processed. Further, each object p satisfying r d ( p , o ) e p s is selected from C q and added to O q in an ascending order of the reachability distance with respect to o;
  • If O q is not empty, jump to 6, and otherwise 3;
  • The reachability distance of the first object p 1 in O q , with respect to the last object in R q , is stored in R d q . Then, p 1 is removed from both O q and C q , and added into R q ;
  • Each object whose reachability distance with respect to the last object r in R q is not greater than e p s is selected from C q . Then, these objects are sorted in ascending order of the reachability distance with respect to r and replace the objects in O q . Finally, jump to 5;
  • The reachability plot is plotted based on R q and R d q .
The OPTICS clustering algorithm [12] is considered more effective than K-means [11] and other clustering algorithms when dealing with high-level data because it can identify changes in density. In terms of clustering parameter selection, OPTICS offers a more reasonable approach compared to clustering algorithms like K-means and DBSCAN [17]. It generates a reachability plot, as depicted in Figure 1, which allows for visual parameter selection. This reduces the impact of randomly setting clustering parameters on the final clustering results. In Figure 1, the horizontal axis represents the sequence of output objects, and the vertical axis stands for the reachability distance of the current object with reference to the previous object, i.e., the reachable distance between object A and object C is 1.5. By setting the value of e p s to 3.9, all the objects can be effectively divided into two clusters: {A,C,D,F} and {B,E}.

2.3. DSM

DSM is a state-of-the-art specification mining tool [8,9], which applies deep learning to specification mining in order to improve the accuracy of the mined automata. However, it also experiences the merging of errors problem between states of the final obtained automata. DSM involves six processes: test case generation and trace collection, model training, trace sampling, feature extraction, clustering, and model selection. As shown in Figure 2, the skeleton of the approach is briefly given as follows:
  • DSM uses Randoop [21] to generate a large number of test cases. Then, it executes the target program with these cases and stores the traces (the methods calls) of the program in a trace set;
  • The LSTM model is trained by all execution traces;
  • In order to improve the efficiency of specification mining, a subset S T r of traces that can cover all adjacent method pairs of execution traces is extracted;
  • Two types of features, i.e.,  F m i and P m i , are extracted in each state S i based on the subset of traces and the LSTM model, where F m i captures information of previously invoked methods before S i , and  P m i captures information about methods immediately after state S i using the LSTM model. For instance, suppose that all methods that appear in the program are in a set M = { m 1 , m 2 , , m n } and a trace is t r = m x 1 , …, m x k , …, m x l , where m x i ( 1 i l and m x i M ) in t r is the invoked method at state S i . At state S k , methods from m x 1 to m x k in t r are invoked. Thus, F m x i k = 1 ( 1 i k ) and for other methods m j in M, F m j k = 0 . All feature values of all states in all traces belonging to the subset S T r are stored in a set O = { o 1 = B 1 , A 1 , …, o l = B h , A h }, where h is the number of states in all traces and for each 1 y h , B y = F m 1 y , …, F m n y , and A y = P m 1 y , …, P m n y ;
  • The K-means [11] and hierarchical clustering algorithms are used to cluster the feature values and generate FSAs. An FSA can be represented as a five-tuple D = (Q, Σ , δ , q 0 , F), where Q is a finite set of states, Σ is a table of input characters, δ is a transition function mapping Q × Σ to Q, q 0 Q is the initial state, and FQ is the set of acceptable states;
  • The FSA with the highest F-measure is selected from the generated FSAs.

3. Specification Mining Based on the OPTICS Clustering Algorithm and Model Checking

Our approach involves three key processes: clustering, model selection, and refinement. As shown in Figure 3, the skeleton of the approach is briefly given as follows:
  • We cluster feature values obtained by DSM based on e p s in the reachability plot, which is drawn according to the OPTICS clustering algorithm [12];
  • The FSAs with the highest F-measure are selected during the model selection;
  • We perform model checking on the selected FSAs and refine the FSAs based on the verification result to obtain the final FSAs.

3.1. Clustering

In order to create an automaton that captures the specification of the target library classes, the feature values in set O generated by DSM are clustered to generate FSAs, where O = { o 1 = B 1 , A 1 , , o h = B h , A h } as defined in Section 2.2.
To do that, some variables used in the clustering algorithm are defined as follows:
  • C q , R q , and e p s are defined as in Section 2.2;
  • K q is used to store objects of the same class as a given object;
  • K is used to mark the category to which the current objects belong;
  • The map L a b e l s is used to map the category K to the corresponding core objects at one clustering;
  • A l l _ l a b e l s is used to store L a b e l s after each clustering;
  • The variable n u m is used to control the increment of e p s at each clustering;
  • P l o t represents the reachability plot generated by the OPTICS clustering algorithm.
In order to alleviate the influence of clustering parameters on the clustering results, we first use the OPTICS clustering algorithm shown in Section 2.2 to generate the reachability plot, and then determine the range of e p s and the value of n u m based on the plot. Further, we cluster objects in O according to e p s and n u m to generate FSAs. Algorithm 1 shows the clustering process. It takes the feature values in O extracted by DSM as the input, and outputs a set A l l _ l a b e l s storing well-classified feature values after multiple clustering. In the algorithm, we first initialize C q , K q , R q , L a b e l s , and A l l _ l a b e l s to ∅, and e p s to i n f (Line 1). Then, the reachability plot P l o t is generated based on OPTICS clustering algorithm O P T I C S _ R P l o t ( O ) , the details of which are shown in Section 2.2 (Line 2). The purpose of generating P l o t is to observe the shape and trend in order to determine the appropriate range [ a , b ] of e p s and the increment value n u m of e p s for each time of clustering (Line 3). In order to obtain A l l _ l a b e l s , which is required to generate FSAs, we cluster the feature values in O with e p s ranging from a to b (Lines 4–22). For each clustering time, we first find all core objects needed for clustering from O and add them to C q (Line 6). Then, we cluster the core objects in C q to obtain L a b e l s . The following is the process of how we generate L a b e l s . We first find objects whose distances to C q [ 0 ] are not greater than e p s in C q (except C q [ 0 ] ), and add them to K q , which stores objects of the same category as C q [ 0 ] . Then, we remove C q [ 0 ] from C q and put it in R q . The distance is the Euclidean distance between two feature values, i.e., two objects (Lines 9–10). Further, we find the core objects in C q that belong to the same class as each object in K q , and add these objects to K q until all objects in K q are processed. Each time the objects added to K q should be deleted from C q to avoid duplication, and the objects that have been processed in K q should also be removed and added to R q (Lines 11–14). When K q is empty, we mark the objects in R q as class K and set R q to empty in order to continue classifying the remaining objects in C q (Lines 15–16). When C q is empty, a clustering of objects in C q is completed. We add L a b e l s to A l l _ l a b e l s and set L a b e l s to empty for generating the next L a b e l s (Lines 19–21). Finally, we construct FSAs based on A l l _ l a b l e s according to the method in [9].
Figure 4 shows an FSA generated based on the OPTICS clustering algorithm. In the FSA, the start state is q 0 = C 1 and the set of acceptable states is F = { C 4 } . The input character table Σ is a collection of methods, Q is a collection of all states, and each transition between method and its connected states makes up the transition function δ .
Algorithm 1: Clustering Process
Input: O
Output: A l l _ l a b e l s
 1Initialize: C q , K q , R q , L a b e l s , A l l _ l a b e l s = ∅, e p s = i n f ;
 2 P l o t = O P T I C S _ R P l o t ( O ) ;
 3Determine the range [ a , b ] of e p s and the value of n u m according to P l o t ;
 4 e p s = a ;
 5while  e p s b  do
 6 C q ← Find all core objects in O based on e p s and m i n _ s a m p l e s ;
 7 K = 0 ;
 8 while  C q ≠ ∅ do
 9 R q ← Take out C q [ 0 ] from C q ;
10 K q ← Take out all objects o satisfying d i s t a n c e ( o , R q [ n 1 ] ) e p s from C q ;
//   ( n represents the number of objects in R q
11 while  K q ≠ ∅ do
12 K q ← Take out all objects o satisfying d i s t a n c e ( o , K q [ 0 ] ) e p s from C q ;
13 R q ← Take K q [ 0 ] out of K q ;
14 end
15 L a b e l s ← Grouping objects in R q into K category;
16 R q = ∅;
17 K = K + 1;
18 end
19 A l l _ l a b e l s L a b e l s ;
20 L a b e l s = ∅;
21 e p s = e p s + n u m ;
22end
23return  A l l _ l a b e l s ;

3.2. Model Selection

DSM [8,9] often chooses the best FSA by predicting the P r e c i s i o n and R e c a l l of all generated FSAs. Here, we define P r e c i s i o n and R e c a l l as follows:
P r e c i s i o n ( D ) = | M P T R f s a _ p a i r s ( D ) | | f s a _ p a i r s ( D ) |
where D is the FSA to be evaluated, M P T R represents all method pairs appearing in the input trace set T R , and  f s a _ p a i r s ( D ) represents the method pairs appearing in D. Note that we call a method pair (x, y) that appears in D if x and y are labeled at two adjacent edges in D, respectively. In fact, we should consider the traces in T R as positive samples and the traces in D as classified into positive samples. However, the number of traces in D usually cannot be calculated. Thus, we consider the method pairs in T R as positive samples and the method pairs in D as classified into positive samples. In Formula (1), | f s a _ p a i r s ( D ) | represents the data that are classified as positive, and | M P T R f s a _ p a i r s ( D ) | represents the data that are actually positive in the classification.
R e c a l l = | a c c e p t e d _ t r a c e | | T R |
where | a c c e p t e d _ t r a c e s | represents the number of traces in the trace set that the generated FSA can accept, and | T R | represents the number of all execution traces. Using just one of P r e c i s i o n or R e c a l l to evaluate an FSA cannot comprehensively evaluate the advantages and disadvantages of the FSA. A higher P r e c i s i o n means an FSA accepts fewer traces that should not be accepted, while a higher R e c a l l means an FSA accepts more traces that should be accepted. Thus, we combine P r e c i s i o n and R e c a l l to obtain the F-measure (Formula (3)) as the actual scoring criteria.
F m e a s u r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l

3.3. FSA Refinement Based on Model Checking

The FSAs generated in Section 3.1 are prone to incorrect paths in the model due to erroneous merging among states.
To mitigate this issue, we refine the FSAs based on model checking. Specifically, we explore the FSA and check if it satisfies the LTL formulas. When a violation is detected, we refine the automaton to eliminate the violation and improve the precision of the FSA. We repeat this process until there is no violation or the number of times for the refinement reaches a threshold value T.
In order to use the partial specification of the software (LTL formulas) to guide the refinement of the overall specification of the software (FSA), we input the traces generated in DSM [8,9], property templates, and confidence thresholds into Texada [6] to generate the following three forms of LTL formulas, which were shown to specify the temporal properties of automata by Beschastnikh et al. [22].
  • A I F ( a , b ) : an occurrence of event a must be immediately followed by event b, i.e., G (aXb);
  • A I P ( a , b ) : an occurrence of event a must be immediately preceded by event b, i.e., F ( a ) → (¬a∪ (bXa));
  • N I F ( a , b ) : an occurrence of event a is never immediately followed by event b, i.e., G (aXb)).
The variables used in the refinement process are defined as follows:
  • D and D represent the FSAs before and after refinement, respectively;
  • K is a set of LTL formulas generated by Texada;
  • T r a n s is a set of tuples S M , q 2 , m 2 , q 3 , M S , where S M is a set of tuples q 1 , m 1 , M S is a set of tuples m 3 , q 4 , q 1 , q 2 , q 3 , and q 4 are states, and m 1 , m 2 , m 3 , and m 4 are methods;
  • C i ( 1 i n u m ) represents newly added states in the FSA.
Algorithm 2 and Figure 5 show the refinement process based on model checking in detail, where the initial FSA D and a set K of LTL formulas are taken as the input. In the process, n u m and t i m e s are initialized to 0 and D is initialized to D (Line 1). We then verify whether the number of times for refinement reaches the threshold value, and whether there exists a property violated by the FSA, i.e., whether the software behavior described by FSA does not match that described by the LTL formula (Line 2). If so, F i n d E r r T r returns the set T r a n s of error segments in the FSA according to the type of violated property as follows:
Algorithm 2: Refinement Process
Input: D = (Q, Σ , δ , q 0 , F), K = { K 1 , …, K t }
Output:  D = ( Q , Σ , δ , q 0 , F)
 1Initialize:  n u m = 0 , t i m e s = 0 , D = D ;
 2while  t i m e s + + < T and there exists a property K j K violated by D  do
 3 T r a n s = F i n d E r r T r ( K j , D ) ;
 4 for each S M , q 2 , m 2 , q 3 , M S T r a n s  do
 5 if  K j is N I F ( a , b )  then
 6 Delete δ ( q 2 , m 2 ) = q 3 ;
 7 Q = Q { C n u m };
 8 Add transition rules to δ : δ ( q 2 , m 2 ) = C n u m and δ ( C n u m , m 3 ) = q 4 for
 each m 3 , q 4 M S ;
 9 n u m = n u m + 1;
10 end
11 if  K j is A I P ( a , b )  then
12 Delete δ ( q 2 , m 2 ) = q 3 ;
13 Q = Q { C n u m } ;
14 Add transition rules to δ : δ ( q 1 , m 1 ) = C n u m for each q 1 , m 1 S M ;
15 n u m = n u m + 1;
16 Q = Q { C n u m } ;
17 Add transition rules to δ : δ ( C n u m 1 , a) = C n u m and δ ( C n u m , m 2 ) = q 3 ;
18 n u m = n u m + 1;
19 end
20 if  K j is A I F ( a , b )  then
21 Delete δ ( q 2 , m 2 ) = q 3 ;
22 Q = Q { C n u m } ;
23 Add a transition rule to δ : δ ( q 2 , m 2 ) = C n u m ;
24 n u m = n u m + 1;
25 Q = Q { C n u m };
26 Add transition rules to δ : δ ( C n u m 1 , b) = C n u m and δ ( C n u m , m 3 ) = q 4
  for each m 3 , q 4 M S ;
27 n u m = n u m + 1;
28 end
29 end
30 D = ( Q , Σ , δ , q 0 , F);
31end
  • If the violated property is N I F ( a , b ) , it returns all segments S M , q 2 , m 2 , q 3 , M S that satisfy m 2 = a , δ ( q 2 , a ) = q 3 , S M containing all tuples q 1 , m 1 satisfying δ ( q 1 , m 1 ) = q 2 , M S containing all tuples m 3 , q 4 satisfying δ ( q 3 , m 3 ) = q 4 , and there is m 3 , q 4 M S satisfying m 3 = b ;
  • If the violated property is A I P ( a , b ) , it returns all segments S M , q 2 , m 2 , q 3 , M S that satisfy m 2 = b , δ ( q 2 , b ) = q 3 , M S containing all tuples m 3 , q 4 satisfying δ ( q 3 , m 3 ) = q 4 , S M containing all tuples q 1 , m 1 satisfying δ ( q 1 , m 1 ) = q 2 , and for each q 1 , m 1 S M , m 1 a ;
  • If the violated property is A I F ( a , b ) , it returns all segments S M , q 2 , m 2 , q 3 , M S that satisfy m 2 = a , δ ( q 2 , a ) = q 3 , S M containing all tuples q 1 , m 1 satisfying δ ( q 1 , m 1 ) = q 2 , M S containing all tuples m 3 , q 4 satisfying δ ( q 3 , m 3 ) = q 4 , and for each m 3 , q 4 M S , m 3 b .
Further, we refine the FSA with different refinement rules depending on the type of formula that the error segment S M , q 2 , m 2 , q 3 , M S violates as follows:
  • If it violates N I F ( a , b ) , m 2 = a , we first delete the transition rule from state q 2 to state q 3 using method a, so that the traces violating N I F ( a , b ) are removed (Line 6). Then, in order to prevent traces that do not violate N I F ( a , b ) from being deleted, we add a state C n u m and some transition rules. First, we add a transition rule from state q 2 to state C n u m using method a. Next, we add transition rules from C n u m to each q 4 in M S using the corresponding method (except b) in M S (Lines 7–8);
  • If it violates A I P ( a , b ) , m 2 = b , we first delete the transition rule from state q 2 to state q 3 using method b, so that the traces violating A I P ( a , b ) are removed (Line 12). Then, in order to increase the precision of the automaton, we add a state C n u m and transition rules from each state q 1 in S M to C n u m using the corresponding method in S M . Next, we update the state subscript n u m and create another new state C n u m . Further, we add a transition rule from state C n u m 1 to state C n u m using method a and a transition rule from state C n u m to state q 3 using method b (Lines 13–17);
  • If it violates A I F ( a , b ) , m 2 = a , we first delete the transition rule from state q 2 to state q 3 using method a, so that the traces violating N I F ( a , b ) are removed (Line 21). Then, in order to increase the precision of the automaton, we add a state C n u m and some transition rules. First, we add a transition rule from state q 2 to state C n u m using method a. Next, we update the state subscript n u m and create a new state C n u m . We then add a transition rule from C n u m 1 to C n u m using method b and transition rules from C n u m to each state q 4 in M S using the corresponding method in M S (Lines 23–26).
After performing a round of refinement, we generate a new FSA D , and the next round of refinement is performed on the basis of D (Line 30).
The example in Figure 6 shows the refinement process for an FSA that violates N I F ( i s E m p t y : T R U E , r e m o v e : T R U E ) . First of all, we use model checking with the LTL formula to find an error segment S M = { C 0 , A r r a y L i s t } , q 2 = C 2 , m 2 = i s E m p t y : T r u e , q 3 = C 3 , M S = { r e m o v e : T R U E , C 4 , a d d : A l l : T R U E , C 4 , i n d e x o f , C 4 } violating N I F ( i s E m p t y : T R U E , r e m o v e : T R U E ) . Next, we delete the transition rule from C 2 to C 3 through i s E m p t y : T R U E , so that the trace that violates N I F ( i s E m p t y : T R U E , r e m o v e : T R U E ) is removed. Further, in order to prevent traces that do not violate N I F ( i s E m p t y : T R U E , r e m o v e : T R U E ) from being deleted, we add a transition rule from C 2 to C 0 through i s E m p t y : T R U E and transition rules from C 0 to C 4 through a d d A l l : T R U E and i n d e x o f . After the above operation, we have refined model A, which contained errors, into model B, which is error-free.

4. Empirical Evaluation

We implemented the proposed approach in a tool named MCLSM and evaluated our approach to answer the following question: How effective is MCLSM compared with the existing specification mining tools?

4.1. Dataset

In the experiment, we used all 13 target library classes from [8,9,13,14] as the benchmark to evaluate MCLSM.
In this benchmark, ArrayList serves as the implementation of a variable-size array of list interfaces, while LinkedList is primarily employed for creating linked list data structures. HashSet, implementing the set interface, ensures the absence of duplicate elements and does not guarantee the order of elements. HashMap, functioning as a Hashtable, facilitates key-value mapping with rapid access speeds. Hashtable, originating from the original java.util, is a concrete implementation of a dictionary. The Signature class in Java is utilized to provide a digital signature algorithm for applications, while the Socket class offers a comprehensive set of network communication methods and properties. ZipOutputStream allows direct content writing to the zip package. The StringTokenizer class implements both the Iterator and Enumeration interfaces. These nine target class libraries are integral components of the Java Development Kit. Additionally, four other class libraries are used, namely, StackAr from the Daikon project, NFST (NumberFormatStringTokenizer) and ToHTMLStream from Apache, and SMTPProtocol from Columba.ristretto.
Table 1 shows the information of these classes. The column “Target Library Class” displays the names of the target classes. The column “M” lists the number of class methods analyzed in the target classes. The column “Generated Test Cases” is the number of test cases generated by Randoop, and the column “Recorded Method Calls” gives the number of method calls recorded in all execution traces of the target classes. The data in Table 1 consist of 13 target classes, which are both complex and simple; thus, we chose to use these data to evaluate the effectiveness of our method on different sizes of data.

4.2. Experience Setting

The experiments were carried out on a 64-bit Ubuntu 18.04 LTS with a 3.20 GHz Intel(R) Core(TM) i5-10505 processor and 16 GB memory. MCLSM uses the OPTICS clustering algorithm to cluster feature values. During the clustering process, we set m i n _ s a m p l e s to 5, and selected the range of e p s according to the reachability plot. In the clustering process, e p s represents the radius of the cluster and is used to determine whether one point can reach another. By adjusting the size of the e p s value, we can change the shape of the cluster. Smaller e p s may result in the formation of more and smaller clusters, while larger e p s may combine more points into one cluster. Therefore, it is very important to select the appropriate e p s in datasets with uneven densities. We determined the range of e p s using the reachability plot and adjusted the e p s in this range to achieve the best clustering effect. In a reachability plot, the boundaries of a cluster usually correspond to points with large variation in reachable distance, that is, high or low points. First, we set a threshold. Then, we found all the high and low points in the reachability plot and removed points larger than the threshold. Next, we determined the range of e p s using the maximum and minimum values of these points. Finally, we selected the appropriate step size to cluster within the defined range and obtained the most suitable e p s value for the current data according to the clustering effect. For example, Figure 7 shows the reachability plot generated for the ArrayList class, where the horizontal axis represents each feature value and the vertical axis stands for the reachability distance of the current feature value with reference to the previous one. For this class, we selected the range of e p s from 0.3 to 4.3, with e p s increasing by 0.5 for each cluster.
The MCLSM uses the LTL formulas to guide Algorithm 2 to refine the FSA. When obtaining the LTL formula using Texada, we set the confidence thresholds in Texada [6] to 1.0 as we were only interested in properties that were never falsified. In the above process, we obtained a total of 94 LTL formulas, and these LTL formulas were used to refine FSAs.

4.3. Experimental Results and Analyses

We compared the effectiveness of MCLSM with the related tools, including k-tails [7], SEKT [13], CONTRACTOR++ [13], trace-enhanced MTS inference (TEMI) [13], and DSM [8,9]. For k-tails and SEKT, we chose k { 1 , 2 } . We followed the F-measure with the DSM method to measure the F-measure of our proposed approach. Table 2 shows the F-measure (%) value of each tool for 13 target library classes. In the table, tools “1-tail” and “2-tail” represent traditional 1-tails and 2-tails, respectively [23], “SEKT 1” and “SEKT 2” stand for State-Enhanced 1-tails and 2-tails, respectively, and “CON++” is short for CONTRACTOR++. The result “-” indicates that the result is unavailable.
From the experimental results in Table 2, we can see that SEKT 2, CONTRACTOR++, and TEMI failed to generate the FSA on 2, 4, 6 classes, respectively, while MCLSM successfully produced the FSA for all classes. The F-measure values of FSAs generated by MCLSM for most classes were higher than those of FSAs generated by other tools except for classes Signature and SringTokenizer. In total, the average F-measure values with reference to MCLSM reached 92.19%, which were 175.85%, 197.67%, 66.95%, 177.17%, 233.41%, and 367.49% higher than 1-tails, 2-tails, SEKT 1, SEKT 2, CONTRACTOR++, and TEMI, respectively.
Specifically, MCLSM was higher than DSM by 28.48%. Because MCLSM uses the OPTICS clustering algorithm to cluster feature values, it is not necessary to set the clustering radius in advance. The OPTICS clustering algorithm generates reachability plots, which adaptively captures density changes between feature values. Then, according to the reachability plot, the appropriate radius is selected for feature value clustering. In contrast, the DSM uses the K-means clustering algorithm, which requires a pre-set clustering radius. This can result in a feature value being incorrectly assigned to an inappropriate cluster. In addition, the FSA generated by the clustering algorithm of DSM may encounter cases of false merging between states. To address this, our approach refines the FSA based on model checking, reducing false merging between states and improving the FSA accuracy. Note that by analyzing two classes, i.e., Signature and SringTokenizer, for which MCLSM did not performs better than DSM, we found that in the the cluster process of MCLSM, outliers are produced by the OPTICS clustering algorithm, so the F-measure is slightly lower than for DSM.
This is because MCLSM generates outliers when clustering data in Signature and SringTokenizer using the OPTICS clustering algorithm, while the K-means clustering algorithm used in DSM does not generate outliers. As a result, the FSA F-measure generated by MCLSM when processing these data was lower than that of DSM.
To alleviate the impact of outliers on the accuracy of the FSA, we can choose larger e p s values, which help to treat outliers as part of a cluster. Therefore, on the premise of ensuring the accuracy of clustering results, it is helpful to choose as large an e p s value as possible to solve the problem of outliers in clustering.
We obtained experimental statistics on the number of different method calls. When experimenting with the HashSet, Signature, Socket, StringTokenizer, StackAr, ToHTMLStream, and SMTPProtocol classes, each class generated more than 100,000 method calls. In this case, the average F-measure value of DSM was 78.52%, and the average F-measure value of MCLSM was 92.03%. When experimenting with the ArrayList, LinkedList, HashMap, Hashtable, ZipOutputStream, and NFST classes, each class generated less than 100,000 method calls. In this case, the average F-measure value of DSM was 63.86% and the average F-measure value of MCLSM was 93.44%. These results indicate that MCLSM has a more significant effect than DSM, and MCLSM exhibits excellent performance on different data scales.

4.4. Threats to Validity

4.4.1. Threats to Internal Validity

Although we dealt with the outliers generated during the clustering process, the accuracy of the FSA generated by clustering various low-density eigenvalues using the OPTICS clustering algorithm can still be further improved. In the future, we will use other clustering algorithms in combination with the OPTICS clustering algorithm in order to deal with possible outliers more efficiently.

4.4.2. Threats to External Validity

As shown in Section 4.1, although we made expanded upon the experimental data compared with previous work, the amount of experimental data are still insufficient. This situation may affect the generalizability of the experimental results. In the future, we will collect more data to test the universality of our approach.

5. Related Work

Specification mining techniques have received a lot of attention due to the increasing complexity of software systems and the high cost of manual analysis for these systems. The proposed specification mining techniques can be broadly classified into two categories: one generates specifications expressed in the temporal logic formulas that all execution traces satisfy by taking execution traces and predefined property templates as inputs [6,24,25,26,27,28,29]; the other produces models expressed by FSAs [2,7,8,9,13,23,30,31], which precisely specify the whole system behavior.

5.1. Mining Specifications Expressed in Temporal Logic Formulas

In this class of methods, an instance of the property is first obtained by replacing each event in the property template with a method in traces, and then all traces are checked to see whether the property instance is satisfied. If so, the property instance is output.
Texada [6] can extract LTL expressions of arbitrary complexity. It takes LTL property templates and traces as the input, and outputs property instances. In addition, Texada supports properties with imperfect confidence by providing two controls: the confidence threshold and support threshold, where the former refers to the minimum proportion of the number of traces satisfying the property instance to the total number of traces, and the latter refers to the minimum number of traces satisfying the property instance.
TRE [32] was developed for real-time systems. It extends the regular input expressions by adding operators specifying the constraint time between events and synthesizes a timing automaton for the given expression. Traces are then checked to see whether they satisfy the timed automaton or not. TREM [27] separates the front and back ends based on TRE, and provides a visual interface in the front end. Alessio Cecconi et al. [33] proposed a comprehensive measurement framework to solve the problems of a lack of feedback and scalability in declarative process mining. Ezio Bartocci et al. [34] summarize the existing methods for extracting specifications from cyber-physical systems (CPS) based on supervised versus unsupervised learning.

5.2. Mining Specifications Expressed as Models Similar to FSAs

This class of specification mining methods generates models similar to FSAs. Usually, an initial model is first generated based on the traces. The model is then refined using different operations to obtain the final model.
The K-tails [7] algorithm first constructs a tree-like deterministic finite automaton (DFA) that is consistent with the input traces and accepts all of them. It then merges the states with the same sequence of calls in the next K step. Variants of K-tails refer to different merge criteria that do not require the exact matching of call sequences [23]. Contractor [30] generates a model like an FSA to describe the behavior of software based on program invariants that need to be manually specified. In order to handle the invariants of dynamic inference, Krka et al. improved it and established a new tool: Contractor++ [13], which can handle inferred invariants and filter out meaningless invariants. They also proposed four methods for dynamically inferencing FSA models for different types of inputs [35]. On the basis of these four methods, the tools SEKT and TEMI are implemented. DSM [8,9] employs a recurrent neural network to extract the feature values and clusters the feature values to generate FSAs. Gao et al. developed dynamic specification mining based on a transformer (DSM-T) [31,36] to improve the accuracy of FSAs generated by DSM. MdRubel Ahmed et al. [37] propose a disruptive method that utilizes the attention mechanism to produce accurate flow specifications from system-on-chip (SoC) IP communication traces.

6. Conclusions

In this work, we propose an approach based on the OPTICS clustering algorithm and model checking to mine specifications from software systems. Unlike previous approaches, we employ the OPTICS clustering algorithm, which does not require the setting of parameters, but rather reads information from a reachability plot and sets e p s . In addition, we refine FSAs based on model checking to mitigate the problem of incorrect merging among states in FSAs generated by clustering feature values. Further, we evaluated the effectiveness of our approach by experimenting on 13 target library classes. Compared with the related tools, our approach can generate better-quality FSAs. In the future, we plan to study more technologies to further improve the accuracy of mined specifications and conduct research on how to mine specifications expressed by more practical formal models.

Author Contributions

Conceptualization, M.W. and Y.F.; methodology, M.W. and Y.F.; validation, M.W.; investigation, Y.F.; writing—original draft preparation, Y.F.; writing—review and editing, M.W.; visualization Y.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no specific grant from any funding agency in the public, commercial or not-for-profit sectors.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in this article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The list of abbreviations and symbols is shown below.
SymbolsDefinition
A I F ( a , b ) An occurrence of event a must be immediately followed by event b
A I P ( a , b ) An occurrence of event a must be immediately preceded by event b
N I F ( a , b ) An occurrence of event a is never immediately followed by event b
T r a n s The error segments in the FSA according to the type of the violated property
QThe finite set of states in FSA
Σ The table of input characters in FSA
δ The transition function mapping Q × Σ to Q
q 0 The initial state in FSA
FThe set of acceptable states in FSA
AcronymsFull Form
FSAFinite-state automaton
OPTICSOrdering points to identify the clustering structure
DBSCANDensity-based spatial clustering of applications with noise
CTLComputing tree logic
LTLLinear-time temporal logic
DSMDeep specification mining
LSTMLong short-term memory
DSM-TDynamic specification mining based on transformer
CPSCyber-physical systems
TEMITrace-enhanced MTS inference
GMMGaussian mixture model
SOCSystem-on-chip

References

  1. Zhong, H.; Su, Z. Detecting API documentation errors. In Proceedings of the 2013 ACM SIGPLAN International Conference on Object Oriented Programming Systems Languages & Applications, Indianapolis, IN, USA, 29–31 October 2013; pp. 803–816. [Google Scholar] [CrossRef]
  2. Beschastnikh, I.; Brun, Y.; Schneider, S.; Sloan, M.; Ernst, M.D. Leveraging existing instrumentation to automatically infer invariant-constrained models. In Proceedings of the 19th ACM SIGSOFT Symposium and the 13th European Conference on Foundations of Software Engineering, Szeged, Hungary, 5–9 September 2011; pp. 267–277. [Google Scholar] [CrossRef]
  3. Lo, D.; Khoo, S.C. SMArTIC: Towards building an accurate, robust and scalable specification miner. In Proceedings of the 14th ACM SIGSOFT International Symposium on Foundations of Software Engineering, Portland, OR, USA, 5–11 November 2006; pp. 265–275. [Google Scholar] [CrossRef]
  4. Lo, D.; Mariani, L.; Pezzè, M. Automatic steering of behavioral model inference. In Proceedings of the 7th Joint Meeting of the European Software Engineering Conference and the ACM SIGSOFT symposium on the Foundations of Software Engineering, Amsterdam, The Netherlands, 24–28 August 2009; pp. 345–354. [Google Scholar] [CrossRef]
  5. Peleg, H.; Shoham, S.; Yahav, E.; Yang, H. Symbolic automata for static specification mining. In Proceedings of the Static Analysis: 20th International Symposium, SAS 2013, Seattle, WA, USA, 20–22 June 2013; Springer: Berlin/Heidelberg, Germany, 2013; pp. 63–83. [Google Scholar]
  6. Lemieux, C.; Park, D.; Beschastnikh, I. General LTL specification mining (T). In Proceedings of the 2015 30th IEEE/ACM International Conference on Automated Software Engineering (ASE), Lincoln, NE, USA, 9–13 November 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 81–92. [Google Scholar] [CrossRef]
  7. Biermann, A.W.; Feldman, J.A. On the synthesis of finite-state machines from samples of their behavior. IEEE Trans. Comput. 1972, 100, 592–597. [Google Scholar] [CrossRef]
  8. Le, T.D.B.; Lo, D. Deep specification mining. In Proceedings of the 27th ACM SIGSOFT International Symposium on Software Testing and Analysis, Amsterdam, The Netherlands, 16–21 July 2018; pp. 106–117. [Google Scholar] [CrossRef]
  9. Le, T.D.B.; Bao, L.; Lo, D. DSM: A specification mining tool using recurrent neural network based language model. In Proceedings of the 2018 26th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, Lake Buena Vista, FL, USA, 4–9 November 2018; pp. 896–899. [Google Scholar] [CrossRef]
  10. Mikolov, T.; Karafiát, M.; Burget, L.; Cernockỳ, J.; Khudanpur, S. Recurrent neural network based language model. In Proceedings of the 11th Annual Conference of the International Speech Communication Association, Makuhari, Japan, 26–30 September 2010; Volume 2, pp. 1045–1048. [Google Scholar]
  11. MacQueen, J. Classification and analysis of multivariate observations. In Proceedings of the 5th Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, CA, USA, 21 June–18 July 1965; University of California Press: Los Angeles, CA, USA, 1967; pp. 281–297. [Google Scholar]
  12. Ankerst, M.; Breunig, M.M.; Kriegel, H.P.; Sander, J. OPTICS: Ordering points to identify the clustering structure. ACM Sigmod Rec. 1999, 28, 49–60. [Google Scholar] [CrossRef]
  13. Krka, I.; Brun, Y.; Medvidovic, N. Automatically Mining Specifications from Invocation Traces and Method Invariants; Technical Report; Citeseer; Center for Systems and Software Engineering, University of Southern California: Los Angeles, CA, USA, 2013. [Google Scholar]
  14. Le, T.D.B.; Le, X.B.D.; Lo, D.; Beschastnikh, I. Synergizing specification miners through model fissions and fusions (t). In Proceedings of the 2015 30th IEEE/ACM International Conference on Automated Software Engineering (ASE), Lincoln, NE, USA, 9–13 November 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 115–125. [Google Scholar] [CrossRef]
  15. Pnueli, A. The temporal logic of programs. In Proceedings of the 18th Annual Symposium on Foundations of Computer Science (sfcs 1977), Providence, RI, USA, 31 October–2 November 1977; IEEE: Piscataway, NJ, USA, 1977; pp. 46–57. [Google Scholar] [CrossRef]
  16. Emerson, E.A.; Clarke, E.M. Characterizing correctness properties of parallel programs using fixpoints. In Proceedings of the Automata, Languages and Programming: Seventh Colloquium, Noordwijkerhout, The Netherlands, 14–18 July 1980; Springer: Berlin/Heidelberg, Germany, 1980; pp. 169–181. [Google Scholar]
  17. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the KDD, Portland, OR, USA, 2–4 August 1996; Volume 96, pp. 226–231. [Google Scholar]
  18. Mahmood, H.; Mehmood, T.; Al-Essa, L.A. Optimizing Clustering Algorithms for Anti-Microbial Evaluation Data: A Majority Score-based Evaluation of K-Means, Gaussian Mixture Model, and Multivariate T-Distribution Mixtures. IEEE Access 2023, 11, 79793–79800. [Google Scholar] [CrossRef]
  19. Lukauskas, M.; Ruzgas, T. A New Clustering Method Based on the Inversion Formula. Mathematics 2022, 10, 2559. [Google Scholar] [CrossRef]
  20. Wang, Z. A new clustering method based on morphological operations. Expert Syst. Appl. 2020, 145, 113102. [Google Scholar] [CrossRef]
  21. Robinson, B.; Ernst, M.D.; Perkins, J.H.; Augustine, V.; Li, N. Scaling up automated test generation: Automatically generating maintainable regression unit tests for programs. In Proceedings of the 2011 26th IEEE/ACM International Conference on Automated Software Engineering (ASE 2011), Lawrence, KS, USA, 6–10 November 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 23–32. [Google Scholar] [CrossRef]
  22. Beschastnikh, I.; Brun, Y.; Abrahamson, J.; Ernst, M.D.; Krishnamurthy, A. Using declarative specification to improve the understanding, extensibility, and comparison of model-inference algorithms. IEEE Trans. Softw. Eng. 2014, 41, 408–428. [Google Scholar] [CrossRef]
  23. Lorenzoli, D.; Mariani, L.; Pezzè, M. Automatic generation of software behavioral models. In Proceedings of the 30th International Conference on Software Engineering, Leipzig, Germany, 10–18 May 2008; pp. 501–510. [Google Scholar] [CrossRef]
  24. Wu, W.; Zhang, Z. Combinatorial Optimization and Applications: 14th International Conference, COCOA 2020, Dallas, TX, USA, 11–13 December 2020, Proceedings; Springer Nature: Cham, Switzerland, 2020; Volume 12577. [Google Scholar]
  25. Bingham, J.; Hu, A.J. Empirically efficient verification for a class of infinite-state systems. In Proceedings of the Tools and Algorithms for the Construction and Analysis of Systems: 11th International Conference, TACAS 2005, Held as Part of the Joint European Conferences on Theory and Practice of Software, ETAPS 2005, Edinburgh, UK, 4–8 April 2005; Proceedings 11. Springer: Berlin/Heidelberg, Germany, 2005; pp. 77–92. [Google Scholar]
  26. Yang, J.; Evans, D.; Bhardwaj, D.; Bhat, T.; Das, M. Perracotta: Mining temporal API rules from imperfect traces. In Proceedings of the 28th International Conference on Software Engineering, Shanghai, China, 20–28 May 2006; pp. 282–291. [Google Scholar] [CrossRef]
  27. Gabel, M.; Su, Z. Javert: Fully automatic mining of general temporal properties from dynamic traces. In Proceedings of the 16th ACM SIGSOFT International Symposium on Foundations of Software Engineering, Atlanta, GA, USA, 9–14 November 2008; pp. 339–349. [Google Scholar] [CrossRef]
  28. Gabel, M.; Su, Z. Symbolic mining of temporal specifications. In Proceedings of the 30th International Conference on Software Engineering, Leipzig, Germany, 10–18 May 2008; pp. 51–60. [Google Scholar] [CrossRef]
  29. Gabel, M.; Su, Z. Online inference and enforcement of temporal properties. In Proceedings of the 32nd ACM/IEEE International Conference on Software Engineering-Volume 1, Cape Town, South Africa, 1–8 May 2010; pp. 15–24. [Google Scholar] [CrossRef]
  30. De Caso, G.; Braberman, V.; Garbervetsky, D.; Uchitel, S. Automated abstractions for contract validation. IEEE Trans. Softw. Eng. 2010, 38, 141–162. [Google Scholar] [CrossRef]
  31. Gao, Y.; Wang, M.; Yu, B. Dynamic Specification Mining Based on Transformer. In Proceedings of the Theoretical Aspects of Software Engineering: 16th International Symposium, TASE 2022, Cluj-Napoca, Romania, 8–10 July 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 220–237. [Google Scholar]
  32. Asarin, E.; Caspi, P.; Maler, O. Timed regular expressions. J. ACM 2002, 49, 172–206. [Google Scholar]
  33. Cecconi, A.; De Giacomo, G.; Di Ciccio, C.; Maggi, F.M.; Mendling, J. Measuring the interestingness of temporal logic behavioral specifications in process mining. Inf. Syst. 2022, 107, 101920. [Google Scholar] [CrossRef]
  34. Bartocci, E.; Mateis, C.; Nesterini, E.; Nickovic, D. Survey on mining signal temporal logic specifications. Inf. Comput. 2022, 289, 104957. [Google Scholar] [CrossRef]
  35. Krka, I.; Brun, Y.; Medvidovic, N. Automatic mining of specifications from invocation traces and method invariants. In Proceedings of the 22nd ACM SIGSOFT International Symposium on Foundations of Software Engineering, Hong Kong, China, 16–22 November 2014; pp. 178–189. [Google Scholar]
  36. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems; Curran Associates Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  37. Rubel Ahmed, M.; Zheng, H. Deep Bidirectional Transformers for SoC Flow Specification Mining. arXiv 2022, arXiv:2203.13182. [Google Scholar]
Figure 1. Reachability plot.
Figure 1. Reachability plot.
Algorithms 17 00028 g001
Figure 2. Overview of DSM approach.
Figure 2. Overview of DSM approach.
Algorithms 17 00028 g002
Figure 3. Overview architecture.
Figure 3. Overview architecture.
Algorithms 17 00028 g003
Figure 4. An FSA obtained by clustering.
Figure 4. An FSA obtained by clustering.
Algorithms 17 00028 g004
Figure 5. FSA refinement method based on model checking.
Figure 5. FSA refinement method based on model checking.
Algorithms 17 00028 g005
Figure 6. Example of detection and refinement.
Figure 6. Example of detection and refinement.
Algorithms 17 00028 g006
Figure 7. Reachability plot generated by clustering algorithm.
Figure 7. Reachability plot generated by clustering algorithm.
Algorithms 17 00028 g007
Table 1. Target library classes.
Table 1. Target library classes.
Target Library Class1MGenerated Test CaseRecorded Method Calls
ArrayList1842,86522,996
LinKedList713,7314847
HashSet823,181257,428
HashMap1153,39667,942
Hashtable879,40389,811
Signature579,096205,386
Socket2180,035130,876
ZipOutputStream5162,97143,626
SringTokenizer5148,649336,924
StackAr7549,648132,826
NFST5158,99895,149
ToHTMLStream17103,562278,631
SMTPProtocol1557,281136,271
Table 2. Experimental results.
Table 2. Experimental results.
Tools1-Tail2-TailSEKT 1SEKT 2CON++TEMIDSMMCLSM
Class
ArrayList13.9613.1336.0313.8613.0716.8722.2781.94
LinKedList27.1525.7286.0226.6724.5207.5132.7698.49
HashSet20.8821.2752.2220.8821.2723.3479.2394.67
HashMap25.4108.7168.94---83.5397.02
Hashtable42.3933.5892.78---76.8291.42
Signature61.5464.2566.8862.0563.9839.06100.0086.44
Socket35.8931.5255.1534.7328.37-51.78100.00
ZipOutputStream46.3647.4262.8047.91--87.6291.80
SringTokenizer52.8852.9721.3052.15--100.0094.19
StackAr16.5416.5434.9116.5416.54-76.2796.55
NFST24.5725.5230.4024.5625.7811.8080.16100.00
ToHTMLStream------69.7089.76
SMTPProtocol------72.6982.63
Average33.4230.9755.2233.2627.6519.7271.7592.19
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fan, Y.; Wang, M. Specification Mining Based on the Ordering Points to Identify the Clustering Structure Clustering Algorithm and Model Checking. Algorithms 2024, 17, 28. https://doi.org/10.3390/a17010028

AMA Style

Fan Y, Wang M. Specification Mining Based on the Ordering Points to Identify the Clustering Structure Clustering Algorithm and Model Checking. Algorithms. 2024; 17(1):28. https://doi.org/10.3390/a17010028

Chicago/Turabian Style

Fan, Yiming, and Meng Wang. 2024. "Specification Mining Based on the Ordering Points to Identify the Clustering Structure Clustering Algorithm and Model Checking" Algorithms 17, no. 1: 28. https://doi.org/10.3390/a17010028

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop