Next Article in Journal
Performance Analysis of Raspberry Pi 3 IP PBX Based on Asterisk
Next Article in Special Issue
Graph Neural Networks for Intelligent Modelling in Network Management and Orchestration: A Survey on Communications
Previous Article in Journal
Selection of Relevant Geometric Features Using Filter-Based Algorithms for Point Cloud Semantic Segmentation
Previous Article in Special Issue
E-Commerce Sales Revenues Forecasting by Means of Dynamically Designing, Developing and Validating a Directed Acyclic Graph (DAG) Network for Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

A Study on the Stability of Graph Edit Distance Heuristics

1
The COBRA Lab, INSA Rouen Normandie, 76800 Rouen, France
2
The COBRA Lab, Université de Rouen Normandie, 76000 Rouen, France
3
The LITIS Lab, INSA Rouen Normandie, 76800 Rouen, France
4
The LITIS Lab, Université de Rouen Normandie, 76000 Rouen, France
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(20), 3312; https://doi.org/10.3390/electronics11203312
Submission received: 12 September 2022 / Revised: 1 October 2022 / Accepted: 8 October 2022 / Published: 14 October 2022
(This article belongs to the Collection Graph Machine Learning)

Abstract

:
Graph edit distance (GED) is a powerful tool to model the dissimilarity between graphs. However, evaluating the exact GED is NP-hard. To tackle this problem, estimation methods of GED were introduced, e.g., bipartite and IPFP, during which heuristics were employed. The stochastic nature of these methods induces the stability issue. In this paper, we propose the first formal study of stability of GED heuristics, starting with defining a measure of these (in)stabilities, namely the relative error. Then, the effects of two critical factors on stability are examined, namely, the number of solutions and the ratio between edit costs. The ratios are computed on five datasets of various properties. General suggestions are provided to properly choose these factors, which can reduce the relative error by more than an order of magnitude. Finally, we verify the relevance of stability to predict performance of GED heuristics, by taking advantage of an edit cost learning algorithm to optimize the performance and the k-nearest neighbor regression for prediction. Experiments show that the optimized costs correspond to much higher ratios and an order of magnitude lower relative errors than the expert cost.

1. Introduction

Graphs provide a flexible representation framework to encode elements as well as the relationships between them, enabling to capture the underlying structural information of the data. Despite providing rich expressiveness, the complexity lying in graph structures becomes its Achilles’ heel when applying machine learning methods for graph data, which are mainly designed to operate on vector representations [1,2]. To leverage this flaw, several approaches have been designed to learn models on graphs, representatives among which include graph embedding strategy [3], graph kernels [4,5], and more recently graph neural networks [6], some of which are closely connected with signal processing on graphs [7,8,9,10,11,12]. Despite their state-of-the-art performances, they seldom operate directly in a graph space, hence reducing the interpretability of the underlying operations. To overcome these issues and preserve the properties of a graph space, some (dis)similarity measure or metric is usually assigned to that space, since most machine learning algorithms rely on (dis)similarity measures between data. One of the most used dissimilarity measures between graphs is the graph edit distance (GED) [13,14]. The GED of two graphs G 1 and G 2 is the minimal amount of distortion required to transform G 1 into G 2 . This transformation includes a series of six elementary edit operations, namely vertex and edge substitutions ( v s and e s ), removals ( v r and e r ), and insertions ( v i and e i ), as shown in  Figure 1. This sequence of edit operations constitutes an edit path π . A non-negative cost c ( e ) can be assigned to each edit operation e. The sum of all edit operation costs included within π is defined as the cost associated with π . The minimal cost among all edit paths defines the GED between G 1 and G 2 .
Evaluating exact GED is NP-hard even for uniform edit costs [15]. In practice, it cannot be done for graphs having more than 12 vertices in general [16]. To avoid this computational burden, strategies to approximate GED in a limited computational time have been proposed [17,18] with acceptable classification or regression performances. Of particular interest are the two famous methods, bipartite [19] and IPFP [20], where upper and lower bounds are estimated as an approximation of GED. The computation of the bounds relies highly on the design of the algorithm, as well as the randomness during the procedure, which leads to a reduction of stability.
As this paper will illustrate, the stability of the GED heuristics is highly relevant to the choice of heuristic method and their prediction performance. As GED is a widely used similarity between graphs, the study of stability can be useful to help promote the performance of GED in various tasks.
In this paper, it is the first time that the stability of GED heuristics has been formally studied. We define the instability of a GED heuristic in terms of the variability of the GED approximations over repeated trials. Methods that can potentially alleviate this problem are proposed. For instance, by carrying out several local searches in parallel, the multi-start counterparts of bipartite and IPFP, named mbipartite and mIPFP, respectively, may acquire better approximation with higher stability [21]. Description and analyses of these approximations and the root of randomness are presented in Section 3 and Section 4.
Another essential ingredient of GED is the underlying edit cost function c ( e ) , which quantifies the distortion carried by any edit operation e. The values of the edit costs for each edit operation have a major impact on the computation of GED and its performance, including its stability [17,22]. Besides fixing costs provided a priori by domain experts for a given dataset, methods are proposed to optimize these costs, e.g., by aligning metrics in the graph to target spaces [23]. Analyzing the optimized edit costs helps further explore its relevance to the stability of GED heuristics.
The remaining part of the paper is organized as follows: Section 2 introduces widely used GED heuristics paradigms. Section 3 defines a measure of the (in)stability of these heuristics, as well as two factors of critical influence. Then, Section 4 gives experiments and analyses. Finally, Section 5 concludes the work and open perspectives.

2. Graph Edit Distances Heuristics

Over the years, many heuristics have been proposed to approximate GED. The authors of [18] categorize these heuristics according to their underlying paradigms. As milestones and baselines to many other heuristics, both bipartite and IPFP achieve high performance [17]. Thereby, in the following sections, we focus on these two heuristics and the related paradigms. First, we provide preliminary definitions of graphs and graph edit distances.

2.1. Graphs and Graph Edit Distances

A graph G = ( V , E ) is an ordered pair of disjoint sets, where V is the vertex set and E V × V is the edge set. A graph can have a label set L from a label space and a labeling function that assigns a label l L to each vertex and/or edge.
Given a set G of N graphs, the Graph Edit Distance (GED) between two graphs G i and G j G is defined as the cost of minimal transformation [19]:
ged ( G i , G j ) = min π Π ( G i , G j ) C ( π , G i , G j ) ,
where π ( G i , G j ) is a mapping between V i ε and V j ε encoding the transformation from G i to G j , and ε represents a dummy element [20]. As described in Section 1, this transformation consists of a series of six elementary operations: removing or inserting a vertex or an edge, and substituting a label of a vertex or an edge by another. C ( π , G i , G j ) measures the cost associated with π :
C ( π , G i , G j ) = v V j π 1 ( v ) = ε c v f i + v V i π ( v ) = ε c v f r + v V i π ( v ) ε c v f s + e = ( v i , v j ) E j | π 1 ( v i ) = ε π 1 ( v j ) = ε ( π 1 ( v i ) , π 1 ( v j ) ) E i c e f i + e = ( v i , v j ) E i | π ( v i ) = ε π ( v j ) = ε ( π ( v i ) , π ( v j ) ) E j c e f r + e = ( v i , v j ) E i | π ( v i ) ε π ( v j ) ε ( π ( v i ) , π ( v j ) ) E j c e f s ,
where c v f r , c v f i , c v f s , c e f r , c e f i , c e f s are the edit costs associated with the six edit operations: respectively, vertex removal, insertion, substitution and edge removal, insertion, and substitution. Without loss of generality, these costs are set to be constant for each edit operation in the following part, denoted, respectively, as c v r , c v i , c v s , c e r , c e i , c e s .

2.2. Paradigm LSAPE-GED and Heuristic Bipartite

GED can be approximated by solving a linear sum assignment problem with edition or error correction (LSAPE). For any two sets V 1 and V 2 , consider a transformation from V 1 to V 2 , with elementary operations on each element i V 1 : substitution ( i j ), insertion ( ε j ), and removal ( i ε ), where j V 2 and ε represents a dummy element. An assignment with edition, also known as the ε -assignment [24], is a bijection between set V 1 ε = V 1 { ε } and set V 2 ε = V 2 { ε } relaxed on ε , namely π : V 1 ε V 2 ε , where | π ( i ) | = 1 for any i V 1 , | π 1 ( j ) | = 1 for any j V 2 , and π ( ε ) = ε . We denote the set of all possible ε -assignments from V 1 ε to V 2 ε as Π ( V 1 , V 2 ) .
Each elementary operation in an ε -assignment π can be associated with a non-negative cost c. Consequently, a cost C is associated with π , namely,
C ( π ) = i V 1 π ( i ) = j c ( i , j ) + j V 2 π 1 ( j ) = ε c ( ε , j ) + i V 1 π ( i ) = ε c ( i , ε ) ,
where each term on the right side successively represents substitutions, insertions, and removals. The costs of all operations induced by π can be represented by a matrix C R ( | V 1 | + 1 ) × ( | V 2 | + 1 ) . LSAPE aims at minimizing this cost over all π Π , namely finding C 🟉 ( π 🟉 ) = min π Π ( V 1 , V 2 ) C ( π ) . We denote the set of all optimal solutions as Π 🟉 ( V 1 , V 2 ) . Variants of the Hungarian algorithm have been used to acquire an optimal solution [25,26].
The GED between two graphs G 1 = ( V 1 , E 1 ) and G 2 = ( V 2 , E 2 ) can be approximated by solving an LSAPE between vertex sets V 1 and V 2 . Each row and column in the cost matrix C respectively correspond to a vertex in V 1 and V 2 ; each entry π 🟉 ( G 1 , G 2 ) = π 🟉 ( V 1 , V 2 ) Π 🟉 ( G 1 , G 2 ) represents a optimal feasible transformation from G 1 to G 2 as in (1), and the optimal cost C 🟉 corresponds to the approximation of GED. This paradigm is named LSAPE-GED.
bipartite is a representative heuristic under the LSAPE-GED paradigm. It constructs the cost matrix C by adding the cost of vertices and the cost of the edges adjacent to them. After that, an optimal LSAPE solution for C and the corresponding cost are computed.

2.3. Paradigm LS-GED and Heuristic IPFP

The local search (LS-GED) paradigm is composed of two steps. First, the transformation π and the cost C ( π ) are initialized randomly or by a heuristic, such as bipartite. Then, starting at these initial results, a refinement procedure is carried out by a local search method to search for improved transformation with a lower cost. With different strategies applied in the second step, various heuristics have been designed. IPFP is a well-known representative one.
GED can be modeled as a quadratic problem. The LSAPE-GED paradigm simplifies this problem by only considering the linear part of GED, namely the costs of vertex transformations, where costs of edge transformations can only be implied as patches, as done by bipartite. In contrast, the IPFP heuristic under the LS-GED paradigm provides a method to extend LSAPE-GED, by including the edge transformations as a quadratic part of GED.
We define a binary matrix X { 0 , 1 } ( | V 1 | + 1 ) × ( | V 2 | + 1 ) equivalent to an ε -assignment π . As a result, the cost of the transformation π can be formalized as
C ( x ) = g x Q x + c x ,
where X is vectorized by the binary vector x = vec ( X ) { 0 , 1 } ( | V 1 | + 1 ) ( | V 2 | + 1 ) by concatenating its rows, c = vec ( C ) is the edit cost vector, and the coefficient g is set to 0.5 if the graphs are undirected and 1 otherwise [1].
The IPFP heuristic approximates the GED by adapting the integer projected fixed point (IPFP) algorithm [27] designed for the quadratic assignment problem (QAP) [20,28]. The algorithm is first initialized randomly or by a heuristic such as bipartite, and then updated by iterations. In each iteration, a linear approximation is computed by a LSAPE solver. Then, the local minimum of the cost and the corresponding binary solution is estimated by a line search [28].

3. Stability of GED Heuristics

The nature of the GED heuristics leads to a drop in computational stability, namely different trials may lead to different results. In the following, we analyze this instability by measuring the variability of the GED approximations over repeated trials. For instance, in the LSAPE-GED paradigm, the cost matrix C may vary given vertex set with different orders, which affects the solution of the LSAPE problem, furthermore causing the instability. Likewise, in the LS-GED paradigm, the instability can be traced back to the initial procedure where a random transformation or a GED heuristic such as bipartite implying stochasticity may be assigned.

3.1. Measure of (In)Stability

To measure the (in)stability of GED heuristics, we define a criterion named relative error. Given a set of graphs G 1 , G 2 , G N , we compute the GED with a heuristic between each pair of graphs N t times (trials). The relative error E r is then defined as
E r = 1 N 2 k = 1 N t i , j = 1 N ( ged ( k ) ( G i , G j ) ged 0 ( G i , G j ) 1 2 i , j = 1 N ged ( k ) ( G i , G j ) + ged 0 ( G i , G j ) ,
where ged ( k ) ( G i , G j ) is the approximation of the GED in the k-th trial using a GED heuristic such as Algorithm 1, and ged 0 ( G i , G j ) is the exact GED between G i and G j . As evaluating ged 0 is normally impractical, we replace it with the minimum approximation over all trials, namely
ged 0 ( G i , G j ) = min 1 k N t ged ( k ) ( G i , G j ) .
The relative error E r measures the average ratio between the offsets and the exact GEDs over trials and pairs of graphs. A smaller value indicates higher stability.

3.2. Influential Factors of Stability

Low stability can degrade the performance of GED heuristics, which implies a broader range of the confidence interval in a prediction task such as regression and classification, or instability of the produced graphs in a pre-image task [29]. A straightforward method to mitigate this problem is repeating the GED computation. The minimum cost over repetitions is then chosen as the GED approximation. Strategies have been proposed to refine this method. Well-known ones are the mbipartite and mIPFP, which are the multi-start counterparts of bipartite and IPFP [21]. These two heuristics start several initial candidates simultaneously to acquire tighter upper bounds. The stability is concurrently ameliorated. As examined in Section 4, the relative error can be reduced by up to four orders of magnitude. Algorithm 1 presents the procedure of mIPFP as an example.
Algorithm 1. Approximation of GED using mIPFP
     Input: Graphs G 1 = ( V 1 , E 1 ) and G 2 = ( V 2 , E 2 ) .
     Vertex edit cost c v , edge edit cost c e . The number of solutions m.
     Output: An approximation C 🟉 of GED between G 1 and G 2 .
1:
Initialize cost C 🟉 = C 0 = .
2:
Let j = 0 .
3:
while  j < m   do
4:
    Approximate a new cost C j + 1 with the IPFP heuristic.
5:
    if  C 🟉 > C j + 1  then
6:
         C 🟉 = C j + 1 .
7:
    end if
8:
     j = j + 1 .
9:
end while
10:
ged ( G i , G j ) = C 🟉 .
Another factor that significantly influences the stability of GED heuristics turns out to be the relative values of vertex and edge edit costs. When vertex costs are markedly larger than edge costs, the GED stability often shows an observable improvement. This phenomenon is detailed in Section 4. When optimized edit cost values are applied, the relative error can be reduced by up to around 30 times compared to using the worst cost values.
Based on the aforementioned information, we analyze the stability with respect to two factors. The first one is the number of random initial candidates of the GED heuristics, namely “# of solutions”. For mbipartite and mIPFP, it is equal to the parameter m, as in Algorithm 1. The second factor is the ratio between vertex and edge edit costs. Let c v f s , c v f i , c v f r , c e f s , c e f i , c e f r R + be the cost functions associated with, respectively, vertex substitutions, insertions, removals and edge substitutions, insertions, and removals. Then, the ratio is defined as
R e c = average ( c v f i , c v f r , c v f s ) average ( c e f i , c e f r , c e f s ) ,
where average ( · ) computes the average value of its inputs.

4. Experiments

In this section, we conduct experimental analyses on the GED stability. First, the influence of the two factors introduced in Section 3.2 is examined, and then the relevance of stability and prediction performance of GED heuristics is verified, taking advantage of a state-of-the-art edit costs learning strategy.
Five well-known public datasets are applied in the experiments: Alkane and Acyclic are composed of acyclic molecules modeled respectively as unlabeled and vertex-labeled graphs, while MAO, Monoterpens, and MUTAG consist of cycles-included molecules represented as graphs labeled on both vertices and edges. The sizes of datasets vary from 68 to 286 (https://brunl01.users.greyc.fr/CHEMISTRY/, accessed on 9 September 2022).
We exploit two multi-start heuristics, mbipartite and mIPFP, to evaluate the stability as described in Section 2. The former induces randomness by permuting vertices and consequently the cost matrix for each graph. The two heuristics employ respectively the implementation in the graphkit-learn [30] and GEDLIB [31] libraries.

4.1. Effects of the Two Factors

 Figure 2 shows the effect of the two factors, namely “# of solutions” and the ratio between vertex and edge edit costs R e c , on the relative error E r defined in (5), considering the mbipartite and mIPFP heuristics on five datasets. The  Figure 2a,c on the left side exhibit how E r drops with the increase of the “# of solutions” m. In most cases, E r drops rapidly when the solution number increases from 1 to around a certain number N , and reaches at a relatively small value; the tendency mitigates afterwards. N is around 20 for mbipartite and 10 for mIPFP. Take datasets Alkane and Acyclic for examples. When using mbipartite, E r ’s on these two datasets drop respectively from 0.23 to 0.08, and from 0.12 to 0.03, as m increases from 1 to 100; when using mIPFP, E r ’s drop respectively from 0.08 to 5 × 10 6 , and from 0.03 to 7.2 × 10 7 . This result indicates that an adequately large number of solutions is necessary, thus a trade-off decision between stability and time complexity needs to be made for different applications.
The  Figure 2b,d on the right side reveal the relation between E r and the ratio R e c . The edge costs are set to 1 and the vertex costs to be the ratio value (for insertions, removals, and substitutions). The removal costs of vertices (resp. edges) are set to 0 if vertices (resp. edges) are not labeled. For both mbipartite and mIPFP, E r is relatively large when the ratio is smaller than 1, namely when edge costs are bigger than vertex costs, and drops with the increase of the ratio. We can observe that a larger ratio leads to higher stability. Take datasets Alkane and Acyclic for examples. When using mbipartite, E r ’s on these two datasets drop respectively from 0.5 to 0.16, and from 0.34 to 0.07, as R e c increases from 0.1 to 10; when using mIPFP, E r ’s drop respectively from 0.03 to 5.6 × 10 3 , and from 0.02 to 6.9 × 10 4 . A possible cause of this phenomenon is that large edge costs amplify the arbitrariness of the edge edit operations. For graphs with n vertices, there are n 2 possible edges that can be inserted, removed, and substituted, which causes more uncertainty when constructing edit paths and computing their costs. Taking IPFP for instance, large edge costs lead to a big cost matrix Q in (4), implying the possibility of more variance on the value of the term g x Q x . Many edit costs given by domain experts are in accordance with this empirical rule, such as the ones in [17]. In the next section, we further validate the relevance of stability and prediction performance benefitting from an edit cost learning method.

4.2. Stability vs. Prediction Performance

As stated in the Introduction, the choice of edit costs has a major impact on the computation of graph edit distance, and thus on the performance associated with a prediction task. To challenge these predefined expert costs with how they can improve the prediction performance, methods to tune the edit costs and thus the GED were proposed in the literature [23,32,33]. These methods provide an opportunity to observe the connection between the prediction performance of GEDs and the choices of edit costs, which further relate to the GED stability, as examined in Section 4.1.
To inspect this relevance, we utilize a state-of-the-art cost learning algorithm [23], where the edit costs are optimized according to a particular prediction task. An alternate iterative procedure is proposed to preserve the distances in both the graph space (GEDs) and target space (Euclidean or Manhattan distances between targets), where the update on edit costs obtained by solving a constrained linear problem and a re-computation of the optimal edit paths according to the newly computed costs are performed alternately. The GEDs with optimized edit costs are then used to train a k-nearest-neighbors (KNN) regression [34] model. KNN predicts the property value or class of an object based on the value or class of its neighbors. It has been widely used for various types of data, such as traffic [35], infrared vision [36], sensors [37], and graphs [38]. Thus, it is suitable for the current experiments. Experiments on Alkane and Acyclic show that optimized costs gain a significant improvement in accuracy compared to random or expert costs [1]. Table 1 summarizes the optimized values of edit costs.
We then examine the stability of the two heuristics when using these edit costs.  Figure 3 demonstrates the relations between the relative errors E r and the ratios R e c between vertex and edge costs (See Section 3.2 and  Figure 2 for more details). Therein, the colors represent datasets (i.e., blue for Alkane and orange for Acyclic), and shapes represent different edit costs, with 🞮, ⬤, and ⯁ respectively for the expert costs, the optimized costs using the Euclidean and Manhattan distances. When applying mIPFP, for the expert costs, R e c ’s for the two datasets are both 1, and E r ’s are 0.02 for Alkane and 0.013 for Acyclic; when using the Euclidean distance, R e c = 34.21 , E r = 0.002 for Alkane, and R e c = 2.6 , E r = 0.002 for Acyclic; when using the Manhattan distance, R e c = 3.11 , E r = 2.33 × 10 3 for Alkane, and R e c = 2.55 , E r = 2.68 × 10 3 for Acyclic. It can be observed that the optimized costs correspond to much higher ratios and an order of magnitude lower relative errors than the expert costs. Similar conclusions can be observed for mbipartite as well. Thus, an empirical conclusion can be derived: the obtained optimized edit costs correspond to higher stability of GEDs, while obtaining a higher performance.

5. Conclusions and Future Work

In this paper, we conducted analyses of the GED heuristics’ stability, which is the first time it is formally investigated in the literature. After defining an (in)stability measure, namely, the relative error, we show the strong connection of the stability with the number of random initial candidates of multi-start GED heuristics and the relation between vertex and edge edit costs. Experiments on five datasets and two GED heuristics indicate that the proper choice of these factors can reduce the relative error by more than an order of magnitude. A further investigation indicates higher stability of GED computation corresponds to the optimized edit costs and thus better prediction performance, where an edit cost learning algorithm is applied to optimize the performance and the k-nearest neighbor regression for prediction.
There are still several challenges to address in future work. First, examining other influential factors and conducting theoretical analyses can help deepen understanding of the stability of the heuristics. Second, it will be helpful to perform more thorough experiments, including other state-of-the-art GED heuristics on datasets from a wider range of fields and statistical properties. Third, higher stability comes with the cost of higher time complexity. Methods that can better balance the stability, time complexity, and prediction performance in practice need to be developed.

Author Contributions

Conceptualization, L.J. (Linlin Jia), B.G. and P.H.; methodology, L.J. (Linlin Jia) and B.G.; software, L.J. (Linlin Jia); validation, L.J. (Linlin Jia), B.G. and P.H.; formal analysis, L.J. (Linlin Jia); investigation, L.J. (Linlin Jia); resources, V.T., L.J. (Laurent Joubert), B.G. and P.H.; data curation, L.J. (Linlin Jia); writing—original draft preparation, L.J. (Linlin Jia); writing—review and editing, B.G. and P.H.; visualization, L.J. (Linlin Jia); supervision, V.T., L.J. (Laurent Joubert), B.G. and P.H.; project administration, V.T., L.J. (Laurent Joubert) and P.H.; funding acquisition, V.T., L.J. (Laurent Joubert) and P.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been partially supported by the University of Rouen Normandy, INSA Rouen Normandy, the “Centre National de la Recherche Scientifique” (CNRS), the European Regional Development Fund (ERDF), Labex SynOrg (ANR-11-LABX-0029), Carnot Institut I2C, the graduate school for research XL-Chem (ANR-18-EURE-0020 XL CHEM), the “Région Normandie”, China Scholarship Council (CSC), and the French national research agency (ANR) under the grant APi (ANR-18-CE23-0014).

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: https://brunl01.users.greyc.fr/CHEMISTRY/, accessed on 9 September 2022.

Acknowledgments

The authors would like also to gratefully acknowledge the “Centre Régional Informatique et d’Applications Numériques de Normandie” (CRIANN) for computing resources.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jia, L. Bridging Graph and Kernel Spaces: A Pre-Image Perspective. Ph.D. Thesis, INSA Rouen Normandie, Saint-Étienne-du-Rouvray, France, 2021. [Google Scholar]
  2. Borgwardt, K.; Ghisu, E.; Llinares-López, F.; O’Bray, L.; Rieck, B. Graph Kernels: State-of-the-Art and Future Challenges. arXiv 2020, arXiv:2011.03854. [Google Scholar]
  3. Goyal, P.; Ferrara, E. Graph embedding techniques, applications, and performance: A survey. Knowl.-Based Syst. 2018, 151, 78–94. [Google Scholar] [CrossRef] [Green Version]
  4. Kriege, N.M.; Johansson, F.D.; Morris, C. A survey on graph kernels. Appl. Netw. Sci. 2020, 5, 1–42. [Google Scholar] [CrossRef] [Green Version]
  5. Gaüzère, B.; Brun, L.; Villemin, D. Graph kernels in chemoinformatics. In Quantitative Graph Theory Mathematical Foundations and Applications; Dehmer, M., Emmert-Streib, F., Eds.; CRC Press: Boca Raton, FL, USA, 2015; pp. 425–470. [Google Scholar]
  6. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Balcilar, M.; Renton, G.; Héroux, P.; Gaüzère, B.; Adam, S.; Honeine, P. Analyzing the Expressive Power of Graph Neural Networks in a Spectral Perspective. In Proceedings of the International Conference on Learning Representations, Virtual Event, Austria, 3–7 May 2021. [Google Scholar]
  8. Ortega, A.; Frossard, P.; Kovačević, J.; Moura, J.M.; Vandergheynst, P. Graph signal processing: Overview, challenges, and applications. Proc. IEEE 2018, 106, 808–828. [Google Scholar] [CrossRef] [Green Version]
  9. Dong, X.; Thanou, D.; Toni, L.; Bronstein, M.; Frossard, P. Graph signal processing for machine learning: A review and new perspectives. IEEE Signal Process. Mag. 2020, 37, 117–127. [Google Scholar] [CrossRef]
  10. Zhang, C.; Florêncio, D.; Chou, P.A. Graph Signal Processing—A Probabilistic Framework; Tech. Rep. MSR-TR-2015-31; Microsoft Research Lab: Redmond, WA, USA, 2015. [Google Scholar]
  11. Richiardi, J.; Achard, S.; Bunke, H.; Van De Ville, D. Machine learning with brain graphs: Predictive modeling approaches for functional imaging in systems neuroscience. IEEE Signal Process. Mag. 2013, 30, 58–70. [Google Scholar] [CrossRef] [Green Version]
  12. Richiardi, J.; Van De Ville, D.; Riesen, K.; Bunke, H. Vector space embedding of undirected graphs with fixed-cardinality vertex sequences for classification. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 902–905. [Google Scholar]
  13. Bunke, H.; Allermann, G. Inexact graph matching for structural pattern recognition. Pattern Recognit. Lett. 1983, 1, 245–253. [Google Scholar] [CrossRef]
  14. Sanfeliu, A.; Fu, K.S. A distance measure between attributed relational graphs for pattern recognition. IEEE Trans. Syst. Man, Cybern. 1983, SMC-13, 353–362. [Google Scholar]
  15. Zeng, Z.; Tung, A.K.; Wang, J.; Feng, J.; Zhou, L. Comparing stars: On approximating graph edit distance. Proc. Vldb Endow. 2009, 2, 25–36. [Google Scholar] [CrossRef] [Green Version]
  16. Neuhaus, M.; Riesen, K.; Bunke, H. Fast suboptimal algorithms for the computation of graph edit distance. In Joint IAPR International Workshops on Statistical Techniques in Pattern Recognition (SPR) and Structural and Syntactic Pattern Recognition (SSPR); Springer: Berlin, Germany, 2006; pp. 163–172. [Google Scholar]
  17. Abu-Aisheh, Z.; Gaüzère, B.; Bougleux, S.; Ramel, J.Y.; Brun, L.; Raveaux, R.; Héroux, P.; Adam, S. Graph edit distance contest: Results and future challenges. Pattern Recognit. Lett. 2017, 100, 96–103. [Google Scholar] [CrossRef]
  18. Blumenthal, D.B.; Boria, N.; Gamper, J.; Bougleux, S.; Brun, L. Comparing heuristics for graph edit distance computation. VLDB J. 2020, 29, 419–458. [Google Scholar] [CrossRef]
  19. Riesen, K. Structural pattern recognition with graph edit distance. In Advances in Computer Vision and Pattern Recognition; Springer: Berlin, Germany, 2015. [Google Scholar]
  20. Bougleux, S.; Gaüzère, B.; Brun, L. Graph edit distance as a quadratic program. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 1701–1706. [Google Scholar]
  21. Daller, É.; Bougleux, S.; Gaüzère, B.; Brun, L. Approximate graph edit distance by several local searches in parallel. In Proceedings of the 7th International Conference on Pattern Recognition Applications and Methods, Funchal, Portugal, 16–18 January 2018. [Google Scholar]
  22. Bunke, H. Error correcting graph matching: On the influence of the underlying cost function. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 917–922. [Google Scholar] [CrossRef]
  23. Jia, L.; Gaüzère, B.; Yger, F.; Honeine, P. A Metric Learning Approach to Graph Edit Costs for Regression. In Joint IAPR International Workshops on Statistical Techniques in Pattern Recognition (SPR) and Structural and Syntactic Pattern Recognition (SSPR); Springer: Berlin, Germany, 2021; pp. 238–247. [Google Scholar]
  24. Bougleux, S.; Brun, L. Linear sum assignment with edition. arXiv 2016, arXiv:1603.04380. [Google Scholar]
  25. Kuhn, H.W. The Hungarian method for the assignment problem. Nav. Res. Logist. Q. 1955, 2, 83–97. [Google Scholar] [CrossRef] [Green Version]
  26. Munkres, J. Algorithms for the assignment and transportation problems. J. Soc. Ind. Appl. Math. 1957, 5, 32–38. [Google Scholar] [CrossRef] [Green Version]
  27. Leordeanu, M.; Hebert, M.; Sukthankar, R. An integer projected fixed point method for graph matching and map inference. Adv. Neural Inf. Process. Syst. 2009, 22, 1114–1122. [Google Scholar]
  28. Bougleux, S.; Brun, L.; Carletti, V.; Foggia, P.; Gaüzere, B.; Vento, M. A quadratic assignment formulation of the graph edit distance. arXiv 2015, arXiv:1512.07494. [Google Scholar]
  29. Jia, L.; Gaüzère, B.; Honeine, P. A graph pre-image method based on graph edit distances. In Joint IAPR International Workshops on Statistical Techniques in Pattern Recognition (SPR) and Structural and Syntactic Pattern Recognition (SSPR); Springer: Berlin, Germany, 2021; pp. 216–226. [Google Scholar]
  30. Jia, L.; Gaüzère, B.; Honeine, P. graphkit-learn: A Python library for graph kernels based on linear patterns. Pattern Recognit. Lett. 2021, 143, 113–121. [Google Scholar] [CrossRef]
  31. Blumenthal, D.B.; Bougleux, S.; Gamper, J.; Brun, L. GEDLIB: A C++ Library for graph edit distance computation. In Proceedings of the International Workshop on Graph-Based Representations in Pattern Recognition, Tours, France, 19–21 June 2019; Springer: Berlin, Germany, 2019; pp. 14–24. [Google Scholar]
  32. Neuhaus, M.; Bunke, H. Automatic learning of cost functions for graph edit distance. Inf. Sci. 2007, 177, 239–247. [Google Scholar] [CrossRef]
  33. Cortés, X.; Conte, D.; Cardot, H. Learning edit cost estimation models for graph edit distance. Pattern Recognit. Lett. 2019, 125, 256–263. [Google Scholar] [CrossRef] [Green Version]
  34. Altman, N.S. An introduction to kernel and nearest-neighbor nonparametric regression. Am. Stat. 1992, 46, 175–185. [Google Scholar]
  35. Cai, L.; Yu, Y.; Zhang, S.; Song, Y.; Xiong, Z.; Zhou, T. A sample-rebalanced outlier-rejected k-nearest neighbor regression model for short-term traffic flow forecasting. IEEE Access 2020, 8, 22686–22696. [Google Scholar] [CrossRef]
  36. Siriborvornratanakul, T. Color and Active Infrared Vision: Estimate Infrared Vision of Printed Color Using Bayesian Classifier and K-Nearest Neighbor Regression. In Pacific Rim Conference on Multimedia; Springer: Berlin, Germany, 2015; pp. 518–527. [Google Scholar]
  37. Naimi, A.; Deng, J.; Shimjith, S.; Arul, A.J. Fault Detection and Isolation of a Pressurized Water Reactor Based on Neural Network and K-Nearest Neighbor. IEEE Access 2022, 10, 17113–17121. [Google Scholar] [CrossRef]
  38. Kang, S. K-nearest neighbor learning with graph neural networks. Mathematics 2021, 9, 830. [Google Scholar] [CrossRef]
Figure 1. An illustration of graph edit operations. v i , v r , v s , e i , e r , and e s denote, respectively, the insertions, removals, and substitutions of vertices and edges. Different letters along with colors illustrate different vertex labels.
Figure 1. An illustration of graph edit operations. v i , v r , v s , e i , e r , and e s denote, respectively, the insertions, removals, and substitutions of vertices and edges. Different letters along with colors illustrate different vertex labels.
Electronics 11 03312 g001
Figure 2. The relative errors of mbipartite and mIPFP on five datasets with respect to the numbers of solutions m and ratios between vertex and edge edit costs R e c . Colors along with the markers indicate different datasets.
Figure 2. The relative errors of mbipartite and mIPFP on five datasets with respect to the numbers of solutions m and ratios between vertex and edge edit costs R e c . Colors along with the markers indicate different datasets.
Electronics 11 03312 g002
Figure 3. The relative errors of mbipartite and mIPFP on datasets Alkane and Acyclic with respect to the ratios R e c between vertex and edge edit costs using different edit costs optimization methods. The colors represent datasets and the shapes of markers represent different edit costs.
Figure 3. The relative errors of mbipartite and mIPFP on datasets Alkane and Acyclic with respect to the ratios R e c between vertex and edge edit costs using different edit costs optimization methods. The colors represent datasets and the shapes of markers represent different edit costs.
Electronics 11 03312 g003
Table 1. Average and standard deviation of fitted edit costs’ values.
Table 1. Average and standard deviation of fitted edit costs’ values.
DatasetEdit CostDistance c ni c nr c ns c ei c er c es
AlkanebipartiteEuclidean26.45 ± 0.4826.24 ± 0.60-0.13 ± 0.060.14 ± 0.09-
Manhattan26.67 ± 0.3726.63 ± 0.58-0.11 ± 0.040.11 ± 0.06-
IPFPEuclidean26.12 ± 0.2425.88 ± 0.25-0.74 ± 0.230.78 ± 0.23-
Manhattan25.94 ± 0.3825.71 ± 0.44-0.89 ± 0.300.77 ± 0.29-
AcyclicbipartiteEuclidean13.81 ± 0.4813.83 ± 0.8010.46 ± 0.401.37 ± 0.461.45 ± 0.461.41 ± 0.09
Manhattan13.76 ± 0.3914.14 ± 0.5710.28 ± 0.441.44 ± 0.201.45 ± 0.191.45 ± 0.07
IPFPEuclidean11.61 ± 0.4511.68 ± 0.4311.07 ± 0.534.49 ± 0.304.46 ± 0.244.48 ± 0.18
Manhattan11.52 ± 0.4011.40 ± 0.4010.61 ± 0.524.50 ± 0.314.50 ± 0.314.50 ± 0.10
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jia, L.; Tognetti, V.; Joubert, L.; Gaüzère, B.; Honeine, P. A Study on the Stability of Graph Edit Distance Heuristics. Electronics 2022, 11, 3312. https://doi.org/10.3390/electronics11203312

AMA Style

Jia L, Tognetti V, Joubert L, Gaüzère B, Honeine P. A Study on the Stability of Graph Edit Distance Heuristics. Electronics. 2022; 11(20):3312. https://doi.org/10.3390/electronics11203312

Chicago/Turabian Style

Jia, Linlin, Vincent Tognetti, Laurent Joubert, Benoit Gaüzère, and Paul Honeine. 2022. "A Study on the Stability of Graph Edit Distance Heuristics" Electronics 11, no. 20: 3312. https://doi.org/10.3390/electronics11203312

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop