Abstract
Graph edit distance (GED) is a powerful tool to model the dissimilarity between graphs. However, evaluating the exact GED is NP-hard. To tackle this problem, estimation methods of GED were introduced, e.g., bipartite and IPFP, during which heuristics were employed. The stochastic nature of these methods induces the stability issue. In this paper, we propose the first formal study of stability of GED heuristics, starting with defining a measure of these (in)stabilities, namely the relative error. Then, the effects of two critical factors on stability are examined, namely, the number of solutions and the ratio between edit costs. The ratios are computed on five datasets of various properties. General suggestions are provided to properly choose these factors, which can reduce the relative error by more than an order of magnitude. Finally, we verify the relevance of stability to predict performance of GED heuristics, by taking advantage of an edit cost learning algorithm to optimize the performance and the k-nearest neighbor regression for prediction. Experiments show that the optimized costs correspond to much higher ratios and an order of magnitude lower relative errors than the expert cost.
1. Introduction
Graphs provide a flexible representation framework to encode elements as well as the relationships between them, enabling to capture the underlying structural information of the data. Despite providing rich expressiveness, the complexity lying in graph structures becomes its Achilles’ heel when applying machine learning methods for graph data, which are mainly designed to operate on vector representations [1,2]. To leverage this flaw, several approaches have been designed to learn models on graphs, representatives among which include graph embedding strategy [3], graph kernels [4,5], and more recently graph neural networks [6], some of which are closely connected with signal processing on graphs [7,8,9,10,11,12]. Despite their state-of-the-art performances, they seldom operate directly in a graph space, hence reducing the interpretability of the underlying operations. To overcome these issues and preserve the properties of a graph space, some (dis)similarity measure or metric is usually assigned to that space, since most machine learning algorithms rely on (dis)similarity measures between data. One of the most used dissimilarity measures between graphs is the graph edit distance (GED) [13,14]. The GED of two graphs and is the minimal amount of distortion required to transform into . This transformation includes a series of six elementary edit operations, namely vertex and edge substitutions ( and ), removals ( and ), and insertions ( and ), as shown in Figure 1. This sequence of edit operations constitutes an edit path . A non-negative cost can be assigned to each edit operation e. The sum of all edit operation costs included within is defined as the cost associated with . The minimal cost among all edit paths defines the GED between and .
Figure 1.
An illustration of graph edit operations. , , , , , and denote, respectively, the insertions, removals, and substitutions of vertices and edges. Different letters along with colors illustrate different vertex labels.
Evaluating exact GED is NP-hard even for uniform edit costs [15]. In practice, it cannot be done for graphs having more than 12 vertices in general [16]. To avoid this computational burden, strategies to approximate GED in a limited computational time have been proposed [17,18] with acceptable classification or regression performances. Of particular interest are the two famous methods, bipartite [19] and IPFP [20], where upper and lower bounds are estimated as an approximation of GED. The computation of the bounds relies highly on the design of the algorithm, as well as the randomness during the procedure, which leads to a reduction of stability.
As this paper will illustrate, the stability of the GED heuristics is highly relevant to the choice of heuristic method and their prediction performance. As GED is a widely used similarity between graphs, the study of stability can be useful to help promote the performance of GED in various tasks.
In this paper, it is the first time that the stability of GED heuristics has been formally studied. We define the instability of a GED heuristic in terms of the variability of the GED approximations over repeated trials. Methods that can potentially alleviate this problem are proposed. For instance, by carrying out several local searches in parallel, the multi-start counterparts of bipartite and IPFP, named mbipartite and mIPFP, respectively, may acquire better approximation with higher stability [21]. Description and analyses of these approximations and the root of randomness are presented in Section 3 and Section 4.
Another essential ingredient of GED is the underlying edit cost function , which quantifies the distortion carried by any edit operation e. The values of the edit costs for each edit operation have a major impact on the computation of GED and its performance, including its stability [17,22]. Besides fixing costs provided a priori by domain experts for a given dataset, methods are proposed to optimize these costs, e.g., by aligning metrics in the graph to target spaces [23]. Analyzing the optimized edit costs helps further explore its relevance to the stability of GED heuristics.
The remaining part of the paper is organized as follows: Section 2 introduces widely used GED heuristics paradigms. Section 3 defines a measure of the (in)stability of these heuristics, as well as two factors of critical influence. Then, Section 4 gives experiments and analyses. Finally, Section 5 concludes the work and open perspectives.
2. Graph Edit Distances Heuristics
Over the years, many heuristics have been proposed to approximate GED. The authors of [18] categorize these heuristics according to their underlying paradigms. As milestones and baselines to many other heuristics, both bipartite and IPFP achieve high performance [17]. Thereby, in the following sections, we focus on these two heuristics and the related paradigms. First, we provide preliminary definitions of graphs and graph edit distances.
2.1. Graphs and Graph Edit Distances
A graph is an ordered pair of disjoint sets, where V is the vertex set and is the edge set. A graph can have a label set L from a label space and a labeling function ℓ that assigns a label to each vertex and/or edge.
Given a set of N graphs, the Graph Edit Distance (GED) between two graphs and is defined as the cost of minimal transformation [19]:
where is a mapping between and encoding the transformation from to , and represents a dummy element [20]. As described in Section 1, this transformation consists of a series of six elementary operations: removing or inserting a vertex or an edge, and substituting a label of a vertex or an edge by another. measures the cost associated with :
where are the edit costs associated with the six edit operations: respectively, vertex removal, insertion, substitution and edge removal, insertion, and substitution. Without loss of generality, these costs are set to be constant for each edit operation in the following part, denoted, respectively, as .
2.2. Paradigm LSAPE-GED and Heuristic Bipartite
GED can be approximated by solving a linear sum assignment problem with edition or error correction (LSAPE). For any two sets and , consider a transformation from to , with elementary operations on each element : substitution (), insertion (), and removal (), where and represents a dummy element. An assignment with edition, also known as the -assignment [24], is a bijection between set and set relaxed on , namely , where for any , for any , and . We denote the set of all possible -assignments from to as .
Each elementary operation in an -assignment can be associated with a non-negative cost c. Consequently, a cost C is associated with , namely,
where each term on the right side successively represents substitutions, insertions, and removals. The costs of all operations induced by can be represented by a matrix . LSAPE aims at minimizing this cost over all , namely finding . We denote the set of all optimal solutions as . Variants of the Hungarian algorithm have been used to acquire an optimal solution [25,26].
The GED between two graphs and can be approximated by solving an LSAPE between vertex sets and . Each row and column in the cost matrix respectively correspond to a vertex in and ; each entry represents a optimal feasible transformation from to as in (1), and the optimal cost corresponds to the approximation of GED. This paradigm is named LSAPE-GED.
bipartite is a representative heuristic under the LSAPE-GED paradigm. It constructs the cost matrix by adding the cost of vertices and the cost of the edges adjacent to them. After that, an optimal LSAPE solution for and the corresponding cost are computed.
2.3. Paradigm LS-GED and Heuristic IPFP
The local search (LS-GED) paradigm is composed of two steps. First, the transformation and the cost are initialized randomly or by a heuristic, such as bipartite. Then, starting at these initial results, a refinement procedure is carried out by a local search method to search for improved transformation with a lower cost. With different strategies applied in the second step, various heuristics have been designed. IPFP is a well-known representative one.
GED can be modeled as a quadratic problem. The LSAPE-GED paradigm simplifies this problem by only considering the linear part of GED, namely the costs of vertex transformations, where costs of edge transformations can only be implied as patches, as done by bipartite. In contrast, the IPFP heuristic under the LS-GED paradigm provides a method to extend LSAPE-GED, by including the edge transformations as a quadratic part of GED.
We define a binary matrix equivalent to an -assignment . As a result, the cost of the transformation can be formalized as
where is vectorized by the binary vector by concatenating its rows, is the edit cost vector, and the coefficient g is set to 0.5 if the graphs are undirected and 1 otherwise [1].
The IPFP heuristic approximates the GED by adapting the integer projected fixed point (IPFP) algorithm [27] designed for the quadratic assignment problem (QAP) [20,28]. The algorithm is first initialized randomly or by a heuristic such as bipartite, and then updated by iterations. In each iteration, a linear approximation is computed by a LSAPE solver. Then, the local minimum of the cost and the corresponding binary solution is estimated by a line search [28].
3. Stability of GED Heuristics
The nature of the GED heuristics leads to a drop in computational stability, namely different trials may lead to different results. In the following, we analyze this instability by measuring the variability of the GED approximations over repeated trials. For instance, in the LSAPE-GED paradigm, the cost matrix may vary given vertex set with different orders, which affects the solution of the LSAPE problem, furthermore causing the instability. Likewise, in the LS-GED paradigm, the instability can be traced back to the initial procedure where a random transformation or a GED heuristic such as bipartite implying stochasticity may be assigned.
3.1. Measure of (In)Stability
To measure the (in)stability of GED heuristics, we define a criterion named relative error. Given a set of graphs , we compute the GED with a heuristic between each pair of graphs times (trials). The relative error is then defined as
where is the approximation of the GED in the k-th trial using a GED heuristic such as Algorithm 1, and is the exact GED between and . As evaluating is normally impractical, we replace it with the minimum approximation over all trials, namely
The relative error measures the average ratio between the offsets and the exact GEDs over trials and pairs of graphs. A smaller value indicates higher stability.
3.2. Influential Factors of Stability
Low stability can degrade the performance of GED heuristics, which implies a broader range of the confidence interval in a prediction task such as regression and classification, or instability of the produced graphs in a pre-image task [29]. A straightforward method to mitigate this problem is repeating the GED computation. The minimum cost over repetitions is then chosen as the GED approximation. Strategies have been proposed to refine this method. Well-known ones are the mbipartite and mIPFP, which are the multi-start counterparts of bipartite and IPFP [21]. These two heuristics start several initial candidates simultaneously to acquire tighter upper bounds. The stability is concurrently ameliorated. As examined in Section 4, the relative error can be reduced by up to four orders of magnitude. Algorithm 1 presents the procedure of mIPFP as an example.
| Algorithm 1. Approximation of GED using mIPFP |
Input: Graphs and . Vertex edit cost , edge edit cost . The number of solutions m. Output: An approximation of GED between and .
|
Another factor that significantly influences the stability of GED heuristics turns out to be the relative values of vertex and edge edit costs. When vertex costs are markedly larger than edge costs, the GED stability often shows an observable improvement. This phenomenon is detailed in Section 4. When optimized edit cost values are applied, the relative error can be reduced by up to around 30 times compared to using the worst cost values.
Based on the aforementioned information, we analyze the stability with respect to two factors. The first one is the number of random initial candidates of the GED heuristics, namely “# of solutions”. For mbipartite and mIPFP, it is equal to the parameter m, as in Algorithm 1. The second factor is the ratio between vertex and edge edit costs. Let , , , , , be the cost functions associated with, respectively, vertex substitutions, insertions, removals and edge substitutions, insertions, and removals. Then, the ratio is defined as
where computes the average value of its inputs.
4. Experiments
In this section, we conduct experimental analyses on the GED stability. First, the influence of the two factors introduced in Section 3.2 is examined, and then the relevance of stability and prediction performance of GED heuristics is verified, taking advantage of a state-of-the-art edit costs learning strategy.
Five well-known public datasets are applied in the experiments: Alkane and Acyclic are composed of acyclic molecules modeled respectively as unlabeled and vertex-labeled graphs, while MAO, Monoterpens, and MUTAG consist of cycles-included molecules represented as graphs labeled on both vertices and edges. The sizes of datasets vary from 68 to 286 (https://brunl01.users.greyc.fr/CHEMISTRY/, accessed on 9 September 2022).
We exploit two multi-start heuristics, mbipartite and mIPFP, to evaluate the stability as described in Section 2. The former induces randomness by permuting vertices and consequently the cost matrix for each graph. The two heuristics employ respectively the implementation in the graphkit-learn [30] and GEDLIB [31] libraries.
4.1. Effects of the Two Factors
Figure 2 shows the effect of the two factors, namely “# of solutions” and the ratio between vertex and edge edit costs , on the relative error defined in (5), considering the mbipartite and mIPFP heuristics on five datasets. The Figure 2a,c on the left side exhibit how drops with the increase of the “# of solutions” m. In most cases, drops rapidly when the solution number increases from 1 to around a certain number , and reaches at a relatively small value; the tendency mitigates afterwards. is around 20 for mbipartite and 10 for mIPFP. Take datasets Alkane and Acyclic for examples. When using mbipartite, ’s on these two datasets drop respectively from 0.23 to 0.08, and from 0.12 to 0.03, as m increases from 1 to 100; when using mIPFP, ’s drop respectively from 0.08 to , and from 0.03 to . This result indicates that an adequately large number of solutions is necessary, thus a trade-off decision between stability and time complexity needs to be made for different applications.
Figure 2.
The relative errors of mbipartite and mIPFP on five datasets with respect to the numbers of solutions m and ratios between vertex and edge edit costs . Colors along with the markers indicate different datasets.
The Figure 2b,d on the right side reveal the relation between and the ratio . The edge costs are set to 1 and the vertex costs to be the ratio value (for insertions, removals, and substitutions). The removal costs of vertices (resp. edges) are set to 0 if vertices (resp. edges) are not labeled. For both mbipartite and mIPFP, is relatively large when the ratio is smaller than 1, namely when edge costs are bigger than vertex costs, and drops with the increase of the ratio. We can observe that a larger ratio leads to higher stability. Take datasets Alkane and Acyclic for examples. When using mbipartite, ’s on these two datasets drop respectively from 0.5 to 0.16, and from 0.34 to 0.07, as increases from 0.1 to 10; when using mIPFP, ’s drop respectively from 0.03 to , and from 0.02 to . A possible cause of this phenomenon is that large edge costs amplify the arbitrariness of the edge edit operations. For graphs with n vertices, there are possible edges that can be inserted, removed, and substituted, which causes more uncertainty when constructing edit paths and computing their costs. Taking IPFP for instance, large edge costs lead to a big cost matrix in (4), implying the possibility of more variance on the value of the term . Many edit costs given by domain experts are in accordance with this empirical rule, such as the ones in [17]. In the next section, we further validate the relevance of stability and prediction performance benefitting from an edit cost learning method.
4.2. Stability vs. Prediction Performance
As stated in the Introduction, the choice of edit costs has a major impact on the computation of graph edit distance, and thus on the performance associated with a prediction task. To challenge these predefined expert costs with how they can improve the prediction performance, methods to tune the edit costs and thus the GED were proposed in the literature [23,32,33]. These methods provide an opportunity to observe the connection between the prediction performance of GEDs and the choices of edit costs, which further relate to the GED stability, as examined in Section 4.1.
To inspect this relevance, we utilize a state-of-the-art cost learning algorithm [23], where the edit costs are optimized according to a particular prediction task. An alternate iterative procedure is proposed to preserve the distances in both the graph space (GEDs) and target space (Euclidean or Manhattan distances between targets), where the update on edit costs obtained by solving a constrained linear problem and a re-computation of the optimal edit paths according to the newly computed costs are performed alternately. The GEDs with optimized edit costs are then used to train a k-nearest-neighbors (KNN) regression [34] model. KNN predicts the property value or class of an object based on the value or class of its neighbors. It has been widely used for various types of data, such as traffic [35], infrared vision [36], sensors [37], and graphs [38]. Thus, it is suitable for the current experiments. Experiments on Alkane and Acyclic show that optimized costs gain a significant improvement in accuracy compared to random or expert costs [1]. Table 1 summarizes the optimized values of edit costs.
Table 1.
Average and standard deviation of fitted edit costs’ values.
We then examine the stability of the two heuristics when using these edit costs. Figure 3 demonstrates the relations between the relative errors and the ratios between vertex and edge costs (See Section 3.2 and Figure 2 for more details). Therein, the colors represent datasets (i.e., blue for Alkane and orange for Acyclic), and shapes represent different edit costs, with 🞮, ⬤, and ⯁ respectively for the expert costs, the optimized costs using the Euclidean and Manhattan distances. When applying mIPFP, for the expert costs, ’s for the two datasets are both 1, and ’s are 0.02 for Alkane and 0.013 for Acyclic; when using the Euclidean distance, , for Alkane, and , for Acyclic; when using the Manhattan distance, , for Alkane, and , for Acyclic. It can be observed that the optimized costs correspond to much higher ratios and an order of magnitude lower relative errors than the expert costs. Similar conclusions can be observed for mbipartite as well. Thus, an empirical conclusion can be derived: the obtained optimized edit costs correspond to higher stability of GEDs, while obtaining a higher performance.
Figure 3.
The relative errors of mbipartite and mIPFP on datasets Alkane and Acyclic with respect to the ratios between vertex and edge edit costs using different edit costs optimization methods. The colors represent datasets and the shapes of markers represent different edit costs.
5. Conclusions and Future Work
In this paper, we conducted analyses of the GED heuristics’ stability, which is the first time it is formally investigated in the literature. After defining an (in)stability measure, namely, the relative error, we show the strong connection of the stability with the number of random initial candidates of multi-start GED heuristics and the relation between vertex and edge edit costs. Experiments on five datasets and two GED heuristics indicate that the proper choice of these factors can reduce the relative error by more than an order of magnitude. A further investigation indicates higher stability of GED computation corresponds to the optimized edit costs and thus better prediction performance, where an edit cost learning algorithm is applied to optimize the performance and the k-nearest neighbor regression for prediction.
There are still several challenges to address in future work. First, examining other influential factors and conducting theoretical analyses can help deepen understanding of the stability of the heuristics. Second, it will be helpful to perform more thorough experiments, including other state-of-the-art GED heuristics on datasets from a wider range of fields and statistical properties. Third, higher stability comes with the cost of higher time complexity. Methods that can better balance the stability, time complexity, and prediction performance in practice need to be developed.
Author Contributions
Conceptualization, L.J. (Linlin Jia), B.G. and P.H.; methodology, L.J. (Linlin Jia) and B.G.; software, L.J. (Linlin Jia); validation, L.J. (Linlin Jia), B.G. and P.H.; formal analysis, L.J. (Linlin Jia); investigation, L.J. (Linlin Jia); resources, V.T., L.J. (Laurent Joubert), B.G. and P.H.; data curation, L.J. (Linlin Jia); writing—original draft preparation, L.J. (Linlin Jia); writing—review and editing, B.G. and P.H.; visualization, L.J. (Linlin Jia); supervision, V.T., L.J. (Laurent Joubert), B.G. and P.H.; project administration, V.T., L.J. (Laurent Joubert) and P.H.; funding acquisition, V.T., L.J. (Laurent Joubert) and P.H. All authors have read and agreed to the published version of the manuscript.
Funding
This work has been partially supported by the University of Rouen Normandy, INSA Rouen Normandy, the “Centre National de la Recherche Scientifique” (CNRS), the European Regional Development Fund (ERDF), Labex SynOrg (ANR-11-LABX-0029), Carnot Institut I2C, the graduate school for research XL-Chem (ANR-18-EURE-0020 XL CHEM), the “Région Normandie”, China Scholarship Council (CSC), and the French national research agency (ANR) under the grant APi (ANR-18-CE23-0014).
Data Availability Statement
Publicly available datasets were analyzed in this study. These data can be found here: https://brunl01.users.greyc.fr/CHEMISTRY/, accessed on 9 September 2022.
Acknowledgments
The authors would like also to gratefully acknowledge the “Centre Régional Informatique et d’Applications Numériques de Normandie” (CRIANN) for computing resources.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Jia, L. Bridging Graph and Kernel Spaces: A Pre-Image Perspective. Ph.D. Thesis, INSA Rouen Normandie, Saint-Étienne-du-Rouvray, France, 2021. [Google Scholar]
- Borgwardt, K.; Ghisu, E.; Llinares-López, F.; O’Bray, L.; Rieck, B. Graph Kernels: State-of-the-Art and Future Challenges. arXiv 2020, arXiv:2011.03854. [Google Scholar]
- Goyal, P.; Ferrara, E. Graph embedding techniques, applications, and performance: A survey. Knowl.-Based Syst. 2018, 151, 78–94. [Google Scholar] [CrossRef]
- Kriege, N.M.; Johansson, F.D.; Morris, C. A survey on graph kernels. Appl. Netw. Sci. 2020, 5, 1–42. [Google Scholar] [CrossRef]
- Gaüzère, B.; Brun, L.; Villemin, D. Graph kernels in chemoinformatics. In Quantitative Graph Theory Mathematical Foundations and Applications; Dehmer, M., Emmert-Streib, F., Eds.; CRC Press: Boca Raton, FL, USA, 2015; pp. 425–470. [Google Scholar]
- Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef] [PubMed]
- Balcilar, M.; Renton, G.; Héroux, P.; Gaüzère, B.; Adam, S.; Honeine, P. Analyzing the Expressive Power of Graph Neural Networks in a Spectral Perspective. In Proceedings of the International Conference on Learning Representations, Virtual Event, Austria, 3–7 May 2021. [Google Scholar]
- Ortega, A.; Frossard, P.; Kovačević, J.; Moura, J.M.; Vandergheynst, P. Graph signal processing: Overview, challenges, and applications. Proc. IEEE 2018, 106, 808–828. [Google Scholar] [CrossRef]
- Dong, X.; Thanou, D.; Toni, L.; Bronstein, M.; Frossard, P. Graph signal processing for machine learning: A review and new perspectives. IEEE Signal Process. Mag. 2020, 37, 117–127. [Google Scholar] [CrossRef]
- Zhang, C.; Florêncio, D.; Chou, P.A. Graph Signal Processing—A Probabilistic Framework; Tech. Rep. MSR-TR-2015-31; Microsoft Research Lab: Redmond, WA, USA, 2015. [Google Scholar]
- Richiardi, J.; Achard, S.; Bunke, H.; Van De Ville, D. Machine learning with brain graphs: Predictive modeling approaches for functional imaging in systems neuroscience. IEEE Signal Process. Mag. 2013, 30, 58–70. [Google Scholar] [CrossRef]
- Richiardi, J.; Van De Ville, D.; Riesen, K.; Bunke, H. Vector space embedding of undirected graphs with fixed-cardinality vertex sequences for classification. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 902–905. [Google Scholar]
- Bunke, H.; Allermann, G. Inexact graph matching for structural pattern recognition. Pattern Recognit. Lett. 1983, 1, 245–253. [Google Scholar] [CrossRef]
- Sanfeliu, A.; Fu, K.S. A distance measure between attributed relational graphs for pattern recognition. IEEE Trans. Syst. Man, Cybern. 1983, SMC-13, 353–362. [Google Scholar]
- Zeng, Z.; Tung, A.K.; Wang, J.; Feng, J.; Zhou, L. Comparing stars: On approximating graph edit distance. Proc. Vldb Endow. 2009, 2, 25–36. [Google Scholar] [CrossRef]
- Neuhaus, M.; Riesen, K.; Bunke, H. Fast suboptimal algorithms for the computation of graph edit distance. In Joint IAPR International Workshops on Statistical Techniques in Pattern Recognition (SPR) and Structural and Syntactic Pattern Recognition (SSPR); Springer: Berlin, Germany, 2006; pp. 163–172. [Google Scholar]
- Abu-Aisheh, Z.; Gaüzère, B.; Bougleux, S.; Ramel, J.Y.; Brun, L.; Raveaux, R.; Héroux, P.; Adam, S. Graph edit distance contest: Results and future challenges. Pattern Recognit. Lett. 2017, 100, 96–103. [Google Scholar] [CrossRef]
- Blumenthal, D.B.; Boria, N.; Gamper, J.; Bougleux, S.; Brun, L. Comparing heuristics for graph edit distance computation. VLDB J. 2020, 29, 419–458. [Google Scholar] [CrossRef]
- Riesen, K. Structural pattern recognition with graph edit distance. In Advances in Computer Vision and Pattern Recognition; Springer: Berlin, Germany, 2015. [Google Scholar]
- Bougleux, S.; Gaüzère, B.; Brun, L. Graph edit distance as a quadratic program. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 1701–1706. [Google Scholar]
- Daller, É.; Bougleux, S.; Gaüzère, B.; Brun, L. Approximate graph edit distance by several local searches in parallel. In Proceedings of the 7th International Conference on Pattern Recognition Applications and Methods, Funchal, Portugal, 16–18 January 2018. [Google Scholar]
- Bunke, H. Error correcting graph matching: On the influence of the underlying cost function. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 917–922. [Google Scholar] [CrossRef]
- Jia, L.; Gaüzère, B.; Yger, F.; Honeine, P. A Metric Learning Approach to Graph Edit Costs for Regression. In Joint IAPR International Workshops on Statistical Techniques in Pattern Recognition (SPR) and Structural and Syntactic Pattern Recognition (SSPR); Springer: Berlin, Germany, 2021; pp. 238–247. [Google Scholar]
- Bougleux, S.; Brun, L. Linear sum assignment with edition. arXiv 2016, arXiv:1603.04380. [Google Scholar]
- Kuhn, H.W. The Hungarian method for the assignment problem. Nav. Res. Logist. Q. 1955, 2, 83–97. [Google Scholar] [CrossRef]
- Munkres, J. Algorithms for the assignment and transportation problems. J. Soc. Ind. Appl. Math. 1957, 5, 32–38. [Google Scholar] [CrossRef]
- Leordeanu, M.; Hebert, M.; Sukthankar, R. An integer projected fixed point method for graph matching and map inference. Adv. Neural Inf. Process. Syst. 2009, 22, 1114–1122. [Google Scholar]
- Bougleux, S.; Brun, L.; Carletti, V.; Foggia, P.; Gaüzere, B.; Vento, M. A quadratic assignment formulation of the graph edit distance. arXiv 2015, arXiv:1512.07494. [Google Scholar]
- Jia, L.; Gaüzère, B.; Honeine, P. A graph pre-image method based on graph edit distances. In Joint IAPR International Workshops on Statistical Techniques in Pattern Recognition (SPR) and Structural and Syntactic Pattern Recognition (SSPR); Springer: Berlin, Germany, 2021; pp. 216–226. [Google Scholar]
- Jia, L.; Gaüzère, B.; Honeine, P. graphkit-learn: A Python library for graph kernels based on linear patterns. Pattern Recognit. Lett. 2021, 143, 113–121. [Google Scholar] [CrossRef]
- Blumenthal, D.B.; Bougleux, S.; Gamper, J.; Brun, L. GEDLIB: A C++ Library for graph edit distance computation. In Proceedings of the International Workshop on Graph-Based Representations in Pattern Recognition, Tours, France, 19–21 June 2019; Springer: Berlin, Germany, 2019; pp. 14–24. [Google Scholar]
- Neuhaus, M.; Bunke, H. Automatic learning of cost functions for graph edit distance. Inf. Sci. 2007, 177, 239–247. [Google Scholar] [CrossRef]
- Cortés, X.; Conte, D.; Cardot, H. Learning edit cost estimation models for graph edit distance. Pattern Recognit. Lett. 2019, 125, 256–263. [Google Scholar] [CrossRef]
- Altman, N.S. An introduction to kernel and nearest-neighbor nonparametric regression. Am. Stat. 1992, 46, 175–185. [Google Scholar]
- Cai, L.; Yu, Y.; Zhang, S.; Song, Y.; Xiong, Z.; Zhou, T. A sample-rebalanced outlier-rejected k-nearest neighbor regression model for short-term traffic flow forecasting. IEEE Access 2020, 8, 22686–22696. [Google Scholar] [CrossRef]
- Siriborvornratanakul, T. Color and Active Infrared Vision: Estimate Infrared Vision of Printed Color Using Bayesian Classifier and K-Nearest Neighbor Regression. In Pacific Rim Conference on Multimedia; Springer: Berlin, Germany, 2015; pp. 518–527. [Google Scholar]
- Naimi, A.; Deng, J.; Shimjith, S.; Arul, A.J. Fault Detection and Isolation of a Pressurized Water Reactor Based on Neural Network and K-Nearest Neighbor. IEEE Access 2022, 10, 17113–17121. [Google Scholar] [CrossRef]
- Kang, S. K-nearest neighbor learning with graph neural networks. Mathematics 2021, 9, 830. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).