Next Article in Journal
Micro-Terrain Recognition Method of Transmission Lines Based on Improved UNet++
Previous Article in Journal
Flow-Based Community Search Approach for Functionally Cohesive Building Group Recognition: A Case Study on Commercial Complexes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GeoFAN: Point Pattern Recognition in Spatial Vector Data

1
State Key Laboratory of Precision Space-Time Information Sensing Technology, Department of Precision Instrument, Tsinghua University, Beijing 100084, China
2
National Key Laboratory of Electromagnetic Space Security, Chengdu 610000, China
3
School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
ISPRS Int. J. Geo-Inf. 2025, 14(6), 214; https://doi.org/10.3390/ijgi14060214
Submission received: 5 April 2025 / Revised: 13 May 2025 / Accepted: 27 May 2025 / Published: 29 May 2025

Abstract

The recognition of point patterns in spatial vector data has important applications in geographic mapping and formation recognition. However, the application of traditional methods to spatial vector data faces two difficulties. Firstly, these data are low signal-to-noise ratio data in which the point patterns are mixed with a large number of normal point clusters; thus, it is difficult to recognize point patterns from these unstructured data using traditional clustering or machine learning methods. Secondly, the lack of edge connectivity relationships in spatial vector data directly hinders the application of graph models. Few studies have systematically solved the above difficulties. In this article, we propose a geometric feature attention scheme to overcome the above challenges. We also present an implementation of the scheme based on the graph method, termed GeoFAN, to extract and classify point patterns simultaneously in spatial vector data. Firstly, the raw data are transformed into a graph structure consisting of adjacency and attribute matrices. Secondly, a geometric feature attention module is proposed to enhance the feature representation of point patterns. Finally, the recognition results of all points are output via GeoFAN. The macro precision, recall, and F 1 score of five simulated point pattern types with different attributes and point numbers are 92.8%, 90.3%, and 91.5%, respectively, and GeoFAN is trained with simulated data to recognize real location-based point patterns successfully. The proposed GeoFAN showed superior performance and generalization ability in point pattern recognition.

1. Introduction

Point pattern recognition in spatial vector point data refers to the automatic recognition of interested point clusters with spatial distribution characteristics, including artificial elements such as building clusters, aircraft and ship formations, and natural elements such as island clusters and river systems. In geographic mapping, point pattern recognition enhances the description of regional geographic features by mining the principal components in space, thus reflecting important geographic features [1,2]. In formation recognition, point pattern recognition discovers battle formations among a large number of points, thus enhancing regional situational awareness and helping safeguard regional rights and interests [3,4,5].
Due to the complexity and specificity of spatial vector data, the current point pattern recognition process is usually extracted first and then classified, which is difficult to complete at one time. The complexity is that these data have low signal-to-noise ratio (SNR), uneven spatial distribution [6,7], and point pattern diversity [8], as shown in Figure 1. The point patterns are mixed with normal points and normal point clusters, which can easily lead to misclassification by traditional clustering or machine learning methods. The specificity is that the distribution of points in spatial vector data has complex non-Euclidean features, and it is difficult to model its attribute and spatial correlation directly by traditional methods. In comparison, the graph-based method expresses the coupling between attributes and space explicitly through the adjacency and attribute matrices, which is a direct and efficient way to represent the spatial vector data. However, spatial vector data contain only point locations without edge relationships [9]; it is necessary to construct the adjacency matrix artificially [10]. The geometric features of the data are as important as the point attributes [11], which, together, constitute the fingerprint of the point pattern. Although the distance threshold parameter can be used to determine neighbors, it results in the initial distance weights of all neighboring nodes being equalized (e.g., binarized 0/1 connections) if only this aggregation condition is used, thus losing the geometric features that are implied by the actual distance differences. In addition, the choice of the threshold parameter presents subjective difficulties. If the threshold is chosen too large, it leads to the aggregation of false targets, and if the threshold is chosen too small, it leads to disconnection within the point pattern. Techniques such as inverse distance weighting and radial basis functions can assign different weights to neighboring points. However, existing research lacks models based on inverse distance weighting or radial basis functions in the integration of point pattern extraction and classification applications.
In earlier studies, point pattern extraction was achieved by designing multiple spatial constraints to delineate point clusters progressively. Basaraner et al. delineate building clusters by strong geospatial separation of elements such as roads, rivers, etc. [12]. However, methods that simply rely on other elements to achieve clustering struggle to obtain more detailed results, and elements used for delineation are often lacking in practical situations. Point clustering methods have been introduced to point pattern extraction for their simplicity [13]. These methods combine the points that satisfy the conditions into new clusters through homogeneity constraints, mostly derived from classical point clustering algorithms, including the K-Means algorithm [14], the DBSCAN algorithm [15], etc. In [16], a Gaussian kernel-based DBSCAN method is introduced to extract ship formations under different clutter ratios efficiently. Cetinkaya et al. compared the effectiveness of four algorithms for grouping buildings, of which DBSCAN performs better in grouping buildings in urban blocks with different distributions [17]. Graph theory has also been applied to point clustering. Yan et al. used multidimensional features of buildings as node attributes, input the spectral domain graph convolutional neural network to predict the center location coordinates of clusters, and used the K-Means method to cluster the buildings [18]. Point pattern recognition methods based on clustering algorithms can simply and quickly extract the point pattern with aggregated distribution. However, they need to set the artificial parameter, which increases application difficulty, and is prone to false extraction when the SNR is low and there are normal point clusters. Other extraction methods, such as neighborhood graph-based methods (e.g., Delaunay Triangulation (DT), Nearest Neighbor Graph (NNG), Minimum Spanning Tree (MST), etc.), are overly dependent on expert experience and difficult to generalize [19,20,21,22].
To classify the extracted point patterns, template matching-based methods were developed [23,24]. They classify point patterns by comparing them with standard templates. In [25], a Hough transform-based method for classifying ship formations is introduced. It detects formation shapes such as straight lines formed by ships through the Hough transform and then compares them with templates to realize ship formation recognition. Cheng et al. processed the binary coded mapping using the convolutional Radon transform and compared formations with three standard templates, classifying “Y”, “T”, and “I” ship formations with different offsets [26]. The logic of the template-based method is simple. However, template matching methods will fail if the point pattern is partially obscured or slightly deviates from the template.
Machine learning methods make point pattern classification intelligent. He et al. proposed a depthwise separable convolutional neural network (DSCNN) to classify ship formations with different point detection errors [16]. Lin et al. proposed the Long Short-Term Memory (LSTM)-based algorithm for aircraft formation classification [27]. Liang et al. designed a formation coding method and combined it with a support vector machine (SVM) to construct a formation recognition model with 0.955 recognition accuracy [28]. A CNN-based aircraft formation recognition method is later proposed, with an improved recognition accuracy of 0.965 [29]. He et al. classified the clustered building patterns using the random forest algorithm and further partitioned the building clusters with no recognized patterns based on the graph partitioning method [30]. Machine learning has received much attention since it can automatically mine data for features. However, spatial vector data have unstructured properties; thus, traditional machine learning methods struggle to capture complex structural information from them.
Traditional methods that simply concatenate point attributes and spatial coordinates limit the ability to deeply integrate heterogeneous information, such as spatial proximity and semantic features. In comparison, graph neural networks (GNNs) can jointly process spatial positions and attribute features by constructing spatial proximity networks. Lin et al. introduced a recognition algorithm based on a graph neural network and the LSTM network to recognize ship formations [31]. Yan et al. classified regular and irregular building patterns using a graph convolutional neural network [32]. Scholars extend the applications of GNN by transforming some scenarios into point pattern recognition problems. In [33], a model that combines the Shape Context (SC) descriptor and a graph convolutional neural network (GCNN) is proposed to classify the interchange patterns. Yu et al. modeled river systems as graph structures and proposed a river system pattern recognition algorithm based on a spectral domain graph convolutional neural network to classify river systems [34]. GNN has significant advantages when working with non-Euclidean data. It requires an adjacency matrix as input. However, since spatial vector data contain only point locations without edge information, GNN cannot be directly applied. Although it is possible to construct the adjacency matrix by setting the aggregation condition, the geometric features of the graph may be lost, and the aggregation condition is cumbersome to set up. Inverse distance weighting and radial basis functions can assign different weights to neighboring points without setting an aggregation condition. However, existing research lacks relevant models directly applied to integrated point pattern extraction–classification scenarios.
Table 1 summarizes the strengths and weaknesses of existing methods in spatial vector point pattern recognition. The traditional point pattern extraction methods based on clustering algorithms face the difficulty of artificial parameter adjustment and low-SNR scenario application. Point pattern classification methods based on template matching and traditional machine learning are difficult to apply in complex scenarios. Graph methods cannot be applied directly due to the lack of edge information. In this article, we propose a graph structure of spatial vector data and a GEOmetric Feature Attention Network (GeoFAN) for point pattern recognition in spatial vector data.
Our main contributions are as follows:
1.
A graph conversion method for raw spatial vector data is proposed to represent the spatial distribution and attributes of all points. The distance between points is calculated as weighted adjacency matrices. The point characteristics are used as attribute matrices to realize the graph conversion, which fully expresses the geometric feature of the raw data. Complete and accurate data conversion is the basis of point pattern recognition.
2.
A geometric feature attention scheme is proposed to enhance the features of point patterns. The scheme can embed the graph geometric features in the attention coefficients and achieve the global weighted aggregation, which avoids setting up the aggregation conditions and solves the problem that the graph neural network cannot be directly applied to spatial vector data. As the geometric features of the spatial vector data are fully utilized, the differences between patterned points and normal points and between different point patterns are increased, which solves the difficulty of the low-SNR scenario application.
3.
A GeoFAN method is proposed to recognize point patterns in low-SNR spatial vector data. It solves the difficulty that extraction and classification cannot be done simultaneously, and it is a general and effective method to recognize point patterns in spatial vector data without setting any artificial parameters. The method is tested and evaluated by simulated point patterns and real location-based point patterns with different patterned point spacing, domain area range (no other points exist within this range of a point pattern), and point pattern types. The test results show that GeoFAN has superior recognition performance and generalization ability in point pattern recognition scenarios.
The rest of the article is organized as follows. Section 2 describes the materials and methods. Section 3 presents the experimental results. Section 4 discusses the findings and future work. Section 5 concludes the article.

2. Materials and Methods

In this section, we study the graph conversion of spatial vector data and the point pattern recognition method. Spatial vector data consists of points’ latitude, longitude, and attributes. By converting raw spatial vector data into a graph structure consisting of adjacency matrices and attribute matrices, the spatial distribution characteristics and point attributes can be extracted to express the raw spatial vector data effectively. Then, we propose a GeoFAN point pattern recognition method. The inputs are the adjacency and attribute matrices of the graph data, and the outputs are the point classification results numbered by point pattern types. We outline the structure of GeoFAN and describe in detail the point pattern recognition procedure.

2.1. Spatial Vector Data Graph Conversion

2.1.1. Raw Spatial Vector Data

Typically, observation instruments are located far from the targets to obtain spatial vector data from a wide area. Therefore, the effect of elevation can be ignored, and the observation area forms a two-dimensional plane.
An example of raw spatial vector data is shown in Figure 1. The data can be decomposed as follows:
1.
Point set: V R n × 1
2.
Latitude and longitude set: P R n × 2 . The latitude and longitude in radians of the ith point are denoted as p i = ( φ i , λ i )
3.
Attribute set: X R n × d . The attributes of the ith point are denoted as x i = ( x i 1 , x i 2 , , x i d )

2.1.2. Graph Conversion

A complete graph includes the adjacency matrix A and the attribute matrix X . In this section, we construct the adjacency matrix of the spatial vector data by the relative distance between points and the attribute matrix by the characteristics of all points. The conversion procedure is shown in Figure 2.
In spatial vector data, the points consist of normal and patterned points. Patterned points tend to be more aggregated and characterized by relative spatial proximity; thus, pairs of points with smaller spatial distances are more likely to be associated. To obtain spatial distance information, we calculate the relative spatial distance between points by latitude and longitude. If the data do not have latitude and longitude information, other methods are used to calculate the relative spatial distance.
The haversine formula is an effective method for calculating distances based on the difference in latitude and longitude between two points on the Earth, which is widely used in geographic information systems. The haversine formula for calculating the distance between two points is as follows:
a i j = 2 R arcsin H ( Δ φ ) + cos ( φ i ) cos ( φ j ) H ( Δ λ )
H ( θ ) = sin 2 ( θ 2 )
Δ φ = φ j φ i
Δ λ = λ j λ i
where H is the haversine function. a i j is the spatial distance between the ith point and the jth point. R is the radius of the Earth.
The distance-weighted adjacency matrix A can be obtained by calculating the haversine formula.
The point characteristics obtained from the spatial vector data contain multiple parameters. We assumed that the points in the raw spatial vector data contain three parameters, such as length, width, and height. Since the instrument has observation errors and points’ parameters may be time-varying, each parameter has maximum, minimum, and average values. The attribute matrix can be constructed as X R n × 9 , and its row vector x i is the attribute vector of the ith point:
x i = ( x i 1 , x i 2 , x i 3 , x i 4 , x i 5 , x i 6 , x i 7 , x i 8 , x i 9 )
where x i 1 , x i 2 , and x i 3 are the maximum, average, and minimum values of the 1st parameter of the ith point, respectively. x i 4 , x i 5 , and x i 6 are the maximum, average, and minimum values of the 2nd parameter of the ith point, respectively. x i 7 , x i 8 , and x i 9 are the maximum, average, and minimum values of the 3rd parameter of the ith point, respectively.
To make different attributes have the same weight, we normalize the columns of the attribute matrix to obtain the normalized attribute matrix X ¯ :
m j = max i x i j
x ¯ i j = x i j m j
where m j is the maximum value of the jth column of the attribute matrix X . x ¯ i j is the element of the ith row and the jth column of the normalized attribute matrix X ¯ .

2.2. Recognize Point Pattern via GeoFAN

2.2.1. Geometric Feature Attention Module

The graph neural network is an effective method for processing spatial vector data. With the development of graph neural networks, a variety of models, such as the graph convolutional network (GCN), graph autoencoder (GAE), graph generative network (GGN), graph recurrent network (GRN), and graph attention network (GAT), have appeared. These models can be adapted to multiple types of tasks at the node, edge, and graph levels.
Spatial vector data have the following characteristics. There are far more normal points in a frame than point patterns. Normal points are unevenly distributed, and some have stronger aggregation characteristics than point patterns. The strong similarity of different point patterns makes them difficult to distinguish. The above issues make point pattern recognition more challenging. Attention mechanisms can effectively face the above challenges, and they have been widely used in deep neural networks. They can enhance the model’s extraction performance of key features from the data and improve its expressive ability by adaptively assigning weights to different neighbors, thus achieving more accurate recognition results.
We calculate the correlation coefficient of the point pairs from the attributes of the points. The correlation matrix E can be represented as
e i j = σ ( x i , x j )
where e i j is the element of the ith row and the jth column of the correlation matrix E . x i is the attribute vector of the ith point. x j is the attribute vector of the jth point. σ ( · ) is a function that calculates the correlation of two points.
We choose a single fully connected layer to represent σ ( · ) . The correlation coefficient can be represented as
e i j = LeakyRelu ( [ x i W x j W ] u T )
where ‖ is a column-splicing operation. W R d × d 1 is the downscale conversion matrix of the point attribute of the trainable fully connected layer. u R 2 d 1 is the weight vector of the layer. We select the activation function with reference to [35], the activation function is LeakyReLU (negative slope is 0.2), thereby introducing a nonlinear capability to the network.
The correlation coefficients are calculated between all points, including the coefficient of each point with itself. According to Equation (9), the correlation matrix E can be obtained by inputting the point attributes.
To enhance the features of point patterns, a geometric feature graph attention (GeoFA) mechanism is proposed, as shown in Figure 3. Taking the inverse of all elements in the adjacency matrix A and multiplying them with the corresponding elements in the correlation matrix E to obtain the geometric feature attention matrix E ˜ ,
a i j = 1 / a i j
e ˜ i j = a i j e i j
where e ˜ i j is the element of the ith row and the jth column of the geometric feature attention matrix E ˜ .
The geometric feature attention module enables points with closer spatial distances to have greater weights, resulting in relatively greater attention coefficients, enhancing the difference between patterned and normal points. After calculating the geometric feature attention coefficients, the aggregation operation is realized, and the new attribute vector of point v i can be output:
x i = j = 1 n e ˜ i j x j W
The above describes the single-head geometric feature attention module.

2.2.2. Structure of GeoFAN

To further enhance the expressive power of the attention layer, a multi-head attention mechanism is introduced. K mutually independent attention modules are invoked, and the output of each module is then spliced together. A graph attention network layer can be represented as
x i ( k ) = j = 1 n e ˜ i j ( k ) x j W ( k )
where e ˜ i j ( k ) is the geometric feature attention coefficient calculated by the kth GeoFA module between the ith and the jth point. W ( k ) is the conversion matrix of the point attributes of the kth GeoFA module.
By adding multiple independent attention modules, the model can allocate attention to multiple relevant features between the center point and neighboring points, which enhances the model’s learning ability. The structure of the proposed GeoFAN method is shown in Figure 4. We used two multi-head graph attention layers to map the attribute network in low dimensions. The output of each GeoFA module in the first layer is obtained by splicing, and the output of each GeoFA module in the second layer is averaged to obtain the prediction results. The first layer output p i and the second layer output q i of the ith point are represented as follows:
p i = | | k = 1 K 1 j = 1 n e ˜ i j 1 ( k ) x j W 1 ( k )
e ˜ i j 1 ( k ) = 1 a i j LeakyRelu ( [ x i W 1 ( k ) | | x j W 1 ( k ) ] u 1 T )
q i = 1 K 2 k = 1 K 2 j = 1 n e ˜ i j 2 ( k ) p j W 2 ( k )
e ˜ i j 2 ( k ) = 1 a i j LeakyRelu ( [ p i W 2 ( k ) | | p j W 2 ( k ) ] u 2 T )
where W 1 ( k ) R d × d 1 is the point attribute conversion matrix of the first layer of GeoFAN, used to map the input attribute matrix X from d-dimensional to d 1 -dimensional. The attribute vector of the ith point output by the first layer is p i R ( K 1 d 1 ) . W 2 ( k ) R ( K 1 d 1 ) × d 2 is the point attribute conversion matrix of the second layer of GeoFAN. The attribute vector of the ith point output by the second layer is q i R d 2 .
The GeoFAN outputs the probability that each point belongs to each type; thus, its column dimension is equal to the number of point types. After taking the type with the largest probability as the recognition result of each point, point pattern recognition is realized.

2.3. Experimental Data and Environment

To validate the effectiveness of our proposed method, we performed qualitative and quantitative assessments of point pattern recognition effects. We constructed a formation recognition scenario. The experiment datasets in the article are synthetic datasets of 242 frames to simulate the satellite observation scenarios of ships at sea, the coordinates of ships in the synthetic datasets are extracted from the global satellite Automatic Identification System (AIS) data in 2016 to simulate the real spatial distribution of the ships, the study area covers the global major shipping lanes and sea areas, and the attributes are artificially generated, which are not directly mapped to the real physical parameters, focusing on verifying the model’s ability to mine point patterns with arbitrary features. The average point number of each frame is 264. Each point has specific attributes, realizing the construction of the spatial vector datasets.
The point attributes contain three main types of parameters: x 1 , x 2 , and x 3 . In practical scenarios, such as electromagnetic satellite signal detection, the target attributes contain the maximum, minimum, and average values of the signal during the observation time, with the maximum/minimum reflecting the fluctuating range of the signal attribute, and the average reflecting the signal attribute under its steady-state operation, which together define the fingerprint of the target. However, due to the sparsity of the actual data and the lack of a priori knowledge of the target, the true average value distribution is difficult to predict. To simulate the randomness of average values in real scenarios and ensure the generalization ability of the model under limited data, we selected nine parameters: x 1 max , x 1 ave , x 1 min , x 2 max , x 2 ave , x 2 min , x 3 max , x 3 ave , and x 3 min . The statistical parameters we set for all the points are shown in Figure 5. The average values of three parameters are randomly generated were between the maximum and minimum values to cover all possible average values in real scenarios.
There are a large number of normal points and only 0–3 point patterns in a single frame in the datasets, resulting in an average SNR lower than 5%. To make the model learn more point pattern features, data expansion is needed to increase the richness of the training datasets. Different numbers of point patterns were added to each frame, expanding the 242 datasets to 1000, significantly increasing the training data for learning. The training and test datasets were divided in a 9:1 ratio. Please note that the point patterns added to the datasets were only used for simulated analysis. In the real location-based point pattern recognition scenario, the datasets are not expanded.
The training was performed with Python 3.9. We trained our model using PyTorch 1.9.0 on NVIDIA GeForce RTX 2060. Since the patterned points in the datasets were fewer than normal points, the learning rate was too high to recognize all points as normal points, and too low to converge too slowly. We set the initial learning rate to 0.003, and after every 1 epoch, the learning rate was reduced to 0.997. The attention heads of both geometric feature attention layers were set to 8. The training process was stopped when the loss was no longer decreasing to prevent overfitting. In our datasets where there were more negative samples and fewer positive samples, we used three metrics to accurately and comprehensively evaluate the recognition effect of the GeoFAN method: macro accuracy, macro recall, and macro F 1 score:
M a c r o P = 1 m i = 1 m P i
M a c r o R = 1 m i = 1 m R i
M a c r o F 1 = 1 m i = 1 m F 1 i
F 1 i = 2 × P i × R i P i + R i
where m denotes the classes of points. P i denotes the precision of the ith label. R i denotes the recall of the ith label. F 1 i denotes the F 1 score of the ith label. Macro-metrics can give the same weight to positive and negative samples, thus effectively evaluating the model’s recognition performance.

3. Results

In this section, we test the recognition performance of GeoFAN using synthetic datasets. The effects of patterned point spacing, domain area range, and point pattern types on the recognition performance are analyzed, respectively, and we designed experiments to compare GeoFAN with other algorithms. The effectiveness of GeoFAN in point pattern recognition is verified by applying the method to real location-based points.

3.1. Performance Under Various Patterned Point Spacing

To analyze the model’s recognition effect under different patterned point spacing, we expanded the datasets by adding point patterns with different point spacing to the original point background.
Firstly, we set the condition that the maximum patterned point spacing is smaller than 100 km, then randomly selected a point cluster from the synthetic datasets that satisfies the condition as a point pattern, and used its point number and the attributes of each point as the fingerprint of this point pattern type. Multiple point pattern types could be generated according to the same method. Then, we obtained the latitude and longitude range of point distribution in each frame, randomly generated the center points of point patterns within this range, and generated the patterned points near each center point. By controlling the maximum spacing of patterned points, we created point patterns with different patterned point spacing. The following constraints were used to control the patterned point spacing:
d max L lim
where d max represents the maximum distance between points in a point pattern. L lim represents the constraint distance.
We used the original datasets as background and added patterned points with a maximum spacing of 20 km, 40 km, 60 km, 80 km, and 100 km as five different recognition scenarios. The 242 datasets were expanded to 1000 in each scenario. The point pattern type was 1. The attributes of the generated point patterns are shown in Table 2. Point index 1–8 indicates that this specific point pattern contains exactly 8 points. The number of point patterns in each frame is 0–6. The domain area range is 50 km. x 1 ave , x 2 ave , and x 3 ave are randomized between the minimum and maximum values.
The recognition results of three frames of the maximum spacing of 100 km are shown in Figure 6. Figure 6a–c show the ground truth of three frames. Figure 6d–f show the recognition results of the three frames. From the results, it can be seen that there are a few false alarms in Figure 6d. Figure 6e,f show that the recognition results of the second and third frames are consistent with the ground truth, which shows that GeoFAN has good recognition ability for point patterns with a maximum patterned point spacing of up to 100 km.
The precision, recall, and F 1 score results are shown in Table 3. It shows that the proposed GeoFAN performs better in recognizing point patterns with smaller point spacing than with larger point spacing. When the maximum patterned point spacing is small, the attention coefficients of patterned points will be larger than those of normal points after being weighted by the distance, increasing the difference of their updated attribute vectors. When the maximum patterned point spacing is large, the attention coefficient between points in point patterns decreases, thus the patterned points and normal points have strong similarity in updated attribute vectors, making it difficult for the model to differentiate between normal points and patterned points, thus the recognition effect of the model decreases.

3.2. Performance Under Various Domain Area Ranges

In geographic mapping and formation recognition, point patterns usually exhibit spatial independence in spatial vector data, i.e., they are spatially distant from other points, and this distance is referred to as the domain area range. Within the domain area range of point patterns, it can be assumed that there are no other points. In the original point background, we added point patterns with different domain area ranges of 100 km, 400 km, 700 km, and 1000 km as four different recognition scenarios. The 242 datasets were expanded to 1000 in each scenario. The point pattern attributes are the same as in Table 2. The maximum patterned point spacing was 100 km. The point pattern type was 1. The number of point patterns in each frame was 0–6.
The recognition results are shown in Table 4. The test results show that the GeoFAN method better recognizes point patterns with a larger domain area range than those with a smaller domain area range. When the domain area range is small, the normal points spatially close to the patterned points have large attention coefficients, which increases the difficulty of model recognition, resulting in the normal points being misclassified as patterned points. When the domain area range is larger, the number of normal points near the patterned points are reduced, including some “normal point clusters” with aggregation characteristics (see Figure 1), which increases the SNR of the datasets and the difference between the attention coefficients of the patterned points and the normal points. Therefore, the patterned points’ recognition accuracy, recall, and F 1 score are improved.

3.3. Performance Comparison with Clustering Algorithm

We specifically added comparison experiments on the classical DBSCAN clustering algorithm and GeoFAN, since only a single point pattern type is involved in Section 3.1 and Section 3.2 (essentially, the detection of spatially aggregated clusters). To verify the key role of attributes in pattern recognition, the experiments were designed in two typical scenarios: scenario A (maximum patterned point spacing of 100 km, domain area range of 50 km) and scenario B (maximum patterned point spacing of 100 km, domain area range of 100 km), which maintain a correspondence with the experimental conditions of Section 3.1 and Section 3.2. Considering the sensitivity of the DBSCAN algorithm to parameter settings, we used parameter optimization seeking on the entire dataset to confirm optimal parameters and perform point pattern extraction with these parameters. Table 5 shows the performance comparison results of the two algorithms in the two scenarios.
The experimental results show that GeoFAN achieves macro F 1 scores of 95.27% and 95.28% in the two test scenarios, which are higher than those of 88.26% and 89.04% for DBSCAN. The main reason is that DBSCAN only relies on spatial features for extraction, and is prone to misjudge normal point clusters with aggregation characteristics as patterned points in low-SNR scenarios, which is reflected in its low precision. In comparison, GeoFAN achieves the synergistic processing of spatial proximity and attributes through the geometric feature attention mechanism, which effectively constructs the joint spatial-attribute fingerprints of the point patterns, mitigating the misclassification of normal point clusters. It is worth noting that the weak advantage of DBSCAN in the recall index reveals the difference between the two algorithms: DBSCAN captures more normal point clusters with dense distribution, and has fewer misses compared to GeoFAN. Since the point patterns in the training datasets have fewer point pattern samples compared to normal points, there is a category imbalance; thus, GeoFAN may omit some patterned points.
In practice, the performance of DBSCAN relies on manual parameter tuning and only enables pattern extraction, not classification. GeoFAN not only circumvents the artificial parameter tuning problem through the end-to-end learning paradigm, but also realizes the integrated extraction–classification processing, which is fully verified in the following experiments with various pattern types.

3.4. Performance Under Various Point Pattern Types

The definition of point pattern type includes both point number characteristics and attribute characteristics, and two point patterns are considered to be of different types when they differ in either of these two characteristics. In the original point background, multiple types of point patterns were added to verify the ability of GeoFAN. The point pattern types for the four scenarios were 2, 3, 4, and 5, respectively. The 242 datasets were expanded to 1000 in each scenario. The attributes of the first type are listed in Table 2, and the attributes of the remaining four types of point patterns are shown in Table 6, Table 7, Table 8 and Table 9. The maximum patterned point spacing was 60 km. The domain area range was 50 km. In the four scenarios, the numbers of point patterns in each frame were 0–12, 0–18, 0–24, and 0–30, respectively, to ensure the balance of point pattern number of each type in different scenarios.
The recognition results are shown in Table 10. When the point pattern types increase from 2 to 4, the precision, recall, and F 1 scores of different types do not change significantly. However, when adding the fifth type, the recognition effect of the first and fifth types decrease, while the recognition effect of the other types is not affected much. This is because the newly added type has a very strong similarity to the first type. Point indexes 1, 2, 3, 5, and 6 of the fifth type are similar to point indexes 1, 4, 3, 8, and 6 of the first type. With such a strong similarity, it is difficult to distinguish between the two point pattern types. Nevertheless, the GeoFAN model still captures the minor attribute differences and point number differences between the two types, with F 1 scores of 87.15% and 79.35% for the two point pattern types, respectively, verifying the robustness and effectiveness of GeoFAN in point pattern recognition. The recognition results illustrate that the number of types does not significantly affect the recognition results, while similar point patterns will.
In the scenario of five point pattern types, we compared the recognition performance of the graph convolutional network (GCN), 1-neighbor graph attention network (1N-GAT), and 2-neighbor graph attention network (2N-GAT) with our proposed GeoFAN. Due to the lack of adjacency information in the raw spatial vector data, GAT could not be directly applied to our scenarios and required the artificial construction of adjacency matrices. We improved GAT to 1N-GAT and 2N-GAT. 1N-GAT and 2N-GAT represent the graph attention networks aggregating one and two nearest neighbor points, respectively. Figure 7 demonstrates a frame of recognition results of GCN, 1N-GAT, 2N-GAT, and GeoFAN. Figure 7a is the ground truth of this frame. As shown in Figure 7b, there are more misclassifications and omissions in the GCN recognition result, which may be caused by the difficulty of GCN in capturing neighboring features. 1N-GAT may cause normal points to be misclassified as patterned points because it only aggregates one nearest neighbor of each point, and many false alarms (FAs) can be seen in Figure 7c. In Figure 7d, 2N-GAT aggregates two nearest neighbors and achieves a better result than 1N-GAT, and still, FAs exist. It shows that increasing the number of aggregated nearest neighbors can reduce FAs. However, the number of aggregated neighbors can’t exceed point number in a single point pattern, otherwise the normal points will be aggregated incorrectly. Due to the variable point number in a point pattern, it is difficult to determine the number of neighbors that need to be aggregated in practical applications. In Figure 7e, although the GeoFAN recognizes a few points of the fifth type as the first type due to the similar attributes, there is no false alarm in the recognition result.
Figure 8 shows the training processes of the four algorithms in the scenario of five types, which are stopped when the losses do not decrease further. As seen from the training processes, our proposed GeoFAN model finally converges with a higher F 1 score of 0.915 than GCN 0.631, 1N-GAT 0.738, and 2N-GAT 0.786, and with a smaller loss value than the other three algorithms, which illustrates that the GeoFAN outperforms the GCN, 1N-GAT, and 2N-GAT in point pattern recognition. A comparison of the recognition performance of the four algorithms is shown in Table 11, which indicates that GeoFAN performs better than other algorithms on all metrics for five types of point patterns.
The following may be the reason why the performance of other methods is limited. GCN is less flexible since the inter-node attribute correlation is not considered during the aggregation process, making it difficult to extract the correlation strength between the central node and its neighboring nodes in point patterns. 1N-GAT and 2N-GAT achieve better results than GCN by aggregating one and two nearest neighbor points around each point. However, due to the uneven distribution of points in point patterns, selecting too few points in the aggregation may lead to inter-disconnection. Although this can be mitigated by increasing the neighbor number used for aggregation, the point numbers of point patterns are variable; thus, the selection of the neighbor number used for aggregation is a great difficulty. In comparison, GeoFAN not only achieved better results, but also does not need to set artificial parameters, which is a practical and high-performance point pattern recognizer.

3.5. Performance Under Real Location-Based Point Patterns

The GeoFAN model has achieved excellent recognition performance in simulated point patterns. In practice, there may be a lack of training datasets of some point patterns, resulting in the model being unable to learn these point pattern types directly. Therefore, we use the GeoFAN model trained on simulated data to recognize the real location-based point patterns. This application can help recognize point patterns with sparse training datasets and predefine the unknown point patterns to be recognized.
The method was applied to five different types of point patterns. Figure 9a–d show the ground truth of the five point pattern types in four frames. Figure 9e–h show the GeoFAN’s recognition results of the five point pattern types. From the recognition results, it can be seen that GeoFAN can accurately recognize the third and fourth types, as shown in Figure 9f,g. In Figure 9f, there is a slight misjudgment for the second type. Figure 9e,h show that recognition errors occur in the results of the first type and the fifth type. The recognition accuracies of the first and fifth types are affected by the similarity of their attributes, leading to crosstalk. Despite the crosstalk in the recognition results, the first and fifth types are not recognized as the other types. It illustrates that when two point patterns are very similar, correctly distinguishing between the two real point patterns using simulated data remains a huge challenge. Nevertheless, the GeoFAN still succeeds in separating the first type and the fifth type from the others and accurately recognizes some points on both types. The recognition results show that the GeoFAN model can be trained with simulated data to recognize real location-based point patterns, which is of high value in practical situations where there is a lack of point pattern datasets. GeoFAN recognizes multiple types of real location-based point patterns from a large number of points, which reaffirms the excellent performance of GeoFAN in low-SNR point pattern recognition scenarios.

4. Discussion

The GeoFAN framework proposed in this study effectively solves the dual challenges of low SNR and missing edge information, realizing point pattern recognition from spatial vector data through the fusion of the geometric feature attention mechanism and graph neural network. The method can realize simultaneous extraction and classification of point patterns without manual parameter tuning, which significantly improves the recognition accuracy of cluster structures (e.g., building and island clusters) in geographic mapping, and provides fast situational awareness for formation recognition (e.g., ship and aircraft formations). Its performance of successfully migrating to real location-based scenarios after training based on simulated data verifies the strong generalization ability of the model in scenarios where annotated data are scarce or expensive to acquire. This result provides a new paradigm for the automated analysis of complex spatial vector data, which is especially applicable to field that needs to mine weakly correlated patterns from massive, noisy data.
The joint spatial-attribute modeling mechanism designed in this study is similar to the idea of Tobler’s First Law, which is to increase the relevance of targets through their proximity. The inverse distance weighting first achieves the quantitative modeling of spatial proximity, effectively preserving spatial features’ continuity and integrity. Then, the geometric attention mechanism deeply integrates the spatial proximity and attribute correlation. This spatial-attribute coupled modeling mechanism shows advantages in the experiments, with the macro F 1 scores improving by 17.69% and 12.91% compared with 1N-GAT and 2N-GAT based on lossy spatial features and attributes, respectively, and the macro F 1 score improving by 6.24% compared with the DBSCAN based on spatial features. The core innovation is that GeoFAN not only retains the spatial features completely through inverse distance weighting, but also constructs a joint fingerprint of spatial proximity and attribute, which not only improves the effect compared to 1N-GAT, 2N-GAT, and DBSCAN, but also eliminates the need for manual parameter tuning and achieves the integration of extraction and classification.
The computational complexity of GeoFAN mainly involves two links. The feature mapping link maps all the node features from d-dimensional to d 1 -dimensional, with a computational complexity of O ( n × d × d 1 ) . The computation of attention link maps the 2 d 1 vectors of the paired points to the real numbers, and there are n points in a graph; thus, its computational complexity is O ( n 2 × d 1 ) . The computational complexity of single-head GeoFAN is O ( n × d × d 1 + n 2 × d 1 ) . In our experiments, the average number of points in a single frame is about 264, and the model training and inference times are within an acceptable range at this scale. However, we observe increased video memory footprint and computation time when working with larger datasets ( n > 264 ). To address this problem, the following optimization strategies can be adopted: (1) A sparse attention mechanism. Constrain the global attention computation to the local neighborhood through K-nearest neighbor filtering, which reduces the complexity to O ( n K ) . (2) Hierarchical processing methods. Filter regions through density-awareness, then compute in the candidate regions. The above methods can significantly improve the computational efficiency while maintaining the performance of the model.
While GeoFAN performs well in point pattern recognition scenarios, distinguishing subtle distinctions among highly similar point patterns presents an opportunity for further refinement. In the future, a hybrid architecture combining attention and contrastive learning can be explored to strengthen the inter-class feature difference. In addition, it is necessary to integrate multimodal data to improve the performance of the point pattern recognition model. Finally, an incremental learning process can be introduced to realize the learning and recognition of unknown point patterns.

5. Conclusions

This article focuses on the problem of spatial vector point pattern recognition, proposing a graph conversion method for converting points based on latitude and longitude coordinates into a graph structure and a GeoFAN recognition model for point pattern recognition. The raw spatial vector data are converted into a graph structure through the inter-point distance and attributes, which are used as the input of the GeoFAN model. A geometric feature attention mechanism is proposed to enhance the inter-correlation of point patterns, and finally, the GeoFAN model effectively and accurately recognizes point patterns from the low-SNR spatial vector data. The experimental results show that GeoFAN can recognize the point patterns with five different types with a macro F 1 score of 91.47%, and it can recognize the real location-based point patterns by the model trained on the simulated data. In military scenarios, GeoFAN can identify high-threat battle formations based on satellite observation of ship positioning and attribute data, providing key decision-making support for battlefield situational awareness; in urban planning scenarios, GeoFAN can automatically identify building cluster patterns with specific spatial distribution and functional/structural characteristics, providing a basis for urban functional zoning and spatial structure optimization.
In subsequent work, we plan to introduce contrastive learning strategies to improve GeoFAN to recognize more similar point patterns, then include more types of point patterns for model training. Further, developing an incremental learning framework to enable continuous recognition of emerging point patterns will fill a critical technology gap in real-time spatial vector data analysis.

Author Contributions

Conceptualization, Zhuoyi Yang, Zeyi Li, Haitao Zhang and Wei Zhang; methodology, Zhuoyi Yang, Zeyi Li and Wei Zhang; software, Zhuoyi Yang, Zeyi Li and Yihang Huang; validation, Zhuoyi Yang; formal analysis, Zhuoyi Yang; investigation, Yanwei Wang; resources, Haitao Zhang; data curation, Haitao Zhang; writing—original draft preparation, Zhuoyi Yang; writing—review and editing, Zhuoyi Yang and Yanwei Wang; visualization, Yihang Huang; supervision, Haitao Zhang and Wei Zhang; project administration, Haitao Zhang. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Dataset available on request from the authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhao, R.; Ai, T.; Yu, W.; He, Y.; Shen, Y. Recognition of building group patterns using graph convolutional network. Cartogr. Geogr. Inf. Sci. 2020, 47, 400–417. [Google Scholar] [CrossRef]
  2. Bai, L.; Huang, W.; Zhang, X.; Du, S.; Cong, G.; Wang, H.; Liu, B. Geographic mapping with unsupervised multi-modal representation learning from VHR images and POIs. ISPRS J. Photogramm. Remote Sens. 2023, 201, 193–208. [Google Scholar] [CrossRef]
  3. Qiao, D.; Liu, G.; Lv, T.; Li, W.; Zhang, J. Marine vision-based situational awareness using discriminative deep learning: A survey. J. Mar. Sci. Eng. 2021, 9, 397. [Google Scholar] [CrossRef]
  4. Liang, M.; Weng, L.; Gao, R.; Li, Y.; Du, L. Unsupervised maritime anomaly detection for intelligent situational awareness using AIS data. Knowl.-Based Syst. 2024, 284, 111313. [Google Scholar] [CrossRef]
  5. Song, G.; Liu, Z.; Liu, T.; Qiang, Y. Research on the method of warship formation threat assessment based on structural contribution degree. J. Phys. Conf. Ser. 2021, 1861, 012066. [Google Scholar] [CrossRef]
  6. Yang, J.; Ma, L.; Liu, J. Modeling and application of ship density based on ship scale conversion and grid. Ocean. Eng. 2021, 237, 109557. [Google Scholar] [CrossRef]
  7. Steiniger, S.; Burghardt, D.; Weibel, R. Recognition of island structures for map generalization. In Proceedings of the 14th Annual ACM International Symposium on Advances in Geographic Information Systems, Arlington, VA, USA, 10–11 November 2006; pp. 67–74. [Google Scholar]
  8. Gaoqian, Z.; Ying, Z. Radar netting technology & its development. In Proceedings of the 2011 IEEE CIE International Conference on Radar, Chengdu, China, 24–27 October 2011; Volume 1, pp. 933–937. [Google Scholar]
  9. Yan, X.; Ai, T. Analysis of irregular spatial data with machine learning: Classification of building patterns with a graph convolutional neural network. arXiv 2018, arXiv:1809.08196. [Google Scholar]
  10. Zhang, Y.; Ren, H.; Ye, J.; Gao, X.; Wang, Y.; Ye, K.; Xu, C.Z. Aoam: Automatic optimization of adjacency matrix for graph convolutional network. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 5130–5136. [Google Scholar]
  11. Han, J.; Cen, J.; Wu, L.; Li, Z.; Kong, X.; Jiao, R.; Yu, Z.; Xu, T.; Wu, F.; Wang, Z.; et al. A survey of geometric graph neural networks: Data structures, models and applications. arXiv 2024, arXiv:2403.00485. [Google Scholar] [CrossRef]
  12. Basaraner, M.; Selcuk, M. A structure recognition technique in contextual generalisation of buildings and built-up areas. Cartogr. J. 2008, 45, 274–285. [Google Scholar] [CrossRef]
  13. Ezugwu, A.E.; Shukla, A.K.; Agbaje, M.B.; Oyelade, O.N.; José-García, A.; Agushaka, J.O. Automatic clustering algorithms: A systematic review and bibliometric analysis of relevant literature. Neural Comput. Appl. 2021, 33, 6247–6306. [Google Scholar] [CrossRef]
  14. MacQueen, J. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Statistics; University of California Press: Berkeley, CA, USA, 1967; Volume 5, pp. 281–298. [Google Scholar]
  15. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the KDD, Portland, OR, USA, 2–4 August 1996; Volume 96, pp. 226–231. [Google Scholar]
  16. He, H.; Wu, L.; Hu, X. Warship formation extraction and recognition based on density-based spatial clustering of applications with noise and improved convolutional neural network. IET Radar Sonar Navig. 2022, 16, 1912–1923. [Google Scholar] [CrossRef]
  17. Cetinkaya, S.; Basaraner, M.; Burghardt, D. Proximity-based grouping of buildings in urban blocks: A comparison of four algorithms. Geocarto Int. 2015, 30, 618–632. [Google Scholar] [CrossRef]
  18. Yan, X.; Ai, T.; Yang, M.; Tong, X.; Liu, Q. A graph deep learning approach for urban building grouping. Geocarto Int. 2022, 37, 2944–2966. [Google Scholar] [CrossRef]
  19. Regnauld, N. Contextual building typification in automated map generalization. Algorithmica 2001, 30, 312–333. [Google Scholar] [CrossRef]
  20. Wang, W.; Du, S.; Guo, Z.; Luo, L. Polygonal clustering analysis using multilevel graph-partition. Trans. GIS 2015, 19, 716–736. [Google Scholar] [CrossRef]
  21. Wu, B.; Yu, B.; Wu, Q.; Chen, Z.; Yao, S.; Huang, Y.; Wu, J. An extended minimum spanning tree method for characterizing local urban patterns. Int. J. Geogr. Inf. Sci. 2018, 32, 450–475. [Google Scholar] [CrossRef]
  22. Yu, W.; Zhou, Q.; Zhao, R. A heuristic approach to the generalization of complex building groups in urban villages. Geocarto Int. 2021, 36, 155–179. [Google Scholar] [CrossRef]
  23. Ruas, A.; Holzapfel, F. Automatic characterisation of building alignments by means of expert knowledge. In Proceedings of the 21th ICC Conference, Durban, South Africa, 10–16 August 2003; pp. 1604–1616. [Google Scholar]
  24. Rainsford, D.; Mackaness, W. Template matching in support of generalisation of rural buildings. In Proceedings of the Advances in Spatial Data Handling: 10th International Symposium on Spatial Data Handling, Ottawa, Canada, 9–12 July 2002; Springer: Berlin/Heidelberg, Germany, 2002; pp. 137–151. [Google Scholar]
  25. Liu, W.; Sun, W.; Zhu, L.; Zhao, Z. Study on Formation Recognition and Target Selection Method of Shipborne Aircraft Against Warship Formation. J. Ordnance Equip. Eng. 2020, 41, 85–89. [Google Scholar]
  26. Cheng, B.; Wang, D.; Niu, J.; Li, X.; Luan, S. Convolutional Radon Transformation Method for Sparse Collective Recognizing. In Proceedings of the 2023 9th International Conference on Control, Automation and Robotics (ICCAR), Beijing, China, 21–23 April 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 394–398. [Google Scholar]
  27. Lin, Z.; Zhang, X.; Hao, N.; He, F. An LSTM-based fleet formation recognition algorithm. In Proceedings of the 2021 40th Chinese Control Conference (CCC), Shanghai, China, 26–28 July 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 8565–8569. [Google Scholar]
  28. Liang, F.; Zhou, Y.; Zhang, Z.; Chen, X.; Tang, X.; Sun, Q. An LSTM-based Method for Recognition and Prediction of Aircraft Formation. In Proceedings of the 2022 International Conference on Pattern Recognition and Intelligent Systems, Montréal, QC, Canada, 21–25 August 2022; pp. 80–84. [Google Scholar]
  29. Liang, F.; Zhou, Y.; Zhang, C.; Song, Z.; Zhao, X. An aircraft formation recognition method based on deep learning. In Proceedings of the 2022 7th International Conference on Communication, Image and Signal Processing (CCISP), Chengdu, China, 18–20 November 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 457–460. [Google Scholar]
  30. He, X.; Zhang, X.; Xin, Q. Recognition of building group patterns in topographic maps based on graph partitioning and random forest. ISPRS J. Photogramm. Remote Sens. 2018, 136, 26–40. [Google Scholar] [CrossRef]
  31. Lin, Z.; Zhang, X.; He, F. A GNN-LSTM-Based Fleet Formation Recognition Algorithm. In Proceedings of the International Conference on Guidance, Navigation and Control, Harbin, China, 5–7 August 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 7272–7281. [Google Scholar]
  32. Yan, X.; Ai, T.; Yang, M.; Yin, H. A graph convolutional neural network for classification of building patterns using spatial vector data. ISPRS J. Photogramm. Remote Sens. 2019, 150, 259–273. [Google Scholar] [CrossRef]
  33. Yang, M.; Cao, M.; Cheng, L.; Jiang, H.; Ai, T.; Yan, X. Classification of urban interchange patterns using a model combining shape context descriptor and graph convolutional neural network. Geo-Spat. Inf. Sci. 2024, 27, 1622–1637. [Google Scholar] [CrossRef]
  34. Yu, H.; Ai, T.; Yang, M.; Huang, L.; Yuan, J. A recognition method for drainage patterns using a graph convolutional network. Int. J. Appl. Earth Obs. Geoinf. 2022, 107, 102696. [Google Scholar] [CrossRef]
  35. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
Figure 1. Raw spatial vector data sample. Point patterns are represented by red dots and normal points are represented by blue dots. Normal point clusters consist of normal points with strong aggregation characteristics, as shown in the three illustrations of normal point clusters below. In comparison to the irregular point clusters, point patterns have a specific number of points with point attributes, as shown in the two illustrations in the upper left corner.
Figure 1. Raw spatial vector data sample. Point patterns are represented by red dots and normal points are represented by blue dots. Normal point clusters consist of normal points with strong aggregation characteristics, as shown in the three illustrations of normal point clusters below. In comparison to the irregular point clusters, point patterns have a specific number of points with point attributes, as shown in the two illustrations in the upper left corner.
Ijgi 14 00214 g001
Figure 2. Flowchart of the graph conversion procedure. The raw spatial data are divided into multiple point sets by frames. The latitude and longitude set of each point set generates the adjacency matrix. The attribute set generates the normalized attribute matrix.
Figure 2. Flowchart of the graph conversion procedure. The raw spatial data are divided into multiple point sets by frames. The latitude and longitude set of each point set generates the adjacency matrix. The attribute set generates the normalized attribute matrix.
Ijgi 14 00214 g002
Figure 3. Structure of geometric feature attention module.
Figure 3. Structure of geometric feature attention module.
Ijgi 14 00214 g003
Figure 4. Structure of GeoFAN.
Figure 4. Structure of GeoFAN.
Ijgi 14 00214 g004
Figure 5. Statistical attributes of all points. The horizontal coordinate is the value of three types of parameters and the vertical coordinate is the corresponding point counter. The insets show the distribution of the point counters for the third parameter within 0–7.
Figure 5. Statistical attributes of all points. The horizontal coordinate is the value of three types of parameters and the vertical coordinate is the corresponding point counter. The insets show the distribution of the point counters for the third parameter within 0–7.
Ijgi 14 00214 g005
Figure 6. Recognition results of maximum patterned point spacing of 100 km. The red dots are patterned points and the blue dots are normal points. FA represents false alarm. (ac) Ground truth of frames 1, 2, and 3, respectively. (df) Recognition results of frames 1, 2, and 3, respectively.
Figure 6. Recognition results of maximum patterned point spacing of 100 km. The red dots are patterned points and the blue dots are normal points. FA represents false alarm. (ac) Ground truth of frames 1, 2, and 3, respectively. (df) Recognition results of frames 1, 2, and 3, respectively.
Ijgi 14 00214 g006
Figure 7. Recognition results of four algorithms for 5 types of point patterns. (a) Ground truth. (b) GCN. (c) 1N-GAT. (d) 2N-GAT. (e) GeoFAN.
Figure 7. Recognition results of four algorithms for 5 types of point patterns. (a) Ground truth. (b) GCN. (c) 1N-GAT. (d) 2N-GAT. (e) GeoFAN.
Ijgi 14 00214 g007
Figure 8. Validation losses and validation F 1 scores over time. (a) GCN. (b) 1N-GAT. (c) 2N-GAT. (d) GeoFAN.
Figure 8. Validation losses and validation F 1 scores over time. (a) GCN. (b) 1N-GAT. (c) 2N-GAT. (d) GeoFAN.
Ijgi 14 00214 g008
Figure 9. Recognition results of 5 types of point patterns based on real location-based data. (ad) Ground truth of frames containing 5 types of point patterns. (eh) Recognition results of frames containing 5 types of point patterns.
Figure 9. Recognition results of 5 types of point patterns based on real location-based data. (ad) Ground truth of frames containing 5 types of point patterns. (eh) Recognition results of frames containing 5 types of point patterns.
Ijgi 14 00214 g009
Table 1. Advantages and disadvantages of different methods in spatial vector point pattern recognition.
Table 1. Advantages and disadvantages of different methods in spatial vector point pattern recognition.
ApproachesAdvantagesDisadvantages
Point pattern
extraction
Segmentation elementsUtilizes natural/artificial elements for pattern separation with intuitive logic.Highly dependent on auxiliary geographic elements.
ClusteringSimple and efficient for large-scale data; effective for explicit spatial aggregation patterns.Require manual parameter tuning; vulnerable to noise/normal clusters in low-SNR scenarios.
Neighborhood graphPattern separation through spatial proximity in the absence of segmentation elements.Dependent on expert experience; difficult to generalize.
Point pattern
classification
Template matchingSimple matching logic; high accuracy for standardized point pattern classification.Partially obscured or small deviation from the template may result in a match failure.
Traditional machine learningAutomatically learn the features in the sample that have been extracted to achieve the extraction of the unknown point pattern.Difficulty in capturing spatial correlation of unstructured data.
Graph neural networkExplicit spatial-attribute coupling via graph structures; handles non-Euclidean data.Cannot be applied directly in space vector data due to lack of adjacency matrices.
Table 2. Attributes of each point in the point pattern.
Table 2. Attributes of each point in the point pattern.
Point Index x 1 x 2 x 3
MaxMinMaxMinMaxMin
186.2986.2016.6816.611.121.08
251.7246.5229.6216.386.581.52
345.2344.3948.6113.626.511.02
482.0281.9928.2725.381.201.09
544.9244.3026.679.509.740.96
651.4246.5217.308.924.333.06
743.5443.3950.7413.046.616.06
813.6713.4123.6916.9432.5431.50
Table 3. Precision, recall, and F 1 score (%) of different maximum patterned point spacings.
Table 3. Precision, recall, and F 1 score (%) of different maximum patterned point spacings.
Point Type20 km40 km60 km80 km100 km
P R F 1 P R F 1 P R F 1 P R F 1 P R F 1
Normal99.8499.8199.8399.5599.7499.6599.6899.3799.5399.2599.5399.3998.8799.4599.21
Patterned97.7998.1897.9996.8394.6795.7493.0496.4294.7095.0192.3193.6493.8688.9391.33
Macro98.8299.0098.9198.1997.2097.6996.3697.8997.1197.1395.9296.5296.4194.1995.27
Table 4. Precision, recall, and F 1 score (%) of different domain area ranges.
Table 4. Precision, recall, and F 1 score (%) of different domain area ranges.
Point Type100 km400 km700 km1000 km
P R F 1 P R F 1 P R F 1 P R F 1
Normal99.0699.3699.2198.9899.4799.2399.0199.5099.2599.4299.6199.52
Patterned92.9289.8591.3694.3889.5891.9295.5891.5093.5096.2894.5395.40
Macro95.9994.6195.2896.6894.5395.5797.2995.5096.3897.8597.0797.46
Table 5. Performance comparison with DBSCAN on point pattern extraction.
Table 5. Performance comparison with DBSCAN on point pattern extraction.
Point TypeDBSCANOurs
P R F 1 P R F 1
Maximum patterned
point spacing of 100 km
Normal96.5299.0397.7698.8799.4599.21
Patterned70.2389.6478.7693.8688.9391.33
Macro83.3894.3488.2696.4194.1995.27
Domain area
range of 100 km
Normal96.6299.2397.9099.0699.3699.21
Patterned71.2091.7480.1892.9289.8591.36
Macro83.9195.4989.0495.9994.6195.28
Table 6. Attributes of each point in the second type of point pattern.
Table 6. Attributes of each point in the second type of point pattern.
Point Index x 1 x 2 x 3
MaxMinMaxMinMaxMin
142.4242.2729.2721.471.581.52
279.7079.6177.1376.942.382.25
380.1980.0147.6947.642.222.20
449.2049.1322.8820.9114.6314.53
515.4915.3567.6965.362.890.72
682.7982.668.568.08599.89.36
Table 7. Attributes of each point in the third type of point pattern.
Table 7. Attributes of each point in the third type of point pattern.
Point Index x 1 x 2 x 3
MaxMinMaxMinMaxMin
183.8183.80124.99124.9112.4111.85
281.2281.11125.03124.920.270.21
383.8983.81125.06124.810.530.24
441.6741.4421.4616.401.460.27
545.7145.6735.7731.201.141.08
Table 8. Attributes of each point in the fourth type of point pattern.
Table 8. Attributes of each point in the fourth type of point pattern.
Point Index x 1 x 2 x 3
MaxMinMaxMinMaxMin
154.0638.9420.0819.941.361.18
249.6147.4059.628.023.521.61
345.1444.22192.3716.103.330.92
413.6913.5323.6216.9931.9531.77
516.4016.29182.2718.000.610.51
651.5146.4912.998.086.681.22
783.7583.6317.5115.700.410.30
882.0281.9728.0325.101.531.04
952.4651.7131.969.833.463.39
Table 9. Attributes of each point in the fifth type of point pattern.
Table 9. Attributes of each point in the fifth type of point pattern.
Point Index x 1 x 2 x 3
MaxMinMaxMinMaxMin
185.1883.6717.6315.740.610.25
282.0281.9828.5124.831.271.10
345.0544.3353.1914.906.440.84
451.6646.5712.237.113.471.38
513.5713.4723.6017.4131.9731.84
652.3951.7418.395.903.463.28
Table 10. Precision, recall, and F 1 score (%) of different point pattern types.
Table 10. Precision, recall, and F 1 score (%) of different point pattern types.
Point Type2 Types3 Types4 Types5 Types
P R F 1 P R F 1 P R F 1 P R F 1
Normal99.2999.4399.3698.9398.8098.8698.0099.1698.5898.1499.2198.67
1st Pattern94.7093.0493.8690.5898.2894.2793.1090.0791.5686.2288.1087.15
2nd Pattern98.6798.9198.7998.9197.5698.2398.2498.6298.4396.5796.2296.40
3rd Pattern---93.7985.6789.5595.0985.7090.1596.0787.4191.53
4th Pattern------96.8593.3795.0895.5395.9495.73
5th Pattern---------84.3474.9179.35
Macro97.5597.1397.3495.5595.0895.2396.2693.3894.7692.8190.3091.47
Table 11. Comparison of the effectiveness of four algorithms in recognizing five point pattern types.
Table 11. Comparison of the effectiveness of four algorithms in recognizing five point pattern types.
Point TypeGCN1N-GAT2N-GATOurs
P R F 1 P R F 1 P R F 1 P R F 1
Normal92.7097.2294.9193.7296.8695.2695.6897.3296.4998.1499.2198.67
1st Pattern64.7157.2360.7461.9268.9865.2666.0072.2268.9786.2288.1087.15
2nd Pattern74.4167.2370.6486.0276.3780.9189.6687.4788.5596.5796.2296.40
3rd Pattern79.1851.9562.7491.9177.2983.9788.2886.1287.1996.0787.4191.53
4th Pattern70.0158.1963.5569.7358.4363.5882.0770.4075.7995.5395.9495.73
5th Pattern40.6618.9825.8858.6449.5553.7158.4850.8454.3984.3474.9179.35
Macro70.2858.4763.0876.9971.2573.7880.0377.4078.5692.8190.3091.47
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Z.; Li, Z.; Zhang, H.; Zhang, W.; Wang, Y.; Huang, Y. GeoFAN: Point Pattern Recognition in Spatial Vector Data. ISPRS Int. J. Geo-Inf. 2025, 14, 214. https://doi.org/10.3390/ijgi14060214

AMA Style

Yang Z, Li Z, Zhang H, Zhang W, Wang Y, Huang Y. GeoFAN: Point Pattern Recognition in Spatial Vector Data. ISPRS International Journal of Geo-Information. 2025; 14(6):214. https://doi.org/10.3390/ijgi14060214

Chicago/Turabian Style

Yang, Zhuoyi, Zeyi Li, Haitao Zhang, Wei Zhang, Yanwei Wang, and Yihang Huang. 2025. "GeoFAN: Point Pattern Recognition in Spatial Vector Data" ISPRS International Journal of Geo-Information 14, no. 6: 214. https://doi.org/10.3390/ijgi14060214

APA Style

Yang, Z., Li, Z., Zhang, H., Zhang, W., Wang, Y., & Huang, Y. (2025). GeoFAN: Point Pattern Recognition in Spatial Vector Data. ISPRS International Journal of Geo-Information, 14(6), 214. https://doi.org/10.3390/ijgi14060214

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop