Next Article in Journal
Joint Embedding of Semantic and Statistical Features for Effective Code Search
Previous Article in Journal
Development of a Machine Learning-Based Framework for Predicting Vessel Size Based on Container Capacity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Skeleton-Line-Based Graph Convolutional Neural Network for Areal Settlements’ Shape Classification

1
Faculty of Geomatics, Lanzhou Jiaotong University, Lanzhou 730070, China
2
National-Local Joint Engineering Research Center of Technologies and Applications for National Geographic State Monitoring, Lanzhou 730070, China
3
Gansu Provincial Engineering Laboratory for National Geographic State Monitoring, Lanzhou 730070, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(19), 10001; https://doi.org/10.3390/app121910001
Submission received: 31 August 2022 / Revised: 26 September 2022 / Accepted: 29 September 2022 / Published: 5 October 2022
(This article belongs to the Section Earth Sciences)

Abstract

:
Among the geographic elements, shape recognition and classification is one of the im portant elements of map cartographic generalization, and the shape classification of an areal settlement is an important part of geospatial vector data. However, there is currently no relatively simple and efficient way to achieve areal settlement classification. Therefore, we combined the skeleton line vector data of an areal settlement and the graph convolutional neural network to propose an areal settlement shape classification method that (1) extracts the skeleton line of the areal settlement to form a dual graph with nodes as edges, (2) extracts multiple features to obtain a graph representation of the shape, (3) extracts and aggregates the shape information represented by the areal settlement skeleton line using the graph convolutional neural network for multiple rounds to extract high-dimensional shape information, and (4) completes the shape classification of the high-dimensional shape information. The experiment used 240 samples, and the classification accuracy was 93.3%, with areal settlement shapes of E-, F-, and H-type achieving F-measures of 96.5%, 92.3%, and 100%, respectively. The result shows that the classification method of the areal settlement shape has high accuracy.

1. Introduction

As a geospatial data type, an areal settlement plays an important role in the field of geographic information systems (GIS), usually appearing as a basic element in map updating [1,2], cartographic generalization [3,4,5], and so on. To improve the readability and quality of areal settlement data, it is necessary to classify the shape of the areal settlement.
As an important attribute of areal settlement elements, shape conveys a greater amount of information in the representation of geographical entities than color, texture, [6], etc. The shape recognition methods that national and international experts use for areal settlement can be broadly divided into two aspects:
1. Shape description and recognition of the two-dimensional surface area of an areal settlement, which is also divided into two parts. One is based on the description of the overall shape after gridding of the vectorial areal settlement. For example, the most straightforward way to describe the shape of a building is by constructing a description of the chain code [7], the shape context [8] descriptor, using the grid as the object. It also converts remotely sensed images into raster images, with wavelet descriptors based on raster data [9], or histogram statistics [10], both of which can provide a good description of image features. The other is the abstraction of areas within the areal settlement in special shapes instead. For example, the areal settlement is matched with a template in the shape of the alphabet [11,12]. This method is effective for the recognition of an areal settlement or the division of a two-dimensional plane into multiple triangular surface areas [13,14], a good way of describing global and local shape features. All of these methods partition the areal settlement into regular geometric shapes, However, the results are not satisfactory for describing some of the more concave, convex, or complex areal settlements.
2. Description of the shape of the contour of the 2D-surface-shaped areal settlement boundary line, quantifying the contour with some special descriptive methods to describe the contour, for example, construction of a measure of the similarity of points on different contours through shape contexts [2], construction of the chordal eigenmatrix of a contour [15], multiscale shape representation of plane curves based on contour curvature [16], rotation functions [17], and description of the contours of building polygons using several spatial indicators [18]. These methods have the advantage of being computationally intuitive, but the two-dimensional surface contour information is too redundant and does not extract key concave and convex information for the classification of areal settlement shapes.
In recent years, deep learning has achieved excellent results in computer vision, natural language processing, and speech recognition [19,20,21]. Commonly used in the earth sciences to understand spatial processes and improve geographical insight, it is primarily combined with regular grid-like data to extract geometric features through convolutional neural networks. For example, convolutional neural networks combined with rasterized remote sensing images [22] are used to extract key information about important overpass structures or road intersections [23] or contour line data of one-dimensional trajectories are processed into regular raster-like data [24]. Both are prepared for the needs of neural networks for regular-like data.
However, the data for convolutional neural networks are supposed to be regularized grid-like data and cannot be applied directly to some irregularly shaped data. For this reason, some scholars have proposed graph convolutional neural networks (GCNs) [25], which use graph structures to construct irregular data, combined with convolutional neural networks. For example, graph data structures are constructed to recognize the shapes of river system clusters [26]. Constructing building group graph data helps to extract the geometric topological features among building groups and classify the internal regularity of building groups [27,28]. However, these methods have achieved good results in the distinction of groups of buildings rather than individual settlements. For large-scale individual areal settlements, based on theories such as the first law of geography and spatial association, the descriptive representation of elements in map space involves a local map structure consisting of neighboring points, neighboring edges, or arcs in a single geographic element [29]. Therefore, the combination of the local graph structure, the global perception of the overall structure, and the method of graph convolutional neural networks classifying the areal settlement by the shape of the alphabet [30]. These graph convolutional–neural–network-based areal settlement recognition algorithms can help accurately distinguish areal settlement shapes. However, the above method’s data source is contour-based, which gives complete shape information on the shape of the areal settlement but has a redundant number of nodes compared to the skeleton lines [31]. Moreover, it is not very effective at distinguishing complex shapes of areal settlement. Inspired by these studies, this paper uses skeleton line data to better balance the characteristics of the overall alignment and local morphological features of a large-scale areal settlement. The graph structure is constructed using one-dimensional skeleton line data, and the low-dimensional shape information represented by the skeleton lines is aggregated and finally classified by a graph convolutional neural network through the extraction of skeleton line features. This solves the problem of classifying the shapes of complex-shaped large-scale areal settlements.

2. Methodology

2.1. Framework

In this paper, the data source is the areal settlement element. First, the skeleton line of the element is extracted and created as a graph data structure. Then, the graphs are annotated according to the shape categories of the residential elements. Next, the high-dimensional structural features of the residential elements are extracted by means of supervised graph learning using a graph convolutional neural network. Finally, a classifier classifies the graph data, thus classifying the areal settlement shapes. This is divided into four processes, and the framework is shown in Figure 1.
(1)
Data preprocessing: Areal settlement vector data were collected from the OSM dataset. Skeleton lines were constructed on the basis of the vector data of the areal settlement.
(2)
Graph representation: A component dual-graph structure was constructed on the basis of the skeleton line graph structure.
(3)
Feature extraction: Features of local and global structures representing the skeletal line dual graph of the areal settlement structure were extracted.
(4)
GCN construction: The model architecture based on the GCN was designed. Then, the GCN was trained and tested, Finally, a fully connected layer as a classifier completed the Areal settlements’ shape classification.

2.2. Data Preprocessing

The data were obtained from the OpenStreetMap open-source dataset and divided into nine alphabetical shapes according to the Gestalt Principle. including E-type, F-type, I-type, and Y-type [32], as shown in Figure 2. Nine types of areal settlement were selected manually.
The first step in shape classification using graph convolutional neural networks is to construct the areal settlement entities as graph data structures to create a graph representation of the vector shape of the areal settlement. We use the skeleton line of the areal settlement entity to describe its shape. The Delaunay triangulation of the areal settlement is shown in Figure 3a, and the shape center of the Type III triangle (red triangular) is used as the end point of the skeleton line [33]. The end points are connected in sequence to form the skeleton line, as shown in Figure 3b.

2.3. Graph Representation

The nodes of a graph neural network generally represent individuals in the network structure, e.g., people in a social network and functions in a software-structured network, and the edges represent the relationships between individuals [34]. Specifically for the areal settlement skeleton line structure, although its natural structure can be regarded as a graph structure, the relative location information within the areal settlement is embedded in the skeleton line and its connection relationship. Therefore, this paper takes the polygonal skeleton line segments of the areal settlement as the minimum shape unit, while using them as nodes of the graph, and the connection relationship between the line segments as edges to construct a graph data structure. As shown in Figure 4, the edge AB has a node D and the line segment DE is an edge.

2.4. Feature Extraction

The node attribute feature extraction directly affects the shape representation of the whole graph. The skeletal lines of the residential area are used to construct the graph, mainly by describing the geometric features of the one-dimensional skeletal line combination and the topological relationships between the line segments, to extract their features of local and global structures and thus achieve an overall overview of the areal settlement.

2.4.1. Local Features Extraction

Large-scale areal settlements are simple two-dimensional surface shapes. The essential local information is the length and the direction of the extension. Therefore, we extract the skeleton line length and azimuth as local feature. The length of the skeleton line segment AB is L1; the azimuth of the skeleton line segment is α as shown in Figure 5.
The length of the line segment consisting of the azimuth angle α denotes the direction of the line segment AB in the two-dimensional plane. In this study. the angle from the northern end of the line due north, turned clockwise to the target line, is noted as the azimuth of the line and takes values in the range 0.360°.
The shape features must be normalized as they should remain constant under translation, rotation, and reduction. The length of the skeleton line segment L1 is normalized by dividing it by the sum of the total lengths of the skeleton lines D. The azimuth angle α is normalized by dividing it by 360° according to the range of values of the azimuth angle α, as follows:
L 1 ¯ = L 1 D
α ¯ = α 360

2.4.2. Global Features Extraction

Since the skeleton lines represent the main and convex features of the areal settlement, the features between adjacent skeleton lines in residential areas cannot be ignored and offer great advantages in the classification of complex shapes. As shown in Figure 6a, the midpoint D of the boundary line segment AB is connected to the center point of the areal settlement. The global features are extracted as follows: The length of the line OD is L 2 . The turning angle of the line segment OD to OE is β . It represents the angle from OD to OE, which is positive in the counterclockwise direction and negative otherwise. Area S1 is denoted by the line connecting the center point of the areal settlement to the points on either side of the skeleton line segment (Figure 6b).
The global features need to be normalized. The length L 2 is normalized by dividing the sum of L sum the angle β is normalized by dividing 360°, and the area S 1 by dividing the sum of the local area S sum , as follows:
L 2 ¯ = L 2 L s u m
β ¯ = β 360
S ¯ = S 1 S s u m

3. Structure of the GCN-Based Residential Classification Model

After gaining the feature vectors for the training set, the validation set, and the test set, the data are inserted into a graph convolutional neural network. When constructing an areal settlement into a graph structure, the skeleton lines of the areal settlement need to be mapped to the graph structure. In this research, the skeleton line characteristics of the surface settlement morphology are embedded in the nodes of the graph and the main morphology of the areal settlement is expressed by the graph structure of the skeleton line. The attributes of the nodes in the graph carry local and overall morphological features of the surface settlement, which are extracted by different methods because the residential areas have different shapes. Thus, when performing convolutional operations on the graph, features of the shape can be extracted from different aspects, and in classification prediction, the shape features are aggregated by a fully connected neural network and their category is predicted according to the obtained probability values of the shape.
Figure 7 displays the GCN architecture used in this study. In the forward propagation process, the initial input signal is the feature matrix X 0 R N × D of the skeleton line, where N is the number of skeleton line nodes and D is the number of skeleton line features. Deep information about the areal settlement is extracted through graph convolution and activation calculation of hidden layers. The number of neurons in each hidden layer is I, that is, the number of convolution kernels. Finally, the probability Y R 1 × 9 of the nine skeleton lines is output in the L + 1 t h layer, or the fully connected layer. The graph convolution and activation calculation formula of the GCN hidden layer is as follows:
H j l + 1 = R e L U i = 1 I D ^ 1 2   A ^   D ^ 1 2   H i l W l + 1 + b j l + 1
where R e L U · represents a rectified linear function, H i l is the output of the i t h neuron in the l t h layer, and H j l + 1 is the output of the j t h neuron in the L + 1 t h layer. H 0 is set as X 0 . W l + 1 R I l × I l + 1 and b j l + 1 R I × I l + 1 are the trainable weight matrix and bias, respectively, which are the training parameters of the GCN. Furthermore, A   ^ = I N + A is the adjacency matrix and D   ^ is the degree matrix of A   ^ .
After the graph convolution operator is applied to L hidden layers, the output of the L + 1 t h layer is H L + 1 R 1 × 9 , or Y. The L + 1 t h layer is the fully connected layer, which is calculated as follows:
Y = S o f t m a x r e d u c e _ m e a n H L W L + 1 + b L + 1
where W L + 1 R I L × 9 and b L + 1 R 1 × 9 represent the learning weights and bias, respectively, of the L + 1 t h layer and H L W L + 1 + b L + 1 R N × 9 calculates the probability of a different classification at each node of the graph. The final classification is judged by a global probability value. Therefore, the r e d u c _ m e a n · function is used to calculate the mean along the first dimension (node dimension). Then, the normalized exponential function S o f t m a x · is used to obtain the output probability Y of the nine classifications. In backpropagation, to maximize the efficiency of the model, the cross-entropy function is used to calculate the model loss, and the Adam algorithm is selected for the optimization.

4. Experiments and Analysis

4.1. Experimental Environment and Data

The experiments were implemented in Python 3.7 and Pytorch 1.6. The platform used was Microsoft Win10 64-bit operating system, CPU Intel(R) Core(TM) i7-3740QM at 2.70 GHz, 8 GB of RAM, and 500 GB of hard disk.
In this experiment, there are 200 of each areal settlement type and a total of 1800 samples of 9 shapes are selected. Firstly, 120 samples are randomly selected from each of the 9 types, i.e., a total of 1080 samples are used as the training set; 40 areal settlement data are randomly selected from each category in the remaining 720 samples; a total of 360 samples are used as the validation set; and the remaining 360 samples are used as the test set. Due to the limitations of the number of datasets and the effect of a more accurate response model, the training set, the validation set, and the test set are divided into a ratio of 6:2:2.
In this experiment, the structure of the constructed skeleton line graph is labeled according to the shape of the areal settlement and a dataset with labels is obtained. For feature extraction of the nodes in the graph, the attributes of the corresponding nodes are extracted according to the method described in Section 2.4, and the extracted attributes are assigned to the corresponding nodes to complete the pre-processing of the graph data.

4.2. Model Structure and Parameter Settings

This experiment uses a 5-layer model with 4 convolutional layers and 1 linear classification layer. Model parameters are shown in Figure 8. The graph is constructed using the method proposed in this paper for constructing residential graph data and extracting node attributes. The constructed graph with labels is input into the model in batches, the weight parameters are initialized, the model parameters are learned, and the weight parameters are updated by back propagation, with a maximum of 350 iterations. After experimentation, the model with a learning rate of 0.005 and a training batch of 10 is selected. To prevent overfitting, training is stopped when the number of successive iterations is 100 and the accuracy of the validation set is no longer increasing.
In the first graph input layer, the input feature dimension is 5 and the embedding vector dimension is 64. After the graph convolution operation, the output value of the first layer is passed to the next layer and the ReLU function is used as the activation function. Layers 2 to 4 of the graph convolutional layers are used as the hidden layers of the graph convolutional neural network. The input and output vectors of the model have a dimension of 64, and the activation function is a ReLU function. The fifth layer is a linear classification layer, which uses a fully connected neural network and a softmax function as the classification layer to classify the output of the hidden layer through the classification layer. The output is the predicted shape classes of the areal settlement, corresponding to the shape classes in the dataset.
Figure 9 presents the validation accuracy and loss of the training and validation sets in this experiment. From the results, it can be concluded that the loss value decreased to 0.069 at the end of the training and the accuracy value of the validation set was 95.3%, while the model was able to converge well after 250 iterations. From the accuracy value of 95.3%, it can be concluded that the model has good generalization ability, has good sensitivity to the skeleton line of the face areal settlement, and can effectively classify the shape of the areal settlement.

4.3. Sensitivity Analysis of Model Parameters

In the training of neural networks, the training batch size has an impact on the performance of the model. The decrease in each gradient is determined by the overall sample data for each training batch. In the experiments, the accuracy of the model training results was further studied by varying the training batch size from 10 to 500, as shown in Figure 10, from which it can be concluded that the accuracy of the model training results did not decrease significantly when the training batch size was increased from 10 to 500. However, the accuracy was the highest when the training batch size was 10, and there was a significant convergence after 300 epoch.
With a learning rate of 0.005 and a training batch size of 10, keeping other conditions constant, this experiment investigated the effect of the number of layers of the model and the dimensionality of the shape-embedding vector on the performance of the model. From Figure 11, it can be concluded that the accuracy of the model classification could be improved by increasing the depth of the model and the dimension of the shape-embedding vector under certain conditions, and it can be seen from the figure that the highest accuracy can be obtained by choosing a model depth of 4 and an embedding vector dimension of 64 as parameters.
After the experimental study to derive the hyperparameters when the model is optimal, the effect of model depth on model performance is further investigated. The optimal hyperparameters of the model were identified when the other hyperparameters of the model were certain, such as a training batch of 10, an embedding vector dimension of 64, a learning rate of 0.005, and a maximum of 350 iterations.

4.4. Classification Accuracy Analysis

To test the effectiveness of the areal settlement shape classification model, it is necessary to give a test set, use the model to predict each sample in the test set, and calculate an evaluation score on the basis of the results of the model’s predicted classification. For classification problems, common evaluation criteria include accuracy, precision, recall, and F-measure values [26]. This paper uses accuracy to evaluate how well the model classifies all classes as a whole, and for each class, the performance estimates are evaluated using precision and recall. The accuracy of a class is the proportion of all samples predicted to be in that class that are predicted correctly. The recall of a class is the proportion of all samples truthfully labeled as being in that class that are predicted correctly. The result is shown in Table 1. The F-measure is the summed mean of the accuracy and recall and represents a metric given when both are equally important.
Accuracy   rate :   P = N u m b e r   o f   C o r r e c t   C l a s s i f i c a t i o n s   N u m b e r   o f   c l a s s i f i c a t i o n s   i n   t h e   M o d e l
Recall   rate :   R = N u m b e r   o f   c o r r e c t   c l a s s i f i c a t i o n s A r t i f i c i a l   n u m b e r   o f   c l a s s i f i c a t i o n s
Measurement   values :   F = 2 × P × R P + R

4.5. Comparison to Other Methods

To compare the differences between the skeleton-line-based graph convolutional neural network method and other similar methods, this paper conducted comparative experiments using the contour-line-based graph convolutional neural network method [29] and the turning angle function method.
On using the contour-line-based graphical convolutional neural network method, four features are extracted from the nodes: 1.the length of contour line AB is L 1 , 2.the azimuth of contoured line segments is α (Figure 12a), 3.the length of the line segment OP is L 2 , and 4.the angle of the turn of the line OP to OQ is β (Figure 12b). The input feature dimension of the model is 128 × 4. A four-layer hidden layer structure with 128 hidden layers is used to finally obtain 128 × 9 dimensional embedding vectors, and the training results of the model are shown in Figure 13.
In the comparison experiment of the turning angle function, the principle is shown in Equation (11). For polygon A, the tangent angle of any point O on its boundary along the reference direction (e.g., x-axis) is v. The turning angle function f(s) means the change in the tangent angle along the boundary of A is inverse or counterclockwise relative to the arc length s. A total of 120 × 9 = 1080 experimental data in 9 classes were subjected to accuracy calculations, yielding a matching accuracy of 0.438.
L P A , B = | f A s f B s | P = 0 1 | f A s f B s | P d s 1 p
A comparison of the experimental results of the skeleton-line-based graph convolutional neural network, the contour-line-based graph convolutional neural network, and the turning angle function is shown in Table 2.
The main reason for this discrepancy is that the skeleton line accurately portrays the overall orientation and convexity of the settlement, whereas the contour line cannot accurately portray the convexity of the settlement. The features of the skeleton line achieve an apparent classification effect after dimensional expansion and multilayer convolution of the graph convolutional neural network. In contrast, the classification of the graph convolutional neural network is not affected by spatial factors such as the starting point and the starting angle of the areal settlement, so the accuracy of the graph convolutional neural network based on skeleton line is higher than that of the turning angle function.

4.6. Applications

To verify the usability of the model, the skeleton-line-based shape classification model for residential maps in this study was applied to identify the shape of some areal settlements in Lanzhou City, with experimental data from 1:5000 Lanzhou City building data. In the vector graph of the areal settlement in Lanzhou City, this paper selected some areal settlements as the test objects, such as the yellow part in Figure 12. The test sample dataset contained 240 samples, the selected areal settlement was manually annotated, and the trained model was used to classify this dataset. The classification result accuracy reaches about 93.3%, the accuracy rate is 93.3%, and the recall rate is 93.3%. The model classification results are shown in Table 3. As can be seen from Figure 14, in practice, buildings are mostly rectangular in shape, of which Type F and Type E are well classified and can all be recognized.
On combining the data from the training phase in the previous section, the classification results of the model in the experimental set are slightly lower than those in the test set, but the accuracy is still strong. In distinguishing complex shapes, such as E, F, and H shapes, the measured values are mainly concentrated between 95% and 100%. The main reason is that the skeleton line is a better representation of the main body and convexity of the settlement than the areal settlements, thus providing a more accurate description of the local features of a complex-shaped settlement. However, for the T- and Y-shaped areal settlements, with similar skeleton line structures, the measured values are 75.9% and 66.7%, respectively. In the next step of the study, the training set and the experimental dataset will be further increased in order to reflect the accuracy of the model classification more objectively.

5. Conclusions

Areal settlements’ shape classification is a classical and challenging issue in large-scale topographic maps. This study proposed a graph convolutional neural network shape classification method based on the skeleton lines of the areal settlement. The method uses a skeleton line of areal settlement elements as the basis for constructing the calculation graph. The geometric features of the skeleton line segments are extracted as the nodal attribute features of the computational graph, using the skeleton line segments as the smallest unit to express the shape features of the faceted areal settlement in a graph structure. To achieve shape classification of an areal settlement, a graph with labels and node attributes is used as input for the GCN classification mode, and the graph convolutional neural network is used to discriminate between high-dimensional shape features through supervised training. Lastly, a fully connected classifier can accurately classify the areal settlement’s shape. The experimental results show that the method has a high discriminative ability for areal settlement shapes.
The method uses areal settlement skeleton lines as input data to construct an end-to-end classification model. Validation with real data sets has demonstrated that the skeleton line can effectively represent both local and global features and has a high discriminative ability for the shape of alphabetical areal settlement. The distinction of complex shapes of areal settlement such as F- and E- and H-types is particularly effective and outperforms other existing methods. However, the objects of the areal settlement in this method are large-scale. Shape variation in small- and medium-scale areal settlements is one of the main areas to be improved and is proposed to be studied in depth in combination with the method of graph convolutional neural networks.

Author Contributions

Y.L. and X.L. conceived the skeleton-line and Graph Convolutional Neural Network together; Y.L. collected the data, designed the experiments, and wrote the manuscript; W.W. and P.L. offered significant contribution to result evaluation; H.Y. modified the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 41930101, 42161066), the Industrial Support and the Program Project of Universities in Gansu Province (grant no.2022CYZC-30).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhou, X.; Wang, H.; Wu, Z. An incremental updating method for land cover database using refined 2-dimensional intersection type. Acta Geod. Cartogr. Sin. 2017, 46, 114–122. [Google Scholar]
  2. Kim, J.-Y.; Yu, K.-Y. Automatic Detection of the Updating Object by Areal Feature Matching Based on Shape Similarity. J. Korean Soc. Surv. Geod. Photogramm. Cartogr. 2012, 30, 59–65. [Google Scholar] [CrossRef] [Green Version]
  3. Zhang, X.; Stoter, J.; Ai, T. Automated evaluation of building alignments in generalized maps. Int. J. Geogr. Inf. Sci. 2013, 27, 1550–1571. [Google Scholar] [CrossRef]
  4. Stoter, J.; Burghardt, D.; Duchêne, C. Methodology for evaluating automated map generalization in commercial software. Comput. Environ. Urban Syst. 2009, 33, 311–324. [Google Scholar] [CrossRef] [Green Version]
  5. Harrie, L.; Stigmar, H.; Djordjevic, M. Analytical estimation of map readability. ISPRS Int. J. Geo-Inf. 2015, 4, 418–446. [Google Scholar] [CrossRef] [Green Version]
  6. Palmer, S.E. Vision Science: Photons to Phenomenology; MIT Press: Cambridge, MA, USA, 1999. [Google Scholar]
  7. Liu, Y.K.; Žalik, B. An efficient chain code with Huffman coding. Pattern Recognit. 2005, 38, 553–557. [Google Scholar]
  8. Belongie, S.; Malik, J.; Puzicha, J. Shape matching and object recognition using shape contexts. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 509–522. [Google Scholar] [CrossRef] [Green Version]
  9. Peter, A.M.; Rangarajan, A. Maximum likelihood wavelet density estimation with applications to image and shape matching. IEEE Trans. Image Process. 2008, 17, 458–468. [Google Scholar] [CrossRef] [Green Version]
  10. Saavedra, J.M. Sketch based image retrieval using a soft computation of the histogram of edge local orientations (s-helo). In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 2998–3002. [Google Scholar]
  11. Yan, X.; Ai, T.; Yang, M. A Simplification of Residential Feature by the Shape Cognition and Template Matching Method. Acta Geod. Cartogr. Sin. 2016, 45, 874–882. [Google Scholar] [CrossRef]
  12. Ai, T.; Cheng, X.; Liu, P.; Yang, M. A shape analysis and template matching of building features by the Fourier transform method. Comput. Environ. Urban Syst. 2013, 41, 219–233. [Google Scholar] [CrossRef]
  13. Alajlan, N.; El Rube, I.; Kamel, M.S.; Freeman, G. Shape retrieval using triangle-area representation and dynamic space warping. Pattern Recognit. 2007, 40, 1911–1920. [Google Scholar] [CrossRef]
  14. Yang, C.; Wei, H.; Yu, Q. A novel method for 2D nonrigid partial shape matching. Neurocomputing 2018, 275, 1160–1176. [Google Scholar] [CrossRef]
  15. Cheng, M.; Sun, Q.; Xu, L.; Cheng, H. Polygon contour similarity and complexity measurement and application in simplification. Acta Geod. Cartogr. Sin. 2019, 48, 489–501. [Google Scholar] [CrossRef]
  16. Mokhtarian, F.; Mackworth, A.K. A theory of multiscale, curvature-based shape representation for planar curves. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 789–805. [Google Scholar] [CrossRef]
  17. Arkin, E.M.; Chew, L.P.; Huttenlocher, D.P.; Kedem, K.; Mitchell, J.S. An Efficiently Computable Metric for Comparing Polygonal Shapes; Cornell University: Ithaca, NY, USA, 1991. [Google Scholar]
  18. Basaraner, M.; Cetinkaya, S. Performance of shape indices and classification schemes for characterising perceptual shape complexity of building footprints in GIS. Int. J. Geogr. Inf. Sci. 2017, 31, 1952–1977. [Google Scholar] [CrossRef]
  19. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 84–90. [Google Scholar] [CrossRef] [Green Version]
  20. Yao, L.; Mao, C.; Luo, Y. Graph convolutional networks for text classification. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 7370–7377. [Google Scholar]
  21. Abdel-Hamid, O.; Mohamed, A.; Jiang, H.; Deng, L.; Penn, G.; Yu, D. Convolutional neural networks for speech recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 1533–1545. [Google Scholar] [CrossRef] [Green Version]
  22. Zhang, H.; Li, C.; Wu, P.; Yin, Y.; Guo, M. A complex junction recognition method based on GoogLeNet model. Trans. GIS 2020, 24, 1756–1778. [Google Scholar]
  23. He, H.; Qian, H.; Xie, L.; Duan, P. Interchange Recognition Method Based on CNN. Acta Geod. Cartogr. Sin. 2018, 47, 385–395. [Google Scholar] [CrossRef]
  24. Yao, D.; Zhang, C.; Zhu, Z.; Hu, Q.; Wang, Z.; Huang, J.; Bi, J. Learning deep representation for trajectory clustering. Expert Syst. 2018, 35, e12252. [Google Scholar] [CrossRef]
  25. Feng, M.; Meunier, J. Skeleton Graph-Neural-Network-Based Human Action Recognition: A Survey. Sensors 2022, 22, 2091. [Google Scholar] [CrossRef] [PubMed]
  26. Yu, H.; Ai, T.; Yang, M.; Huang, L.; Yuan, J. A recognition method for drainage patterns using a graph convolutional network. Int. J. Appl. Earth Obs. Geoinf. 2022, 107, 102696. [Google Scholar] [CrossRef]
  27. Yan, X.; Ai, T.; Yang, M.; Tong, X.; Liu, Q. A graph deep learning approach for urban building grouping. Geocarto Int. 2022, 37, 2944–2966. [Google Scholar] [CrossRef]
  28. Yan, X.; Ai, T.; Yang, M.; Yin, H. A graph convolutional neural network for classification of building patterns using spatial vector data. ISPRS J. Photogramm. Remote Sens. 2019, 150, 259–273. [Google Scholar] [CrossRef]
  29. Yan, X.; Ai, T.; Yang, M.; Tong, X. Graph convolutional autoencoder model for the shape coding and cognition of buildings in maps. Int. J. Geogr. Inf. Sci. 2021, 35, 490–512. [Google Scholar] [CrossRef]
  30. Yu, Y.; He, K.; Wu, F.; Xu, K. Graph Convolution Neural Network Method for Shape Classification of Areal Settlements. Acta Geod. Et Cartogr. Sin. 2021, 65, 1–20. Available online: http://kns.cnki.net/kcms/detail/11.2089.P.20211208.2334.008.html (accessed on 21 August 2022).
  31. Chen, G.; Qian, H. Extracting Skeleton Lines from Building Footprints by Integration of Vector and Raster Data. ISPRS Int. J. Geo-Inf. 2022, 11, 480. [Google Scholar] [CrossRef]
  32. Yan, X.; Ai, T.; Zhang, X. Template matching and simplification method for building features based on shape cognition. ISPRS Int. J. Geo-Inf. 2017, 6, 250. [Google Scholar] [CrossRef] [Green Version]
  33. Luo, D.; Qian, H.; He, H.; Duan, P.; Guo, X.; Zhong, J. A Method of Extracting Multi-Scale Skeleton Lines for Polygon Buildings. J. Geomat. Sci. Technol. 2019, 3603, 324–330. [Google Scholar]
  34. Zhao, X.; Wang, S.; Wang, H. Organizational Geosocial Network: A Graph Machine Learning Approach Integrating Geographic and Public Policy Information for Studying the Development of Social Organizations in China. ISPRS Int. J. Geo-Inf. 2022, 11, 318. [Google Scholar] [CrossRef]
Figure 1. Framework.
Figure 1. Framework.
Applsci 12 10001 g001
Figure 2. The nine shape classes of areal settlements.
Figure 2. The nine shape classes of areal settlements.
Applsci 12 10001 g002
Figure 3. Extraction of areal settlement skeleton lines: (a) Delaunay triangulation; (b) The skeleton line of the areal settlement.
Figure 3. Extraction of areal settlement skeleton lines: (a) Delaunay triangulation; (b) The skeleton line of the areal settlement.
Applsci 12 10001 g003
Figure 4. Dual graph of skeleton lines.
Figure 4. Dual graph of skeleton lines.
Applsci 12 10001 g004
Figure 5. Local features extraction of node D.
Figure 5. Local features extraction of node D.
Applsci 12 10001 g005
Figure 6. Global features extraction of node D: (a) length and turning angle; (b) local area.
Figure 6. Global features extraction of node D: (a) length and turning angle; (b) local area.
Applsci 12 10001 g006
Figure 7. Architecture of the GCN.
Figure 7. Architecture of the GCN.
Applsci 12 10001 g007
Figure 8. Parameters of the areal settlement classification model.
Figure 8. Parameters of the areal settlement classification model.
Applsci 12 10001 g008
Figure 9. Changing curve of loss and validation accuracy in GCN model training.
Figure 9. Changing curve of loss and validation accuracy in GCN model training.
Applsci 12 10001 g009
Figure 10. The validation accuracy of the model with training batches of different sizes; (a) batch 10, accuracy 0.9361, (b) batch 50, accuracy 0.9278, (c) batch 100, accuracy, 0.9222, and (d) batch 500, accuracy 0.9028.
Figure 10. The validation accuracy of the model with training batches of different sizes; (a) batch 10, accuracy 0.9361, (b) batch 50, accuracy 0.9278, (c) batch 100, accuracy, 0.9222, and (d) batch 500, accuracy 0.9028.
Applsci 12 10001 g010aApplsci 12 10001 g010b
Figure 11. Test of the number of hidden layers and vector dimension.
Figure 11. Test of the number of hidden layers and vector dimension.
Applsci 12 10001 g011
Figure 12. Feature extraction; (a) local feature extraction of node P; (b) global feature extraction of node P.
Figure 12. Feature extraction; (a) local feature extraction of node P; (b) global feature extraction of node P.
Applsci 12 10001 g012
Figure 13. Changing curve of loss and validation accuracy in GCN model training.
Figure 13. Changing curve of loss and validation accuracy in GCN model training.
Applsci 12 10001 g013
Figure 14. Experimental data of Lanzhou City.
Figure 14. Experimental data of Lanzhou City.
Applsci 12 10001 g014
Table 1. The results of the areal settlement classification test.
Table 1. The results of the areal settlement classification test.
ClassArtificial Number of ClassificationsNumber of Classifications in the ModelNumber of Correct ClassificationsP/(%)R/(%)F
I-shape404040100100100
L-shape40403792.592.592.5
T-shape40443988.697.592.8
U-shape40393794.992.593.7
Z-shape40393910097.598.7
E-shape40433888.49591.6
Y-shape40403997.597.597.5
F-shape40363597.287.592.1
H-shape40393997.597.597.5
Total36036034395.395.395.3
Table 2. Comparison of methods.
Table 2. Comparison of methods.
MethodsGCN Based on Skeleton LinesGCN Based on Contour LinesTurning Function
Accuracy rate95.3%64.4%47.8%
Table 3. Classified statistics of some settlements in Lanzhou.
Table 3. Classified statistics of some settlements in Lanzhou.
ClassArtificial Number of Classifications Number of Classifications in the ModelNumber of Correct ClassificationsP/(%)R/(%)F
I-shape88848410095.597.7
L-shape40424095.210097.5
T-shape17121191.764.775.9
U-shape38383694.794.794.7
Z-shape914857.188.969.5
E-shape15141410093.396.5
Y-shape69555.683.366.7
F-shape13131292.392.392.3
H-shape141414100100100
total24024022493.393.393.3
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Y.; Lu, X.; Yan, H.; Wang, W.; Li, P. A Skeleton-Line-Based Graph Convolutional Neural Network for Areal Settlements’ Shape Classification. Appl. Sci. 2022, 12, 10001. https://doi.org/10.3390/app121910001

AMA Style

Li Y, Lu X, Yan H, Wang W, Li P. A Skeleton-Line-Based Graph Convolutional Neural Network for Areal Settlements’ Shape Classification. Applied Sciences. 2022; 12(19):10001. https://doi.org/10.3390/app121910001

Chicago/Turabian Style

Li, Yiyan, Xiaomin Lu, Haowen Yan, Wenning Wang, and Pengbo Li. 2022. "A Skeleton-Line-Based Graph Convolutional Neural Network for Areal Settlements’ Shape Classification" Applied Sciences 12, no. 19: 10001. https://doi.org/10.3390/app121910001

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop