Next Article in Journal
Analyzing Thermal Environment Contributions and Driving Factors of LST Heterogeneity in Different Urban Development Zones
Previous Article in Journal
Pose Estimation Based on Bidirectional Visual–Inertial Odometry with 3D LiDAR (BV-LIO)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cartographic Generalization of Islands Using Remote Sensing Images for Multiscale Representation

School of Geospatial Engineering and Science, Sun Yat-sen University, Zhuhai 519082, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2024, 16(16), 2971; https://doi.org/10.3390/rs16162971
Submission received: 30 June 2024 / Revised: 12 August 2024 / Accepted: 12 August 2024 / Published: 14 August 2024

Abstract

:
The multi-scale representation of remote sensing images provides various levels of image information crucial for decision-making in GIS applications and plays a significant role in information processing, data analysis, and geographic modeling. Traditional methods for multi-scale representation of remote sensing images often struggle to simplify local details of individual targets while preserving the overall characteristics of target groups. These methods also encounter issues such as transitional texture distortion and rough final boundaries. This paper proposes a novel multi-scale representation method for remote sensing images based on computer vision techniques, which effectively maintains the overall characteristics of target groups. Initially, the K-means algorithm is employed to distinguish between islands and oceans. Subsequently, a superpixel segmentation algorithm is used to aggregate island groups and simplify the generated boundaries. Finally, texture synthesis and transfer are applied based on the original image to produce the aggregated island images. Evaluation metrics demonstrate that this method can generate multi-scale aggregated images of islands, effectively eliminate redundant information, and produce smooth boundaries.

1. Introduction

As a crucial carrier of geographic information, remote sensing imagery’s multi-scale representation has been a key focus in geographic information science research. Multi-scale representation refers to providing data processing results at various scales according to different user requirements for features, locations, attributes, and relationships of geographic information. Currently, the main methods for multi-scale representation include image pyramids and filtering transformations. Image pyramids achieve multi-resolution representation by continuously downsampling to obtain images with varying levels of detail [1]. Specifically, images at different levels are derived from the original image by reducing its resolution. Due to their simple structure, significant enhancement of computational efficiency, and noise removal capabilities, image pyramids are widely used in fields such as feature extraction [2], optical flow tracking, and image enhancement [3].
Image filtering is one of the core operations in the field of computer vision. As another technique for multi-scale image representation, its primary function is to preserve the overall characteristics of target groups while removing unnecessary local details and other redundant information. Due to its similarity to the core objectives of cartographic generalization, image filtering will be utilized as an experimental method in this study. Its effects will be compared with other techniques such as multi-scale aggregation and simplification. Image filtering is widely used in areas such as noise reduction, edge detection, and image sharpening. Buades et al. (2005) [4] proposed a non-local means filtering method, which effectively achieves noise reduction while preserving the original features of the image by associating the similarity between pixels with weights. Additionally, there are mean filtering techniques (linear filtering) and median filtering techniques (non-linear filtering). Although these methods can smooth boundaries, they also tend to blur the image content, resulting in a loss of information.
Islands constitute the primary focus of this study. In geological terms, islands are defined as landmasses surrounded by water, typically with an area of no less than 500 square meters. They play significant roles in natural ecosystems by contributing to the formation of features such as atolls, coastlines, and volcanic island chains, while also serving as abundant resources for local fisheries. Moreover, islands are essential components in both cartographic mapping and remote sensing applications. Given their distinctive characteristics compared to mainland features, islands necessitate specialized treatment in cartographic generalization processes. For instance, in the cartographic generalization of sparsely distributed islands, the large disparity in area between the sea and islands means that islands displayed at the same scale can appear overly detailed. Therefore, it is necessary to employ the principle of variable scaling to separately determine the comprehensive scales for islands and marine areas [5].
The aggregation of planar elements is a fundamental aspect of cartographic generalization. With the development of GIS and the continuous accumulation of geographic spatial data, conflicts between different targets inevitably arise at the same mapping scale. Therefore, a series of operations are required to remove redundant local information while maintaining overall characteristics. Traditional vector-based methods, such as boundary extraction in ArcGIS, aggregate planar features but often lose original textures, resulting in geospatial information loss and lack of scale diversity. Raster-based aggregating methods, like filtering, cannot eliminate gaps between islands, failing to achieve global simplification. To address these challenges, this paper proposes a new approach for the multi-scale representation of remote sensing imagery, aiming to achieve a balance between simplification efficiency and visualization efficacy. This method ensures the preservation of overall target group characteristics while generating aggregated outputs at varying scales through the adjustment of parameters.
The structure of this paper is as follows: The first section provides a comprehensive review of relevant research on aggregation and simplification algorithms. The second section details the procedural steps of the proposed method, illustrated using small island samples: Firstly, the Kmeans algorithm from computer vision theory is used to separate islands and seawater based on color. Secondly, superpixel segmentation is employed to aggregate discrete islands and smooth the generated boundaries. Finally, texture synthesis and migration are conducted using the original image. The third section presents selected experimental data, experimental methodologies, results, and evaluation of metrics. The fourth section concludes with a summary of the experiment’s findings and a discussion of its limitations.

2. Related Works

The process of cartographic generalization involves the abstraction and generalization of geospatial information in accordance with specific conditions, thereby reflecting geographical laws. The objective of cartographic generalization is to reconcile the disparity between map models and complex realities, facilitating the conversion of data map content into novel representations.
With the advancement of computer technology, automatic cartographic generalization technology is also progressing rapidly. This progress is attributed to the collaborative efforts of numerous cartographic generalization operators specialized in different tasks. The subprocess of complex cartographic generalization, known as the cartographic generalization operator, independently describes the transformation process of map elements. Various models for generalization operators have been developed in cartographic generalization, including the seven-operator generalization model [6] and nine-operator generalization model [7], among others. These models encompass essential algorithmic operations such as simplification, selection, shifting, merging and smoothing that form the foundation for cartographic generalization algorithms. In this paper, two specific types of cartographic generalization operators are comprehensively described.

2.1. Aggregation

Due to the ongoing decrease in map scale, both the dimensions of the depicted targets and the gaps between these illustrations are progressively shrinking. As the spacing between these graphics becomes excessively congested, rendering them indistinguishable, there arises a necessity to substitute the smaller graphics with a singular, enlarged representation. This process is referred to as aggregation.
Elements that are physically adjacent and have similar properties are perceived as a whole [8]. During the merging process, the target element loses some or even most of its typical characteristics [9]. Therefore, it is important to preserve the global characteristics of objects before and after merging.
The target elements of aggregation operators are typically categorized into artificial architectural elements and natural planar elements. Research on the aggregation of artificial planar elements primarily relies on raster data and vector data, with vector data being the predominant one. Delaunay triangulation skeletons are extensively employed for merging vector data. Ai and Zhang (2007) [10] present a cluster distribution analysis model based on the Delaunay triangulation skeleton, which utilizes a specialized geometric structure resembling a Voronoi diagram to calculate the structure variables of each partition while considering conflict detection, displacement, and the geometry’s impact on merging. Guo et al. (2016) [11] present an approach to automatically identify right angles and aggregate buildings based on six parameters in Delaunay triangulation. Zhang et al. (2023) [12] introduced a polygonal building aggregation method based on the bridging modes and minimum loop search, which retains the characteristic of right angles while demonstrating improved visual clarity in results through validation advantages. Considering urban architecture’s orthogonality, Guo and Ai (2000) [13] proposed a novel method for aggregating polygons with shared arcs using clipped and extended links that overcome drawbacks associated with organizing arc link sequences in traditional aggregation methods without necessitating unnecessary detection of polygon islands. He et al. (2018) [14] introduced a progressive aggregation method for residential areas based on scale subgroups, which effectively maintains orthogonal characteristics and ensures area balance while also considering cartographic bundle reduction.
To address the issue of the aggregation of natural planar elements, an iterative approach is commonly employed to select an object from the target group and merge it with neighboring objects. Li et al. (2018) [15] proposed an automated processing method for agglomeration areas based on parameters such as the distributed pattern index (DPI). This method effectively resolves the problem of subjective judgment dependency in previous agglomeration methods, enabling intelligent computer processing. Subsequently, Li et al. (2021) [9] introduced a novel approach for merging structured geographic objects by classifying and identifying them according to seven spatial structure parameters, while utilizing miter buffer transform to extract their complete boundaries. Compared to traditional methods, this technique ensures balanced area changes across different land cover categories.
However, the current aggregation methods still have notable drawbacks. During the aggregation process, detailed geographic information can be overly simplified or omitted, which can impact the accuracy of geographic analysis. The aggregated data might distort the true situation, leading to misleading representations [16].

2.2. Simplification

Simplification is a crucial operator extensively employed in cartographic generalization. Simplification typically refers to the process of eliminating intricate details from external contours and internal structures of geographical elements while retaining their main features.
Based on shape characteristics, simplification targets can be categorized into two groups: linear elements and planar elements. As a map contains the most abundant elements, linear elements have garnered significant attention from researchers. The challenge lies in removing unnecessary bends while preserving local bending features, domain features, and global features of linear elements.
Vector data-based simplification methods were initially applied to the simplification of linear elements. Douglas and Peucker (1973) [17] proposed a recursive approach that effectively reduces the number of points required for drawing lines during automatic generalization. Ai et al. (2013) [18] proposed a coastline simplification method based on Delaunay triangulation that considers estuarine tree modeling to capture dendrite patterns along with specific morphology and domain constraints related to coastlines. Subsequently, Ai et al. (2016) [19] present a polyline simplification algorithm based on the ε-circle rolling method, which utilizes Delaunay triangulation for detecting bending characteristics, constructing an envelope structure around a polyline, and generating abstract results accordingly. This method inherits Perkal’s algorithm’s scale-specificity while maintaining the primary shape of polylines amidst large-scale changes and effectively avoiding self-intersection issues after linear element simplifications occur. Visvalingam and Whyatt (1993) [20] proposed a progressive linear element simplification method based on calculating “effective area”.
The representation of polygonal boundaries is an approximation method used to simplify linear elements in raster data, but its applicability for cartographic generalization is limited due to the constraints imposed by the raster data type, resulting in a scarcity of related research.
In the field of simplification, planar elements can be categorized into natural and artificial types. Simplifying natural surface elements is relatively straightforward as the curvature of their boundaries allows the application of linear element simplification methods. However, simplifying settlement areas poses additional complexities due to their unique orthogonality. Zhang et al. (2006) [21] proposed a least square adjustment-based method for residential area simplification, while Chen et al. (2011) [22] introduced a constrained Delaunay triangulation nets approach to identify local concave–convex structures within residential areas.
In addition to vector data, scholars have extensively investigated simplification methods based on raster data. The techniques for simplifying residential land primarily encompass mathematical morphology [23,24] and neural network technology. Cheng et al. (2013) [25] proposed a simplified method for locating residents using a backpropagation neural network (BPNN) model. Yang et al. (2022) [26] introduced a hybrid approach to simplify all buildings by employing a backpropagation neural network to construct an evaluator through supervised learning that measures building parameters. This method demonstrates superior performance in terms of corner squareness and preservation of regional architectural characteristics compared to traditional approaches.
However, existing simplification methods still have certain limitations. During the simplification process, errors can be introduced and subsequently propagate through further analyses, potentially leading to inaccurate results [27]. And if the algorithm is not precise enough, it may lead to topological errors, such as line crossings or overlaps, which can undermine the integrity and reliability of the map data [16].

3. Methods

The aggregation of planar elements primarily encompasses object grouping and geometric form integration [10]. Elements adjacent in physical space and possessing similar attributes are perceived as cohesive wholes [8], with target elements often forfeiting a portion or even the majority of their distinctive traits during the aggregation process. Natural planar element aggregation, a subset of planar element aggregation methods, constitutes the principal focus of this study. Initially, islands are delineated, followed by classification and comprehensive simplification through superpixel algorithms to emphasize the boundary-smoothing characteristics essential for the aggregation of natural planar elements. Subsequently, texture synthesis and migration are employed using the Efros-Leung algorithm [28]. The schematic diagram depicting the methodology of this study is presented in Figure 1 below.

3.1. Island-Cover Extraction

The initial step involves utilizing the K-means clustering algorithm for color segmentation to classify islands and water bodies. The K-means algorithm stands as one of the most renowned and commonly employed clustering methodologies. However, it is crucial to note that its utilization entails an initialization phase where the number of clusters must be predetermined, thus rendering it not strictly an unsupervised clustering approach [29]. The fundamental aim of the K-means algorithm is to partition a dataset into K distinct clusters, striving to maximize the intra-cluster similarity while minimizing the inter-cluster similarity. As depicted in Figure 2a, the original image has dimensions of 205 × 190 pixels. By setting the cluster parameter to a value of 2, the classification outcome is illustrated in Figure 2b. Increasing the cluster value leads to a greater variety of features in the classification results. The original image is segmented based on color, resulting in two mask images, as shown in Figure 2c,d.

3.2. Islands Aggregation

The purpose of the superpixel algorithm is to find an over-segmented set of images [30]. Superpixel segmentation is an image segmentation method designed to partition an image into regions with homogeneous characteristics. After the segmentation of planar elements, typically six types of superpixels are generated [31], as depicted in Figure 3.
The recognition rules of six types of superpixels are as follows:
Type 1 These superpixels represent the original island planar elements, displayed in light green. They are treated as baseline pixels;
Type 2 These superpixels, which are solely connected to a single polygon and are first-order neighbors to Type 1 superpixels, are indicated in dark green and treated as type 2;
Type 3 These superpixels, which are linked to two or more polygons and are first-order neighbors to Type 1 superpixels, are shown in yellow and treated as type 3;
Type 4 These superpixels denote internal superpixels, first-order neighbors to Type 1 superpixels, and are connected to only one polygon, depicted in light blue. They are treated as type 4;
Type 5 These superpixels signify internal superpixels distinct from Type 4 superpixels, presented in dark blue. They are treated as type 5;
Type 6 These superpixels encompass all superpixels other than the aforementioned types, represented in white. They are treated as type 6;
Next, we conducted superpixel segmentation on the island planar elements depicted in the sample image and subsequently performed global simplification on the results. Figure 4a,b display the globally simplified images for superpixel counts of N = 1500 and N = 500, respectively. Generally, the complexity of texture and structure within an image correlates with the required number of superpixels. Additionally, median filtering is utilized to process the simplified images. As demonstrated in Figure 4c,d, median filtering effectively smooths the boundaries of aggregated polygons, ensuring a favorable merging effect while preserving the global characteristics of the planar elements and simplifying local details. This highlights the influence of different parameter settings on the aggregating effects of planar elements within the image.

3.3. Texture Migration

Firstly, extract the T-boundary from the aggregated image. Secondly, traverse the image to identify all enclosed white pixels bounded by the T-boundary, forming closed regions. Apply color assignments to these regions, as demonstrated in Figure 5b. Subtract the T-boundary fill image from the classified image yields, as shown in Figure 5c. Overlaying the blue pixels onto the original image, with their RGB values set to white, produces Figure 5d. The white areas are designated for texture filling.
It is apparent that further issues remain to be resolved: the excessive width between the original island boundary and the aggregated boundary due to the superpixel size setting results in overly rough texture filling post-filling. To address this, we employ morphological erosion operations to diminish the boundary width.
As erosion operations are exclusively applicable to binary images, we first undertake color transformation on the T-boundary fill image, as illustrated in Figure 6a. Subsequently, employing experimentation, we ascertain that setting the erosion iterations to 6 yields the most favorable outcome, as evidenced in Figure 6b. The superimposition of the eroded boundary onto the original image, as illustrated in Figure 6c, reveals a noticeable inward contraction of the white area post-erosion, closely aligning with the island edge.
The next task involves filling the vacant areas, necessitating the utilization of texture from the original island planar elements. Here, we employ the Efros–Leung algorithm for texture synthesis: leveraging statistical properties of local regions within the texture, new texture is generated by identifying the most similar local region from the original island planar elements to the area awaiting synthesis. The computational complexity of this method can be expressed as O(n × w2 × h2 × k), where n represents the total number of image pixels, w represents the filter kernel size, h represents the number of times the filter kernel slides, and k represents the number of channels. A small segment of island planar elements near the white area is extracted as a sample, as depicted in Figure 7a; texture synthesis based on the sample image using the Efros–Leung algorithm yields the result shown in Figure 7b; segments representing the aggregated island shape are excavated from the synthesized texture image, as illustrated in Figure 7c; finally, the synthesized texture is applied to the blank areas, resulting in the effect depicted in Figure 7d. It is evident that the filled texture closely resembles that of the original island planar elements, facilitating a smooth transition.
The following illustration, Figure 8, exemplifies the application of this methodology to other island planar elements:

4. Experiments

4.1. Dataset

The original experimental image data are sourced from Google Maps and organized into RGB map tiles using an image pyramid structure. On the server, the tiling strategy manages and loads large-scale geospatial data by dividing it into smaller, manageable chunks [32]. Each map tile is uniquely identified by its level, row number, and column number. Different resolution tiles are retrieved as users browse and zoom the map. In this study, 300 map tiles at level 8 with a scale of 1:611.5 were downloaded and mosaicked to form two regions containing different quantities of polygonal islands, as shown in Figure 9.
The Philippines and Sulawesi, located in Southeast Asia, exhibit a tropical maritime climate with consistently high temperatures, humidity, and significant rainfall. These regions experience distinct wet and dry seasons, influenced by monsoons, and are frequently affected by typhoons. The geography includes mountainous terrain and numerous archipelagic formations.

4.2. Experimental Procedure

Utilizing MATLAB 9.12.0.1927505 (R2022a) for data classification, we implement the K-means clustering algorithm to segment colors, effectively separating the islands and water bodies, as shown in Figure 10a,b. Following this segmentation process, island masks are individually extracted (Figure 10c,d), with the islands distinguished in blue against a white background.
Nonetheless, two outstanding issues remain unaddressed: Firstly, there are voids within the interior of the classified island segments, stemming from the misclassification of shadows by the K-means clustering segmentation algorithm. Secondly, certain small islands from the original dataset persist in the classification output. These scattered islands are predominantly located far from major islands, which could potentially impede subsequent aggregating procedures, resulting in aesthetically and practically compromised mapping outcomes.
For the first issue, no intervention may be necessary as these interior voids will be filled by the synthetic textures during the aggregation process, thereby not affecting the experimental results.
Concerning the second issue, we mitigate it by employing a method involving setting area thresholds to eliminate fragmented islands. Initially, the blue connected components representing islands are labeled, and the area of each connected component is computed. Subsequently, a threshold is established, and connected components with areas smaller than the threshold are deleted.
Then, we recorded the positions of removed connected components, excavated tiny islands on the original image, and filled the resulting holes with a texture resembling seawater.
The figure below, Figure 11, illustrates the aggregation effects of the experimental method at three different scales. This study is capable of producing results at various scales by controlling the number of superpixels. From the figure, it can be observed that this method effectively aggregates multiple polygonal islands while preserving the overall characteristics of the entire area.
Although certain aggregated results effectively preserve the overall characteristics, the resulting boundaries are not smooth and exhibit noticeable jaggedness. Consequently, filtering is introduced to smooth these boundaries. The comparative effects before and after smoothing are shown in Figure 12a,b, respectively. Subsequently, the closed regions enclosed by the T boundaries are filled, as depicted in Figure 12d. The comparison of the final effects of aggregation at different scales is shown in Figure 13.
To better illustrate the characteristics of the aggregating method utilized in this experiment, we employed the ArcGIS aggregating algorithm to process island planar features and conduct a comparative analysis. In Figure 14, the top three images display the aggregated results obtained from ArcGIS, whereas the bottom three present the results produced by our experimental method. It is evident that our method effectively maintains the overall characteristics of the aggregated surface features while simplifying local details and generating smooth aggregation boundaries. Conversely, the ArcGIS results demonstrate insufficient simplification of local details, leading to notably rough aggregation boundaries.
Following this, the filled image is subjected to a binary operation, followed by a suitable erosion process to inwardly contract the aggregated boundary. The resultant image is then superimposed onto the original image, yielding the effect illustrated in Figure 15.
Subsequently, we employ the Efros–Leung algorithm for texture synthesis. Initially, a small patch of island planar elements is sampled from the vicinity of the white areas to be filled. Texture expansion and generation are then conducted based on this sample image. Finally, the synthesized texture is applied to the regions aggregated at different scales, resulting in the observed effect shown in Figure 16 and Figure 17.

4.3. Experimental Results and Evaluation of Metrics

Figure 18 illustrates the aggregated outcomes of island planar elements under varied parameter configurations. The ultimate shape of the aggregated island is governed by two parameters: the count of superpixels, denoted as N, and the filter kernel size, denoted as len. Specifically, the superpixel count predominantly impacts the resultant island’s area, while the filter kernel size modulates the degree of boundary smoothness in the aggregated outcome.
Next, a similarity assessment will be conducted. Median filtering and mean filtering techniques, as classic grid-based data methods, will be applied to the raw images. The processed results will then be compared with those obtained from the proposed method. The fast adaptive bilateral filtering was originally proposed for sharpening and noise removal, now it can also be used for other applications, such as artifact removal and texture filtering [33]. To simulate smooth texture transitions, we employ a combination of deep neural network filtering (CNN filtering) with robust denoising capabilities [34] and Gaussian blur on the original images. These two advanced filters will also serve as comparative methods in our study. The computational complexities of the above four methods are O(n × w2 log(w2)), O(n × w2), O(nlog(n)), and O(n × w2 × k2), respectively, where n represents the total number of image pixels, w represents the filter kernel size, and k represents the number of channels. The results obtained from all methods are shown in Figure 19.
The evaluation of similarity between the processed result image and the original one will be based on five assessment metrics: Mean Squared Error (MSE), Peak Signal-to-Noise Ratio (PSNR), Normalized Mutual Information (NMI), Structural Similarity Index (SSIM), and Spatial Correlation Coefficient (SCC).
Mean Squared Error (MSE) computes the average of the squared pixel-wise differences between the two images, representing a fundamental metric for similarity evaluation. A lower MSE value signifies a higher degree of similarity between the images.
M S E = 1 M N i = 1 M j = 1 N [ x ( i , j ) y ( i , j ) ] 2
The Peak Signal-to-Noise Ratio (PSNR) quantifies the similarity between two images by employing the maximum pixel value, denoted as MAX, and the Mean Squared Error (MSE). Here, ‘inf’ denotes infinity. A higher PSNR signifies that the quality of the processed image approaches that of the original image.
P S N R = 10 log M A X 2 M S E
Normalized Mutual Information (NMI) is primarily employed for assessing the consistency of classification labels. Its computation incorporates mutual information and entropy from information theory, facilitating equitable comparisons across diverse datasets and varying numbers of clusters. A higher NMI value, approaching 1, indicates a greater similarity between two images.
N M I = 2 I ( A ; B ) H ( A ) + H ( B )
The Structural Similarity Index (SSIM) evaluates the similarity between different images by analyzing their brightness, contrast, and structure. A higher SSIM value, nearing 1, denotes a greater similarity between two images. The Spatial Correlation Coefficient (SCC) combines visual features and spatial features to measure the similarity between two images. It can also be used to describe whether the variation patterns of one dataset are correlated with those of another dataset.
The evaluation metrics for all processing results are shown in Table 1 and Table 2.
From the table, it is evident that the performance of various metrics is closely linked to changes in island area. The two traditional filters, mean and median filters, result in relatively minor changes and do not introduce new island pixels, thus demonstrating strong performance across all metrics. In the results obtained through our experimental method, the greater the number of superpixels (N), the smaller the variation in island area, leading to superior performance across metrics. Using MSE as an example, the calculation formula clearly shows that MSE measures similarity between two images by calculating the point-by-point difference between corresponding pixels. This metric straightforwardly emphasizes visual similarity, indicating that the fewer the number of pixels modified in an image or the smaller the change in pixel values, the more closely the modified image resembles the original. The proposed method, however, involves the addition of numerous synthetic pixels during the aggregating process, resulting in significant deviations in overall pixel values from the original image and consequently poorer MSE performance.
NMI stands out as the most comprehensive metric among all, as it evaluates the average amount of information and the consistency of category distributions by considering mutual information and entropy [35], thereby enabling fair comparisons across datasets with varying cluster numbers. Thus, a higher NMI value reflects the reasonableness and effectiveness of our experimental method in aggregating islands. Moreover, a high and stable SCC indicates that the pattern of involving island pixels in our experimental method aligns well with the trend of aggregating planar features.
Image feature points and sharpness are also crucial factors influencing similarity. Fast adaptive filtering tends to lose edge details during the denoising process, resulting in a performance that is not as strong as that of traditional filtering. Moreover, incorporating Gaussian blur in CNNs leads to the loss of significant image details, resulting in the weakest performance across all metrics.
Since this experiment explores the multi-scale representation of islands, similarity serves only as a reference metric. Due to inherent algorithmic limitations, traditional filtering methods cannot fill the gaps between islands, which means they cannot achieve multi-scale representation. Our method, however, can produce aggregating effects at different scales and exhibits superior performance in similarity metrics compared to various filtering techniques.

5. Summary

In addressing issues of traditional aggregating methods such as the inability to maintain simplification efficiency while preserving overall characteristics, coarse overall boundaries after aggregating and difficulty meeting diverse scale requirements, this research proposes an integrative innovative approach based on computer vision theory. Initially, the K-means clustering analysis is applied to distinguish islands from seawater. Subsequently, superpixel segmentation is utilized to generate aggregated results at varying scales through parameter adjustments. Finally, the Efros–Leung algorithm is employed to synthesize island-like pixels and fill in gaps. The experiment employs four metrics—MSE, PSNR, NMI, and SSIM—to assess both visual similarity to the original image and structural consistency in category distribution.
Compared to traditional approaches in multi-scale representation of remote sensing imagery, the novel method proposed in this study exhibits several advantages:
  • Traditional vector-based methods often lose original textures when aggregating planar features, and raster-based aggregation methods cannot eliminate gaps between islands. However, our proposed method can preserve the original textures while filling the gaps between islands during the aggregating process.
  • In contrast to traditional aggregating techniques, our method achieves smoother boundaries and sharper textures, thereby enhancing visualization quality;
  • Unlike traditional methods limited to a single scale, our approach allows for the generation of aggregated results at varying scales tailored to specific requirements.
However, this method still presents certain limitations in the multi-scale representation of remote sensing imagery. During texture synthesis, the approach utilizes only a small subset of pixels from the planar elements of islands in the original image. In cases when there is a significant difference in pixel values between the two ends of the area to be aggregated, it may result in non-smooth transitions in texture migration. And this section occupies the majority of the computational complexity of the entire algorithm. Therefore, there is a need for further research into a comprehensive method that considers all pixels while increasing computational efficiency. Moreover, in the classification step, the sharp edges at the junction between the island and the seawater will affect the boundary generation, which means the accuracy needs to be further improved.

Author Contributions

Conceptualization, R.L. and Y.S.; data curation, R.L.; formal analysis, R.L.; investigation, R.L.; methodology, R.L. and Y.S.; project administration, Y.S.; resources, W.D.; supervision, Y.S.; visualization, R.L.; writing—original draft, R.L.; writing—review and editing, R.L. and Y.S. R.L. and Y.S. equal contributions. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Natural Science Foundation of China [Grant Number 42001402].

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dollár, P.; Appel, R.; Belongie, S.; Perona, P. Fast Feature Pyramids for Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1532–1545. [Google Scholar] [CrossRef] [PubMed]
  2. Simoncelli, E.P.; Freeman, W.T. The steerable pyramid: A flexible architecture for multi-scale derivative computation. Proc. Int. Conf. Image Process. 1995, 3, 444–447. [Google Scholar]
  3. Freeman, W.T.; Adelson, E.H. The design and use of steerable filters. IEEE Trans. Pattern Anal. Mach. Intell. 1991, 13, 891–906. [Google Scholar] [CrossRef]
  4. Buades, A.; Coll, B.; Morel, J.M. A non-local algorithm for image denoising. Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. 2005, 2, 60–65. [Google Scholar]
  5. Liu, L.; Li, W.; Qiu, H. Research on Principles and Methods of Sparse Island Cartographic Generalization. Sci. Surv. Mapp. 2013, 38, 117–119. [Google Scholar] [CrossRef]
  6. Beard, M.K. Knowledge-based generalization of geologic maps. Comput. Geosci. 1991, 17, 1073–1083. [Google Scholar]
  7. Ruas, A.; Lagrange, J.P. Data and knowledge modeling for generalization. In GIS and Generalization: Methodology and Practice; Muller, J.C., Lagrange, J.P., Weibel, R., Eds.; Taylor & Francis: London, UK, 1995; pp. 73–90. [Google Scholar]
  8. Wertheimer, M. Laws of Organization in Perceptual Forms. In A Source Book of Gestalt Psychology; Ellis, W.D., Ed.; Harcourt, Brace: New York, NY, USA, 1938; pp. 71–88. [Google Scholar]
  9. Li, C.; Yin, Y.; Wu, P.; Wu, W. An area merging method in map generalization considering typical characteristics of structured geographic objects. Cartogr. Geogr. Inf. Sci. 2021, 48, 210–224. [Google Scholar] [CrossRef]
  10. Ai, T.; Zhang, X. The aggregation of urban building clusters based on the skeleton partitioning of gap space. In The European Information Society; Springer: Berlin/Heidelberg, Germany, 2007; pp. 153–170. [Google Scholar]
  11. Guo, P.; Li, C.; Yin, Y. Classification and filtering of constrained Delaunay triangulation for automated building aggregation. ISPRS J. Photogramm. Remote Sens. 2021, 176, 282–294. [Google Scholar]
  12. Zhang, F.; Sun, Q.; Ma, J.; Lyu, Z.; Huang, Z. A polygonal buildings aggregation method considering obstacle elements and visual clarity. Geocarto Int. 2023, 38, 2266672. [Google Scholar] [CrossRef]
  13. Guo, R.; Ai, T. Simplification and aggregation of building polygon in automatic map generalization. J. Wuhan Tech. Univ. Surv. Mapp. 2000, 25, 25–30. [Google Scholar]
  14. He, X.; Zhang, X.; Yang, J. Progressive amalgamation of building clusters for map generalization based on scaling subgroups. ISPRS Int. J. Geo-Inf. 2018, 7, 116. [Google Scholar] [CrossRef]
  15. Li, C.; Yin, Y.; Liu, X.; Wu, P. An automated processing method for agglomeration areas. ISPRS Int. J. Geo-Inf. 2018, 7, 204. [Google Scholar] [CrossRef]
  16. Feng, Y.; Thiemann, F.; Sester, M. Learning Cartographic Building Generalization with Deep Convolutional Neural Networks. ISPRS Int. J. Geo-Inf. 2019, 8, 258. [Google Scholar] [CrossRef]
  17. Douglas, D.H.; Peucker, T.K. Algorithms for the reduction of the number of points required to represent a digitized line or its caricature. Cartogr. Int. J. Geogr. Inf. Geovisualization 1973, 10, 112–122. [Google Scholar] [CrossRef]
  18. Ai, T.; Zhou, Q.; Zhang, X.; Huang, Y.; Zhou, M. A simplification of ria coastline with geomorphologic characteristics preserved. Mar. Geod. 2014, 37, 167–186. [Google Scholar] [CrossRef]
  19. Ai, T.; Ke, S.; Yang, M.; Li, J. Envelope generation and simplification of polylines using Delaunay triangulation. Int. J. Geogr. Inf. Sci. 2017, 31, 297–319. [Google Scholar] [CrossRef]
  20. Visvalingam, M.; Whyatt, J.D. Line generalisation by repeated elimination of points. Cartogr. J. 1993, 30, 46–51. [Google Scholar] [CrossRef]
  21. Zhang, J.; Yuan, C.; Tong, X. A least-squares-adjustment-based method for automated map generalization of settlement areas in GIS. In Proceedings of the Geoinformatics 2006: Geospatial Information Science, Wuhan, China, 28–30 October 2006; SPIE: Bellingham, WA, USA, 2006; Volume 6420, p. 64200K. [Google Scholar]
  22. Chen, W.; Long, Y.; Shen, J.; Li, W. Structure Recognition and Progressive Simplification of the Concaves of Building Polygon Based on Constrained D-TIN. Geomat. Inf. Sci. Wuhan Univ. 2011, 36, 584–587. [Google Scholar]
  23. Meijers, M. Building simplification using offset curves obtained from the straight skeleton. In Proceedings of the 19th ICA Workshop on Generalisation and Multiple Representation, Helsinki, Finland, 14 June 2016. [Google Scholar]
  24. Kada, M. Aggregation of 3D buildings using a hybrid data approach. Cartogr. Geogr. Inf. Sci. 2011, 38, 153–160. [Google Scholar] [CrossRef]
  25. Cheng, B.; Liu, Q.; Li, X.; Wang, Y. Building simplification using backpropagation neural networks: A combination of cartographers’ expertise and raster-based local perception. GIScience Remote Sens. 2013, 50, 527–542. [Google Scholar] [CrossRef]
  26. Yang, M.; Yuan, T.; Yan, X.; Ai, T.; Jiang, C. A hybrid approach to building simplification with an evaluator from a backpropagation neural network. Int. J. Geogr. Inf. Sci. 2022, 36, 280–309. [Google Scholar] [CrossRef]
  27. Stillwell, J.; Daras, K.; Bell, M. Spatial Aggregation Methods for Investigating the MAUP Effects in Migration Analysis. Appl. Spat. Anal. 2018, 11, 693–711. [Google Scholar] [CrossRef]
  28. Efros, A.A.; Leung, T.K. Texture Synthesis by Non-parametric Sampling. In Proceedings of the IEEE International Conference on Computer Vision (ICCV’99), Corfu, Greece, 20–25 September 1999; pp. 1033–1038. [Google Scholar] [CrossRef]
  29. MacQueen, J.B. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, CA, USA, 21 June–18 July 1965; Volume 1: Statistics. Le Cam, L.M., Neyman, J., Eds.; University of California Press: Berkeley, CA, USA, 1967; pp. 281–297. [Google Scholar]
  30. Wang, P.; Zeng, G.; Gan, R.; Wang, J.; Zha, H. Superpixel Segmentation Based on Multiple Seed Growth. In Proceedings of the 2017 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS), Xiamen, China, 6–9 November 2017; pp. 1–6. [Google Scholar]
  31. Shen, Y.; Ai, T.; Li, W.; Yang, M.; Feng, Y. A polygon aggregation method with global feature preservation using superpixel segmentation. Comput. Environ. Urban Syst. 2019, 75, 117–131. [Google Scholar] [CrossRef]
  32. Auer, M.; Zipf, A. 3D WebGIS: From Visualization to Analysis. An Efficient Browser-Based 3D Line-of-Sight Analysis. ISPRS Int. J. Geo-Inf. 2018, 7, 279. [Google Scholar] [CrossRef]
  33. Gavaskar, R.G.; Chaudhury, K.N. Fast Adaptive Bilateral Filtering. Trans. Image Process. 2019, 28, 779–790. [Google Scholar] [CrossRef]
  34. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef]
  35. Wachowiak, M.P.; Smoliková, R.; Tourassi, G.D.; Elmaghraby, A.S. Similarity Metrics Based on Nonadditive Entropies for 2D–3D Multimodal Biomedical Image Registration. In Medical Imaging 2003: Image Processing; Sonka, M., Fitzpatrick, J.M., Eds.; SPIE: Bellingham, WA, USA, 2003; Volume 5032, pp. 1090–1100. [Google Scholar]
Figure 1. Framework of our proposed multi-scale representation method.
Figure 1. Framework of our proposed multi-scale representation method.
Remotesensing 16 02971 g001
Figure 2. Image preprocessing of island samples. (a) Original image, (b) the classified outcome, (c) land mask, (d) water mask.
Figure 2. Image preprocessing of island samples. (a) Original image, (b) the classified outcome, (c) land mask, (d) water mask.
Remotesensing 16 02971 g002
Figure 3. An illustration of superpixel segmentation.
Figure 3. An illustration of superpixel segmentation.
Remotesensing 16 02971 g003
Figure 4. Superpixel segmentation and global simplification results. (a) N = 1500, (b) N = 500, (c) N = 1500, (d) N = 500.
Figure 4. Superpixel segmentation and global simplification results. (a) N = 1500, (b) N = 500, (c) N = 1500, (d) N = 500.
Remotesensing 16 02971 g004
Figure 5. Image processing flowchart of islands. (a) T-boundary, (b) filled map, (c) subtraction, (d) overlay.
Figure 5. Image processing flowchart of islands. (a) T-boundary, (b) filled map, (c) subtraction, (d) overlay.
Remotesensing 16 02971 g005
Figure 6. Boundary refinement processing. (a) Binarization, (b) result of erosion, (c) overlay.
Figure 6. Boundary refinement processing. (a) Binarization, (b) result of erosion, (c) overlay.
Remotesensing 16 02971 g006
Figure 7. Texture generation and migration. (a) Sample, (b) texture generation, (c) excavation of texture, (d) migration.
Figure 7. Texture generation and migration. (a) Sample, (b) texture generation, (c) excavation of texture, (d) migration.
Remotesensing 16 02971 g007
Figure 8. Aggregation of other planar elements. (a) Original image, (b) aggregation.
Figure 8. Aggregation of other planar elements. (a) Original image, (b) aggregation.
Remotesensing 16 02971 g008
Figure 9. Experimental area. (a) Sulawesi Island, (b) Philippine Archipelago.
Figure 9. Experimental area. (a) Sulawesi Island, (b) Philippine Archipelago.
Remotesensing 16 02971 g009
Figure 10. Classification and extraction. (a) Classification 1, (b) classification 2, (c) extraction 1, (d) extraction 2.
Figure 10. Classification and extraction. (a) Classification 1, (b) classification 2, (c) extraction 1, (d) extraction 2.
Remotesensing 16 02971 g010
Figure 11. Different scales of aggregation. (a) N = 1500, (b) N = 2500, (c) N = 5000, (d) N = 500, (e) N = 1500, (f) N = 9000.
Figure 11. Different scales of aggregation. (a) N = 1500, (b) N = 2500, (c) N = 5000, (d) N = 500, (e) N = 1500, (f) N = 9000.
Remotesensing 16 02971 g011
Figure 12. Smoothing and filling. (a) Before smoothing, (b) after smoothing, (c) before filling, (d) after filling.
Figure 12. Smoothing and filling. (a) Before smoothing, (b) after smoothing, (c) before filling, (d) after filling.
Remotesensing 16 02971 g012
Figure 13. Multiscale representation. (a) N = 6000, (b) N = 9000, (c) the original scale.
Figure 13. Multiscale representation. (a) N = 6000, (b) N = 9000, (c) the original scale.
Remotesensing 16 02971 g013
Figure 14. Comparison of two aggregating algorithms. (a) Arcgis 1, (b) Arcgis 2, (c) Arcgis 3, (d) improvement 1, (e) improvement 2, (f) improvement 3.
Figure 14. Comparison of two aggregating algorithms. (a) Arcgis 1, (b) Arcgis 2, (c) Arcgis 3, (d) improvement 1, (e) improvement 2, (f) improvement 3.
Remotesensing 16 02971 g014
Figure 15. Erosion and superimposition. (a) erosion 1, (b) land-cover 1, (c) water-cover 1, (d) erosion 2, (e) land-cover 2, (f) water-cover 2.
Figure 15. Erosion and superimposition. (a) erosion 1, (b) land-cover 1, (c) water-cover 1, (d) erosion 2, (e) land-cover 2, (f) water-cover 2.
Remotesensing 16 02971 g015
Figure 16. Sulawesi Island aggregation result. (a) N = 1800, (b) N = 7000, (c) the original scale.
Figure 16. Sulawesi Island aggregation result. (a) N = 1800, (b) N = 7000, (c) the original scale.
Remotesensing 16 02971 g016
Figure 17. Philippine Archipelago aggregation result. (a) N = 9000, (b) region b in the original scale, (c) region c in the original scale.
Figure 17. Philippine Archipelago aggregation result. (a) N = 9000, (b) region b in the original scale, (c) region c in the original scale.
Remotesensing 16 02971 g017
Figure 18. Aggregated results of different parameters. (a) N = 2500 len = 17, (b) N = 7000 len = 30, (c) N = 1800 len = 40, (d) N = 1800 len = 65, (e) N = 6400 len = 75, (f) N = 9000 len = 50.
Figure 18. Aggregated results of different parameters. (a) N = 2500 len = 17, (b) N = 7000 len = 30, (c) N = 1800 len = 40, (d) N = 1800 len = 65, (e) N = 6400 len = 75, (f) N = 9000 len = 50.
Remotesensing 16 02971 g018
Figure 19. Comparison of the different methods based on the raster data. (a) Original image, (b) median filtering, (c) mean filtering, (d) fast adaptive bilateral filtering, (e) CNN + Gaussian blur, (f) the proposed method.
Figure 19. Comparison of the different methods based on the raster data. (a) Original image, (b) median filtering, (c) mean filtering, (d) fast adaptive bilateral filtering, (e) CNN + Gaussian blur, (f) the proposed method.
Remotesensing 16 02971 g019
Table 1. Evaluation metrics (1).
Table 1. Evaluation metrics (1).
Sulawesi IslandMSEPSNRNMISSIMSCCArea (Pixels)
Original0inf11172,716
Mean Filtering4.06296666542.040.58530.9706038870.9479
Median Filtering3.37413215642.850.65550.974185110.9501
Fast Adaptive
Bilateral Filtering
11.61022885640462337.480.42580.90688593770763540.9576
CNN + Gaussian Blur19.70034281412760535.190.27860.84112784365730860.8788
N = 2500, len = 178.24051284838.970.8370.9212239140.8612122,520
N = 1800, len = 4010.8256950437.790.79510.9057799650.8605136,301
N = 7000, len = 306.94340515139.720.85820.9265344990.8639114,136
Table 2. Evaluation metrics (2).
Table 2. Evaluation metrics (2).
Philippine ArchipelagoMSEPSNRNMISSIMSCCArea (Pixels)
Original0inf111657,615
Mean Filtering8.83084936538.670.54840.9457349310.9440
Median Filtering7.66745600939.280.60130.9496958020.9487
Fast Adaptive
Bilateral Filtering
17.139298639849335.790.40970.8678396974595230.9533
CNN + Gaussian Blur25.918529024953433.990.32450.77775734600865130.8912
N = 1800, len = 6532.7671303132.980.47880.7941312970.89531,441,493
N = 6400, len = 7524.062964234.320.53610.8414121550.90621,224,822
N = 9000, len = 5016.8869555835.860.5810.8723601660.92311,055,782
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, R.; Shen, Y.; Dai, W. Cartographic Generalization of Islands Using Remote Sensing Images for Multiscale Representation. Remote Sens. 2024, 16, 2971. https://doi.org/10.3390/rs16162971

AMA Style

Li R, Shen Y, Dai W. Cartographic Generalization of Islands Using Remote Sensing Images for Multiscale Representation. Remote Sensing. 2024; 16(16):2971. https://doi.org/10.3390/rs16162971

Chicago/Turabian Style

Li, Renzhu, Yilang Shen, and Wanyue Dai. 2024. "Cartographic Generalization of Islands Using Remote Sensing Images for Multiscale Representation" Remote Sensing 16, no. 16: 2971. https://doi.org/10.3390/rs16162971

APA Style

Li, R., Shen, Y., & Dai, W. (2024). Cartographic Generalization of Islands Using Remote Sensing Images for Multiscale Representation. Remote Sensing, 16(16), 2971. https://doi.org/10.3390/rs16162971

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop