Next Article in Journal
Systematic Review of Factors Influencing Students’ Performance in Educational Buildings: Focus on LCA, IoT, and BIM
Previous Article in Journal
Experimental and Numerical Investigation on Stress Concentration Factors of Offshore Steel Tubular Column-to-Steel Beam (STCSB) Connections
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Prior-Guided Dual Branch Multi-Feature Fusion Network for Building Segmentation in Remote Sensing Images

1
School of Computer Science, Qinghai Normal University, Xining 810016, China
2
School of Mathematics and Information Technology, Yuncheng University, Yuncheng 044000, China
3
State Key Laboratory of Tibetan Intelligent Information Processing and Application, Xining 810016, China
*
Author to whom correspondence should be addressed.
Buildings 2024, 14(7), 2006; https://doi.org/10.3390/buildings14072006
Submission received: 4 June 2024 / Revised: 24 June 2024 / Accepted: 27 June 2024 / Published: 2 July 2024
(This article belongs to the Section Construction Management, and Computers & Digitization)

Abstract

:
The domain of remote sensing image processing has witnessed remarkable advancements in recent years, with deep convolutional neural networks (CNNs) establishing themselves as a prominent approach for building segmentation. Despite the progress, traditional CNNs, which rely on convolution and pooling for feature extraction during the encoding phase, often fail to precisely delineate global pixel interactions, potentially leading to the loss of vital semantic details. Moreover, conventional CNN-based segmentation models frequently neglect the nuanced semantic differences between shallow and deep features during the decoding phase, which can result in subpar feature integration through rudimentary addition or concatenation techniques. Additionally, the unique boundary characteristics of buildings in remote sensing images, which offer a rich vein of prior information, have not been fully harnessed by traditional CNNs. This paper introduces an innovative approach to building segmentation in remote sensing images through a prior-guided dual branch multi-feature fusion network (PDBMFN). The network is composed of a prior-guided branch network (PBN) in the encoding process, a parallel dilated convolution module (PDCM) designed to incorporate prior information, and a multi-feature aggregation module (MAM) in the decoding process. The PBN leverages prior region and edge information derived from superpixels and edge maps to enhance edge detection accuracy during the encoding phase. The PDCM integrates features from both branches and applies dilated convolution across various scales to expand the receptive field and capture a more comprehensive semantic context. During the decoding phase, the MAM utilizes deep semantic information to direct the fusion of features, thereby optimizing segmentation efficacy. Through a sequence of aggregations, the MAM gradually merges deep and shallow semantic information, culminating in a more enriched and holistic feature representation. Extensive experiments are conducted across diverse datasets, such as WHU, Inria Aerial, and Massachusetts, revealing that PDBMFN outperforms other sophisticated methods in terms of segmentation accuracy. In the key segmentation metrics, including mIoU, precision, recall, and F1 score, PDBMFN shows a marked superiority over contemporary techniques. The ablation studies further substantiate the performance improvements conferred by the PBN’s prior information guidance and the efficacy of the PDCM and MAM modules.

1. Introduction

The evolution of remote sensing technology has seen it become an indispensable tool in the acquisition of geographic information, thanks to its expansive reach and wealth of detail. The extraction and segmentation of buildings from remote sensing images have significant application value in practice [1]. Building segmentation facilitates the precise extraction of key data points, including the exact location and spatial boundaries of the structures. This capability is crucial for understanding the built environment. The data obtained from this process lays the groundwork for subsequent analytical work and practical application of the information, which is vital for making informed decisions and executing strategic plans across various fields [2]. These applications encompass urban planning, the development of smart cities, the management of land resources, and the assessment of disaster risks, all of which rely on accurate and detailed insights derived from building segmentation [3].
Over the years, a multitude of approaches have been developed for the segmentation of buildings in remote sensing images. Predominantly, these traditional approaches hinge on the extraction of various features such as texture, geometric properties, spectral information, and shadows to distinguish buildings. For example, Partovi et al. [4] utilized parametric and tracking algorithms to pinpoint the building boundary points, which were then assembled into polygons outlining building contours. Ghaffarian [5] applied a dual-threshold shadow detection technique, leveraging the shadow shapes and solar illumination directions for building detection. Dikmen et al. [6] combined shadow detection with spectral characteristics for the segmentation of buildings. Yuan et al. [7] incorporated both spectral and textural data in their segmentation methodology, focusing on the extraction of building structures. Abdullah et al. [8] employed LIDAR data grouped with a digital elevation model (DEM) for automated segmentation, while Noyel et al. [9] introduced a general morphological segmentation approach for hyperspectral images using the watershed transformation. Yamashita et al. [10] applied the discriminant theory to classify LIDAR-derived polygons as either buildings or non-buildings. Despite their contributions, these methods are constrained by the need for multi-source data and exhibit a high degree of complexity, which can limit their broader application.
The traditional methods for building segmentation, as previously discussed, are heavily dependent on manual feature engineering and selection processes. This reliance may inadvertently overlook the critical information or intrinsic correlations within the dataset, which can diminish the models’ generalization capabilities. Consequently, these models may produce subpar segmentation outcomes when confronted with the intricacies of complex remote sensing images. In recent years, the advent of deep learning has revolutionized the domain of image processing. Convolutional neural networks (CNNs), in particular, have demonstrated remarkable advancements in a spectrum of computer vision tasks, including semantic segmentation [11]. This holistic learning approach significantly streamlines the classification process, transcending the limitations of traditional feature engineering that relies on human expertise. Notable CNN architectures such as UNet [12], DeepLabV3 [13], and EncNet [14] have garnered substantial acclaim for their achievements in semantic segmentation across diverse domains, encompassing natural images, medical imaging, and remote sensing imagery. Given that convolutional operations are inherently focused on local receptive fields, the aforementioned CNN-based methods have innovated by incorporating large kernel convolutions, dilated convolutions, and the stacking of convolutional and pooling layers. These enhancements are designed to expand the receptive field, thereby capturing a broader context and facilitating the extraction of more comprehensive and nuanced features from the data.
In contrast to traditional convolutional operations, non-local neural networks [15] extend the concept of non-local mean operations in computer vision. They introduce a mechanism that captures the relationships between pixels over long distances, irrespective of their spatial separation. DNLNet (Disentangled Non-Local Neural Networks) [16] builds upon this non-local mechanism by introducing the DNL block, which is designed to replace the conventional non-local block. This innovation further enhances the precision of semantic segmentation tasks. Fu et al. [17] proposed the Dual Attention Network (DaNet), which integrates channel and spatial attention mechanisms based on self-attention. Following the success of the Transformer model [18] in natural language processing, where it employs a multi-head self-attention mechanism to supplant the convolutional modules, there has been a surge in research adapting this model to the visual domain. Notably, Dosovitskiy et al. [19] pioneered the application of the Vision Transformer (ViT) architecture to image recognition tasks, substantiating its efficacy. This was followed by Strudel et al. [20], who innovatively adapted the ViT-based Segmenter network for image segmentation tasks. However, the computational expense of self-attention within the Transformer framework can be exceedingly high [21], and certain operations may prove to be redundant [22]. Additionally, while self-attention-based transformer-structured networks excel on extensive datasets, they can encounter difficulties when trained on smaller, more limited datasets [23]. These challenges necessitate further research and innovation to optimize the application of transformer models within the visual domain.
The task of semantic segmentation for buildings in remote sensing images presents unique and heightened challenges when juxtaposed with the segmentation of natural images. Beyond the multi-scale features that are prevalent in the natural images, the remote sensing imagery is characterized by intricate backgrounds filled with a multitude of objects that are unrelated to buildings. This complexity can lead to interference from the background with the segmentation of the targets [24]. Furthermore, there are instances where the buildings and non-building elements may bear a resemblance in appearance, such as the scenarios where the roofs of certain buildings blend in with the roads, making them indistinguishable [25]. Current models for building segmentation in remote sensing images have predominantly drawn upon the architectures that have demonstrated efficacy in the segmentation of natural images. For example, Qian et al. [26] harnessed the capabilities of DeepLabV3 for semantic segmentation of buildings. Benchabana et al. [27] employed the superpixels to aggregate image pixels and subsequently utilized a variational autoencoder coupled with a CNN for building identification and detection. Wang et al. [28] introduced a dual-path Vision Transformer (ViT) network tailored for building segmentation. Cui et al. [29] proposed an enhanced Transformer model to segment densely packed urban buildings from remote sensing images with intricate backgrounds. Despite these advancements, existing methods have not fully considered the distinctive attributes of buildings, such as their relatively regular boundaries. These features represent crucial prior information and constitute a significant distinction between buildings and other non-building elements, including land, vegetation, and roadways. Incorporating such characteristics could potentially enhance the accuracy and reliability of building segmentation in remote sensing images.
Recently, researchers including Liu et al. [30], Li et al. [31], and Ding et al. [32] have redesigned and refined convolutional operations and the underlying structures of CNN networks. These efforts have significantly bolstered the representational capacity of CNNs. Motivated by such progress, we have re-examined the CNN-based UNet network model [12], which was originally crafted as an end-to-end solution for medical image segmentation on modestly scaled cellular datasets. The UNet model’s outstanding segmentation capabilities have inspired a plethora of scholars to extend its framework, giving rise to a variety of enhanced iterations. Notable among these are nnU-Net with a self-configuring approach [33], dynamic U-Net [34] and AgileFormer [35]. These adaptations integrate dynamic convolutions or dynamic deformable attention mechanisms into the UNet’s original design, yielding remarkable outcomes in the domain of medical image segmentation. Beyond the medical sector, as research has progressed, the UNet model has been progressively adapted for the segmentation of buildings in remote sensing images, with equally remarkable results. For instance, Daranagama, S. et al. [36] proposed a UNet architecture based on image enhancement and polygon regularization for building detection. Pan et al. [37] introduced a deep learning architecture based on UNet for the segmentation and classification of urban village buildings. Su et al. [38] refined the UNet-based building segmentation technique by integrating DenseNet, UNet, dilated convolutions, and DecovNet. Lei et al. [39] presented an attention-based dual hybrid attention UNet tailored for pixel-level building segmentation. Ahmadi, S.A. et al. [40] designed a two-stage UNet architecture for building damage detection. These are all improvements made to the UNet architecture. In this paper, we have taken a step further by not only refining and optimizing the architecture of UNet but also integrating crucial prior knowledge into our network design. Specifically, we propose a prior-guided dual-branched multi-feature fusion network that has infused regional aggregation information derived from superpixels and the boundary details of buildings to obtain more accurate segmentation results.
The specific contributions of this paper are as follows:
  • Novel Network Architecture: We design a novel building segmentation network termed the prior-guided dual branch multi-feature fusion network (PDBMFN). This framework encompasses a prior-guided branch network (PBN) in the encoding process, a parallel dilated convolution module (PDCM) designed to assimilate prior information, and a multi-feature aggregation module (MAM) in the decoding process. The integration and encoding of PBN’s prior information within the PDCM are further augmented by the MAM during the decoding phase, refining the feature representation.
  • Prior-guided Branch Network: We propose the prior-guided branch network (PBN), which integrates prior knowledge to enhance segmentation accuracy. PBN capitalizes on the regional aggregation properties of superpixels and the boundary cues from edge images, improving the accuracy of the building boundaries in the segmentation output.
  • Parallel Dilated Convolution Module: We present the parallel dilated convolution module (PDCM), which initially consolidates encoded prior feature information from dual branches. Subsequently, it employs parallel dilated convolution kernels with varying dilation rates to apprehend multi-scale contextual information. This strategy enables the network to adapt to targets of varying dimensions, enhancing its scene comprehension. We also incorporate the convolutional block attention module (CBAM) to integrate and refine features derived from different dilated convolutions.
  • Multi-feature Aggregation Module: We construct the multi-feature aggregation module (MAM), which progressively aggregates feature information from assorted layers during the network’s decoding stage. By integrating aggregated skip connections, the MAM deeply merges shallow, fine-grained features with profound semantic information, capitalizing on their synergistic interplay to construct an enriched and holistic feature representation, thereby optimizing the network’s segmentation efficacy.
The subsequent sections of this paper are structured as follows: Section 2 delineates the structure of the proposed network model. Section 3 demonstrates the model’s merits through empirical analyses and comparative studies. Section 4 concludes the paper with a summarizing overview.

2. Proposed Method

In this section, we first present the overall structure of PDBMFN. Subsequently, we introduce the image superpixel segmentation and edge detection methods used in the PDBMFN network, along with the PBN, PDCM, and MAM modules. Finally, we discuss the loss function employed by the network.

2.1. The Architecture of PDBMFN

The PDBMFN is a deep convolutional neural network, employing an encoder-decoder architecture akin to the UNet. The foundational operations of this architecture include downsampling, convolution, upsampling, and feature concatenation. However, the inherently limited receptive field of standard convolutions restricts the network’s capacity to capture expansive global features. Moreover, the max pooling operation, while reducing dimensionality, can inadvertently discard critical feature information. Additionally, the upsampling and downsampling primarily consider local features within individual channels. The feature fusion facilitated by skip connections is also somewhat limited, as it only concatenates features along the channel dimension, without actively exploring the interrelationships between features across different channels.
In response to the aforementioned challenges, we have engineered the PDBMFN network model, which extends the capabilities of traditional convolutional networks and refines the precision of global feature extraction and segmentation outcomes. The PDBMFN network primarily comprises three essential components: the PBN, the PDCM, and the MAM. The architecture of this network model is visually represented in Figure 1, which illustrates the data flow within the network and delineates the interrelationships among its various components.
Enhancing Segmentation Accuracy with PBN. Firstly, we introduce the PBN structure (Regional Prior-guided Branch and Edge Prior-guided Branch in Figure 1), which takes advantage of the region aggregation properties of superpixels and the boundary prior information from images. This enhances the accuracy of building detection during the encoding process.
Expanding Receptive Field with PDCM. Then, we incorporate the PDCM to counteract the small receptive field problem associated with convolutions. The PDCM is designed to expand the receptive field, allowing for a more comprehensive capture of contextual information. We then utilize the convolutional block attention module (CBAM) to refine features by enriching channel and spatial feature representations. We derive attention-enhanced features that retain essential building information while filtering out noise and irrelevant details.
Feature Aggregation with MAM. Lastly, the MAM module is designed to aggregate feature information progressively from different decoding levels during the decoding stage. By incorporating comprehensive skip connections, the MAM effectively intertwines superficial, fine-grained features with deeper semantic insights. As a result, the network’s segmentation performance is significantly enhanced.
The operational mechanisms of each module in the PDBMFN are to be thoroughly elucidated in the ensuing sections. It is important to note that, apart from the PDCM and MAM integral to the PDBMFN, the network’s upsampling, downsampling, and convolutional structures are preserved in accordance with the original UNet.

2.2. Deriving Prior Information from Remote Sensing Images

In PDBMFN, we initiate the process by conducting superpixel segmentation on the original image to extract regional prior information. Concurrently, we perform edge detection on the original image to procure an accurate edge prior information image. Consequently, the input to the PDBMFN network model encompasses not only the original image but also its corresponding superpixel and edge detection images. The prior-guided network is conducive to feature extraction, enhancing the segmentation accuracy of building structures in remote sensing images.

2.2.1. Deriving Prior Information of Regions through Superpixel Segmentation

A superpixel is an image segmentation technique that groups adjacent pixels with similar textures, colors, brightness, and other characteristics into a single superpixel, which is then filled with a uniform color. Superpixels efficiently represent image features using fewer pixels, greatly reducing the complexity of post-processing. They are widely used in the field of computer vision. The SLIC [41] algorithm (Simple Linear Iterative Clustering), as its name suggests, is a simple iterative clustering superpixel segmentation algorithm with linear complexity. Due to its excellent performance in both computational speed and segmentation effect, this paper adopts SLIC as the superpixel segmentation algorithm.
The SLIC algorithm generates superpixels based on the k-means clustering algorithm. Given an image with N pixels, the algorithm divides it into K superpixels of approximately the same size, with each superpixel containing approximately N / K pixels. During the initial clustering, the adjacent distance between the centers of two neighboring superpixels is approximately S = N / K . The overall process of the SLIC algorithm is similar to k-means. During clustering, it iteratively calculates the cluster with the minimum distance, updates the center point of the superpixel, and reassigns the membership of other pixels.
In the traditional k-means algorithm, the distance from each cluster center to every pixel in the image needs to be calculated. In stark contrast, the SLIC algorithm limits the search area by only searching for similar pixels within a 2 S × 2 S region around the superpixel center. This significantly reduces the number of distance calculations required, resulting in a clear speed advantage compared to conventional k-means clustering.
Figure 2 illustrates the outcomes of SLIC superpixel segmentation with varying initializations of the control parameter K . The original image is sourced from the Inria Aerial Dataset. As K increases, the size of the superpixels decreases, and the segmentation results become closer to the original image with faster algorithmic computation. Conversely, when K decreases, the superpixels become larger, resulting in greater loss of regional and boundary information, greater deviation from the original image, and slower computation speed. Therefore, it is necessary to strike a balance in selecting the value of K in experiments.
We set the number of SLIC superpixels to be one percent of the number of pixels in the original image, i.e., K = N / 100 . Consequently, for images from the Inria Aerial Dataset and the Massachusetts Buildings Dataset, the number of superpixels in their superpixel images is 2500. Similarly, for images from the WHU dataset, the number of superpixels in their superpixel images is 2621. We will provide a detailed introduction to the configuration of the databases in Section 3.1.

2.2.2. Deriving Edge Prior Information through Edge Detection

Edge detection is a fundamental technique within the domain of computer vision, designed to pinpoint regions of abrupt brightness variation—regions that define the edges within an image. Given the inherent discontinuities in pixel brightness, conventional edge detection algorithms, including the Robert, Prewitt, and Sobel operators (refer to [42] for a detailed exposition of these methods), apply differential operators to identify edges. Among these, the Sobel operator stands out for its enhanced precision in edge localization and its superior noise suppression capabilities. Leveraging these benefits, we have chosen to employ the Sobel edge detection algorithm to generate an image that captures prior edge information. As illustrated in Figure 3, the Sobel operator offers a clearer delineation of edge features and a reduced presence of noise when juxtaposed with the Robert and Prewitt operators, thereby providing a more refined edge detection outcome.

2.3. The Design of PBN Network Architecture

The PBN, depicted in Figure 1, is a network with a dual-branch encoding architecture, comprising two distinct pathways: one guided by regional priors and the other by edge priors. This dual-branch structure is predominantly located in the encoding part of the PDBMFN, where the region prior-guided branch uses the prior information derived from superpixel segmentation to guide the segmentation of building areas, and the edge information from the edge prior-guided branch is further utilized to enhance the building edge features extracted by the edge prior-guided branch.

2.3.1. The Region Prior-guided Branch

The region prior-guided branch, as illustrated in Figure 1, undertakes the encoding of the concatenated original image and its superpixel representation. This process involves a series of convolutional and downsampling operations that progressively refine the feature representation. Furthermore, this branch is designed to integrate edge features from the edge prior-guided branch via the Parallel Dilated Convolution Module (PDCM). This continuous feature fusion allows the region prior-guided branch to capture a richer semantic context, thereby achieving a more nuanced understanding of the building structures within the imagery.
Let F 1 i represent the feature map obtained after the ith downsampling operation followed by a convolution in the region prior-guided branch. The definition of F 1 i is as follows:
F 1 i = f 3 × 3 , 3 × 3 ( D s ( f 3 × 3 , 3 × 3 ( C a t ( I , S P ) ) ) ) , i = 1 f 3 × 3 , 3 × 3 ( D s ( P D C M i - 1 ) ) , i > 1
In this context, f 3 × 3 , 3 × 3 represents two consecutive 3 × 3 convolutions, D s denotes the downsampling operations, C a t stands for concatenate, I and S P refer to the original image and its corresponding superpixel image, respectively, and i is the number of downsampling operations. PDCM will be introduced in detail in Section 2.4, and we will delve into the details of how the edge features from the edge prior-guided branch are incorporated into the features of the region prior-guided branch, enhancing the edge characteristics.
These dual branch networks facilitate a gradual compression of the input data, distilling all essential information into a condensed abstract form. Throughout the encoding phase, the dimensionality of the feature maps is effectively reduced by half via downsampling, while the richness of feature representation is enhanced by doubling the number of channels through consecutive convolution operations.

2.3.2. The Edge Prior-Guided Branch

The edge prior-guided branch involves encoding the edge map from the original image through multiple convolutions and downsampling operations to derive deeper semantic edge information from the input image, as shown in Figure 1. Let F 2 i represent the feature map obtained after the ith downsampling operation followed by a convolution in the edge prior-guided branch. The definition of F 2 i is as follows:
F 2 i = f 3 × 3 , 3 × 3 ( D s ( f 3 × 3 , 3 × 3 ( E ) ) ) , i = 1 f 3 × 3 , 3 × 3 ( D s ( F 2 i - 1 ) , i > 1
In this context, f 3 × 3 , 3 × 3 represents two consecutive 3 × 3 convolutions, D s denotes the downsampling operations, E refers to the original edge image, and i is the number of downsampling operations.

2.3.3. Decoding Process

The decoding process aims to gradually transform the encoded abstract representation into meaningful outputs, namely the final network predictions of building or non-building information. Each step of the decoding process may rely on the output of its previous step as well as contextual information passed from the encoder. In the original UNet decoding process, the size of the feature map is doubled through upsampling, concatenated with the corresponding feature map from the encoding layer, and then the number of channels is halved through a double convolution operation. Building upon this approach, we employ the multi-feature aggregation module (MAM) to concurrently assimilate feature information from adjacent decoding layers. These neighboring feature maps encapsulate semantic insights extracted at varying depths by the network, each with distinct spatial dimensions. The MAM adeptly intertwines the superficial, fine-grained features from the preceding level with the profound semantic attributes of the subsequent level, culminating in an enriched aggregation of features. The MAM will be described in detail in Section 2.5.
In summary, the PBN effectively merges the superpixel image, enriched with regional information, with the original image during the encoding phase, subsequently channeling this composite input into the encoding network. Concurrently, an edge image, derived from an edge detection algorithm, is directed into a separate encoding network. Post downsampling within the region prior-guided branch, the downsampled data from the edge prior-guided branch is seamlessly integrated via the PDCM, thereby refining the building edges. Expanding upon the UNet’s decoding paradigm, we incorporate the MAM to deeply amalgamate semantic information across various decoding strata. This aggregation of multi-level features significantly bolsters the network’s segmentation capabilities, yielding enhanced predictive accuracy in discerning building structures within remote sensing imagery.

2.4. The Design of PDCM Module

With a fixed kernel size, dilated convolution can increase the receptive field of output units without increasing computational complexity. PDCM utilizes parallel convolutions with different dilation rates to capture multi-scale contextual information under different receptive fields, adapting to targets of various sizes and enhancing the network’s understanding of the scene.
As illustrated in Figure 4, the two inputs, F 1 i (from the region prior-guided branch) and F 2 i (from the edge prior-guided branch) are the downsampled feature maps obtained through double convolution from different branches. These two inputs are added together and then concatenated with the outputs of four convolutions with different dilation rates. Subsequently, batch normalization and ReLU activation are applied, followed by the introduction of the CBAM attention mechanism module. The specific calculation is as follows:
F i = F 1 i + F 2 i
G _ c a t i = C a t ( D i l C o n v 1 ( F i ) , D i l C o n v 2 ( F i ) , D i l C o n v 3 ( F i ) , D i l C o n v 4 ( F i ) )
P D C M i = C B A M ( σ ( B N ( G _ c a t i ) ) )
In the equations, D i l C o n v i denotes a dilated convolution operation with a dilation rate of i and a kernel size of 3 × 3 . In this paper, we choose four parallel convolutions with dilation rates of 1, 2, 3, and 4, respectively. C a t denotes the concatenation operation along the tensor’s channel dimension. B N stands for batch normalization, σ is the activation function ReLU, and C B A M is the attention module. In the CBAM, each channel of the features represents a specialized feature detector; therefore, the channel attention focuses on what kind of features are meaningful. Following the channel attention mechanism, the Convolutional Block Attention Module (CBAM) further incorporates a spatial attention module designed to hone in on salient spatial features. This dual-attention approach allows CBAM to synergize channel-wise and spatial information, thereby enriching the network’s capacity to integrate and refine features across various layers. For an in-depth examination of the CBAM’s architectural intricacies, one may refer to [43].

2.5. The Design of the MAM

Expanding upon UNet’s original decoding process, the MAM is meticulously crafted to systematically and iteratively consolidate feature information between contiguous layers, denoted as F i 1 and F i . Here, F i 1 is the shallow-level feature map in the decoding process, while F i is the deep-level feature map. As illustrated in Figure 5, the deeper-level features F i are first deconvolved to derive F i 1 , which is then added to the shallower-level features F i 1 from the previous layer. Upon the application of the activation function Sigmoid, a set of weight coefficients are obtained and multiplied with F i 1 . The calculation is as follows:
F i 1 = D e C o n v ( F i )
O u t p u t = F i 1 × σ ( F i 1 + F i 1 )
In the equations, D e C o n v denotes the deconvolution operation, × denotes element-wise multiplication, + denotes element-wise addition, and O u t p u t is the output feature of MAM.
The integration of this aggregated skip connection facilitates a profound fusion of the superficial, fine-grained features from the preceding layer with the profound semantic insights from the subsequent layer. This integration leverages the symbiotic nature of these features, capitalizing on their complementary attributes to forge a more enriched and holistic feature representation. As a result, the network’s segmentation outcomes are further refined and optimized.

2.6. Loss Function

Given that our segmentation outcomes are confined to two distinct classes—namely, buildings and non-buildings, we adopt the commonly used Binary Cross Entropy Loss (BCELoss) as the loss function for our model. The BCEloss is utilized to assess the discrepancy between the model’s output and the actual labels in binary classification tasks, specifically distinguishing between buildings and non-buildings in this study. For each individual sample, the BCELoss computes the cross-entropy loss by comparing the probability distribution generated by the model with the ground truth labels. Subsequently, the losses across all samples are averaged to obtain the overall loss value. The calculation is as follows:
B C E ( x i , y i ) = 1 N i = 0 N 1 y i l o g ( x i ) + ( 1 y i ) l o g ( 1 x i )
In Equation (8), x is the model’s predicted label, y is the actual label, and i denotes a specific pixel in the image. When there are N sample pixels during the training process, Equation (8) calculates the average loss.

3. Experimental Results and Analyses

This section delineates the deployment of the PDBMFN across three representative datasets for building segmentation: the Inria Aerial, WHU, and Massachusetts datasets. We proceed by juxtaposing the performance of PDBMFN with several seminal network models that have garnered acclaim in the realm of semantic segmentation in recent years. This includes models such as UNet [12], DeepLabV3 [13], EncNet [14], DaNet [17], DnlNet [16], and Segmenter [20]. Through these comparative experiments, we aim to substantiate the efficacy of our proposed network model. At the end of this section, we conduct an array of ablation studies. These studies are designed to affirm the individual contributions of the PBN, PDCM, and MAM to the overall performance of the network. By systematically evaluating the impact of each component, we demonstrate their integral roles in enhancing the segmentation accuracy and robustness of PDBMFN.
All experiments in this paper were conducted on a Windows workstation with the following main configurations: CPU: Intel Core i5-2.50 GHz, Main Memory: 32 GB, GPU: NVIDIA GeForce RTX 4080S. The primary software environment consists of Python 3.9 and PyTorch 2.1. In our experimental training phase, we selected the RMSprop optimizer, known for its effectiveness in gradient descent optimization. We commenced with an initial learning rate of 0.0001. The training was conducted with a batch size of eight samples. Additionally, we executed 80 epochs for each training session.

3.1. Datasets

(1)
Inria Aerial Dataset [44]. This is an urban building segmentation dataset that only labels buildings and the background. The images are 0.3 m-resolution orthorectified RGB images, encompassing both densely populated areas (such as the financial district of San Francisco) and mountainous towns (like Lienz in Austrian Tyrol). The original image size is 5000 × 5000, but due to experimental computational limitations, we cropped them into ten images of 500 × 500 each, resulting in a total of 18,000 images. Following the protocol established in [45], we randomly split the dataset into a training set of 12,600 images, a validation set of 2700 images, and a test set of 2700 images.
(2)
WHU Dataset [46]. The original aerial photography data comes from the Land Information Service website of New Zealand. The authors manually downsampled the images to a ground resolution of 0.3 m, with each image measuring 512 × 512 pixels. Adhering to the original dataset partitioning, we discarded images with significant blank regions, resulting in a training set of 9366 images, a validation set of 1537 images, and a test set of 3770 images.
(3)
Massachusetts Buildings Dataset [47]. This dataset consists of 151 aerial images of Boston, each measuring 1500 × 1500 pixels. Similarly, we cropped the original images into 500 × 500 sizes and discarded images with significant blank areas. Following the protocol established in [44], we randomly reassigned the images into training, validation, and test sets, with 810, 170, and 170 images in each set, respectively.

3.2. Model Accuracy Metrics

In the semantic segmentation of buildings, it is necessary to determine whether each pixel in each image belongs to a building or the background. We employ commonly used metrics such as precision, recall, F1-score and mean intersection over union (mIoU) to evaluate the accuracy of the model in segmenting buildings.
Precision refers to the proportion of samples predicted to belong to the building category that actually belong to the building category in the segmentation results. The calculation is formally presented in Equation (9). In this equation, T P represents the number of true buildings, F P represents the number of samples where the background is incorrectly identified as a building, and T P + F P represents the total number of samples predicted by the model as belonging to the building category.
P r e c i s i o n = T P T P + F P × 100 %
Recall represents the proportion of samples that truly belong to the building category and are correctly classified as such. The calculation is formally presented in Equation (10), in which, the F N is the number of false negatives where the buildings are misclassified as background.
R e c a l l = T P T P + F N × 100 %
F1-Score is the harmonic mean of precision and recall. The calculation is formally presented in Equation (11).
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l × 100 %
The mIoU is one of the most commonly used evaluation metrics in image semantic segmentation. It assesses the degree of overlap between the predicted label area and the actual label area by calculating the ratio of their intersection to their union. The value of mIoU ranges from 0 to 1, with a higher value indicating a higher degree of overlap between the prediction and the true annotation. The calculation is formally presented in Equation (12), which takes into account both the building and nonbuilding categories. T N represents the number of samples where the background is correctly identified as a background in Equation (12).
m I o U =   ( ( T P T P + F P + F N + T N T N + F P + F N ) / 2 ) × 100 %

3.3. Experiments and Result Analyses on the Inria Aerial Dataset

To substantiate the efficacy of the PDBMFN, we deployed it on the Inria Aerial dataset. In parallel, we conducted experiments with other state-of-the-art network architectures, including UNet, DeepLabV3, EncNet, DaNet, DNLNet, and Segmenter. The visual results of these segmentation experiments are presented in Figure 6.
Figure 6 presents a comparative analysis of the segmentation results on representative images from the Inria Aerial dataset. Column (a) shows images with large building targets, columns (b) and (d) contain images with both large and small targets; column (c) exclusively features small target images; and column (e) displays images with regularly shaped and elongated building boundaries. This well-designed presentation allows for a comprehensive visual assessment of each model’s segmentation performance across varying complexities and target sizes.
In column (a), where the distinction between building bodies and background roads is challenging, models such as UNet and DeepLabV3 result in fragmented buildings with incomplete structures. DaNet and EncNet also struggle with incomplete building delineation, while DNLNet faces difficulties in discerning boundaries between buildings and roads. Segmenter shows a tendency to leave holes within building interiors. In stark contrast, our PDBMFN method adeptly differentiates between buildings and the background, yielding clearer and more regular boundaries. In column (b), most methods demonstrate issues with merging adjacent buildings and occasionally overlook small targets. However, our approach excels by detecting all building targets and effectively separating adjacent structures. In column (c), our method’s superiority is evident in its superior boundary regularity. In column (d), other methods exhibit internal hollowness and irregular boundaries, often missing subtle protrusions or edges. Our method, in contrast, effectively captures these details. Finally, in column (e), UNet and DeepLabV3 produce irregular edges, while EncNet, DNLNet, and DaNet introduce false detections. Segmenter not only suffers from false positives but also misidentifies connected regions as separate. Our method, however, significantly outperforms these issues, demonstrating exceptional performance in detecting subtle building features.
As illustrated in Table 1, our method achieves the highest scores in terms of the mIoU metric, surpassing UNet, DeepLabV3, EncNet, DaNet, DNLNet, and Segmenter by margins of 3.09%, 3.54%, 1.58%, 0.96%, 1.15%, and 0.89%, respectively. In the recall metric, our method also leads, with improvements of 3.21%, 5%, 3.34%, 1.74%, 1.7%, and 1.65% over the compared models. Similarly, in the F1-Score metric, our method shows the best performance, outperforming UNet, DeepLabV3, EncNet, DaNet, DNLNet, and Segmenter by 5.2%, 5.4%, 1.82%, 0.09%, 0.5%, and 0.43%, respectively. The results in Table 1 clearly demonstrate that our method surpasses the other six network models in mIoU, recall, and F1-score metrics, thereby corroborating the effectiveness of our approach.

3.4. Experiments and Result Analyses on the WHU Dataset

We further applied the network models, including PDBMFN, UNet, DeepLabV3, EncNet, DaNet, DnlNet, and Segmenter, to the WHU dataset. The visual results of the segmentation experiments are presented in Figure 7. As depicted, columns (a) and (b) present images characterized by large building targets, while columns (c) and (d) are dedicated to images with small targets. Column (e) showcases a diverse set of images, encompassing a spectrum of building sizes ranging from large to medium, small, and even minute targets. It is noteworthy that the images in column (b) not only include two instances of small target buildings but also a large target building with unique features, such as multiple small protrusions along its edges.
In column (a), which features buildings with circular boundaries, only DNLNet, Segmenter, and our PDBMFN method demonstrate accurate segmentation capabilities. While DNLNet and Segmenter are successful, they have their limitations; DNLNet does not match our method in terms of edge neatness and detection precision, and Segmenter shows gaps in the detection of buildings on the right side, as indicated by the red-marked areas. For the image presented in column (b), UNet and EncNet do not fully segment the large target building, and other methods such as DeepLabV3, DaNet, DNLNet, and Segmenter introduce false detections and noises. In stark contrast, our method achieves complete segmentation for both large and small targets, is devoid of noise, and presents well-defined segmentation of protruding edges for the large targets. In column (c), our method provides the closest match to the true label in the red-marked area, where other methods fall short and deliver incomplete results. For the scenario depicted in column (d), our method once again aligns most accurately with the true label, distinguishing two separate targets in the red-marked area, whereas other methods often misidentify them as a single entity. Finally, in column (e), the complexity of the image, with a mix of large, medium, small, and even tiny targets, leads to false positives across all methods, erroneously tagging non-buildings as buildings. However, in the red-marked area with multiple distinct targets, our method offers the most accurate approximation of the true label compared to the other methods. The competing methods either yield incomplete results, display unclear boundaries, or are plagued by a combination of these issues.
Table 2 presents a comparative analysis of the performance metrics, highlighting the superior results achieved by our method. In the mIoU metric, our method surpasses UNet, DeepLabV3, EncNet, DaNet, DNLNet, and Segmenter by margins of 2.59%, 0.96%, 0.48%, 0.20%, 0.33%, and 0.53%, respectively. When examining the precision metric, our method is narrowly surpassed only by DaNet. In the recall metric, our method secures the top position, outperforming the aforementioned models by 2.8%, 0.35%, 0.92%, 0.32%, 0.7%, and 0.4%, respectively. Furthermore, our method achieves the highest F1-Score, outperforming the others by 3.54%, 1.78%, 0.51%, 0.14%, 1.20%, and 0.54%, respectively. It can be observed from Table 2 that our method outperforms the other six network models in mIoU, recall, and F1-Score metrics, further validating the effectiveness of our approach.

3.5. Experiments and Result Analyses on the Massachusetts Dataset

Compared to the Inria Aerial dataset and the WHU dataset, the Massachusetts dataset is characterized by a smaller scale and introduces a unique set of challenges. These include a higher degree of noise, reduced image clarity, and the presence of smaller targets, which collectively contribute to a more demanding segmentation task. As a result, the performance of all evaluated models on the Massachusetts dataset tends to be lower than their performance on the other two datasets. The visual results of the segmentation experiments are displayed in Figure 8.
In Figure 8a, which highlights a typical building with distinct boundaries marked by the red box, our method uniquely achieves precise segmentation. While UNet successfully segments two buildings, it leaves incomplete fragments within the interiors, and other methods fall short of fully separating the two distinct parts. In column (b), our method yields results that more closely align with the true label, particularly regarding the edges of the large target building. Within the red box, our method also more accurately captures the size and boundaries of the small target building compared to other methods. In column (c), for the large target image enclosed by the red box, our method, along with DaNet, DNLNet, and Segmenter, effectively identifies its protruding edge portions. In contrast, other methods either fail to detect these details or incorrectly fragment the building. However, for several small targets, DaNet, DNLNet, and Segmenter demonstrate over-segmentation or adhesion issues. In column (d), for the large target image marked by the red box, our method and UNet provide results that more closely approximate the true label, whereas other methods exhibit noticeable divergences from the actual outcomes. Nonetheless, for a few small targets, the edges detected by UNet are less refined than those identified by our method. Finally, in column (e), for the multiple target buildings within the red box, our method offers a representation that is more faithful to the true situation in terms of both boundaries and regions. In contrast, other methods either overlook some targets or present incomplete results. The comparative analysis underscores the robustness and accuracy of our method in handling segmentation tasks, even in the presence of challenging factors such as noise, unclear imagery, and small targets.
As detailed in Table 3, our method secured the top position in terms of the mIoU metric, with improvements of 3.34%, 4.13%, 2.68%, 4.03%, 0.38%, and 2.05% over UNet, DeepLabV3, EncNet, DaNet, DNLNet, and Segmenter, respectively. Our method also demonstrated the best performance in the precision metric, outperforming UNet, DeepLabV3, EncNet, DaNet, DnlNet, and Segmenter by 4.39%, 0.84%, 1.68%, 3.28%, 0.02%, and 1.49%, respectively. Additionally, our method also demonstrated the best performance in the recall metric, outperforming UNet, DeepLabV3, EncNet, DaNet, DnlNet, and Segmenter by 1.61%, 1.73%, 0.80%, 0.74%, 0.27%, and 1.41%, respectively. Furthermore, our method achieved the highest F1-Score, exceeding UNet, DeepLabV3, EncNet, DaNet, DnlNet, and Segmenter by 2.96%, 1.31%, 1.23%, 1.97%, 0.15%, and 1.45%, respectively. As evident from Table 3, our method excelled in mIoU, precision, recall, and F1-score metrics compared to the other six network models, validating the effectiveness of our approach in the context of building segmentation.

3.6. Ablation Experiments

The proposed PDBMFN is primarily composed of three crucial modules: a branch structure that incorporates prior information, including superpixels and edge information; a PDCM module; and a MAM module. In this section, the ablation experiments are conducted to substantiate the individual contributions of these modules to the PDBMFN network model.
The experimental results are presented in Table 4. An analysis of the data reveals that the introduction of superpixel images into the original UNet, leveraging its regional prior information, led to an increase in mIoU by 0.59%, 0.66%, and 2.53% in the Inria Aerial, WHU, and Massachusetts datasets, respectively. Further integration of the PDCM module, which facilitates multi-scale parallel convolutions, yielded additional improvements in mIoU, with increments of 0.91%, 0.31%, and 0.06% across the same datasets. Building upon this foundation, the incorporation of the edge prior-guided branch encoding network, which seamlessly merges edge information derived from edge detection with the encoding information from the region prior-guided branch, followed by the PDCM module, led to further refinement of the mIoU, with increases of 0.27%, 0.94%, and 0.31% in the respective datasets. Ultimately, the introduction of the MAM module during the decoding phase, which orchestrates multiple rounds of feature aggregation at each layer, served to augment the network’s performance even more. The experimental results across the three datasets demonstrated that the incorporation of MAM led to an improvement in mIoU of 1.32%, 0.8%, and 0.44%, respectively, compared to the model without MAM. When compared to the baseline UNet model, our PDBMFN achieved a cumulative improvement of 3.09%, 2.59%, and 3.34% in the mIoU metric for the Inria Aerial, WHU, and Massachusetts datasets, respectively. These results underscore the synergistic impact of the PDBMFN’s modular design on the overall performance of the network in the task of building segmentation.
In summary, the introduction of prior information enhances the accuracy of the network’s segmentation edge results, while the PDCM and MAM modules further improve the network’s performance during encoding and decoding, effectively capturing building features.

4. Discussions and Conclusions

We propose the prior-guided dual branch multi-feature fusion network (PDBMFN), a novel approach to building segmentation in remote sensing images. This network is designed to overcome the limitations of traditional CNNs, which often underutilize critical prior information related to building regions and boundaries. Our solution integrates regional and edge-prior information into the encoding phase. Specifically, the original image and its superpixel segmentation image are fed into the region prior-guided branch of the PBN, which then merges with the features from the edge prior-guided branch of the PBN that takes the edge image as input during the downsampling phase of the encoding process. To address the shortcomings of traditional convolution and pooling operations, which may fail to accurately capture global pixel relationships and thus result in semantic information loss, we propose the Parallel Dilated Convolution Module (PDCM). This module leverages dilated convolutions across various scales, enabling the network to capture a broader semantic context within larger image regions. Furthermore, in response to the conventional decoding stage’s suboptimal feature fusion techniques, we introduce the multi-feature aggregation module (MAM). This module aggregates shallow and deep semantic information in a staged and iterative manner, enriching the feature representation and yielding more precise building features.
Our experiments on the WHU, Inria Aerial, and Massachusetts datasets confirm the effectiveness of PDBMFN, showcasing superior performance over other state-of-the-art semantic segmentation models, particularly in terms of the mIoU metric and other quantitative measures. Through meticulously designed ablation experiments on the three classical datasets, we have substantiated the enhancements in network segmentation performance conferred by the introduction of branch structures and the incorporation of prior information, as well as the PDCM and MAM modules. These experiments have not only validated the improvements but also provided the individual contributions of the network’s components.
Our PDBMFN addresses the limitations of CNNs in feature acquisition by introducing PBN with prior information, as well as PDCM and MAM, to enhance the network’s feature representation capabilities. This approach has achieved promising segmentation results. Recently, Wang et al. [48] and Ruan et al. [49] have proposed using visual state space (VSS) as an alternative to convolutions in CNNs and have constructed novel, efficient encoder-decoder segmentation networks, achieving excellent results. These works have provided us with new inspiration, and utilizing VSS to further enhance the feature representation capabilities and improve segmentation performance in the PDBMFN network structure will be an important direction for our future work. Furthermore, our validation of the method’s effectiveness has, to this point, been confined to three representative datasets. The generalizability of our approach to remote sensing image data that exhibits varying resolutions and building scene distributions necessitates additional empirical validation. Another future project will be directed towards assessing the adaptability and robustness of our method across a broader spectrum of datasets, ensuring its applicability to diverse real-world scenarios in remote sensing image analysis.

Author Contributions

Conceptualization, Y.W., P.Z. and D.Z.; methodology, Y.W.; software, Y.W.; validation, Y.W., P.Z. and F.W.; formal analysis, Y.W., P.Z., M.Z., S.G. and D.Z.; investigation, Y.W. and F.W.; resources, Y.W., P.Z. and F.W.; data curation, P.Z. and F.W.; writing—original draft preparation, Y.W. and P.Z.; writing—review and editing, Y.W., P.Z., F.W., M.Z., S.G. and D.Z.; visualization, Y.W., P.Z. and F.W.; supervision, M.Z., S.G. and D.Z.; project administration S.G. and D.Z.; funding acquisition, S.G. and D.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by National Natural Science Foundation of China (62102213, 62262056); Qinghai Provincial Natural Science Foundation of China (2023-ZJ-94); Natural Science Youth Foundation of Qinghai Province (No. 2023-ZJ-947Q); 2022 Annual Technological Innovation Project of Higher Education Institutions in Shanxi Province (491).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The Inria Aerial Dataset is provided by Maggiori et al. (https://project.inria.fr/aerialimagelabeling/ (accessed on 21 April 2024)); WHU Dataset provided by J. Liu et al. from Wuhan University. (http://gpcv.whu.edu.cn/data/building_dataset.html (accessed on 21 April 2024)); The Massachusetts Dataset was provided by Dr. Volodymyr Mnih. (https://www.cs.toronto.edu/~vmnih/data/ (accessed on 21 April 2024)).

Acknowledgments

We sincerely thank the authors of the three datasets mentioned above for providing free datasets.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhao, W.; Persello, C.; Stein, A. Building Instance Segmentation and Boundary Regularization from High-Resolution Remote Sensing Images. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium(IGARSS), Waikoloa Village, HI, USA, 26 September–2 October 2020; pp. 3916–3919. [Google Scholar]
  2. Aslam, R.W.; Shu, H.; Naz, I.; Quddoos, A.; Yaseen, A.; Gulshad, K.; Alarifi, S.S. Machine Learning-Based Wetland Vulnerability Assessment in the Sindh Province Ramsar Site Using Remote Sensing Data. Remote Sens. 2024, 16, 928. [Google Scholar] [CrossRef]
  3. Yu, A.; Quan, Y.; Yu, R.; Guo, W.; Wang, X.; Hong, D.; Zhang, H.; Chen, J.; Hu, Q.; He, P. Deep learning methods for semantic segmentation in remote sensing with small data: A survey. Remote Sens. 2023, 15, 4987. [Google Scholar] [CrossRef]
  4. Partovi, T.; Bahmanyar, R.; Kraus, T.; Reinartz, P. Building Outline Extraction Using a Heuristic Approach Based on Generalization of Line Segments. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 933–947. [Google Scholar] [CrossRef]
  5. Ghaffarian, S. Automatic Building Detection based on Supervised Classification using High Resolution Google Earth Images. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2014, XL-3, 101–106. [Google Scholar] [CrossRef]
  6. Dikmen, M.; Halici, U. A Learning-Based Resegmentation Method for Extraction of Buildings in Satellite Images. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2150–2153. [Google Scholar] [CrossRef]
  7. Yuan, J.; Wang, D.; Li, R. Remote Sensing Image Segmentation by Combining Spectral and Texture Features. IEEE Trans. Geosci. Remote Sens. 2014, 52, 16–24. [Google Scholar] [CrossRef]
  8. Abdullah, S.; Mohammad, A.; Lu, G. Automatic segmentation of LiDAR point cloud data at different height levels for 3D building extraction. In Proceedings of the IEEE International Conference on Multimedia and Expo Workshops (ICMEW), Chengdu, China, 14–18 July 2014; pp. 1–6. [Google Scholar]
  9. Noyel, G.; Angulo, J.; Jeulin, D. Morphological Segmentation of Hyperspectral Images. Image Anal. Stereol. 2020, 26, 101–109. [Google Scholar] [CrossRef]
  10. Yamashita, T.; Wester, D.; Tewes, M.; Young, J.; Lombardi, J. Distinguishing Buildings from Vegetation in an Urban-Chaparral Mosaic Landscape with LiDAR-Informed Discriminant Analysis. Remote Sens. 2023, 15, 1703. [Google Scholar] [CrossRef]
  11. Wu, Y.; Wang, X.; Chen, T.; Dou, Y. DA-ResNet: Dual-stream ResNet with attention mechanism for classroom video summary. Pattern Anal. Appl. 2024, 27, 32. [Google Scholar] [CrossRef]
  12. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical image computing and computer-assisted intervention(MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  13. Chen, L.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar] [CrossRef]
  14. Zhang, H.; Dana, K.; Shi, J.; Zhang, Z.; Wang, X.; Tyagi, A.; Agrawal, A. Context Encoding for Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition(CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7151–7160. [Google Scholar]
  15. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition(CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7794–7803. [Google Scholar]
  16. Yin, M.; Yao, Z.; Cao, Y.; Li, X.; Zhang, Z.; Lin, S.; Hu, H. Disentangled non-local neural networks. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020. [Google Scholar]
  17. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual Attention Network for Scene Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 3141–3149. [Google Scholar]
  18. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  19. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar] [CrossRef]
  20. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021; pp. 7262–7272. [Google Scholar]
  21. Zhang, X.; Zeng, H.; Guo, S.; Zhang, L. Efficient long-range attention network for image super-resolution. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 649–667. [Google Scholar]
  22. Yu, W.; Luo, M.; Zhou, P.; Si, C.; Zhou, Y.; Wang, X.; Feng, J.; Yan, S. MetaFormer is Actually What You Need for Vision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 10809–10819. [Google Scholar]
  23. Lin, T.; Wang, Y.; Liu, X.; Qiu, X. A survey of transformers. AI Open 2022, 3, 111–132. [Google Scholar] [CrossRef]
  24. Zheng, Z.; Zhong, Y.; Wang, J.; Ma, A. Foreground-Aware Relation Network for Geospatial Object Segmentation in High Spatial Resolution Remote Sensing Imagery. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 4095–4104. [Google Scholar]
  25. Hu, G.; Yang, C.; Xu, L.; Shang, H.; Wang, Z.; Qin, Z. Improved U-Net remote sensing image semantic segmentation method. Acta Geod. Cartogr. Sin. 2023, 52, 980–989. [Google Scholar]
  26. Qian, Z.; Cao, Y.; Shi, Z.; Qiu, L.; Shi, C. A Semantic Segmentation Method for Remote Sensing Images based on Deeplab v3. In Proceedings of the International Conference on Big Data & Artificial Intelligence & Software Engineering (ICBASE), Zhuhai, China, 24–26 September 2021; pp. 396–400. [Google Scholar]
  27. Benchabana, A.; Kholladi, M.K.; Bensaci, R.; Khaldi, B. Building detection in high-resolution remote sensing images by enhancing superpixel segmentation and classification using deep learning approaches. Buildings 2023, 13, 1649. [Google Scholar] [CrossRef]
  28. Wang, L.; Fang, S.; Meng, X.; Li, R. Building Extraction With Vision Transformer. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5625711. [Google Scholar] [CrossRef]
  29. Cui, L.; Jing, X.; Wang, Y.; Huan, Y.; Xu, Y.; Zhang, Q. Improved Swin Transformer-Based Semantic Segmentation of Postearthquake Dense Buildings in Urban Areas Using Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 369–385. [Google Scholar] [CrossRef]
  30. Liu, Z.; Mao, H.; Wu, C.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11966–11976. [Google Scholar]
  31. Li, D.; Li, L.; Chen, Z.; Li, J. Shift-ConvNets: Small Convolutional Kernel with Large Kernel Effects. arXiv 2024, arXiv:2401.12736. [Google Scholar] [CrossRef]
  32. Ding, X.; Zhang, Y.; Ge, Y.; Zhao, S.; Song, L.; Yue, X.; Shan, Y. Unireplknet: A universal perception large-kernel convnet for audio, video, point cloud, time-series and image recognition. arXiv 2023, arXiv:2311.15599. [Google Scholar] [CrossRef]
  33. Isensee, F.; Paul, F.; Jaeger, S.; Jens, P.; Klaus, H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef]
  34. Yang, J.; Marcus, D.; Sotiras, A. Dynamic u-net: Adaptively calibrate features for abdominal multi-organ segmentation. arXiv 2024, arXiv:2403.07303. [Google Scholar] [CrossRef]
  35. Qiu, P.; Yang, J.; Kumar, S.; Ghosh, S.S.; Sotiras, A. AgileFormer: Spatially Agile Transformer UNet for Medical Image Segmentation. arXiv 2024, arXiv:2404.00122. [Google Scholar] [CrossRef]
  36. Daranagama, S.; Witayangkurn, A. Automatic building detection with polygonizing and attribute extraction from high-resolution images. ISPRS Int. J. Geo-Inf. 2021, 10, 606. [Google Scholar] [CrossRef]
  37. Pan, Z.; Xu, J.; Guo, Y.; Hu, Y.; Wang, G. Deep Learning Segmentation and Classification for Urban Village Using a Worldview Satellite Image Based on U-Net. Remote Sens. 2020, 12, 1574. [Google Scholar] [CrossRef]
  38. Su, Z.; Li, W.; Ma, Z.; Gao, R. An improved U-Net method for the semantic segmentation of remote sensing images. Appl. Intell. 2022, 52, 3276–3288. [Google Scholar] [CrossRef]
  39. Lei, J.; Liu, X.; Yang, H.; Zeng, Z.; Feng, J. Dual Hybrid Attention Mechanism-Based U-Net for Building Segmentation in Remote Sensing Images. Appl. Sci. 2024, 14, 1293. [Google Scholar] [CrossRef]
  40. Ahmadi, S.A.; Mohammadzadeh, A.; Yokoya, N.; Ghorbanian, A. BD-SKUNet: Selective-kernel UNets for building damage assessment in high-resolution satellite images. Remote Sens. 2024, 16, 182. [Google Scholar] [CrossRef]
  41. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC Superpixels Compared to State-of-the-Art Superpixel Methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef]
  42. Dari, N.W. Identifikasi Deteksi Tepi Pada Pola Wajah Menerapkan Metode Sobel, Roberts dan Prewitt. Bull. Inf. Technol. 2022, 3, 85–91. [Google Scholar]
  43. Woo, S.; Park, J.; Lee, J.; Kweon, I. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  44. Emmanuel, M.; Yuliya, T.; Guillaume, C.; Pierre, A. Can Semantic Labeling Methods Generalize to Any City? The Inria Aerial Image Labeling Benchmark. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium(IGARSS), Fort Worth, TX, USA, 23–28 July 2017; pp. 3226–3229. [Google Scholar]
  45. Guo, S.; Liu, L.; Gan, Z.; Wang, Y.; Zhang, W.; Wang, C.; Jiang, G.; Zhang, W.; Yi, R.; Ma, L.; et al. Isdnet: Integrating shallow and deep networks for efficient ultra-high resolution segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 4361–4370. [Google Scholar]
  46. Liu, J.; Ji, S. A novel recurrent encoder-decoder structure for large-scale multi-view stereo reconstruction from an open aerial dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 6050–6059. [Google Scholar]
  47. Mnih, V. Machine Learning for Aerial Image Labeling. Ph.D. Thesis, University of Toronto (Canada), Toronto, ON, Canada, 2013. [Google Scholar]
  48. Wang, Z.; Zheng, J.; Zhang, Y.; Cui, G.; Li, L. Mamba-unet: Unet-like pure visual mamba for medical image segmentation. arXiv 2024, arXiv:2402.05079. [Google Scholar] [CrossRef]
  49. Ruan, J.; Xiang, S. VM-UNet: Vision Mamba UNet for Medical Image Segmentation. arXiv 2024, arXiv:2402.02491. [Google Scholar] [CrossRef]
Figure 1. The architecture of PDBMFN.
Figure 1. The architecture of PDBMFN.
Buildings 14 02006 g001
Figure 2. The results of SLIC superpixel segmentation with different K values.
Figure 2. The results of SLIC superpixel segmentation with different K values.
Buildings 14 02006 g002
Figure 3. The comparison of edge detection results between Robert, Prewitt, and Sobel.
Figure 3. The comparison of edge detection results between Robert, Prewitt, and Sobel.
Buildings 14 02006 g003
Figure 4. The architecture of Parallel Dilated Convolution Module (PDCM).
Figure 4. The architecture of Parallel Dilated Convolution Module (PDCM).
Buildings 14 02006 g004
Figure 5. The architecture of Multi-feature Aggregation Module (MAM).
Figure 5. The architecture of Multi-feature Aggregation Module (MAM).
Buildings 14 02006 g005
Figure 6. Comparison of experimental results on typical images from the Inria Aerial dataset.
Figure 6. Comparison of experimental results on typical images from the Inria Aerial dataset.
Buildings 14 02006 g006
Figure 7. Comparison of experimental results on the typical images from the WHU dataset.
Figure 7. Comparison of experimental results on the typical images from the WHU dataset.
Buildings 14 02006 g007
Figure 8. Comparison of experimental results on the typical images from the Massachusetts dataset.
Figure 8. Comparison of experimental results on the typical images from the Massachusetts dataset.
Buildings 14 02006 g008
Table 1. Comparison of accuracy metrics among different methods on the typical images from the Inria Aerial Dataset.
Table 1. Comparison of accuracy metrics among different methods on the typical images from the Inria Aerial Dataset.
MethodsmIoU (%)Precision (%)Recall (%)F1-Score (%)
UNet83.0381.0282.6681.83
DeepLabV382.5882.4280.8781.63
EncNet84.5488.0882.5385.21
DaNet85.1689.9584.1386.94
DnlNet84.9789.0384.1786.53
Segmenter85.2389.1284.2286.60
Ours86.1288.2385.8787.03
Table 2. Comparison of accuracy metrics among different methods on the typical images from the WHU dataset.
Table 2. Comparison of accuracy metrics among different methods on the typical images from the WHU dataset.
MethodsmIoU (%)Precision (%)Recall (%)F1-Score (%)
UNet90.1691.5590.1590.84
DeepLabV391.7992.6092.6092.60
EncNet92.2795.8092.0393.87
DaNet92.5595.9192.6394.24
DnlNet92.4294.1492.2593.18
Segmenter92.2295.1892.5593.84
Ours92.7595.8692.9594.38
Table 3. Comparison of accuracy metrics among different methods on the typical images from the Massachusetts dataset.
Table 3. Comparison of accuracy metrics among different methods on the typical images from the Massachusetts dataset.
MethodsmIoU (%)Precision (%)Recall (%)F1-Score (%)
UNet75.1277.1676.2376.69
DeepLabV374.3780.7176.1178.34
EncNet75.7879.8777.0478.42
DaNet74.4378.2777.1077.68
DnlNet78.0881.5377.5779.50
Segmenter76.4180.0676.4378.20
Ours78.4681.5577.8479.65
Table 4. Comparison of the effect of sequentially adding different modules on the network performance.
Table 4. Comparison of the effect of sequentially adding different modules on the network performance.
MethodsChoicesMetrics (mIoU)
SuperPixelPDCM EdgeMAMInria AerialWHUMassachusetts
Baseline (UNet)××××83.0390.1675.12
+SuperPixel×××83.6290.7077.65
+PDCM××84.5391.0177.71
+Edge×84.8091.9578.02
+MAM (Ours)86.1292.7578.46
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, Y.; Zhao, P.; Wang, F.; Zhou, M.; Geng, S.; Zhang, D. A Prior-Guided Dual Branch Multi-Feature Fusion Network for Building Segmentation in Remote Sensing Images. Buildings 2024, 14, 2006. https://doi.org/10.3390/buildings14072006

AMA Style

Wu Y, Zhao P, Wang F, Zhou M, Geng S, Zhang D. A Prior-Guided Dual Branch Multi-Feature Fusion Network for Building Segmentation in Remote Sensing Images. Buildings. 2024; 14(7):2006. https://doi.org/10.3390/buildings14072006

Chicago/Turabian Style

Wu, Yingbin, Peng Zhao, Fubo Wang, Mingquan Zhou, Shengling Geng, and Dan Zhang. 2024. "A Prior-Guided Dual Branch Multi-Feature Fusion Network for Building Segmentation in Remote Sensing Images" Buildings 14, no. 7: 2006. https://doi.org/10.3390/buildings14072006

APA Style

Wu, Y., Zhao, P., Wang, F., Zhou, M., Geng, S., & Zhang, D. (2024). A Prior-Guided Dual Branch Multi-Feature Fusion Network for Building Segmentation in Remote Sensing Images. Buildings, 14(7), 2006. https://doi.org/10.3390/buildings14072006

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop