Next Article in Journal
Aerial-Ground Cross-View Vehicle Re-Identification: A Benchmark Dataset and Baseline
Previous Article in Journal
Mechanisms of Ocean Acidification in Massachusetts Bay: Insights from Modeling and Observations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GOFENet: A Hybrid Transformer–CNN Network Integrating GEOBIA-Based Object Priors for Semantic Segmentation of Remote Sensing Images

1
Southern Marine Science and Engineering Guangdong Laboratory (Guangzhou), Guangzhou 511458, China
2
Ocean College, Zhejiang University, Zhoushan 316021, China
3
State Key Laboratory of Satellite Ocean Environment Dynamics, Second Institute of Oceanography, Ministry of Natural Resources, Hangzhou 310012, China
4
School of Oceanography, Shanghai JiaoTong University, Shanghai 200030, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(15), 2652; https://doi.org/10.3390/rs17152652 (registering DOI)
Submission received: 18 June 2025 / Revised: 20 July 2025 / Accepted: 28 July 2025 / Published: 31 July 2025

Abstract

Geographic object-based image analysis (GEOBIA) has demonstrated substantial utility in remote sensing tasks. However, its integration with deep learning remains largely confined to image-level classification. This is primarily due to the irregular shapes and fragmented boundaries of segmented objects, which limit its applicability in semantic segmentation. While convolutional neural networks (CNNs) excel at local feature extraction, they inherently struggle to capture long-range dependencies. In contrast, Transformer-based models are well suited for global context modeling but often lack fine-grained local detail. To overcome these limitations, we propose GOFENet (Geo-Object Feature Enhanced Network)—a hybrid semantic segmentation architecture that effectively fuses object-level priors into deep feature representations. GOFENet employs a dual-encoder design combining CNN and Swin Transformer architectures, enabling multi-scale feature fusion through skip connections to preserve both local and global semantics. An auxiliary branch incorporating cascaded atrous convolutions is introduced to inject information of segmented objects into the learning process. Furthermore, we develop a cross-channel selection module (CSM) for refined channel-wise attention, a feature enhancement module (FEM) to merge global and local representations, and a shallow–deep feature fusion module (SDFM) to integrate pixel- and object-level cues across scales. Experimental results on the GID and LoveDA datasets demonstrate that GOFENet achieves superior segmentation performance, with 66.02% mIoU and 51.92% mIoU, respectively. The model exhibits strong capability in delineating large-scale land cover features, producing sharper object boundaries and reducing classification noise, while preserving the integrity and discriminability of land cover categories.

1. Introduction

Advancements in sensor technology have propelled the development of high-resolution remote sensing imagery. Owing to its rich detail and spatial information, such imagery is widely used in ecological monitoring [1], land cover mapping [2,3], change detection [4], and urban planning [5]. Extracting meaningful information and identifying target objects from remote sensing images, or the intelligent interpretation of such images, remains a key research focus. Image segmentation, as a subtask of intelligent interpretation, aims to extract the desired objects from remote sensing imagery. Traditional methods in remote sensing image analysis, exemplified by object-based techniques (GEOBIA), have been widely used over recent years [6,7]. These algorithms can be broadly categorized into threshold-based [8], edge-based [9], region-based [10,11], and hybrid approaches [12]. The core principle of these algorithms is to partition the image into numerous high-homogeneity segments with high heterogeneity between segments. However, a key limitation is that these geographic entities, as components of semantic information, lack intrinsic semantic meaning [13]. While multi-scale optimization segmentation algorithms [14,15] have improved the semantic relevance of segments and created more meaningful geographic entities (geo-objects), many segments still remain in their original state, exhibiting significant over-segmentation [16]. Consequently, traditional methods face substantial challenges in optimizing segmentation parameters and addressing the complexity and large scale of remote sensing images.
The rapid advancement of deep learning in recent years has resulted in substantial breakthroughs in image segmentation, particularly through the application of convolutional neural networks (CNNs). The introduction of Fully Convolutional Networks (FCNs) [17] marked a historic breakthrough in semantic segmentation tasks. Building on this, ResNet [18] demonstrated powerful feature extraction capabilities through residual connections in deep convolutional networks. Concurrently, UNet [19] addressed the issue of spatial information loss caused by successive convolutions by employing skip connections for feature fusion. Network architectures based on ResNet or incorporating UNet’s skip connections have become widely adopted in semantic segmentation tasks [20,21].
Remote sensing imagery poses unique challenges for semantic segmentation arising from the intricate diversity of land cover types (intra-class spectral variability and inter-class spectral similarity) and scale variability. While CNNs leverage local connectivity and weight sharing to exhibit locality and translation invariance—thereby reducing model parameters and enhancing feature learning efficiency—they still have limited capability in capturing global semantic information [22]. On the one hand, DeepLabv3+ [23] and PSPNet [24] expand the receptive field to obtain multi-scale global context, but their global information is still constrained by local regions. On the other hand, attention mechanisms such as DANet [25], which uses parallel channel attention mechanism (CAM) and spatial attention mechanism (SAM), and CBAM [26], which serially combines CAM and SAM, offer efficient approaches for modeling global information. Unfortunately, they are constrained by the number of modules and computational overhead. Therefore, there is an urgent need for a purely attention-based architecture to facilitate comprehensive global information extraction.
Transformers, renowned in natural language processing for their ability to capture global relationships, offer a novel approach to semantic segmentation [27]. The Vision Transformer (ViT) was the first to employ a pure Transformer architecture for image recognition tasks [28]. Pretrained on large datasets, ViT slightly outperforms CNN-based models, demonstrating the Transformer model’s robust feature extraction capabilities and effectiveness in image processing. Zheng et al. introduced the first semantic segmentation network with a pure Transformer backbone, SETR [29]. Subsequently, Liu et al. [30] introduced the Swin Transformer, a hierarchical Transformer architecture that applies self-attention within small windows and establishes cross-window connections through shifted windows, significantly enhancing the efficiency of self-attention calculations. Additionally, the hierarchical nature of the Swin Transformer can be selectively integrated into existing semantic segmentation models. While Transformers effectively capture long-range global features, local feature information is often overlooked. Therefore, creating a model that leverages both Transformer and CNN advantages is crucial for effectively segmenting remote sensing images, allowing for the capture of both global context and fine spatial details.
To achieve the goal of intelligent interpretation of remote sensing images, we propose GOFENet (Geospatial Object Feature Enhancement Network), a hybrid semantic segmentation network that integrates CNN and Transformer architectures. This approach effectively addresses the limitations of CNNs in capturing global context and the tendency of Transformer models to overlook fine-grained details. Built upon a ResNet backbone and integrated with Swin Transformer blocks, the proposed network leverages UNet-style skip connections for hierarchical feature fusion. While multi-scale segmentation often produces objects at varying meaningful levels, it effectively enhances pixel connectivity and global context representation. To leverage this, we introduce an auxiliary encoding branch that extracts high-level structural cues from GEOBIA-derived segmentation maps. The experimental results on three benchmark datasets demonstrate that GOFENet achieves outstanding segmentation performance, exhibiting excellent capabilities in extracting both local and global information.
The main contributions of this study are summarized as follows:
(1) A hybrid encoder architecture with feature enhancement modules. GOFENet combines the local detail modeling strength of CNNs with the long-range dependency modeling of Swin Transformers. To further enhance feature representation, we introduce a cross-channel selection module (CSM), which applies channel-wise attention to highlight informative features and suppress irrelevant ones. Additionally, we design a feature enhancement module (FEM) that incorporates both channel and spatial attention mechanisms to integrate global contextual cues into CNN features. Together, these modules refine the feature representations and improve both the accuracy and robustness of the segmentation model.
(2) Incorporation of object-level priors from GEOBIA into a DL framework. We introduce an auxiliary encoder that employs cascaded atrous convolutions to extract global structural information from GEOBIA-derived segmentation maps. A shallow–deep feature fusion module (SDFM) is then designed to integrate feature representations from both the main and auxiliary encoders at aligned semantic levels. This design enables the network to effectively combine multi-scale object-level and pixel-level information, thereby enhancing its capability for scene understanding and contextual reasoning.

2. Related Work

2.1. Semantic Segmentation of Remote Sensing Images Based on CNN, Transformer, and Hybrid Models

2.1.1. CNN-Based Semantic Segmentation Models

Techniques such as multi-scale feature extraction, expanding network receptive fields, integrating local and global contexts, and enhancing global or detailed features have proven effective in improving segmentation performance on remote sensing images [31,32,33,34]. However, high-resolution remote sensing images often contain intricate patterns and man-made structures. CNN-based segmentation networks, which primarily extract local semantic features, exhibit limitations in modeling global information across the entire image. Consequently, relying solely on local information can make it challenging to accurately identify and segment these complex patterns and objects. To overcome this limitation, integrating global contextual understanding is crucial for improving segmentation accuracy in such intricate settings.

2.1.2. Transformer-Based Semantic Segmentation Model

Leveraging its unique self-attention mechanism, the Transformer architecture effectively captures long-range dependencies and excels in handling both the details and structure of images, rendering it especially promising for remote sensing applications. Transformer-based encoders and decoders, collectively referred to as pure Transformer structures, include notable models such as Segmenter [35], SegFormer [36], and SwinUNet [37]. These models have demonstrated significant advancements in image segmentation applications for remote sensing data. While Transformers are adept at global modeling, they often exhibit inadequate localization capabilities, which can undermine segmentation accuracy in scenarios requiring precise spatial information [38]. This lack of fine-grained spatial resolution is a critical challenge, as accurate segmentation of complex remote sensing images often demands both robust global context understanding and detailed local feature extraction.

2.1.3. CNN–Transformer Hybrid Model

Researchers have shown significant interest in combining CNNs’ proficiency in extracting local features with Transformers’ capability to capture long-range dependencies and model sequential images to create synergistic network architectures. For instance, DC-Swin employs the Swin Transformer as the encoder and adopts a convolutional decoder with dense connectivity specifically designed for the segmentation of high-resolution remote sensing imagery, achieving superior performance compared to CNN-based methods [39]. Wang et al. [40] proposed a dual-path perception network that integrates a ResT-based dependency path with a CNN-based texture path, enabling the simultaneous extraction of global contextual information and fine-grained local details. In many studies, CNNs and Transformers are configured as parallel branches [41,42,43], forming dual-encoder frameworks that effectively fuse global semantics with spatial detail. These approaches underscore the potential of hybrid architectures to harness the complementary strengths of CNNs and Transformers, enhancing segmentation accuracy in complex remote sensing scenarios. By contrast, UNetFormer [44] adopts a ResNet encoder paired with a Transformer-based decoder, also demonstrating promising segmentation performance.

2.2. Deep Learning Framework Coupled with GEOBIA

Image segmentation and classification constitute the core processes of GEOBIA, where segments contain rich spectral, textural, morphological, and remote sensing index attributes that facilitate the recognition of complex targets in remote sensing imagery. However, the irregular shapes of these segments pose significant challenges for their integration into deep learning frameworks [45]. Based on an extensive literature review, existing integration strategies can be broadly categorized as follows (see Figure 1):
(a) Integration with CNN-based Segmentation Results. This approach refines or merges segmentation outcomes produced by CNNs [46,47,48]. The major drawback of this method is the requirement for manually constructed rule sets.
(b) Object-based patch classification. In this method, irregular objects are expanded into fixed-size rectangular patches and classified using CNNs or Transformer models [49]. Two padding strategies are commonly used: inserting placeholder values (e.g., 0 or 255) or incorporating contextual information surrounding the object. The latter has been shown to improve classification accuracy [50,51]. Nevertheless, this method is highly dependent on segmentation accuracy, with under-segmentation potentially leading to significant performance degradation [52].
(c) Deep feature extraction from object attributes. Deep learning models are utilized to obtain deep and abstract features from object attributes [53,54,55]. However, such representations may disregard spatial relationships. To address this, recent work has incorporated spatial context into tabular models to reduce misclassification errors [56], or combined tabular attributes with deep features within dual-branch architectures [55].
(d) Heterogeneous patch filtering. This strategy involves preprocessing image patches—such as applying variance-based filtering—prior to CNN classification, aiming to enhance model input quality and downstream performance [50,51,57].

2.3. Attention Mechanism

Attention mechanisms enhance CNNs’ ability to capture global information by using scene- or image-level data. CAMs assess and weight the importance of each channel to enhance key features [31,58], while SAMs focus on significant spatial locations via pixel-wise weighting [25,26]. Some approaches combine both to form pixel-level global attention maps [32,43,44], and others leverage cross-attention to strengthen interactions between different feature levels or modalities [59].
The self-attention module was first introduced to computer vision by Wang et al. in a non-local fashion to extract long-range dependencies between features, known as the dot-product attention mechanism [60,61]. This concept is frequently utilized in network designs for remote sensing images [58,62]. However, its computational complexity increases quadratically with spatial resolution, resulting in a substantial rise in computational cost. To mitigate the memory consumption, approaches such as sparse attention [63,64], linear attention [65], and query-key value reduction strategies [66,67] have been proposed. Transformers extend self-attention to image patch sequences, enabling global context modeling for vision tasks.

3. Method

3.1. Multi-Scale Segmentation Optimization Algorithm—EIODA

The Edge-guided Image Object Detection Approach (EIODA) is a local-scale optimization segmentation algorithm within multi-scale optimization methods. Its key innovation lies in considering both spectral and edge information from images [14]. Specifically, during the segment growth phase, the algorithm performs multi-scale analysis while leveraging edge information as a constraint to determine the optimal segmentation stage for segments. This method aims to find the optimal segmentation of real objects of various scales in the image. The algorithm comprises four main stages: First, edge detection algorithms are used to extract the image edges. Next, a segmentation algorithm, typically Multi-Resolution Segmentation (MRS) [10], is employed for pre-segmentation of high-resolution remote sensing images, ensuring highly homogeneous over-segmented objects. Then, an R-tree and a region adjacency graph (RAG) are constructed to store the edge information and the over-segmented image objects, respectively. The final and critical step involves calculating edge completeness curves during the image growth process to identify the optimal segmentation scale. An example of EIODA segmentation is shown in Figure 2. The original image (Figure 2a) is segmented using EIODA to obtain the segmentation result (Figure 2b), and then the spectral mean values of the RGB channels in each segment are calculated to generate the segmented image (Figure 2c). This integration of spectral and edge information enhances the detection capability and reliability of geo-objects [16].

3.2. Overall Network Structure

The proposed GOFENet architecture is a typical encoder–decoder structure composed of three main components: the primary encoder, the auxiliary encoder, and the decoder. Overall, the network integrates CNN and Swin Transformer. It incorporates the effective design of UNet, utilizing skip connections between the encoder and the decoder to transfer low-level features. To preserve the network’s ability to extract global context information, Swin Transformer basic blocks (SWTB) are added, as shown in Figure 3. Notably, we also introduce an auxiliary branch based on optimal segmentation results. This branch employs a series of dilated convolutions to capture object-based image features, thereby enhancing the network’s multi-scale and global perceptual capabilities. The design of the auxiliary encoder will be explored in the Section 5. Additionally, we have developed CSMs and FEMs to further enhance the performance of CNN. The overall architecture of GOFENet is depicted in Figure 4.
Unlike the conventional ResNet50 main encoder, our approach employs CSM before each down-sampling operation (except for the first layer). It strengthens the relationships between channels and improves the model’s capacity to capture and express critical information. After the first ResBlock in the original remote sensing image R h × w × 3 , the resolution of the feature map is halved, and the number of channels is increased to 64. The output feature map of the n-th ResBlock, where n 2 and C1 = 128, is denoted as T n R h / 2 n × w / 2 n × 2 n 1 C 1 . The SWTB, which is central to the Swin Transformer architecture, is introduced in the third, fourth, and fifth layers of the backbone network. For efficient modeling, Swin Transformer employs both the conventional window-based multi-head self-attention (W-MSA) and shifted window-based multi-head self-attention (SW-MSA) mechanisms. W-MSA partitions the feature map into non-overlapping windows of size S × S and computes self-attention within these local windows. SW-MSA, on the other hand, improves inter-window information flow by offsetting the windows by half their size ([S/2]), and then computing attention within these shifted windows. In the experiment, the window size used is S = 7. In the SWTB, W-MSA and SW-MSA are connected in sequence and executed alternately, with the computation process as follows:
t ^ l = W - MSA ( LN ( t l 1 ) ) + t l 1 t l = MLP ( LN ( t ^ l ) ) + t ^ l t ^ l + 1 = SW - MSA ( LN ( t l ) ) + t l t l + 1 = MLP ( LN ( t ^ l + 1 ) ) + t ^ l + 1
where tl−1 signifies the input feature of the SWTB and tl+1 indicates the output feature of the SWTB. The feature output size of the SWTB remains consistent with the input size, denoted as W n ; the size of feature map is W n     R h / 2 n × w / 2 n × 2 n 3 C 2 , where n  3, C2 = 512. Additionally, the FEM, guided by SWTB, is designed to enhance the representativeness and effectiveness of the low-level features from the CNN through skip connections by highlighting the importance of CNN channels.
In the auxiliary branch, after the original remote sensing image R h × w × 3 is processed by the EIODA segmentation algorithm, a segmented image Z h × w × 3 is obtained. The auxiliary encoder shares the weights of the first layer with the main encoder. It is then processed through three dilated convolutions. The output of the n-th stage can be represented as S n     Z h / 2 n × w / 2 n × 2 n 3 C 3 , where n  2, C3 = 128.
After the encoding stage, the low-level features— T n processed by the FEM and S n from the auxiliary branch network—are fused by summation. These fused features are then up-sampled to increase resolution and passed through 3 × 3 convolutional layers to decrease the number of channels. Each convolutional layer is succeeded by batch normalization and ReLU activation. This procedure is repeated four times, gradually expanding the feature map F to F     R h / 2 × w / 2 × 64 . Ultimately, bilinear interpolation up-sampling and two additional 3 × 3 convolutional layers are implemented to the F to produce the final predicted mask.

3.3. Cross-Channel Selective Module

While ResNet networks excel at extracting deep-level features from images, they do not adequately address multi-scale and contextual information. This is essential for the semantic segmentation of remote sensing imageries, which often display complex features and intra-class variability. Figure 5 depicts the structure of the CSM, and its computational process can be expressed as follows:
Y   =   X + X Softmax ( © { CA m ( GConv m ( X m ) ) } ) m [ 1 , 4 ]
X denotes the input features and X represents the output features. CAm refers to the channel attention block, while GConvm and X m denote the m-th group convolution block and feature input, respectively. The © means concatenation on the channel dimension, and ʘ stands for element-wise product. First, the feature map X obtained from the Resblock is divided into channels according to a specified ratio. In this work, an equal division is employed, with the ratio r1:r2:r3:r4 = 1:1:1:1, resulting in feature maps X 1 , X 2 , X 3 , and X 4 . Next, each of these feature maps undergoes group convolutions with different numbers of groups (g1, g2, g3, g4 = 1, 4, 8, 16) and kernel sizes (3 × 3, 5 × 5, 7 × 7, 9 × 9), producing outputs X 1 , X 2 , X 3 , and X 4 , respectively. Subsequently, each X m is processed through a channel attention block that uses adaptive-sized 1D convolutional kernels to improve cross-channel communication within each group. The adaptive convolutional kernel size K s is calculated using the following formula:
K s = log 2 C + b γ o d d
where C represents the number of input channels, | t | o d d denotes the nearest odd number to t. b is the bias coefficient, and γ is the scaling factor, with default values of 1 and 2, respectively. After concatenating the obtained X ~ m along the channel dimension, the softmax function is utilized to boost inter-group correlation. Finally, the channel weights are multiplied by the original feature matrix X to generate the attention map, which is then added to X to produce the output feature Y.

3.4. Feature Enhancement Module

In terms of the overall network architecture, a dual-branch structure with both CNN decoders and Transformer decoders is not utilized, as this would significantly increase the model complexity. Instead, the SWTB is integrated into the intermediate and deep layers of the network to ensure the preservation of global information. It is worth noting that the features of skip connections are not enhanced by CSM, as this may interfere with the capability of the SWTB to capture global information. To effectively apply global contextual features to the channel attention mechanism of CNN feature maps, we designed the FEM. It emphasizes the channels crucial for capturing global semantic information while suppressing irrelevant features. This design improves the model’s generalization ability and robustness across different data distributions. The structure is illustrated in Figure 6. As previously mentioned, Tn and Wn denote the outputs of the n-th layer Resblock and SWTB, respectively. FEM adopts a dual-path structure to enhance feature representations by capturing both spatial and channel dependencies. The spatial path applies a 3 × 3 dilated convolution followed by directional pooling to extract horizontal and vertical contexts, which are fused into a spatial attention map. The channel path reshapes the input and employs Swin Transformer blocks to extract global context, followed by mean and std pooling across spatial dimensions. The pooled statistics are concatenated and passed through a group-wise 1D convolution and sigmoid to generate channel attention. The two attention maps are fused and applied to the input via element-wise multiplication to refine features. This entire process can be encapsulated by the following equation:
A s = Conv 1 × 1 ( P h ( T n ) P v ( T n ) )
A c = δ ( Conv 1 D group = c ( © { μ ( W n ) , σ ( W n ) } ) )
T n = T n ( A s A c )
where P h ·   and P v ( · ) are directional pooling along horizontal and vertical axes, μ ·   and σ ( · ) denote mean and standard deviation pooling. T n denotes the output features after passing through FEM, GConv1D signifies the 1D grouped convolution with a kernel size of 2 × 2, and δ represents the sigmoid function.

3.5. Shallow–Deep Feature Fusion Module

Object-based image features aggregate similar pixels, which simplifies remote sensing images by describing meaningful segments and reflecting the overall properties of the image. However, this approach has the drawback of losing detailed information. Since the overall information is obtained directly from the raw image, we consider it to be a form of shallow image features. To effectively merge pixel-based features (local features) with object-based features (global features), the design of the SDFM is illustrated in Figure 7. As depicted in Figure 4, the channel count in the main encoder at each layer (excluding the first layer) is four times that of the auxiliary encoder. Therefore, to enable the addition of deep features (DF) and shallow features (SF), the DF need to undergo channel reduction via a 1 × 1 convolution. The fused features obtained by adding DF and SF are referred to as additive features (AF). The pixel attention module (PAM) comprises two components. First, the fused features undergo maximum and average pooling along the channel dimension to capture spatial statistics. Second, adaptive average pooling is applied to generate channel-wise attention. These channel and spatial attentions are then combined to produce the attention map (AM). This process is mathematically expressed by the following formula:
A M = MLP ( AvgPool ( A F ) ) Conv 7 ( { GAP ( A F ) , GMP ( A F ) } )
where MLP denotes the fully connected layer, while GAP and GMP represent the calculation of the mean and maximum values of pixel space along the channel dimension, respectively. The AF and AM are cross-arranged along the channel dimension. Pixel-wise attention maps (PM) are then generated by applying a 7 × 7 grouped convolution followed by a sigmoid function. The PM is then multiplied by the SF and DF using the weight factor α and summed to produce the fused features (MF), as described in Equation (6). The channels are subsequently restored to the size of DF using a 1 × 1 convolution. The default value for α is 0.5.
M F = A F + α × P M S F + ( 1 α ) × P M D F

4. Results

4.1. Datasets

4.1.1. GID

The Gaofen Image Dataset (GID) is a large-scale land cover classification dataset [68], which includes 150 high-quality GF-2 images collected from over 60 different cities in China. It contains a fine land cover classification (FLC) dataset, which comprises 15 annotated classes known as GID15, with each image measuring 6800 × 7200 pixels. Currently, only 110 annotated images from the training and validation sets are publicly available, and the official test server is inaccessible. As a result, the dataset is partitioned into training, validation, and test sets with a ratio of 9:1:1.

4.1.2. LoveDA Dataset

The LoveDA dataset consists of 5987 high-resolution images, categorized into 2522 training images, 1669 validation images, and 1796 test images [69]. These images are collected from both urban and rural environments, with each image sized at 1024 × 1024 pixels. The dataset encompasses seven land cover types: roads, buildings, lakes, bare land, forests, farmland, and background, with the highest number of annotations for buildings. The challenges of the dataset arise from factors such as the heterogeneous spatial distribution of urban and rural landscapes, complex backgrounds, and multi-scale objects, making LoveDA one of the most challenging datasets for land cover classification in the realm of remote sensing semantic segmentation.

4.2. Implementation Details

4.2.1. Experimental Settings

The models are trained using the PyTorch 2.0 framework, with all experiments performed on an NVIDIA RTX 3090 GPU with 24 GB of memory. To expedite convergence, the AdamW optimizer is employed with an initial learning rate of 0.0001, adjusted according to the yolox_warm_cos_lr schedule [70]. To facilitate efficient model, images from the GID and LoveDA datasets are cropped to a non-overlapping size of 512 × 512 pixels, discarding any excess portions. Various data augmentation techniques are applied during training, including random scaling within the range of [0.8–1.2], random vertical and horizontal flips, random rotations, and color enhancements, with a batch size of 8. The maximum number of training epochs is established at 300 for the GID and 100 for the LoveDA dataset.
For the EIODA algorithm, the MRS algorithm is used as the initial segmentation method, utilizing fixed segmentation parameters: a scale of 18, with compactness and smoothness set to the default value of 0.5. The EIODA segmentation method is applied to all image tiles in the dataset to obtain the complete segmentation results. It is implemented in a C++ environment on a computer with an Intel i7-12700F CPU (2.1 GHz) and 32 GB of RAM.

4.2.2. Loss Function

As depicted in Figure 8, the class imbalance present in the GID and the LoveDA dataset leads to a model that predominantly emphasizes the majority classes during training, frequently overlooking the minority classes. To tackle this issue, we use a combination of dice loss and cross-entropy loss as the final loss, which is formulated as
L = L C E + L D i c e

4.2.3. Evaluation Metrics

Overall Accuracy (OA), mean Pixel Accuracy (mPA), and mean Intersection over Union (mIoU) are employed to quantitatively assess the segmentation performance of the model. OA denotes the ratio of accurately predicted pixels to the total number of pixels, ranging from 0 to 1, and reflects the overall performance of the model across all classes. The evaluation index mPA is calculated as the average of pixel accuracies across all land cover categories. Intersection over Union (IoU) is a commonly adopted measure in segmentation tasks that evaluates the overlap between predicted and ground truth regions for each class. mIoU signifies the average IoU across all classes and offers a comprehensive evaluation of the model’s performance, particularly in addressing class imbalance. It provides more detailed insights than OA. The formulas for these metrics are as follows:
O A = T P + T N T P + T N + F P + F N
m P A = 1 N i = 1 N T P i T P i + F P i
I o U = T P T P + F N + F P
m I o U = 1 N i = 1 N I o U = 1 K i = 1 N T P i T P i + F N i + F P i
Notably, in the above equations, TP, TN, FP, and FN refer to the counts of true positives, true negatives, false positives, and false negatives, respectively, with N indicating the number of classes.

4.3. Ablation Study

Ablation studies are performed on the GID and the LoveDA dataset to assess the proposed network architecture and the efficacy of the three key modules—CSM, FEM, and SDFM—using a baseline network for comparison. This approach enables a thorough analysis of the significance and effectiveness of each component, with all experiments conducted under the same settings. The baseline network employs a complete ResNet50 architecture with SWTB introduced at the third, fourth, and fifth layers. The configuration aligns with the “Swin-Tiny” version, with a window size of 7 and block repetitions of {2, 6, 2}. To ensure compatibility with the output channels {512, 1024, 2048} of the last three layers of the encoder, the number of heads per layer is adjusted to {4, 8, 16}. In the ablation studies, the presence of the SDFM signifies the inclusion of the auxiliary encoder branch.

4.3.1. Effect of Integrating SWTB (Baseline)

To address the issue of global context loss as the depth of convolutional layers increases, our baseline network incorporates the SWTB module into the middle and deeper layers (the third, fourth, and fifth layers) of the ResNet50 backbone. A comparison of Table 1, Table 2 and Table 3 reveals that incorporating the SWTB module enhances segmentation performance on the GID. Specifically, mIoU increases from 59.73% to 62.73%, and OA rises from 80.98% to 82.60%, compared to the original UNet network. Similarly, for the LoveDA dataset, mIoU increases from 47.8% to 48.8%. This demonstrates that the SWTB module proficiently captures long-range dependencies within the data, aggregating more information beneficial for semantic prediction.

4.3.2. Effect of Cross-Channel Selective Module

Table 1 reports that the inclusion of CSM results in improvements of 0.74% and 1.01% in mIoU for the two datasets, respectively. In the LoveDA dataset, the Barren category exhibits the most significant enhancement in segmentation accuracy, achieving an increase of 2.93%. This is closely followed by the Agriculture category, which shows an improvement of 2.79%. Figure 9 presents the segmentation outcomes of the baseline before and after the incorporation of CSM. The figure demonstrates that it enables the model to adaptively adjust channel weights, thereby enhancing significant features while suppressing less critical ones. This adjustment reduces certain patches in the building and forest categories, which improves the model’s capacity for feature representation.

4.3.3. Effect of Feature Enhancement Module

Table 1 demonstrates that integrating FEM into the baseline framework yields improvements of 1.34% and 1.31% in mIoU for the GID and the LoveDA dataset, respectively, highlighting its effectiveness in the network. Notably, segmentation accuracy for various land cover types has been improved, particularly in the building and agriculture classes, with IoU increases of 3.97% and 2.17%, respectively. Figure 10 visually compares segmentation results before and after applying FEM to the baseline model. The results illustrate that it not only maintains robust segmentation performance but also tends to merge discontinuous and closely spaced ground objects into cohesive units. This suggests that incorporating global context information into CNNs can significantly improve segmentation accuracy by better aggregating similar regions.

4.3.4. Effect of Shallow–Deep Feature Fusion Module

Table 1 demonstrates that the introduction of an auxiliary encoding branch, using optimally segmented input images, results in improvements in mIoU, OA, and mPA on the GID by 0.56%, 0.53%, and 0.54%, respectively. On the LoveDA dataset, mIoU increases by 0.83%. Figure 11a provides a clear visual comparison of the segmentation results before and after integrating the SDFM into the baseline model. In the first and second rows, it successfully preserves the boundaries and structural integrity of various land cover types. The third and fourth rows demonstrate that, unlike the FEM, SDFM maintains the separability of distinct land cover classes—such as buildings and water bodies—preventing their erroneous merging into a single entity.
To further assess its contribution, we compare model variants with and without this module under various configurations on the GID and the LoveDA dataset (Figure 11b). The results consistently indicate that its inclusion improves overall segmentation performance. Additionally, as shown in Table 1, it enhances the model’s capacity to capture global features. This improvement is particularly evident in the segmentation performance of large-scale land cover categories such as water bodies, barren land, and agricultural areas in the LoveDA dataset.

4.3.5. Joint Effects of Different Modules

Furthermore, the joint effects of different modules on segmentation results within the baseline framework are explored on both datasets, as shown in Table 1. For the GID and the LoveDA dataset, introducing both CSM and FEM together improve segmentation results by 2.55% and 1.98% mIoU, respectively. Notably, even without incorporating the SDFM, the Baseline + CSM + FEM configuration achieves consistently strong results across both datasets. This demonstrates that the CNN–Swin Transformer hybrid design (CSM + FEM) alone can effectively learn discriminative and robust representations, making it potentially applicable to other vision tasks. When both CSM and SDFM are used, segmentation results increase by 1.84% and 2.36% mIoU, respectively. Considering both FEM and SDFM together result in mIoU increases of 1.66% and 2.66%, respectively. Based on the baseline framework, the proposed GOFENet, which integrates the three key modules (CSM, FEM, and SDFM), achieve mIoU improvements of 3.29% and 3.08% on the respective datasets.
Table 1. Ablation experiment of the proposed modules on the GID and the LoveDA dataset.
Table 1. Ablation experiment of the proposed modules on the GID and the LoveDA dataset.
Model NameModuleGIDLoveDA
CCMFEMSDFMmIoU (%)mPA (%)OA (%)IoU per Category (%)mIoU (%)
BackgroundBuildingRoadWaterBarrenForestAgriculture
Baseline 62.7374.8982.6043.5951.6755.0675.8015.6243.8456.2748.84
Baseline + CSM 63.4774.0182.9944.1452.8754.0674.6718.5545.5759.0649.85
Baseline + FEM 64.0774.7583.4443.3055.6455.0876.8116.7845.0158.4450.15
Baseline + SDFM 63.2975.4383.1342.7953.5254.0476.9119.7943.7656.8849.67
Baseline + CSM + FEM 65.2876.4983.9944.3656.5555.6475.5316.0645.0262.5550.82
Baseline + CSM + SDFM 64.5775.6983.6343.2655.4654.2376.2820.5345.4663.1951.20
Baseline + FEM + SDFM 64.3975.3783.5244.3455.0455.7377.6119.9944.3063.5151.50
Baseline + CSM + FEM + SDFM66.0277.7884.0745.2055.6356.7578.1319.2844.8863.6051.92
Figure 9. Comparison of segmentation results with and without the integration of CSM in the baseline architecture.
Figure 9. Comparison of segmentation results with and without the integration of CSM in the baseline architecture.
Remotesensing 17 02652 g009
Figure 10. Comparison of segmentation results with and without the integration of FEM in the baseline architecture.
Figure 10. Comparison of segmentation results with and without the integration of FEM in the baseline architecture.
Remotesensing 17 02652 g010
Figure 11. (a) Comparison of segmentation results with and without the integration of SDFM in the baseline architecture. (b) Ablation study on the effect of integrating SDFM into different model variants. The plots show mIoU (%) performance on the GID (left) and the LoveDA dataset (right) before and after incorporating the SDFM.
Figure 11. (a) Comparison of segmentation results with and without the integration of SDFM in the baseline architecture. (b) Ablation study on the effect of integrating SDFM into different model variants. The plots show mIoU (%) performance on the GID (left) and the LoveDA dataset (right) before and after incorporating the SDFM.
Remotesensing 17 02652 g011

4.4. Performance Evaluation and Comparisons with Other Models

We evaluate our proposed model against several state-of-the-art semantic segmentation methods on the GID and the LoveDA dataset to highlight its superiority. These methods are categorized into three main types:
(1) CNN-based methods: UNet [19], PSPNet [24], DeepLabV3+ [23], DANet [25], MANet [71], SegNet [72], FactSeg [73], LANet [74], HRNet [75], C-PNet [76].
(2) Transformer-based methods: SegFormer [36], Segmenter [35], Mask2Former [77].
(3) CNN–Transformer hybrid networks: BANet [40], TransUNet [78], DC-Swin [39], SwinUperNet [30], UNetFormer [44], LMA-Swin [41], ESDINet [79].
Table 2. Comparison of segmentation results on the GID. The bold text highlights the highest values in each column, while the second-highest values are marked with underlining.
Table 2. Comparison of segmentation results on the GID. The bold text highlights the highest values in each column, while the second-highest values are marked with underlining.
MethodBackboneIoU per Category (%)mIoU (%)mPA
(%)
OA (%)
BGINLURRUTLPFILDCGPAWSLNGAGRILAPD
UNet [19]ResNet5064.0757.7268.6859.6251.6072.4877.8667.4330.5372.5568.7338.5132.5367.5681.6644.2259.7375.2380.98
PSPNet [24]ResNet5064.7358.3568.2659.4343.6072.0376.4269.7432.4373.3468.1943.9635.6571.4985.6747.0660.6571.9581.25
SegNet [72]ResNet5062.3955.4167.0055.8043.0166.3675.0565.2231.8571.0767.1645.8529.2465.7375.8145.9857.6871.5379.32
DeeplabV3+ [23]ResNet5065.1557.6967.1258.7348.0671.4275.8870.0734.4372.4569.4145.8032.7968.1280.5248.2060.3774.0880.91
DANet [25]ResNet5065.8258.2567.9359.5148.0971.9975.4971.1029.3072.4567.7942.7137.8673.2285.4448.9360.9974.2981.32
SegFormer [36]MIT-B166.9158.3568.5459.7247.4872.3077.5371.8530.6274.0069.4050.0532.7472.7185.4350.2361.7473.7082.23
DC-Swin [39]Swin-Tiny65.9157.3467.3859.4146.3569.2976.3670.2932.5173.1068.0848.0133.0672.3886.9350.9461.0872.9881.51
MANet [71]ResNet5068.7760.8170.2362.0054.2873.1378.8473.7537.8275.6072.2151.0035.7977.5688.2854.4164.6976.2383.59
UNetFormer [44]ResNet1867.8261.2070.2261.3253.8572.5677.5671.7336.4773.5970.0652.8636.4675.1987.2852.5463.7975.7482.80
LMA-Swin [41]Swin-Small68.5059.1070.6162.6355.4670.8377.6873.3237.1974.7670.9345.9937.6276.8488.3750.6163.7874.7983.12
GOFENet (ours)ResNet5069.2461.7771.3363.6855.9074.7979.6674.8240.3875.2472.8955.2540.4377.0687.5556.3566.0277.7884.07
Note: BG (Background), INL (Industrial Land), UR (Urban Residential), RU (Rural Residential), TL (Traffic Land), PF (Paddy Field), IL (Irrigated Land), DC (Dry Cropland), GP (Garden Plot), AW (Arbor Woodland), SL (Shrub Land), NG (Natural Grassland), AG (Artificial Grassland), RI (River), LA (Lake), PD (Pond).

4.4.1. Results on GID

Table 2 provides a quantitative comparison of various semantic segmentation methods on the GID. Our proposed model, GOFENet, outperforms other networks in overall metrics, achieving mIoU, mPA, and OA values of 66.02%, 77.78%, and 84.07%, respectively. Among the 16 land cover categories, including the background, GOFENet almost secures both the highest and second-highest IoU values, demonstrating impressive performance. Notably, compared to the second-best values, the IoU improvements for the AG, GP, and NG categories exceed two percentage points, while the IoU increases for the RU, PF, DC, and PD categories are over one percentage point. This underscores GOFENet’s superior capability in capturing multi-scale information. In contrast to traditional CNN-based models like UNet and SegNet, which lack mechanisms for global context, DeeplabV3+ and PSPNet improve performance by expanding the receptive field through dilated convolutions and pyramid pooling. Attention-based methods such as DANet and MANet leverage dual- or multi-attention mechanisms for enhanced global dependency modeling, with MANet achieving the second-best results in many categories. Moreover, Transformer-based architectures like SegFormer, DC-Swin, LMA-Swin, and UNetFormer exhibit strong performance due to their long-range modeling capabilities, with UNetFormer balancing segmentation quality and model efficiency.
The performance variations in GOFENet across different land cover categories can be attributed to the alignment between its architectural components and the intrinsic characteristics of each land type. For large-area, regular, and spatially continuous classes such as irrigated land, dry cropland, arbor woodland, lake, and artificial water, the combination of global modeling capability and the SDFM helps maintain region-level consistency and significantly enhances classification accuracy. For structurally complex categories such as urban residential and industrial land, SWTB embedded within the FEM effectively captures long-range dependencies and spatial structures, resulting in more accurate boundary delineation. In contrast, for small-area categories like Greenhouse and paddy field, the model’s multi-scale fusion strategy and local detail preservation significantly improve the segmentation of small targets. However, for spectrally heterogeneous and boundary-ambiguous categories such as artificial grassland, garden plot, and natural grassland, subtle inter-class differences and feature overlaps may lead to misclassification and relatively lower accuracy.
In addition to quantitative improvements, GOFENet also demonstrates clear qualitative advantages, as illustrated in Figure 12. First, GOFENet demonstrates superior multi-scale object recognition capabilities, effectively identifying small-scale targets such as arbor woodland in the first image and ponds in the third image, which are missed by other models. Second, the boundaries delineated by GOFENet closely align with the reference outlines, as observed with buildings in the first image, irrigated land in the second image, and ponds in the fifth and seventh image. Finally, its segmentation results exhibit remarkable continuity and smoothness, evident in the connectivity of roads and the minimal presence of holes and patch noise in the segmentation output. These advantages confirm the model’s ability to balance local detail preservation with global structural consistency. Figure 13 presents a visual comparison between the object contours and ground truth boxes identified by the current state-of-the-art methods and GOFENet. It is evident that GOFENet not only achieves high segmentation accuracy but also demonstrates strong consistency with the ground truth contours.

4.4.2. Results on the LoveDA Dataset

We submitted the results to the LoveDA semantic segmentation challenge website to acquire benchmark results for the test data. Table 3 presents the comparison results of GOFENet against a range of state-of-the-art semantic segmentation models on the LoveDA dataset. GOFENet achieves the highest overall segmentation performance, with a mean IoU of 51.9%, outperforming both CNN-based methods (e.g., UNet, HRNet, C-PNet) and Transformer-based approaches (e.g., Mask2Former, SwinUperNet, ESDINet). These results demonstrate the strong generalization capability of GOFENet in handling diverse land cover types across complex urban and rural scenarios. Notably, GOFENet achieves the highest IoU in the Agriculture category (63.6%) and demonstrates strong performance in Water (78.1%) and Barren (19.3%). This improvement can be attributed to the SDFM and multi-scale feature aggregation techniques, which enhance the model’s ability to distinguish regions with high spectral homogeneity. However, the model shows relatively lower performance on small-scale targets (e.g., Building) and categories with complex textures and ambiguous boundaries (e.g., Forest). These challenges are primarily due to the inherent structural complexity and edge uncertainty of such classes, which hinder the accurate extraction of local spatial details and fine-grained boundary representations. Comparatively, HRNet slightly surpasses GOFENet in the Road category (57.4% vs. 56.7%), benefiting from its parallel multi-resolution representation and continuous feature fusion, which are particularly effective for preserving fine spatial details and segmenting elongated structures. In the Building category, Mask2Former achieves higher accuracy (56.8%) through its capacity to model both local and global feature dependencies, thereby improving the segmentation of complex architectural shapes and boundaries. For the Forest class, DC-Swin delivers the best performance (47.2%) by leveraging densely connected feature aggregation modules, which enhance the model’s ability to capture rich texture and contextual information. Overall, the performance variations across categories closely reflect each model’s architectural strengths in extracting specific spatial and spectral features.
Figure 14 qualitatively presents the segmentation results of different models. Observations from the first, second, and fifth rows show that the proposed method effectively preserves the structural integrity and separability of individual buildings and forest regions. In contrast, other models tend to merge adjacent buildings into a single connected component, failing to capture clear boundaries between them. Moreover, in the third row, GOFENet provides a more precise segmentation of the Forest category, maintaining its spatial continuity. In the fourth and sixth rows, the delineation of agricultural areas also more closely aligns with the ground truth. These qualitative findings are further supported by Figure 15, where the features identified by GOFENet exhibit pronounced spatial coherence. The model effectively mitigates fragmented segmentation and significantly reduces the presence of small, scattered patches, highlighting its robustness in capturing continuous land cover patterns.
Table 3. Comparison of segmentation results on the LoveDA dataset. The bold text highlights the highest values in each column, while the second-highest values are marked with underlining.
Table 3. Comparison of segmentation results on the LoveDA dataset. The bold text highlights the highest values in each column, while the second-highest values are marked with underlining.
MethodBackboneIoU per Category (%)mIoU (%)
BackgroundBuildingRoadWaterBarrenForestAgriculture
UNet [19]ResNet5043.152.752.873.110.343.159.947.8
PSPNet [24]ResNet5044.452.153.576.59.744.157.948.3
SegNet [72]ResNet5043.453.052.976.012.744.054.848.1
DeeplabV3+ [23]ResNet5043.050.952.074.410.444.258.547.6
Segmenter [35]Vit-Tiny38.050.748.777.413.343.558.247.1
BANet [40]ResT-Lite43.751.551.176.916.644.962.549.6
DANet [25]ResNet5044.855.553.075.517.645.160.150.2
LANet [74]ResNet5040.050.651.178.013.043.256.947.6
SegFormer [36]MIT-B143.052.253.268.610.345.453.146.5
FactSeg [73]ResNet5042.653.652.876.916.242.957.548.9
MANet [71]ResNet5038.751.742.672.015.342.157.745.7
HRNet [75]W3244.655.357.474.011.145.360.949.8
DC-Swin [39]Swin-Tiny41.354.556.278.114.547.262.450.6
TransUNet [78]ViT-R5043.056.153.778.09.344.956.948.9
SwinUperNet [30]Swin-Tiny43.354.354.378.714.945.359.650.0
C-PNet [76]-44.055.255.378.816.046.458.051.8
Mask2Former [77]Swin-Small44.856.855.578.617.846.360.051.5
ESDINet [79]ResNet1841.653.854.878.719.544.258.050.1
GOFENet (ours)ResNet5045.255.656.778.119.344.963.651.9
Figure 14. Examples of semantic segmentation results on the LoveDA dataset.
Figure 14. Examples of semantic segmentation results on the LoveDA dataset.
Remotesensing 17 02652 g014
Figure 15. Visualization comparison of object contours and ground truth boxes identified by the current state-of-the-art methods and GOFENet on the LoveDA dataset.
Figure 15. Visualization comparison of object contours and ground truth boxes identified by the current state-of-the-art methods and GOFENet on the LoveDA dataset.
Remotesensing 17 02652 g015

5. Discussion

5.1. Design of Auxiliary Branch

The segmentation results of the EIODA algorithm, which integrates spectral and edge information, reflect high homogeneity within segments and high heterogeneity between segments. This approach elevates image processing from the pixel level to the level of meaningful geo-objects, enhancing the overall perception of remote sensing images. This study explores the use of segmentation results as input to a neural network, leveraging object-level features to enhance the global information of the model, and designs an auxiliary branch network. The auxiliary encoder shares parameters with the primary encoder in the first layer, motivated by two main considerations: (1) reducing computational load, and (2) strengthening the pixel-to-object-level correspondence in image features. When the object scale is larger than the convolution kernel size, the features extracted by both the primary and auxiliary encoders correspond to pixel-level and object-level features within the window, respectively. Conversely, when the object scale is smaller than the convolution kernel size, meaning multiple objects are present within the window, the pixels in the feature maps obtained from the auxiliary encoder represent cross-object features, which may introduce additional disruptive information. Specifically, if segmented images are directly input into the CNN as if they were original images, the stacking of convolutional layers may progressively amplify the boundary information between objects due to the pixel discontinuities at object boundaries, leading to information distortion and interference with the fused data. In this scenario, the model is unable to learn any valuable information, potentially even degrading its segmentation performance.
Additionally, we explored the use of deformable convolutions to capture features from segmented images. However, the geometric and spatial differences between objects across different images made learning the offset parameters difficult. Following the recommendations for the hybrid dilated convolution (HDC) [80], a series of consecutive dilated convolutions with appropriate dilation rates is constructed to gradually expand the receptive field of the network. It is noteworthy that, to avoid the interference of non-semantic features on the final results, the SDFM is introduced exclusively in the last three layers of the network decoder.

5.2. Model Efficiency Analysis

To evaluate model performance across two datasets, we utilized computation complexity (GFLOPS) and inference speed (FPS) as metrics to analyze model efficiency under consistent operating conditions. Table 4 and Table 5 present the evaluation results of the models on the two datasets, noting that the input image size in Table 4 is 512 × 512, whereas in Table 5 it is 1024 × 1024. A comprehensive analysis reveals that the hybrid models based on CNNs and Transformers exhibit higher computational complexity compared to pure CNN architectures, as well as typically slower inference speeds. Our proposed GOFENet achieves an inference speed of approximately 17 FPS for 512 × 512 pixel images. The high model complexity is attributed to the uncompressed output channels of the ResNet in the skip connection structure at each decoder layer. Consequently, we considered utilizing ResNet18/ResNet34 as backbones and implementing structural simplifications, termed GOFENet-t and GOFENet-s. Table 4 and Table 5 report a substantial reduction in complexity and a significant enhancement in forward inference speed for the simplified models.
Figure 16 visually illustrates the performance trade-offs across the three metrics. Specifically, on the GID, GOFENet-s reduces GFLOPs by approximately 48.5% while only sacrificing 1.3% mIoU, and achieves a 2.5× increase in inference speed. GOFENet-t further reduces GFLOPs by 77.9%, with a 2.7% mIoU drop, and improves speed by approximately 4.5×. On the LoveDA dataset, GOFENet-s achieves comparable mIoU, while reducing GFLOPs by 46.4% and doubling the inference speed. GOFENet-t maintains similar accuracy (−0.7% mIoU), but with a 77.2% reduction in complexity and 3.8× faster inference. These comparisons demonstrate that the lightweight variants provide practical alternatives depending on deployment constraints: GOFENet-s is suitable for tasks that require a balance of accuracy and moderate acceleration, while GOFENet-t is ideal for real-time applications prioritizing speed and efficiency.
Notably, the GOFENet-s model demonstrates a slight improvement on the LoveDA dataset, with mIoU values for both GOFENet and its simplified versions maintaining between 51% and 52%. This is likely due to the lower class count in the dataset, allowing shallower models to effectively learn the primary characteristics of land cover. However, for the GID, the simplified models strike a balance between complexity and segmentation accuracy, likely because the lightweight model finds it challenging to comprehensively capture the relevant features of multiple categories. These experimental results further validate the efficacy of the network architecture and three key modules, confirming that the integration of optimal segmentation results enhances global information acquisition (Table 6).
Before training the network, the EIODA algorithm is employed to perform multi-scale segmentation on each 512 × 512 × 3 image in the datasets. Table 7 presents statistics on object sizes (pixel counts) for the training, validation, and test sets of both datasets, including minimum, maximum, and mean values, along with the average execution time of the EIODA algorithm. Under the same segmentation parameters, images from the GID contain significantly more objects than those from the LoveDA dataset, suggesting that the GID has more complex and discriminative land cover types. This, in turn, reflects the spectral similarity of different land cover classes in the LoveDA dataset, which renders the distinction between various land cover types more difficult. Notably, without considering GPU acceleration and parallel processing, the EIODA algorithm still exhibits relatively fast processing speed.

5.3. Grad-CAM Visualization

To showcase the benefits of our proposed network and assess the effectiveness of the global and local context fusion mechanism, Grad-CAM [81] visualization is used to analyze the attention maps at the ultimate classification layers of the network across both datasets. Two target land cover categories are randomly selected for evaluation. The first row of images presents the attention maps of GOFENet for these categories at different layers (GOFENet-Ln denotes the feature map after up-sampling at layer n of the network). As observed in Figure 17, with the forward pass of decoder, the attention map of roads and buildings becomes increasingly refined and enhanced. Compared to other models, GOFENet achieves a better balance between global and local features, resulting in clearer land cover boundaries and a higher level of attention.

5.4. Advantages and Limitations of Embedded Object Features

According to the experimental results, the deep learning semantic segmentation framework integrating GEOBIA can elevate segmentation accuracy and refine boundary information to some extent. We explore the effect of integrating segmentation maps via the SDFM on the classification of large-scale land cover categories. As illustrated in Figure 8, our analysis focuses on the dominant classes with the largest pixel coverage in the GID and the LoveDA dataset. Specifically, these include Irrigated Land, Dry Cropland, Arbor Woodland, Lake, and River in GID (Figure 18a), and Agriculture, Forest, Water, and Barren in LoveDA (Figure 18b). Overall, the integration of the SDFM enhances segmentation accuracy for large-scale land cover types. In the GID, although a few categories showed slight performance declines due to external factors, these decreases were limited and did not offset the overall improvement. Notably, the Forest class in the LoveDA dataset exhibited a consistent decline across all four model configurations, likely due to its fragmented and discontinuous spatial distribution, which hindered effective spatial fusion. These results indicate that integrating GEOBIA-derived segmentation priors significantly enhances the model’s capacity for global feature extraction, thereby improving its ability to accurately identify and delineate extensive and spatially coherent land cover categories.
However, while object-based image features capture the overall characteristics of an image, they tend to lose detailed information. Furthermore, issues such as object occlusion, discontinuous linear features, and image noise can significantly affect multi-scale segmentation results, thereby impacting the network’s ability to focus on fine-scale objects.

6. Conclusions

To address the challenges of limited global context modeling and the underutilization of GEOBIA-derived object priors in semantic segmentation, we propose GOFENet, a hybrid dual-encoder network that combines CNN and Swin Transformer to jointly capture local spatial details and long-range dependencies. To further enhance the encoding process, the CSM selectively emphasizes informative channels, while the FEM integrates global semantics from the Transformer into the CNN pathway, improving the discriminative power of skip-connected features. In addition, considering that GEOBIA-based segmented objects, although lacking explicit semantics, provide valuable structural context and global understanding of the scene, we introduce an auxiliary encoding branch. This branch employs cascaded atrous convolutions to extract global structural features from the segmentation priors. These features are then integrated with those from the main encoder through the SDFM, facilitating the joint representation of object-level and pixel-level information. This design improves the model’s ability to capture global contextual dependencies and semantic consistency, ultimately leading to more accurate and robust segmentation outcomes. Extensive experiments on two large-scale datasets—GID and LoveDA—demonstrate the effectiveness and generalization ability of GOFENet, achieving mIoU values of 66.02% and 51.92%, respectively. The resulting segmentation maps display clear object boundaries, reduced background noise, and improved preservation of semantic integrity. Notably, GOFENet exhibits a strong advantage in accurately segmenting large, continuous land cover regions, further highlighting its capability in handling high-resolution and complex scene interpretation tasks.
In future work, we aim to further explore strategies to mitigate the uncertainty introduced when integrating segmented images into deep neural networks. Furthermore, to assess the generalizability of our method, we will improve the model and extend its application to other remote sensing tasks.

Author Contributions

Conceptualization, T.H. and J.C.; methodology, T.H.; software, J.C.; validation, T.H. and J.C.; formal analysis, T.H.; writing—original draft preparation, T.H.; writing—review and editing, T.H., J.C., and D.P.; visualization, T.H.; supervision, J.C. and D.P.; funding acquisition, J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the PI Project of Southern Marine Science and Engineering Guangdong Laboratory (Guangzhou) (GML2021GD0809), the National Key Research and Development Program of China (2022YFC3103101), the NSFC-Zhejiang Joint Fund for the Integration of Industrialization and Informatization (U1609202), and the National Natural Science Foundation of China (41376184 and 40976109).

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lei, P.; Yi, J.; Li, S.; Li, Y.; Lin, H. Agricultural Surface Water Extraction in Environmental Remote Sensing: A Novel Semantic Segmentation Model Emphasizing Contextual Information Enhancement and Foreground Detail Attention. Neurocomputing 2025, 617, 129110. [Google Scholar] [CrossRef]
  2. Liu, Y.; Zhong, Y.; Shi, S.; Zhang, L. Scale-Aware Deep Reinforcement Learning for High Resolution Remote Sensing Imagery Classification. ISPRS J. Photogramm. Remote Sens. 2024, 209, 296–311. [Google Scholar] [CrossRef]
  3. Wang, Y.; Sun, Y.; Cao, X.; Wang, Y.; Zhang, W.; Cheng, X. A Review of Regional and Global Scale Land Use/Land Cover (LULC) Mapping Products Generated from Satellite Remote Sensing. ISPRS J. Photogramm. Remote Sens. 2023, 206, 311–334. [Google Scholar] [CrossRef]
  4. Zhu, Q.; Guo, X.; Deng, W.; Shi, S.; Guan, Q.; Zhong, Y.; Zhang, L.; Li, D. Land-Use/Land-Cover Change Detection Based on a Siamese Global Learning Framework for High Spatial Resolution Remote Sensing Imagery. ISPRS J. Photogramm. Remote Sens. 2022, 184, 63–78. [Google Scholar] [CrossRef]
  5. Zhang, T.; Huang, X. Monitoring of Urban Impervious Surfaces Using Time Series of High-Resolution Remote Sensing Images in Rapidly Urbanized Areas: A Case Study of Shenzhen. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 2692–2708. [Google Scholar] [CrossRef]
  6. Chen, B.; Qiu, F.; Wu, B.; Du, H. Image Segmentation Based on Constrained Spectral Variance Difference and Edge Penalty. Remote Sens. 2015, 7, 5980–6004. [Google Scholar] [CrossRef]
  7. Zhou, Y.; Feng, L.; Chen, Y.; Li, J. Object-Based Land Cover Mapping Using Adaptive Scale Segmentation From ZY-3 Satellite Images. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; IEEE: New York, NY, USA; pp. 63–66. [Google Scholar]
  8. Thenkabail, P.S. Remotely Sensed Data Characterization, Classification, and Accuracies; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar]
  9. Derivaux, S.; Lefevre, S.; Wemmert, C.; Korczak, J. Watershed Segmentation of Remotely Sensed Images Based on a Supervised Fuzzy Pixel Classification. In Proceedings of the 2006 IEEE International Symposium on Geoscience and Remote Sensing, Denver, CO, USA, 31 July–4 August 2006; IEEE: New York, NY, USA, 2006; pp. 3712–3715. [Google Scholar]
  10. Baatz, M.; Schäpe, A. Multiresolution Segmentation: An Optimization Approach for High Quality Multi-Scale Image Segmentation. In Angewandte Geographische Informationsverarbeitung XII; Wichmann Verlag: Karlsruhe, Germany, 2000; pp. 12–23. [Google Scholar]
  11. Tzotsos, A.; Argialas, D. MSEG: A Generic Region-Based Multi-Scale Image Segmentation Algorithm for Remote Sensing Imagery. In Proceedings of the ASPRS 2006 Annual Conference, Reno, NV, USA, 1–5 May 2006. [Google Scholar]
  12. Yang, J.; He, Y.; Caspersen, J. Region Merging Using Local Spectral Angle Thresholds: A More Accurate Method for Hybrid Segmentation of Remote Sensing Images. Remote Sens. Environ. 2017, 190, 137–148. [Google Scholar] [CrossRef]
  13. Blaschke, T.; Hay, G.J.; Kelly, M.; Lang, S.; Hofmann, P.; Addink, E.; Feitosa, R.Q.; Van der Meer, F.; Van der Werff, H.; Van Coillie, F.; et al. Geographic Object-Based Image Analysis–Towards a New Paradigm. ISPRS J. Photogramm. Remote Sens. 2014, 87, 180–191. [Google Scholar] [CrossRef]
  14. Hu, Y.; Chen, J.; Pan, D.; Hao, Z. Edge-Guided Image Object Detection in Multiscale Segmentation for High-Resolution Remotely Sensed Imagery. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4702–4711. [Google Scholar] [CrossRef]
  15. Shen, Y.; Chen, J.; Xiao, L.; Pan, D. Optimizing Multiscale Segmentation with Local Spectral Heterogeneity Measure for High Resolution Remote Sensing Images. ISPRS J. Photogramm. Remote Sens. 2019, 157, 13–25. [Google Scholar] [CrossRef]
  16. He, T.; Chen, J.; Kang, L.; Zhu, Q. Evaluation of Global-Scale and Local-Scale Optimized Segmentation Algorithms in GEOBIA With SAM on Land Use and Land Cover. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 6721–6738. [Google Scholar] [CrossRef]
  17. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  18. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  19. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015. proceedings, part III 18. pp. 234–241. [Google Scholar]
  20. Xiang, X.; Gong, W.; Li, S.; Chen, J.; Ren, T. TCNet: Multiscale Fusion of Transformer and CNN for Semantic Segmentation of Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 3123–3136. [Google Scholar] [CrossRef]
  21. Liu, B.; Li, B.; Sreeram, V.; Li, S. MBT-UNet: Multi-Branch Transform Combined with UNet for Semantic Segmentation of Remote Sensing Images. Remote Sens. 2024, 16, 2776. [Google Scholar] [CrossRef]
  22. Ren, X.; Deng, Z.; Ye, J.; He, J.; Yang, D. FCN+: Global Receptive Convolution Makes Fcn Great Again. Neurocomputing 2025, 631, 129655. [Google Scholar] [CrossRef]
  23. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  24. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  25. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual Attention Network for Scene Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3146–3154. [Google Scholar]
  26. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  27. Khan, S.; Naseer, M.; Hayat, M.; Zamir, S.W.; Khan, F.S.; Shah, M. Transformers in Vision: A Survey. ACM Comput. Surv. 2022, 54, 1–41. [Google Scholar] [CrossRef]
  28. Alexey, D. An Image is Worth 16x16 Words: Transformers for Image Recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  29. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.H.; et al. Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 6881–6890. [Google Scholar]
  30. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  31. Bai, Q.; Luo, X.; Wang, Y.; Wei, T. DHRNet: A Dual-Branch Hybrid Reinforcement Network for Semantic Segmentation of Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 4176–4193. [Google Scholar] [CrossRef]
  32. Gao, Y.; Luo, X.; Gao, X.; Yan, W.; Pan, X.; Fu, X. Semantic Segmentation of Remote Sensing Images Based on Multiscale Features and Global Information Modeling. Expert Syst. Appl. 2024, 249, 123616. [Google Scholar] [CrossRef]
  33. Yang, X.; Li, S.; Chen, Z.; Chanussot, J.; Jia, X.; Zhang, B.; Li, B.; Chen, P. An Attention-Fused Network for Semantic Segmentation of very-High-Resolution Remote sensing Imagery. ArXiv Comput. Vis. Pattern Recognit. 2021, 177, 238–262. [Google Scholar] [CrossRef]
  34. Zhou, N.; Hong, J.; Cui, W.; Wu, S.; Zhang, Z. A Multiscale Attention Segment Network-Based Semantic Segmentation Model for Landslide Remote Sensing Images. Remote Sens. 2024, 16, 1712. [Google Scholar] [CrossRef]
  35. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 7262–7272. [Google Scholar]
  36. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  37. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-unet: Unet-Like pure Transformer for Medical Image Segmentation. In Proceedings of the European Conference on Computer Vision; Springer: Cham, Switzerland, 2022; pp. 205–218. [Google Scholar]
  38. Zhang, Y.; Liu, H.; Hu, Q. Transfuse: Fusing Transformers and Cnns for Medical Image Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Springer: Berlin/Heidelberg, Germany, 2021. proceedings, Part I 24. pp. 14–24. [Google Scholar]
  39. Wang, L.; Li, R.; Duan, C.; Zhang, C.; Meng, X.; Fang, S. A Novel Transformer Based Semantic Segmentation Scheme for Fine-Resolution Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  40. Wang, L.; Li, R.; Wang, D.; Duan, C.; Wang, T.; Meng, X. Transformer Meets Convolution: A Bilateral Awareness Network for Semantic Segmentation of Very Fine Resolution Urban Scene Images. Remote Sens. 2021, 13, 3065. [Google Scholar] [CrossRef]
  41. Ren, D.; Li, F.; Sun, H.; Liu, L.; Ren, S.; Yu, M. Local-Enhanced Multi-Scale Aggregation Swin Transformer for Semantic Segmentation of High-Resolution Remote Sensing Images. Int. J. Remote Sens. 2024, 45, 101–120. [Google Scholar] [CrossRef]
  42. Yao, M.; Zhang, Y.; Liu, G.; Pang, D. SSNet: A Novel Transformer and CNN Hybrid Network for Remote Sensing Semantic Segmentation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 3023–3037. [Google Scholar] [CrossRef]
  43. He, X.; Zhou, Y.; Zhao, J.; Zhang, D.; Yao, R.; Xue, Y. Swin Transformer Embedding UNet for Remote Sensing Image Semantic Segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  44. Wang, L.; Li, R.; Zhang, C.; Fang, S.; Duan, C.; Meng, X.; Atkinson, P.M. UNetFormer: A UNet-like Transformer for Efficient Semantic Segmentation of Remote Sensing Urban Scene Imagery. ISPRS J. Photogramm. Remote Sens. 2022, 190, 196–214. [Google Scholar] [CrossRef]
  45. Johnson, B.A.; Ma, L. Image Segmentation and Object-Based Image Analysis for Environmental Monitoring: Recent Areas of Interest, Researchers’ Views on the Future Priorities. Remote Sens. 2020, 12, 1772. [Google Scholar] [CrossRef]
  46. Wei, S.; Luo, M.; Zhu, L.; Yang, Z. Using Object-Oriented Coupled Deep Learning Approach for Typical Object Inspection of Transmission Channel. Int. J. Appl. Earth Obs. Geoinformation 2023, 116, 103137. [Google Scholar] [CrossRef]
  47. Timilsina, S.; Aryal, J.; Kirkpatrick, J.B. Mapping Urban Tree Cover Changes Using Object-Based Convolution Neural Network (OB-CNN). Remote Sens. 2020, 12, 3017. [Google Scholar] [CrossRef]
  48. Guirado, E.; Blanco-Sacristán, J.; Rodríguez-Caballero, E.; Tabik, S.; Alcaraz-Segura, D.; Martínez-Valderrama, J.; Cabello, J. Mask R-CNN and OBIA Fusion Improves the Segmentation of Scattered Vegetation in Very High-Resolution Optical Sensors. Sensors 2021, 21, 320. [Google Scholar] [CrossRef]
  49. Luo, C.; Li, H.; Zhang, J.; Wang, Y. OBViT: A High-Resolution Remote Sensing Crop Classification Model Combining Obia and Vision Transformer. In Proceedings of the 2023 11th International Conference on Agro-Geoinformatics (Agro-Geoinformatics), Wuhan, China, 25–28 July 2023; IEEE: New York, NY, USA, 2023; pp. 1–6. [Google Scholar]
  50. Liu, T.; Abd-Elrahman, A. An Object-Based Image Analysis Method for Enhancing Classification of Land Covers Using Fully Convolutional Networks and Multi-View Images of Small Unmanned Aerial System. Remote Sens. 2018, 10, 457. [Google Scholar] [CrossRef]
  51. Liu, T.; Abd-Elrahman, A.; Morton, J.; Wilhelm, V.L. Comparing Fully Convolutional Networks, Random Forest, Support Vector Machine, and Patch-Based Deep Convolutional Neural Networks for Object-Based Wetland Mapping Using Images from Small Unmanned Aircraft System. GIScience Remote Sens. 2018, 55, 243–264. [Google Scholar] [CrossRef]
  52. Fu, T.; Ma, L.; Li, M.; Johnson, B.A. Using Convolutional Neural Network to Identify Irregular Segmentation Objects from Very High-Resolution Remote Sensing Imagery. J. Appl. Remote Sens. 2018, 12, 025010. [Google Scholar] [CrossRef]
  53. Abdollahi, A.; Pradhan, B.; Shukla, N. Road Extraction from High-Resolution Orthophoto Images Using Convolutional Neural Network. J. Indian Soc. Remote Sens. 2021, 49, 569–583. [Google Scholar] [CrossRef]
  54. Lam, O.H.Y.; Dogotari, M.; Prüm, M.; Vithlani, H.N.; Roers, C.; Melville, B.; Zimmer, F.; Becker, R. An Open Source Workflow for Weed Mapping in Native Grassland Using Unmanned Aerial Vehicle: Using Rumex Obtusifolius As A Case Study. Eur. J. Remote Sens. 2021, 54, 71–88. [Google Scholar] [CrossRef]
  55. Tang, Z.; Li, M.; Wang, X. Mapping Tea Plantations from VHR Images Using OBIA and Convolutional Neural Networks. Remote Sens. 2020, 12, 2935. [Google Scholar] [CrossRef]
  56. Hossain, M.D.; Chen, D. Target-Based Building Extraction from High-Resolution RGB Imagery Using GEOBIA Framework and Tabular Deep Learning Model. Geomatica 2024, 76, 100007. [Google Scholar] [CrossRef]
  57. Fu, Y.; Liu, K.; Shen, Z.; Deng, J.; Gan, M.; Liu, X.; Lu, D.; Wang, K. Mapping Impervious Surfaces in Town–Rural Transition Belts Using China’s GF-2 Imagery and Object-Based Deep CNNs. Remote Sens. 2019, 11, 280. [Google Scholar] [CrossRef]
  58. Li, X.; Xu, F.; Li, L.; Xu, N.; Liu, F.; Yuan, C.; Chen, Z.; Lyu, X. AAFormer: Attention-attended Transformer for Semantic Segmentation of Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2024, 21, 1–5. [Google Scholar] [CrossRef]
  59. Wang, C.; He, S.; Wu, M.; Lam, S.-K.; Tiwari, P.; Gao, X. Looking Clearer with Text: A Hierarchical Context Blending Network for Occluded Person Re-Identification. IEEE Trans. Inf. Forensics Secur. 2025, 20, 4296–4307. [Google Scholar] [CrossRef]
  60. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  61. Wang, C.; Cao, R.; Wang, R. Learning Discriminative Topological Structure Information Representation for 2D Shape and Social Network Classification Via Persistent Homology. Knowl. Based Syst. 2025, 311, 113125. [Google Scholar] [CrossRef]
  62. Zeng, Q.; Zhou, J.; Tao, J.; Chen, L.; Niu, X.; Zhang, Y. Multiscale Global Context Network for Semantic Segmentation of High-Resolution Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–13. [Google Scholar] [CrossRef]
  63. Kitaev, N.; Kaiser, Ł.; Levskaya, A. Reformer: The Efficient Transformer. arXiv 2020, arXiv:2001.04451. [Google Scholar] [CrossRef]
  64. Huang, L.; Yuan, Y.; Guo, J.; Zhang, C.; Chen, X.; Wang, J. Interlaced Sparse Self-Attention for Semantic Segmentation. arXiv 2019, arXiv:1907.12273. [Google Scholar] [CrossRef]
  65. Yin, P.; Zhang, D.; Han, W.; Li, J.; Cheng, J. High-Resolution Remote Sensing Image Semantic Segmentation via Multiscale Context and Linear Self-Attention. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 9174–9185. [Google Scholar] [CrossRef]
  66. Huang, Z.; Wang, X.; Huang, L.; Huang, C.; Wei, Y.; Liu, W. Ccnet: Criss-Cross Attention for Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 603–612. [Google Scholar]
  67. Wang, S.; Li, B.Z.; Khabsa, M.; Fang, H.; Ma, H. Linformer: Self-Attention with Linear Complexity. arXiv 2020, arXiv:2006.04768. [Google Scholar] [CrossRef]
  68. Tong, X.-Y.; Xia, G.-S.; Lu, Q.; Shen, H.; Li, S.; You, S.; Zhang, L. Land-Cover Classification with High-Resolution Remote Sensing Images Using Transferable Deep Models. Remote Sens. Environ. 2020, 237, 111322. [Google Scholar] [CrossRef]
  69. Wang, J.; Zheng, Z.; Ma, A.; Lu, X.; Zhong, Y. LoveDA: A Remote Sensing Land-Cover Dataset for Domain Adaptive Semantic Segmentation. arXiv 2021, arXiv:2110.08733. [Google Scholar]
  70. Ge, Z. Yolox: Exceeding Yolo Series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar] [CrossRef]
  71. Li, R.; Zheng, S.; Zhang, C.; Duan, C.; Su, J.; Wang, L.; Atkinson, P.M. Multiattention Network for Semantic Segmentation of Fine-Resolution Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–13. [Google Scholar] [CrossRef]
  72. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  73. Ma, A.; Wang, J.; Zhong, Y.; Zheng, Z. FactSeg: Foreground Activation-Driven Small Object Semantic Segmentation in Large-Scale Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
  74. Ding, L.; Tang, H.; Bruzzone, L. LANet: Local Attention Embedding to Improve the Semantic Segmentation of Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 426–435. [Google Scholar] [CrossRef]
  75. Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep High-Resolution Representation Learning for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3349–3364. [Google Scholar] [CrossRef]
  76. Sun, L.; Li, L.; Shao, Y.; Jiao, L.; Liu, X.; Chen, P.; Liu, F.; Yang, S.; Hou, B. Which Target to Focus on: Class-Perception for Semantic Segmentation of Remote Sensing. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–13. [Google Scholar] [CrossRef]
  77. Cheng, B.; Misra, I.; Schwing, A.G.; Kirillov, A.; Girdhar, R. Masked-Attention Mask Transformer for Universal Image Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1290–1299. [Google Scholar]
  78. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. Transunet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar] [CrossRef]
  79. Zhang, X.; Weng, Z.; Zhu, P.; Han, X.; Zhu, J.; Jiao, L. ESDINet: Efficient Shallow-Deep Interaction Network for semantic Segmentation of High-Resolution Aerial Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–15. [Google Scholar] [CrossRef]
  80. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding Convolution for Semantic Segmentation. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; IEEE: New York, NY, USA, 2018; pp. 1451–1460. [Google Scholar]
  81. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual Explanations from Deep Networks Via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
Figure 1. Schematic diagram of the integrated framework of GEOBIA and deep learning for image classification. (a) Integration of GEOBIA with CNN-based segmentation results. (b) Object-based patch classification. (c) Feature extraction using deep learning models. Path (1) displays using DL models to extract deep abstract features of segment attributes for classification. Path (2) illustrates combining segment attributes with deep features of images for feature classification. (d) Heterogeneous patch filtering.
Figure 1. Schematic diagram of the integrated framework of GEOBIA and deep learning for image classification. (a) Integration of GEOBIA with CNN-based segmentation results. (b) Object-based patch classification. (c) Feature extraction using deep learning models. Path (1) displays using DL models to extract deep abstract features of segment attributes for classification. Path (2) illustrates combining segment attributes with deep features of images for feature classification. (d) Heterogeneous patch filtering.
Remotesensing 17 02652 g001
Figure 2. Processing flow of the EIODA algorithm. (a) Original remote sensing image; (b) overlay of the EIODA segmentation result on the original image; (c) display of the segmented image, where the pixel values in the R, G, and B channels represent the spectral mean values of the segment.
Figure 2. Processing flow of the EIODA algorithm. (a) Original remote sensing image; (b) overlay of the EIODA segmentation result on the original image; (c) display of the segmented image, where the pixel values in the R, G, and B channels represent the spectral mean values of the segment.
Remotesensing 17 02652 g002
Figure 3. The structure of SWTB. SWTB applies window-based and shifted window-based multi-head self-attention to capture local and global dependencies. Each attention layer is followed by a feed-forward network, with layer normalization and residual connections to enhance stability and feature representation.
Figure 3. The structure of SWTB. SWTB applies window-based and shifted window-based multi-head self-attention to capture local and global dependencies. Each attention layer is followed by a feed-forward network, with layer normalization and residual connections to enhance stability and feature representation.
Remotesensing 17 02652 g003
Figure 4. The overall architecture of GOFENet.
Figure 4. The overall architecture of GOFENet.
Remotesensing 17 02652 g004
Figure 5. Structure of the CSM. The input feature map is split into multiple channel groups and processed by parallel group convolutions with different kernel sizes. Channel attention is applied within each group, followed by concatenation and softmax fusion to integrate both intra-group and cross-group information.
Figure 5. Structure of the CSM. The input feature map is split into multiple channel groups and processed by parallel group convolutions with different kernel sizes. Channel attention is applied within each group, followed by concatenation and softmax fusion to integrate both intra-group and cross-group information.
Remotesensing 17 02652 g005
Figure 6. Structure of the FEM. It enhances feature representations by integrating both local and global contextual information. Local spatial cues are captured via horizontal and vertical pooling, while global dependencies are modeled through a SWTB-guided channel attention mechanism. The final output is obtained through adaptive fusion of the two branches.
Figure 6. Structure of the FEM. It enhances feature representations by integrating both local and global contextual information. Local spatial cues are captured via horizontal and vertical pooling, while global dependencies are modeled through a SWTB-guided channel attention mechanism. The final output is obtained through adaptive fusion of the two branches.
Remotesensing 17 02652 g006
Figure 7. Structure of the SDFM. It adaptively fuses shallow features (SF) and deep features (DF) via a pixel attention module (PAM). DF is first compressed and added to SF to obtain the initial fused feature AF. In PAM, AF is refined through parallel channel and spatial attention branches to generate the attention map (AM). AF and AM are then fused to produce the pixel-level attention map (PM).
Figure 7. Structure of the SDFM. It adaptively fuses shallow features (SF) and deep features (DF) via a pixel attention module (PAM). DF is first compressed and added to SF to obtain the initial fused feature AF. In PAM, AF is refined through parallel channel and spatial attention branches to generate the attention map (AM). AF and AM are then fused to produce the pixel-level attention map (PM).
Remotesensing 17 02652 g007
Figure 8. Percentage of each semantic label in the GID and the LoveDA dataset.
Figure 8. Percentage of each semantic label in the GID and the LoveDA dataset.
Remotesensing 17 02652 g008
Figure 12. Examples of semantic segmentation results on the GID.
Figure 12. Examples of semantic segmentation results on the GID.
Remotesensing 17 02652 g012
Figure 13. Visualization comparison of object contours and ground truth boxes identified by the current state-of-the-art methods and GOFENet on the GID.
Figure 13. Visualization comparison of object contours and ground truth boxes identified by the current state-of-the-art methods and GOFENet on the GID.
Remotesensing 17 02652 g013
Figure 16. Quantitative comparison of GOFENet variants on the GID (left) and the LoveDA dataset (right), showing mIoU, GFLOPs, and FPS. The figure illustrates how each model balances accuracy and efficiency, enabling flexible deployment under varying computational constraints.
Figure 16. Quantitative comparison of GOFENet variants on the GID (left) and the LoveDA dataset (right), showing mIoU, GFLOPs, and FPS. The figure illustrates how each model balances accuracy and efficiency, enabling flexible deployment under varying computational constraints.
Remotesensing 17 02652 g016
Figure 17. Grad-CAM visualization results from the final classification layer of several models, showing the attention maps for the road class in the GID and the urban residential class in the LoveDA dataset. Regions with higher attention values, depicted by warmer colors, correspond to areas where the network demonstrates greater confidence in its classification.
Figure 17. Grad-CAM visualization results from the final classification layer of several models, showing the attention maps for the road class in the GID and the urban residential class in the LoveDA dataset. Regions with higher attention values, depicted by warmer colors, correspond to areas where the network demonstrates greater confidence in its classification.
Remotesensing 17 02652 g017
Figure 18. Comparison of large-scale land cover classification accuracy after introducing SDFM across different models (a) on the GID, and (b) on the LoveDA dataset. Results from both datasets demonstrate that SDFM effectively improves the classification accuracy of large-scale and spatially continuous land cover types.
Figure 18. Comparison of large-scale land cover classification accuracy after introducing SDFM across different models (a) on the GID, and (b) on the LoveDA dataset. Results from both datasets demonstrate that SDFM effectively improves the classification accuracy of large-scale and spatially continuous land cover types.
Remotesensing 17 02652 g018
Table 4. Comparison of model complexity, speed, and performance on the GID, with complexity and speed evaluated using a 512 × 512 input.
Table 4. Comparison of model complexity, speed, and performance on the GID, with complexity and speed evaluated using a 512 × 512 input.
MethodBackboneComplexity (GFLOPS)mIoU
(%)
Speed
(FPS)
UNet [19]ResNet50184.5759.7365.56
PSPNet [24]ResNet50118.4460.6589.24
DeeplabV3+ [23]ResNet50153.2860.3777.31
DANet [25]ResNet5069.7060.9991.45
SegFormer [36]MIT-B126.6061.7473.07
DC-Swin [39]Swin-Tiny100.9361.0846.17
MANet [71]ResNet50157.0264.6965.05
UNetFormer [44]ResNet1823.5163.7972.75
GOFENet-tResNet18112.9863.2974.75
GOFENet-sResNet34262.7264.7242.49
GOFENetResNet50510.6866.0216.72
Table 5. Comparison of model complexity, speed, and performance on the LoveDA dataset, with complexity and speed evaluated using a 1024 × 1024 input.
Table 5. Comparison of model complexity, speed, and performance on the LoveDA dataset, with complexity and speed evaluated using a 1024 × 1024 input.
MethodBackboneComplexity (GFLOPS)mIoU
(%)
Speed
(FPS)
UNet [19]ResNet50373.147.822.9
PSPNet [24] *ResNet50105.748.352.2
DeeplabV3+ [23] *ResNet5095.847.653.7
Segmenter [35] *ViT-Tiny26.847.114.7
DANet [25]ResNet50278.850.239.7
SegFormer [36]MIT-B1106.146.526.2
MANet [71]ResNet50322.645.721.4
BANet [40] *ResT-Lite52.649.611.5
DC-Swin [39] *Swin-Tiny183.850.623.6
TransUNet [78] *ViT-R50803.448.913.4
SwinUperNet [30] *Swin-Tiny349.150.019.5
GOFENet-tResNet18223.951.233.9
GOFENet-sResNet34525.452.018.3
GOFENetResNet50980.851.98.9
Note: the parameters of the models marked with an asterisk (*) are derived from reference [51].
Table 6. A comparative analysis of the segmentation performance between GOFENet and simplified models.
Table 6. A comparative analysis of the segmentation performance between GOFENet and simplified models.
ModelBackboneLoveDAGID
BackgroundBuildingRoadWaterBarrenForestAgriculturemIoU (%)mPA (%)OA (%)mIoU (%)
GOFENet-tResNet1845.2052.6256.1578.2820.8643.6261.9151.2475.3882.3563.29
GOFENet-sResNet3444.3155.0556.4379.0421.3645.1362.8252.0276.4083.3864.72
GOFENetResNet5045.2055.6356.7578.1319.2844.8863.6051.9277.7884.0766.02
Table 7. Object size statistics and algorithm execution time on two datasets.
Table 7. Object size statistics and algorithm execution time on two datasets.
DatasetsObj. MinObj. MaxObj. MeanEIODA Average
Execution Time
GIDTrain2271310604.35 s/image
Val526551100
Test326331080
LoveDATrain117386231.81 s/image
Val11461606
Test11467584
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

He, T.; Chen, J.; Pan, D. GOFENet: A Hybrid Transformer–CNN Network Integrating GEOBIA-Based Object Priors for Semantic Segmentation of Remote Sensing Images. Remote Sens. 2025, 17, 2652. https://doi.org/10.3390/rs17152652

AMA Style

He T, Chen J, Pan D. GOFENet: A Hybrid Transformer–CNN Network Integrating GEOBIA-Based Object Priors for Semantic Segmentation of Remote Sensing Images. Remote Sensing. 2025; 17(15):2652. https://doi.org/10.3390/rs17152652

Chicago/Turabian Style

He, Tao, Jianyu Chen, and Delu Pan. 2025. "GOFENet: A Hybrid Transformer–CNN Network Integrating GEOBIA-Based Object Priors for Semantic Segmentation of Remote Sensing Images" Remote Sensing 17, no. 15: 2652. https://doi.org/10.3390/rs17152652

APA Style

He, T., Chen, J., & Pan, D. (2025). GOFENet: A Hybrid Transformer–CNN Network Integrating GEOBIA-Based Object Priors for Semantic Segmentation of Remote Sensing Images. Remote Sensing, 17(15), 2652. https://doi.org/10.3390/rs17152652

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop