Next Article in Journal
YOLO-TARC: YOLOv10 with Token Attention and Residual Convolution for Small Void Detection in Root Canal X-Ray Images
Previous Article in Journal
A Variable Reduction Approach for Microbeams on Elastic Foundation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MCFNet: Multi-Scale Contextual Fusion Network for Salient Object Detection in Optical Remote Sensing Images

1
School of Information and Electrical Engineering, Hangzhou City University, Hangzhou 310015, China
2
School of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 3035; https://doi.org/10.3390/s25103035
Submission received: 31 March 2025 / Revised: 29 April 2025 / Accepted: 7 May 2025 / Published: 12 May 2025
(This article belongs to the Section Optical Sensors)

Abstract

:
The rapid advancement of deep learning has catalyzed progress in salient object detection (SOD), extending its impact to the domain of optical remote sensing images (ORSIs). Despite increasing attention, salient object detection for optical remote sensing images (ORSI-SOD) remains highly challenging due to the intrinsic complexities of remote sensing scenes. In particular, severe variations in object scale and quantity, cluttered backgrounds, and irregular object morphologies significantly hinder accurate target localization and boundary delineation. In response to these challenges, we introduce the Multi-scale Contextual Fusion Network (MCFNet) for ORSI-SOD. MCFNet incorporates a Semantic-Aware Attention Module (SAM), which provides explicit semantic guidance during feature extraction. By producing preliminary semantic masks, SAM enables the network to capture long-range contextual dependencies, thereby enhancing localization accuracy for salient objects exhibiting substantial scale variation and structural complexity. In addition, MCFNet integrates a Contextual Interconnection Module (CIM), which promotes effective fusion of local and global contextual features. By facilitating cross-layer interactions and adopting a multiscale refinement strategy, CIM enriches texture representations while suppressing background interference, leading to smoother object boundaries and more precise delineation of salient regions. Extensive evaluations conducted on three standard ORSI-SOD benchmark datasets demonstrate the superior performance of MCFNet compared to existing methods, highlighting its robustness and efficiency in handling challenging remote sensing scenarios.

1. Introduction

Salient Object Detection (SOD) [1] is a fundamental task in computer vision, aimed at emulating the human visual system’s ability to identify and localize the most visually prominent regions within a scene. In recent years, SOD has found widespread applications in various domains, including image segmentation [2,3], image quality assessment [4], visual target tracking [5,6], and content-aware image editing [7]. Optical Remote Sensing Images (ORSIs) are captured by satellites and onboard antenna sensors. Salient object detection in remote sensing images (ORSI-SOD) [8] can effectively identify the most salient target areas in ORSIs with the help of in-depth learning and detailed analysis of image features in both spatial and spectral dimensions. However, the ORSI-SOD task faces several challenges due to the high-altitude and long-distance nature of image capture. In particular, the variability in target numbers and scales is a significant issue, as a single image may contain numerous targets of varying sizes. Furthermore, the complexity of the background environment poses further challenges, as it can obscure the detection of salient targets. Irregular topological features within ORSIs also contribute to the difficulty of accurately identifying prominent targets.
Recent years have witnessed notable advances in salient object detection for natural scene images (NSI-SOD), leading to the development of numerous effective methods. These methods can be broadly divided into main types: traditional methods and deep learning-based methods. Traditional methods [9,10] rely primarily on model-driven feature extraction [11] and classical machine learning algorithms [12]. Although grounded in solid theoretical foundations, these approaches frequently encounter difficulties in delivering precise detection results when faced with intricate or cluttered scenes because of limited generalization capabilities and poor adaptability to diverse background conditions. In contrast, the advent of convolutional neural networks (CNNs) has led to the development of data-driven deep learning models [13,14] that have demonstrated remarkable performance in NSI-SOD tasks. These approaches leverage large-scale annotated datasets to learn robust and discriminative feature representations. For instance, the PA-KRN [15] adopts a hierarchical architecture to enhance salient object detection in natural images, while the SUCA [16] further improves multiscale target perception through stacked interaction networks. Such deep learning methods have consistently outperformed traditional methods in accuracy and robustness. However, directly applying NSI-SOD models to ORSI-SOD tasks is not trivial [17]. ORSIs differ markedly from NSIs in terms of spatial distribution, viewpoint, resolution, and overall scene composition. In addition, they exhibit greater variability in object scale, shape, and appearance. These domain-specific challenges limit the effectiveness of NSI-SOD models when transferred to remote sensing scenarios. As a result, there is a critical need for the development of specialized and adaptive methods tailored to the unique demands of ORSI-SOD.
In response to the limitations of transferring NSI-SOD methods to remote sensing scenarios, several approaches [18,19] specifically designed for ORSI-SOD have been proposed. For example, nested networks [20] generate high-level attention maps by integrating low-level cues with deeper features through a dense attention mechanism. However, these methods often overlook the impact of high-level semantic information on low-level feature representation during attention transfer, which compromises the refinement of object boundaries. ACCoNet [21] improves saliency detection by enabling cross-layer feature interaction and effectively extracting contextual semantic information. Despite its effectiveness in contextual modeling, it incurs substantial computational overhead. AESINet [22] employs separate modules to process high- and low-level features extracted from multiple CNN layers and captures adaptive edge information through perceptual mapping of self-pooled differences. Although these approaches are specifically tailored for ORSI-SOD, they remain limited when confronted with the compounded challenges of varying object quantities and scales, intricate background environments, and irregular object topologies.
Motivated by the limitations of existing approaches, we aim to tackle the persistent challenges in ORSI-SOD, such as background clutter, variable object scales and densities, and inherently low contrast between targets and their surroundings. In this study, we introduce a novel Multi-scale Contextual Fusion Network (MCFNet), which is specifically engineered to achieve robust and accurate saliency detection under complex and diverse conditions. MCFNet is constructed based on two complementary modules: the Semantic Attention Module (SAM) and the Contextual Interconnection Module (CIM). SAM plays a pivotal role in precise target localization. It integrates coarse semantic masks as initial priors, providing robust and explicit semantic guidance to subsequent layers. In contrast to conventional methods, which often falter in capturing long-range dependencies, SAM introduces a global semantic perception mechanism. This mechanism effectively overcomes the constraints imposed by local receptive fields, allowing the network to accurately localize salient regions, even when dealing with objects of varying scales and spatially dispersed distributions. Working in tandem with SAM, CIM is crucial for enhancing object saliency and suppressing background noise. Given that RSIs frequently contain background elements that resemble foreground targets in texture or color, the risk of false positives remains high. CIM addresses this issue by employing a multiscale refinement strategy, enabling deep, cross-layer feature interactions. Through the mining of inter-scale contextual dependencies, CIM effectively integrates high-level semantic cues with low-level texture details. To comprehensively evaluate the effectiveness of MCFNet, we perform extensive experiments on three widely used public ORSI-SOD benchmarks. Across all datasets and most evaluation metrics, MCFNet consistently outperforms 20 state-of-the-art methods, demonstrating clear superiority in both detection accuracy and robustness.
The main contributions of this work can be outlined as follows.
1.
We introduce a Multi-scale Contextual Fusion Network (MCFNet) built upon an encoder–decoder framework, designed to effectively extract high-level semantic cues while promoting the integration of low-level detailed features with high-level contextual information via a cross-level interconnection strategy.
2.
We introduce a Semantic-Aware Attention Module (SAM), which leverages preliminary semantic masks to guide the extraction of deep semantic information in ORSIs, thereby enhancing the ability to localize salient targets under complex conditions.
3.
We design a Contextual Interconnection Module (CIM) that enriches texture representations across multiple scales by refining local features and adaptively fusing contextual information from adjacent layers.
4.
Comprehensive evaluations conducted on three datasets, ORSSD, EORSSD, and ORSI4199, show that our approach achieves superior performance compared to existing methods. Furthermore, ablation studies confirm the contribution and effectiveness of each proposed module.

2. Related Work

2.1. Salient Object Detection Methods in NSIs

The SOD task was originally developed for NSIs analysis. Over the past few years, NSI-SOD has witnessed substantial progress, with a wide range of methods proposed, spanning both traditional and deep learning approaches. Traditional NSI-SOD approaches are generally grouped into three categories: unsupervised methods [11,23,24,25], semi-supervised methods [26,27], and supervised methods [28]. Among these, unsupervised methods have historically dominated, typically relying on interpretable handcrafted features extracted through image preprocessing techniques, such as the sparse graph [11], the saliency tree [23], the maximal entropy random walk [24], and the regularized random walks ranking [25]. In contrast, relatively few traditional methods adopt semi-supervised or supervised strategies. For instance, Zhou et al. [26] developed a boundary-guided approach for pseudo-label generation and designed a semi-supervised learning framework based on a linear feedback control mechanism for iterative optimization. Liang et al. [28] applied supervised learning via support vector machines to select salient features and eliminate redundancies, thereby accelerating model training. Despite their contributions, traditional NSI-SOD methods struggle to generalize across the diverse and complex scenarios encountered in ORSIs.
CNN-based methods for NSI-SOD have overcome many of the limitations inherent in traditional approaches. Unlike traditional techniques, most CNN-based methods adopt a supervised learning paradigm, relying on large-scale annotated datasets to significantly enhance detection accuracy. Over the past years, several influential feature processing strategies have been introduced, including deep supervision [29], multi-scale interaction [14,30], and edge-aware feature enhancement [13]. Huang et al. [29] were among the first to incorporate deep supervision into NSI-SOD frameworks, inspiring a wave of subsequent research exploring its potential. PoolNet [30] utilizes a combination of top-down and bottom-up pathways to extract and integrate multi-scale contextual information. GateNet [14] introduces collapsed spatial pyramid pooling to aggregate spatial feature maps at varying receptive fields, thereby enhancing multi-scale representation. Meanwhile, EGNet [13] focuses on capturing edge features, improving the precision of object localization by highlighting object boundaries more effectively. Although deep learning-based NSI-SOD methods have achieved remarkable success, applying them directly to ORSI-SOD tasks poses considerable challenges. The fundamental differences between NSIs and ORSIs pose substantial challenges, limiting the generalizability and effectiveness of NSI-SOD models in remote sensing contexts.

2.2. Salient Object Detection Methods in ORSIs

Compared with the NSI-SOD task, ORSI-SOD faces more complex and unique challenges. To tackle these challenges, CNN-based models for ORSI-SOD offer promising solutions owing to their strong abilities in feature representation and learning. Specifically, Ma et al. explored efficient segmentation solutions for SAR images from different technical perspectives, proposing a fast segmentation method based on deep task-specific superpixel sampling and soft graph convolution [31] as well as a task-specific region merging strategy for rapid segmentation [32]. Li et al. [20] designed an LV-shaped network architecture, where an L-shaped pyramid module processes inputs at different resolutions, and a V-shaped nested connections structure precisely identifies the position and boundaries of salient objects. Luo et al. [33] further enhanced feature learning by adopting convolution kernels of diverse shapes, which enables it to capture a wide range of salient target features, including topology and orientation, thereby improving the model’s adaptability to complex target morphologies. Tu et al. [34] combined local features extracted via pyramid pooling with global context to construct a strong multi-scale regional representation. Gu et al. [35] introduced an edge-guided branching structure to generate rich boundary information, leading to accurate localization and highlighting of salient regions under boundary label supervision. Despite the remarkable progress made using deep neural networks in ORSI-SOD, CNN-based models face inherent limitations in global modeling capabilities, which often results in difficulty detecting objects across varying scales simultaneously. This problem is especially prominent in RSIs, which usually contain richer and more diverse scales and target morphologies, putting forward higher requirements on the global perception ability of the model.
Transformers [36] excel at capturing long-range interactions, enabling effective extraction of global context from visual data. Consequently, recent studies have increasingly incorporated Transformers into the ORSI-SOD task to enhance global modeling capabilities. Specifically, Zhang et al. [37] designed a dense attention flow structure, which facilitates the efficient connection and fusion of features at multiple levels through an attention mechanism. Zhou et al. [38] introduced a U-shaped architecture, incorporating edge-aware positional attention units to strengthen edge cue extraction, thereby improving the detection accuracy of salient targets. Inspired by this, we propose an efficient global–local perception network based on contextual interactions. Within this network, global information, which is extracted through semantic-aware attention mechanisms, is progressively integrated with local texture details via contextual interactions at every layer. This novel architecture enhances the model’s ability to model complex global dependencies while enabling accurate localization and refinement of fine-grained local features.

3. Method

In this study, we present a comprehensive analysis of the proposed Multi-scale Contextual Fusion Network (MCFNet), as shown in Figure 1. MCFNet adopts an encoder–decoder framework and integrates two key components: the Semantic-aware Attention Module (SAM) and the Contextual Interconnection Module (CIM). In Section 3.1, we provide a brief overview of the MCFNet framework. Subsequently, in Section 3.2 and Section 3.3, we offer an in-depth discussion of these two core components. Finally, in Section 3.4, we elaborate on the design of the decoder and the hybrid loss function.

3.1. Network Overview

Figure 1 presents the architecture of MCFNet, which adopts a generalized encoder–decoder framework comprising an encoder, a semantic-aware attention module (SAM), a contextual interconnection module (CIM), and a decoder. For feature extraction, we employ Res2Net-50 [36] as the backbone. The encoder is structured into five blocks, which capture multi-scale features encompassing both low-level and high-level information. The output of each block can be represented as f e i R C i × H i × W i , i { 1 , 2 , 3 , 4 , 5 } , where H i and W i are 256 2 i and C i { 64 , 128 , 256 , 512 , 1024 } . Notably, the high-level feature encapsulates rich semantic information crucial for ORSI-SOD. For high-level feature f e 5 , we employ SAM, which enhances semantic detail and suppresses background noise, thereby improving the localization of salient regions in complex scenes. In contrast, the low-level features f e 1 , f e 2 , f e 3 , and f e 4 possess abundant texture information but lack global context. To address this, the CIM aggregates cross-scale features from adjacent blocks, facilitating enhanced interaction between low-level and high-level representations. Finally, the decoder consists of five blocks that progressively reconstruct the resolution. The hybrid loss function is employed to enable deep supervision and produce the final output prediction maps.

3.2. Semantic-Aware Attention Module

When processing remote sensing images, salient objects exhibit substantial variability in number and scale and are often obscured by complex backgrounds, presenting significant challenges for accurate target identification and localization. To address these issues, most CNN-based methods typically employ deep convolutional stacks to extract features from increasingly large receptive fields. However, these methods tend to lose critical target information during progressive downsampling, largely due to the inherent limitations of CNNs in global context modeling. Although CNNs effectively capture local image content, their extracted features are insufficient to suppress background noise and achieve precise target localization. In contrast, the self-attention mechanism has gained recognition for its superior ability to model global dependencies [39]. Motivated by this insight, we propose the Semantic-aware Attention Module (SAM), which leverages a preliminary semantic mask to guide feature extraction. SAM enhances critical global information, thereby facilitating robust and accurate target localization by efficiently capturing long-range dependencies and overcoming the limitations of local feature extraction.
The high-level feature f e 5 R C × H × W extracted from the encoder’s fifth layer contains essential semantic information crucial for object detection. To enhance the localization of objects, we first process f e 5 through a series of operations—including a 1 × 1 convolution, batch normalization, ReLU activation, and sigmoid function—to generate a preliminary semantic mask f m . In this process, we preserve the preliminary semantic mask as continuous values within the [ 0 , 1 ] interval rather than binarizing them. Subsequently, we input the preliminary mask f m along with the original f e 5 into our proposed SAM. As depicted in Figure 2, within the SAM, a bilinear fusion scheme is first employed to compute the attention matrix. Specifically, the input feature f e 5 is efficiently decomposed into two intermediate features x 1 R 1 × H × W and x 2 R 1 × H × W via pointwise convolution. Then, the depthwise separable convolution layer is used to generate a query vector Q R C × ( H × W ) and a key vector K R C × ( H × W ) . This effective combination of pointwise and depthwise separable convolutions preserves powerful feature extraction while significantly reducing computational complexity. To emphasize salient targets in the remote sensing image, we apply the preliminary semantic mask f m to both Q and K via element-wise multiplication. This operation accentuates the target region while suppressing extraneous information, yielding the semantic-aware query Q and key K , which can be formulated as:
Q = f m D S c o n v ( P W C o n v ( f e 5 ) ) ,
K = ( f m D S c o n v ( P W C o n v ( f e 5 ) ) ) T
where D S C o n v ( ) is the depthwise separable convolution, P W C o n v ( ) is the pointwise convolution, and ⊙ represents the element-wise multiplication. Subsequently, Q and the transposed K undergo matrix multiplication to produce an attention map A R ( H × W ) × ( H × W ) , which quantifies the association strength between different spatial locations. The computational results are presented below.
A = S o f t m a x ( Q K )
where ⊗ is the matrix multiplication, and S o f t m a x ( ) denotes the softmax layer. Based on these operations, the resulting attention matrix identifies effectively encoded spatial information from a global perspective, thereby capturing and leveraging semantic cues.
To construct the value matrix V in the self-attention mechanism, we apply a depthwise separable convolution layer to the high-level feature f e 5 to efficiently refine the target feature representation. Owing to its inherently small receptive field, depthwise separable convolution effectively emphasizes local feature details. In parallel, feature f e 5 is processed through three sequential dilated convolution layers with expansion rates of 2, 4, and 2, respectively, to broaden the receptive field and capture long-range contextual information. By integrating these two strategies, the value matrix V simultaneously captures local details and global context, which is computed as:
V = D S C o n v ( f e 5 ) D C o n v r = 2 ( D C o n v r = 4 ( D C o n v r = 2 ( f e 5 ) ) )
where ⊕ represents the element-wise summation and D C o n v r = 2 and D C o n v r = 4 denote dilated convolutions with dilation rates of 2 and 4. In the final stage, we multiply the attention matrix A with the value matrix V and obtain the semantic-aware feature vector f s 5 . Then, the semantic-aware feature f s is fused with the original feature vector f e 5 to yield the final output f t 5 of SAM, which can be formulated as follows:
f t 5 = D S C o n v ( f e 5 V α · ( A V ) )
where α is a learnable scaling parameter. Within the SAM, we employ a semantic-aware attention mechanism designed to amplify the salient object’s semantic representation while attenuating noise interference, thereby enabling high-precision target localization.

3.3. Contextual Interconnection Module

By employing SAM, we improve salient object localization by emphasizing global semantic cues. However, it may overlook fine-grained details essential for precise segmentation. In contrast, low-level features f e 1 , f e 2 , f e 3 , and f e 4 encode high-resolution textures and subtle spatial details crucial for delineating accurate object boundaries. Recognizing that effective segmentation requires a balanced interplay between semantic-rich high-level context and detailed low-level features, we propose the Contextual Interaction Module (CIM). CIM integrates adjacent-level features, facilitating multi-scale information interaction via a dedicated Hierarchical Refinement Unit, which restores detailed feature representations. Additionally, an adaptive weighted unit is introduced to further enhance local feature representations, contributing to improved segmentation precision and stability.
In Figure 3, CIM comprises three primary branches, CIM-2, CIM-3, and CIM-4. Each of these branches integrates contextual information derived from its corresponding local and adjacent feature scales. Notably, CIM-1 incorporates only two branches. Accordingly, the operation of CIM can be abstracted as a generalized function F ( ) , which can be formulated as follows:
f t i = F ( f e i , f e i + 1 ) , i = 1 F ( f e i 1 , f e t , f e i + 1 ) , i = 2 , 3 , 4
where f t i ( i { 1 , 2 , 3 , 4 } ) denotes the output features from CIM and represents the fusion of cross-level features from the previous, current, and subsequent feature blocks.
The localization branch incorporates two core components: the Hierarchical Refinement Unit (HRU) and the Adaptive Weighted Adjustment Unit (WAU). Initially, the input feature f e i is processed through the HRU, which utilizes asynchronous convolution techniques to extract features and optimize multi-scale details. To comprehensively capture local features at multiple scales, we construct four parallel branches, each tailored to a specific receptive scale. The first branch employs a single 1 × 1 convolutional layer with residual connection to preserve critical input information. The subsequent three branches expand the receptive fields to capture broader spatial contexts by employing asymmetric convolutions with kernel sizes of t × 1 and 1 × t , where t is set to 3, 5, and 7. Specifically, the t × 1 convolution enhances sensitivity to vertically oriented structures by extending the kernel height, while the 1   ×   t convolution captures horizontally oriented details by increasing kernel width. This dual-directional receptive field expansion enables each branch to comprehensively encode contextual information at its respective scale. Following these processing steps, the four branches produce output feature maps h 1 , h 2 , h 3 , and h 4 . The specific computational procedure is shown below:
h r = C o n v 1 × t ( C o n v t × 1 ( C o n v 1 × 1 ( f e i ) ) ) , t = 3 , 5 , 7
where C o n v 1 × 1 ( ) denotes the convolution operation using a 1 × 1 kernel, while C o n v t × 1 ( ) and C o n v 1 × t ( ) indicate convolutions performed separately with kernels of dimensions t × 1 and 1 × t , respectively. Subsequently, the outputs h 1 , h 2 , h 3 , and h 4 generated by the HRU are fed into WAU. For each feature h r , a channel attention mechanism independently calculates the channel attention c r , enhancing the expressiveness along the channel dimension. The attention maps c r , r { 1 , 2 , 3 , 4 } are then used to derive corresponding weights w i , which are subsequently applied through element-wise multiplication to their respective feature maps, enabling adaptive weighting and refinement. All weighted feature maps are aggregated through summation to produce the integrated output feature f w i . The computational procedure is summarized as follows:
f w i = r = 1 4 exp ( Linear ( CA ( h p ) ) ) p = 1 4 exp ( Linear ( CA ( h p ) ) ) CA ( h r ) h r
where Linear ( ) is a linear layer and CA ( ) denotes channel attention operation. Thereafter, spatial information from f w i is extracted via the spatial attention mechanism, yielding the spatially enhanced feature f m i .
The two adjacent branches effectively capture features at different levels. Specifically, the lower-level feature f e i 1 is first rescaled to match the spatial dimensions of the current feature f e i . In order to enrich the local representations, we obtain the refined feature through the depthwise separable convolution and spatial attention mechanism to realize the accurate segmentation. The process can be formulated as:
f m i 1 = f w i SA ( D S C o n v ( D o w n × 2 ( f e i 1 ) ) i { 2 , 3 , 4 }
where f m i 1 is the output of the adjacent lower-level branch and D o w n × 2 ( ) represents the max-pooling layer with a kernel size of 3 × 3 and stride of 2. Similarly, the higher-level feature f e i + 1 employs the same operation to refine the global information, which is formulated as follows:
f m i + 1 = f w i SA ( D S C o n v ( U p × 2 ( f e i + 1 ) ) i { 1 , 2 , 3 , 4 }
where f m i 1 is the output of the adjacent higher-level branch and U p × 2 ( ) is the upsampling operation. Following these operations, we merge the output features from each branch with the current feature. The adjacent lower-level branch provides detailed texture information, whereas higher-level branches introduce global semantic context. The fusion process is described as follows:
f t i = f m i + 1 ( f m i f m i + 1 ) , i = 1 f m i 1 ( f m i 1 f m i f m i + 1 ) f m i + 1 , i = 2 , 3 , 4
In summary, CIM effectively connects and integrates contextual semantic information to facilitate accurate salient object segmentation.

3.4. Decoder and Loss Function

To obtain accurate salient objects, we employ the widely used deep supervision strategy during the decoding phase, enabling effective supervision [29] of intermediate salient maps. As shown in Figure 1, the decoder comprises five modules, each consisting of convolution layers followed by a deconvolution layer. Through this design, spatial resolution is progressively recovered, and feature representations are iteratively refined. Ultimately, this decoding process generates a salient projection map S i with dimensions 1 × 256 × 256 , where pixel values range between 0 and 1.
During training, pixel-level supervision is applied to each decoder stage. We employed a hybrid loss ( L t o t a l ) function combining pixel-level weighted binary cross-entropy (BCE) loss with map-level weighted intersection-over-union (IoU) loss. Specifically, the pixel-level weighted BCE loss captures the discrepancy between the predicted and actual masks at each pixel, while the map-level weighted IoU loss evaluates the spatial agreement by calculating the intersection-over-union between the predicted and actual masks. The combined hybrid loss can be formally expressed as:
L t o t a l = i = 1 5 ( L b c e ( S i , G ) + L i o u ( S i , G ) )
where G denotes the ground truth, and L b c e and L i o u represent the BCE loss and IoU loss, respectively.

4. Experiments and Results

4.1. Experiment Protocol

(1) Datasets. To assess the effectiveness of our MCFNet, extensive experiments were carried out on three benchmark ORSI-SOD datasets, i.e., ORSSD [20], EORSSD [37], and ORSI4199 [34]. The ORSSD dataset includes 800 images with pixel-level annotations, split into 600 for training and 200 for testing. The EORSSD dataset, an extension of ORSSD, contains 2000 high-resolution images with corresponding ground truths, typically divided into 1400 training and 600 testing samples. The ORSI4199 dataset is the latest and most challenging dataset in the ORSI-SOD field, comprising 2000 training and 2199 testing images. Its rich diversity and complex scenes pose significant challenges for detection models. For each dataset, training and evaluation are conducted using their designated splits.
(2) Implementation Details. Our MCFNet was trained and evaluated on a workstation equipped with an NVIDIA RTX 3090 GPU with 24 GB of memory. All input ORSIs were uniformly resized to 256 × 256 for both the training and testing phases. To improve model robustness and dataset variability, standard augmentation techniques such as horizontal flipping and rotation were employed. The network was optimized using the Adam optimizer [40], with an initial learning rate of 1 × 10 4 , which was decayed by a factor of 0.1 after 50 epochs. Training was conducted for 100 epochs with a batch size of 16.
(3) Evaluation Metrics. We thoroughly evaluated the proposed MCFNet against leading state-of-the-art approaches on three benchmark ORSI-SOD datasets, utilizing six standard evaluation metrics: S-measure ( S α ) [41], Mean Absolute Error (MAE, M ), mean E-measures ( E ξ m e a n ) [42], maximum E-measure ( E ξ m a x ), mean F-measure ( F β m e a n ) [43], and maximum F-measure ( F β m a x ). Specifically, the S-measure captures the structural similarity between region-aware and object-aware predictions, which can be expressed as:
M = 1 H × W i = 1 H j = 1 W S ( i , j ) G ( i , j ) ,
where S denotes the predicted salient map, G denotes the ground truth, and H and W denote the height and width, respectively. The MAE quantifies the average pixel-wise discrepancy between predicted saliency maps and ground truth. The mathematical expression is formulated as follows:
S α = α S o + ( 1 α ) S r , α = 0.5
where S o and S r denote the object-aware and region-aware structural similarity measures, respectively. Both E-measure and F-measure are evaluated in terms of their mean and maximum values, offering a more complete view of performance. The E-measure combines pixel-level precision and object-level consistency to assess structural coherence and is defined as follows:
E ξ = 1 H × W i = 1 H j = 1 W ϕ F M ( i , j )
where ϕ F M ( i , j ) represents the enhanced-alignment matrix. The F-measure offers a harmonic trade-off between precision and recall to reflect the overall segmentation accuracy and is defined as follows:
F β = ( 1 + β 2 ) precision × recall β 2 precision + recall , β 2 = 0.3
Additionally, we analyzed model performance using precision–recall (PR) curves, where curves closer to the top-right corner (1,1) indicate better performance in terms of both precision and recall.

4.2. Comparison with State-of-the-Art Methods

(1) Competing Methods. In this study, we present MCFNet and evaluate its effectiveness by benchmarking it against 20 leading methods on three widely used datasets, i.e., ORSSD, EORSSD, and ORSI4199, to comprehensively assess both accuracy and generalization capability. Our evaluation incorporates both qualitative and quantitative analyses. For comparative purposes, we selected eight prominent methods from the natural image salient object detection, including R3Net [44], PoolNet [30], EGNet [13], SUCA [16], U2Net [45], GateNet [14], MINet [46], and PA-KRN [15]. Meanwhile, we compared our approach with twelve specialized ORSI-SOD methods, including DAFNet [37], EMFINet [47], AGNet [48], MJRBM [34], ACCoNet [21], CorrNet [49], BAFS-Net [35], ERPNet [38], AESINet [22], SFANet [50], ADSTNet [51], and LSHNet [17]. To ensure fairness and consistency, all models were trained and tested under identical experimental conditions and parameter configurations. Furthermore, predictions from competing models were generated using publicly available implementations provided by their respective authors, ensuring the reliability and reproducibility of our results.
(2) Quantitative Comparison. Table 1 presents a comprehensive quantitative comparison of our method against 20 state-of-the-art approaches on the EORSSD, ORSSD, and ORSI4199 datasets. On the ORSSD dataset, our proposed method achieved superior overall performance, ranking first in five evaluation metrics and second in the remaining metric. Notably, PA-KRN previously demonstrated the highest performance among NSI-SOD methods. However, our method outperforms PA-KRN, with significant improvements of 2.52%, 1.80%, 1.60%, 3.93%, and 3.51% in S α , E ξ m e a n , E ξ m a x , F β m e a n , and F β m a x , respectively. While our approach shows slightly lower performance in the S-measure metric compared to the leading ORSI-SOD method, LSHNet, it achieves notable enhancements in all other five metrics, particularly improvements of 0.30% and 0.18% in E ξ m e a n and F β m e a n , respectively.
On the larger EORSSD dataset, our method also achieved outstanding performance. Compared with the PA-KRN, our approach demonstrated improvements of 2.45%, 4.24%, 1.59%, 2.10%, and 2.00% in terms of F β m a x , F β m e a n , E ξ m a x , E ξ m e a n , and S α , respectively. When compared to the second-best method, LSHNet, our method demonstrated improvements of 0.42% and 0.80% in terms of E ξ m e a n and F β m e a n , respectively. Moreover, we observed a substantial improvement in MAE, achieving a notable reduction of 9.38% compared to LSHNet, demonstrating the superiority of our proposed method.
On the challenging ORSI4199 dataset, our method exhibited state-of-the-art performance, achieving the highest rank across all six evaluation metrics. Specifically, our model substantially surpassed PA-KRN by 4.03%, 3.04%, 2.45%, 5.30%, and 3.89% in terms of S α , E ξ m e a n , E ξ m a x , F β m e a n , and F β m a x , respectively. Compared to the second-best model, our approach improved critical metrics of M and S α by 3.0% and 0.10%, respectively. Figure 4 illustrates the PR curves on the three benchmark datasets, demonstrating that our MCFNet achieves superior precision and recall. Notably, the PR curves for our method are positioned closest to the upper-right corner, indicating consistently higher performance compared with all of the other evaluated methods.
Overall quantitative analyses confirm that our proposed model achieves new state-of-the-art performance on the ORSI-SOD task, striking a favorable balance between accuracy and computational efficiency. Furthermore, a comparative evaluation between specialized ORSI-SOD and general NSI-SOD methods reveals a significant advantage of ORSI-specific models. These findings underscore the critical importance and urgency of developing dedicated approaches tailored explicitly for remote sensing imagery.
(3) Qualitative Analysis. In Figure 5, we present representative and challenging ORSI examples to qualitatively compare MCFNet with nine state-of-the-art approaches. These scenarios encompass cases such as interference from surrounding elements, clusters of small objects, shadow-covered targets, complex background textures, low-contrast regions, and irregularly shaped structures. In remote sensing imagery, non-salient regions frequently cause distractions, which can result in inaccurate detection results. As illustrated in the first and second rows of Figure 5, other methods such as LSHNet, CorrNet, and EMFINet exhibit limitations, either fragmenting target shapes or misidentifying background objects as salient targets. In contrast, our proposed method effectively addresses these challenges by preserving the integrity of salient objects while mitigating interference. MCFNet accurately differentiates salient targets from distractors by leveraging contextual modeling. Furthermore, the integration of SAM guides the model to focus on discriminative features relevant to the salient target while suppressing the influence of non-salient objects. This design enables our model to robustly isolate and highlight salient objects even in complex scenes with potential interferences.
In ORSI, the variability in shooting distances and angles often results in objects exhibiting tiny features, particularly in scenes containing multiple small targets, which presents a significant challenge for SOD. The third and fourth rows of Figure 5 illustrate qualitative results in such complex scenarios. As evident from the visualizations, existing state-of-the-art methods (i.e., LSHNet, ADSTNet, and GateNet) often lead to omissions. Meanwhile, methods such as SFANet and PA-KRN erroneously identify non-target regions as salient objects. In contrast, our method demonstrates superior performance in multiple tiny object scenarios, accurately detecting all targets without false positives or omissions.
Shadows present a significant challenge in SOD, as they are often misclassified as salient regions, leading to inaccurate detection. As illustrated in the fifth and sixth rows of Figure 5, many existing methods erroneously highlight shadowed areas, which affects the completeness of target detection. In contrast, our proposed approach accurately identifies target objects while preserving their structural integrity, effectively mitigating the interference of redundant shadow regions.
Effectively mitigating background interference is a critical challenge in salient object detection, particularly in complex environments. As shown in row 7 of Figure 5, all comparative methods, except for our proposed MCFNet, fail to suppress background interference, resulting in excessive background inclusion and incomplete object detection. In row 8, only our method successfully isolates the salient target from the background, achieving precise and robust detection. These results strongly demonstrate the effectiveness and resilience of our model in handling complex background conditions.
Low-contrast object detection remains a challenging aspect of SOD, as salient targets often exhibit visual similarity to the background, making them difficult to distinguish. As shown in the low-contrast bridge detection example in row 9 of Figure 5, existing methods struggle to accurately delineate the complete structure of the bridge, resulting in incomplete predictions. Similarly, in the scenario presented in row 10, methods such as ADSTNet, EMFINet, and PA-KRN fail to correctly identify the true target object, mistakenly segmenting visually similar background regions instead. In contrast, our method not only accurately detects the target but also precisely delineates its boundaries, demonstrating superior robustness in low-contrast settings.
Salient objects in remote sensing images often exhibit complex and irregular topologies, making accurate edge delineation particularly challenging. In practice, methods such as LSHNet, SFANet, and MJRBM frequently fail to capture the full extent of rivers and buildings, producing fragmented or jagged boundaries that poorly represent the true geometry of the targets. In contrast, our method demonstrates superior performance by accurately detecting rivers with well-defined contours and preserving the structural integrity of buildings. This advantage stems from the integration of a semantic-aware attention mechanism and an adjacent contextual information fusion strategy. The semantic-aware attention mechanism enhances detection completeness by guiding the model to focus more holistically on key target features. Simultaneously, the contextual interaction strategy significantly improves boundary sharpness and enables the model to more effectively handle irregular geometric structures.

4.3. Ablation Experiment

In this study, we conducted a series of ablation experiments to systematically evaluate the effectiveness of each component within MCFNet. Specifically, our analysis focused on four key aspects: (1) the individual contribution of each module, (2) the effectiveness of branches in CIM, (3) the effectiveness of semantic-aware attention, and (4) the model’s sensitivity to different backbone architectures. To ensure experimental fairness, each ablation variant was retrained under identical configurations, and their performance was rigorously evaluated and compared.
Effectiveness of the proposed module. To assess the practical contribution of the proposed modules, we designed four model variants for ablation studies. Variant 1 serves as the baseline model, incorporating the Res2Net backbone and the decoder. Variant 2 integrates SAM into the baseline, while variant 3 includes the CIM. Variant 4 combines both SAM and CIM, representing the full version of our model. Experiments were conducted on the ORSSD and EORSSD datasets, with quantitative results summarized in Table 2. With the integration of SAM, variant 2 achieved notable performance gains on the ORSSD dataset, with improvements of 1.51%, 0.50%, 0.46%, and 0.0019 in terms of S α , E ξ m a x , F β m a x , and M , respectively. With the addition of SAM, the F β m a x improved by 0.67% and 1.21% on the EORSSD and ORSI4199 datasets, respectively. Similarly, the addition of the CIM results in consistent and significant enhancements across all evaluation metrics. When SAM and CIM are incorporated, the model demonstrates the most substantial performance improvements in S α , E ξ m a x , F β m a x , and M surpassing the baseline by 1.91%, 0.96%, 1.91%, and 0.0039 on the EORSSD dataset, respectively. These results indicate that each module plays an indispensable role, as removing any one of them results in a noticeable decline in performance. As illustrated in the visual results in Figure 6, the incorporation of SAM can accurately strengthen the localization ability of objects, allowing it to more accurately identify salient regions and suppress background interference. Meanwhile, CIM facilitates effective fusion and interaction between high-level semantic features and low-level texture details, leading to a more coherent and natural delineation of object contours.
The effectiveness of branches in CIM. To further assess the contribution of individual branches within the CIM module, we designed two ablated variants: one excluding the adjacent branches (w/o AB) and the other omitting the current branch (w/o CB). As shown in Table 3, both variants exhibit noticeable drops in performance relative to the full CIM configuration. These results underscore the importance of both branch types in boosting the overall effectiveness of the model. These findings confirm the effectiveness of the dual-branch design in facilitating cross-scale information fusion and underscore the importance of incorporating both current and neighboring contextual features. Specifically, removal of the current branch leads to a substantial performance drop on the ORSSD dataset, with declines of 1.03%, 0.68%, and 1.05% in S α , E ξ m e a n , and F β m e a n on ORSSD, respectively. A similar trend is observed on the EORSSD dataset, where w/o CB causes substantial drops of 1.40%, 0.89%, and 0.87% in S α , E ξ m e a n , and F β m e a n , respectively. The performance on the ORSI4199 dataset drops significantly when either the current branch or adjacent branches are removed, with decreases of 1.08% and 0.67% in S α , and 1.20% and 1.05% in F β mean , respectively. Although the removal of the adjacent branches also leads to a measurable decline, its effect is less pronounced, further highlighting the critical role of the current branch in maintaining performance.
The effectiveness of semantic-aware attention. To further assess the effectiveness of semantic-aware attention (SAM), we conducted two variant experiments by replacing semantic-aware attention with either self-attention [52] or gated attention [53]. As shown in Figure 7, the corresponding heat maps offer an intuitive visualization of their respective performances. In the self-attention variant, we observe a more uniform attention distribution across the feature map, reflecting a stronger global capacity for capturing overall contextual information. However, compared with semantic-aware attention, self-attention is less adept at extracting fine-grained features in regions requiring precise semantic cues. This limitation is particularly evident when objects share similar appearances, where self-attention often struggles to differentiate foreground from background, leading to reduced performance. In contrast, the gated attention variant concentrates attention more narrowly through a gating mechanism, thereby highlighting critical features and improving feature selection accuracy. Despite this advantage, gated attention lacks the deeper semantic comprehension inherent to semantic-aware attention. Consequently, when dealing with optical remote sensing images containing complex backgrounds, gated attention underperforms due to its more superficial handling of implicit semantic relationships.
The sensitivity to various backbones. In evaluating our module’s sensitivity to different backbone networks, we have chosen three well-known encoders, VGG16, PVTv2, and ResNet-50. For each encoder, we constructed a corresponding variant model (MCFNet-VGG, MCFNet-PVT, and MCFNet-ResNet) that preserves our model’s core structure while substituting the backbone, facilitating a comprehensive assessment of performance. Table 4 provides detailed metrics for each variant. Across all three backbones, our proposed model always has a stable performance, which demonstrates robustness and flexibility. Further analysis revealed that the Res2Net backbone yielded superior results. Consequently, Res2Net was chosen as the default backbone for the final model configuration.

5. Conclusions

ORSI-SOD is notably challenging due to the diversity in target number and scale, complex background environments, and irregular topologies. To tackle these issues, we proposed a novel Multi-scale Contextual Fusion Network (MCFNet), comprising a Semantic-Aware Attention Module (SAM) and a Contextual Interconnection Module (CIM). Specifically, SAM introduces preliminary semantic masks to provide accurate semantic guidance during feature extraction, enabling the model to capture long-range dependencies and localize targets robustly. Meanwhile, CIM fuses contextual information across adjacent feature levels through multi-scale refinement, thereby enriching and optimizing texture representations at different scales. This cross-level interaction not only bolsters the model’s ability to discern salient target features but also improves its capacity to suppress background interference, thus increasing robustness. Extensive evaluations and ablation studies on three widely used datasets demonstrate that MCFNet surpasses 20 competing methods by producing more complete and accurate salient regions while maintaining robust detection performance even in complex scenarios.

Author Contributions

J.D. was responsible for the methodology and software conceptualization, as well as the writing of the initial draft. Y.Q. contributed to data curation, formal analysis, and project administration. H.X. supervised the study and was responsible for providing resources, conducting the investigation, and validation. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62406285.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Deng, B.; Liu, D.; Cao, Y.; Liu, H.; Yan, Z.; Chen, H. CFRNet: Cross-Attention-Based Fusion and Refinement Network for Enhanced RGB-T Salient Object Detection. Sensors 2024, 24, 7146. [Google Scholar] [CrossRef] [PubMed]
  2. Peng, Y.; Zhai, Z.; Feng, M. SLMSF-Net: A Semantic Localization and Multi-Scale Fusion Network for RGB-D Salient Object Detection. Sensors 2024, 24, 1117. [Google Scholar] [CrossRef]
  3. Zhao, J.; Wen, X.; He, Y.; Yang, X.; Song, K. Wavelet-Driven Multi-Band Feature Fusion for RGB-T Salient Object Detection. Sensors 2024, 24, 8159. [Google Scholar] [CrossRef]
  4. Chen, C.; Mo, J.; Hou, J.; Wu, H.; Liao, L.; Sun, W.; Yan, Q.; Lin, W. TOPIQ: A Top-Down Approach From Semantics to Distortions for Image Quality Assessment. IEEE Trans. Image Process. 2024, 33, 2404–2418. [Google Scholar] [CrossRef] [PubMed]
  5. Huang, W.; Feng, H.; Xu, H.; Liu, X.; He, J.; Gan, L.; Wang, X.; Wang, S. Surface Vessels Detection and Tracking Method and Datasets with Multi-Source Data Fusion in Real-World Complex Scenarios. Sensors 2025, 25, 2179. [Google Scholar] [CrossRef]
  6. Liu, S.; Shen, X.; Xiao, S.; Li, H.; Tao, H. A Multi-Scale Feature-Fusion Multi-Object Tracking Algorithm for Scale-Variant Vehicle Tracking in UAV Videos. Remote Sens. 2025, 17, 1014. [Google Scholar] [CrossRef]
  7. Cho, D.; Park, J.; Oh, T.H.; Tai, Y.W.; So Kweon, I. Weakly-and Self-Supervised Learning for Content-Aware Deep Image Retargeting. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4558–4567. [Google Scholar]
  8. Zheng, J.; Quan, Y.; Zheng, H.; Wang, Y.; Pan, X. ORSI Salient Object Detection via Cross-Scale Interaction and Enlarged Receptive Field. IEEE Geosci. Remote Sens. Lett. 2023, 20, 6003205–6003209. [Google Scholar] [CrossRef]
  9. Tong, N.; Lu, H.; Ruan, X.; Yang, M.H. Salient Object Detection via Bootstrap Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1884–1892. [Google Scholar]
  10. Song, H.; Liu, Z.; Du, H.; Sun, G.; Le Meur, O.; Ren, T. Depth-Aware Salient Object Detection and Segmentation via Multiscale Discriminative Saliency Fusion and Bootstrap Learning. IEEE Trans. Image Process. 2017, 26, 4204–4216. [Google Scholar] [CrossRef] [PubMed]
  11. Zhou, L.; Yang, Z.; Zhou, Z.; Hu, D. Salient Region Detection Using Diffusion Process on a Two-Layer Sparse Graph. IEEE Trans. Image Process. 2017, 26, 5882–5894. [Google Scholar] [CrossRef]
  12. Ding, L.; Wang, X.; Li, D. Visual Saliency Detection in High-Resolution Remote Sensing Images Using Object-Oriented Random Walk Model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 4698–4707. [Google Scholar] [CrossRef]
  13. Zhao, J.X.; Liu, J.J.; Fan, D.P.; Cao, Y.; Yang, J.; Cheng, M.M. EGNet: Edge Guidance Network for Salient Object Detection. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8779–8788. [Google Scholar]
  14. Zhao, X.; Pang, Y.; Zhang, L.; Lu, H.; Zhang, L. Suppress and Balance: A Simple Gated Network for Salient Object Detection. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 35–51. [Google Scholar]
  15. Xu, B.; Liang, H.; Liang, R.; Chen, P. Locate Globally, Segment Locally: A Progressive Architecture With Knowledge Review Network for Salient Object Detection. In Proceedings of the 25th AAAI Conference on Artificial Intelligence (AAAI 21), Online, 2–9 February 2021; pp. 3004–3012. [Google Scholar]
  16. Li, J.; Pan, Z.; Liu, Q.; Wang, Z. Stacked U-Shape Network With Channel-Wise Attention for Salient Object Detection. IEEE Trans. Multimed. 2020, 23, 1397–1409. [Google Scholar] [CrossRef]
  17. Lee, S.; Cho, S.; Park, C.; Park, S.; Kim, J.; Lee, S. LSHNet: Leveraging Structure-Prior With Hierarchical Features Updates for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5642516–5642531. [Google Scholar] [CrossRef]
  18. He, J.; Zhao, L.; Hu, W.; Zhang, G.; Wu, J.; Li, X. TCM-Net: Mixed Global–Local Learning for Salient Object Detection in Optical Remote Sensing Images. Remote Sens. 2023, 15, 4977. [Google Scholar] [CrossRef]
  19. Huo, L.; Hou, J.; Feng, J.; Wang, W.; Liu, J. Global and multiscale aggregate network for saliency object detection in optical remote sensing images. Remote Sens. 2024, 16, 624. [Google Scholar] [CrossRef]
  20. Li, C.; Cong, R.; Hou, J.; Zhang, S.; Qian, Y.; Kwong, S. Nested Network with Two-Stream Pyramid for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9156–9166. [Google Scholar] [CrossRef]
  21. Li, G.; Liu, Z.; Zeng, D.; Lin, W.; Ling, H. Adjacent Context Coordination Network for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Cybern. 2022, 60, 5607913–5607926. [Google Scholar] [CrossRef]
  22. Zeng, X.; Xu, M.; Hu, Y.; Tang, H.; Hu, Y.; Nie, L. Adaptive Edge-Aware Semantic Interaction Network for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5617416–5617431. [Google Scholar] [CrossRef]
  23. Liu, Z.; Zou, W.; Le Meur, O. Saliency tree: A novel saliency detection framework. IEEE Trans. Image Process. 2014, 23, 1937–1952. [Google Scholar]
  24. Yu, J.G.; Zhao, J.; Tian, J.; Tan, Y. Maximal entropy random walk for region-based visual saliency. IEEE Trans. Cybern. 2013, 44, 1661–1672. [Google Scholar]
  25. Yuan, Y.; Li, C.; Kim, J.; Cai, W.; Feng, D.D. Reversion correction and regularized random walk ranking for saliency detection. IEEE Trans. Image Process. 2017, 27, 1311–1322. [Google Scholar] [CrossRef]
  26. Zhou, Y.; Huo, S.; Xiang, W.; Hou, C.; Kung, S.Y. Semi-Supervised Salient Object Detection Using a Linear Feedback Control System Model. IEEE Trans. Cybern. 2019, 49, 1173–1185. [Google Scholar] [CrossRef] [PubMed]
  27. Pan, C.; Liu, J.; Yan, W.Q.; Cao, F.; He, W.; Zhou, Y. Salient Object Detection Based on Visual Perceptual Saturation and Two-Stream Hybrid Networks. IEEE Trans. Image Process. 2021, 30, 4773–4787. [Google Scholar] [CrossRef] [PubMed]
  28. Liang, M.; Hu, X. Feature selection in supervised saliency prediction. IEEE Trans. Cybern. 2014, 45, 914–926. [Google Scholar] [CrossRef]
  29. Jang, S.W.; Yan, L.; Kim, G.Y. Deep Supervised Attention Network for Dynamic Scene Deblurring. Sensors 2025, 25, 1896. [Google Scholar] [CrossRef] [PubMed]
  30. Liu, J.J.; Hou, Q.; Cheng, M.M.; Feng, J.; Jiang, J. A Simple Pooling-Based Design for Real-Time Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3917–3926. [Google Scholar]
  31. Ma, F.; Zhang, F.; Yin, Q.; Xiang, D.; Zhou, Y. Fast SAR Image Segmentation With Deep Task-Specific Superpixel Sampling and Soft Graph Convolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5214116–5214131. [Google Scholar] [CrossRef]
  32. Ma, F.; Zhang, F.; Xiang, D.; Yin, Q.; Zhou, Y. Fast Task-Specific Region Merging for SAR Image Segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5222316–5222331. [Google Scholar] [CrossRef]
  33. Luo, H.; Liang, B. Semantic-Edge Interactive Network for Salient Object Detection in Optical Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 6980–6994. [Google Scholar] [CrossRef]
  34. Tu, Z.; Wang, C.; Li, C.; Fan, M.; Zhao, H.; Luo, B. ORSI Salient Object Detection via Multiscale Joint Region and Boundary Model. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5607913–5607925. [Google Scholar] [CrossRef]
  35. Gu, Y.; Xu, H.; Quan, Y.; Chen, W.; Zheng, J. ORSI Salient Object Detection via Bidimensional Attention and Full-stage Semantic Guidance. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5603213–5603225. [Google Scholar] [CrossRef]
  36. Li, G.; Bai, Z.; Liu, Z.; Zhang, X.; Ling, H. Salient object detection in optical remote sensing images driven by transformer. IEEE Trans. Image Process. 2023, 32, 5257–5269. [Google Scholar] [CrossRef]
  37. Zhang, Q.; Cong, R.; Li, C.; Cheng, M.M.; Fang, Y.; Cao, X.; Zhao, Y.; Kwong, S. Dense Attention Fluid Network for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Image Process. 2021, 30, 1305–1317. [Google Scholar] [CrossRef] [PubMed]
  38. Zhou, X.; Shen, K.; Weng, L.; Cong, R.; Zheng, B.; Zhang, J.; Yan, C. Edge-Guided Recurrent Positioning Network for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Cybern. 2023, 53, 539–552. [Google Scholar] [CrossRef] [PubMed]
  39. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All You Need. In Proceedings of the International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  40. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1–15. [Google Scholar]
  41. Fan, D.P.; Cheng, M.M.; Liu, Y.; Li, T.; Borji, A. Structure-Measure: A New Way to Evaluate Foreground Maps. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4548–4557. [Google Scholar]
  42. Fan, D.P.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.M.; Borji, A. Enhanced-Alignment Measure for Binary Foreground Map Evaluation. In Proceedings of the International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 1698–1704. [Google Scholar]
  43. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-Tuned Salient Region Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1597–1604. [Google Scholar]
  44. Deng, Z.; Hu, X.; Zhu, L.; Xu, X.; Qin, J.; Han, G.; Heng, P.A. R3Net: Recurrent Residual Refinement Network for Saliency Detection. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 684–690. [Google Scholar]
  45. Qin, X.; Zhang, Z.; Huang, C.; Dehghan, M.; Zaiane, O.R.; Jagersand, M. U2-Net: Going deeper with nested U-structure for salient object detection. Pattern Recognit. 2020, 106, 107404–107416. [Google Scholar] [CrossRef]
  46. Pang, Y.; Zhao, X.; Lu, H. Multi-Scale Interactive Network for Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 9410–9419. [Google Scholar]
  47. Zhou, X.; Shen, K.; Liu, Z.; Gong, C.; Zhang, J.; Yan, C. Edge-Aware Multiscale Feature Integration Network for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5605315–5605330. [Google Scholar] [CrossRef]
  48. Lin, Y.; Sun, H.; Liu, N.; Bian, Y.; Cen, J.; Zhou, H. Attention Guided Network for Salient Object Detection in Optical Remote Sensing Images. In Proceedings of the International Conference on Artificial Neural Networks, Bristol, UK, 6–9 September 2022; pp. 25–36. [Google Scholar]
  49. Li, G.; Liu, Z.; Bai, Z.; Lin, W.; Ling, H. Lightweight Salient Object Detection in Optical Remote Sensing Images via Feature Correlation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5617712–5617733. [Google Scholar] [CrossRef]
  50. Quan, Y.; Xu, H.; Wang, R.; Guan, Q.; Zheng, J. ORSI Salient Object Detection via Progressive Semantic Flow and Uncertainty-Aware Refinement. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5608013–5608025. [Google Scholar] [CrossRef]
  51. Zhao, J.; Jia, Y.; Ma, L.; Yu, L. Adaptive Dual-Stream Sparse Transformer Network for Salient Object Detection in Optical Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 5173–5192. [Google Scholar] [CrossRef]
  52. Gu, Y.; Wang, L.; Wang, Z.; Liu, Y.; Cheng, M.M.; Lu, S.P. Pyramid Constrained Self-Attention Network for Fast Video Salient Object Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 10869–10876. [Google Scholar]
  53. Schlemper, J.; Oktay, O.; Schaap, M.; Heinrich, M.; Kainz, B.; Glocker, B.; Rueckert, D. Attention gated networks: Learning to leverage salient regions in medical images. Med. Image Anal. 2019, 53, 197–207. [Google Scholar] [CrossRef]
Figure 1. The proposed Multi-scale Contextual Fusion Network (MCFNet) for ORSI-SOD, which is structured around two core modules: the semantic-aware attention module (SAM) and the contextual interconnection module (CIM).
Figure 1. The proposed Multi-scale Contextual Fusion Network (MCFNet) for ORSI-SOD, which is structured around two core modules: the semantic-aware attention module (SAM) and the contextual interconnection module (CIM).
Sensors 25 03035 g001
Figure 2. Illustration of semantic-aware attention module (SAM).
Figure 2. Illustration of semantic-aware attention module (SAM).
Sensors 25 03035 g002
Figure 3. Illustration of contextual interconnection module (CIM).
Figure 3. Illustration of contextual interconnection module (CIM).
Sensors 25 03035 g003
Figure 4. Precision–recall curves of eight advanced models evaluated on three datasets.
Figure 4. Precision–recall curves of eight advanced models evaluated on three datasets.
Sensors 25 03035 g004
Figure 5. Visual comparisons with nine representative state-of-the-art methods, including two NSI-SOD methods and seven ORSI-SOD methods.
Figure 5. Visual comparisons with nine representative state-of-the-art methods, including two NSI-SOD methods and seven ORSI-SOD methods.
Sensors 25 03035 g005
Figure 6. Visual comparisons of the key components. ‘w/o’ denotes ‘WITHOUT’.
Figure 6. Visual comparisons of the key components. ‘w/o’ denotes ‘WITHOUT’.
Sensors 25 03035 g006
Figure 7. Visual comparisons of semantic-aware attention. (a) Optical RSIs. (b) Ground truth. (c) Self-attention variant. (d) Gated attention variant. (e) Our MCFNet.
Figure 7. Visual comparisons of semantic-aware attention. (a) Optical RSIs. (b) Ground truth. (c) Self-attention variant. (d) Gated attention variant. (e) Our MCFNet.
Sensors 25 03035 g007
Table 1. Quantitative comparison of our MCFNet with 20 other state-of-the-art methods on EORSSD, ORSSD, and ORSI4199 datasets. The best three results are sequentially marked in red, blue, and green. ‘CN’ is the CNN-based NSI-SOD method, and ‘CR’ is the CNN-based ORSI-SOD method.
Table 1. Quantitative comparison of our MCFNet with 20 other state-of-the-art methods on EORSSD, ORSSD, and ORSI4199 datasets. The best three results are sequentially marked in red, blue, and green. ‘CN’ is the CNN-based NSI-SOD method, and ‘CR’ is the CNN-based ORSI-SOD method.
MethodsTypeParamsORSSDEORSSDORSI4199
( M ) S α M E ξ mean E ξ max F β mean F β max S α M E ξ mean E ξ max F β mean F β max S α M E ξ mean E ξ max F β mean F β max
R3Net18 [44]CN56.20.81410.03990.86810.89130.73830.74560.81840.01710.82940.94830.63020.74980.83920.04010.90210.91410.81270.8250
PoolNet19 [30]CN53.60.84030.03580.86500.93430.69990.77060.82070.02100.81930.92920.64060.75450.81840.05730.80280.81590.73320.7457
EGNet19 [13]CN108.10.87210.02160.90130.97310.75000.83320.86010.01100.87750.95700.69670.78800.83620.04240.89160.90080.82230.8368
SUCA20 [16]CN115.60.92850.01020.96110.96980.87230.88850.91260.00790.93960.96440.90290.85350.82940.04280.82710.83400.78130.7927
U2Net20 [45]CN44.00.91620.01660.93870.95390.84920.97380.91990.00760.93730.96490.83290.87320.83790.03910.89880.90340.82010.8325
GateNet20 [14]CN128.60.92040.01100.95600.97080.87410.90350.90710.00810.93640.96340.82940.86460.85010.03770.91550.92640.83470.8489
MINet20 [46]CN47.60.90400.01440.94540.95450.85740.87610.90400.00930.93460.94420.81740.83440.84980.03670.90980.91630.83220.8473
PA-KRN21 [15]CN141.10.92390.01390.96200.96800.87270.88900.91920.01040.95360.96160.83580.86390.84280.03850.91210.92410.82570.8432
DAFNet21 [37]CR29.40.91910.01130.95390.97710.85110.89280.91660.00600.92900.96590.78420.86120.86530.03440.91670.93650.82440.8470
EMFINet22 [47]CR95.10.94320.00950.97260.98130.90000.91550.93190.00750.95980.97120.85050.87420.85910.04520.90220.91160.81000.8169
AGNet22 [48]CR24.60.93890.00910.97280.98110.89560.90980.92870.00670.96560.97520.85160.87580.86270.03370.92760.93860.85360.8614
MJRBM22 [34]CR43.50.91930.01460.94720.96310.85440.88500.91800.01070.93390.96310.82740.86380.85820.03720.90710.93430.83050.8511
ACCoNet22 [21]CR127.00.94370.00880.97540.97960.89710.91490.92900.00740.96530.97270.85520.88370.86750.03140.93420.94120.86100.8646
CorrNet22 [49]CR4.10.93800.00980.97640.97900.90020.91290.92890.00830.96460.96960.86200.87780.86230.03660.92060.93300.85130.8560
BAFS-Net23 [35]CR31.00.93780.00830.97730.98200.90160.91060.92500.00610.96970.97290.85640.86530.86610.03140.93390.93990.85880.8633
ERPNet23 [38]CR77.20.93520.01140.96040.97380.87980.90360.92520.00820.93660.96650.82690.87430.86520.03670.91670.92840.83870.8538
AESINet23 [22]CR51.60.94550.00850.97410.98140.89620.91600.93470.00640.96470.97570.84960.87920.87020.03090.93570.94230.86260.8676
SFANet24 [50]CR25.10.94530.00770.97890.98300.90630.91920.93490.00580.97260.97690.86800.88330.87610.02920.93850.94470.86590.8710
ADSTNet24 [51]CR62.10.93790.00860.97400.98070.90420.91240.93110.00650.97090.97690.87160.88040.87100.03180.93560.94330.86530.8698
LSHNet24 [17]CR-0.94910.00750.97640.98240.90540.92000.93700.00640.96920.97610.86430.88440.87590.02990.93920.94620.86900.8758
MCFNet (Ours)CR29.60.94720.00750.97930.98350.90700.92020.93740.00580.97330.97690.87120.88510.87680.02900.93980.94670.86950.8760
Table 2. Ablation analysis measuring the overall contribution of SAM and CIM in MCFNet on the ORSSD, EORSSD, and ORSI4199 datasets. The best results are marked in bold.
Table 2. Ablation analysis measuring the overall contribution of SAM and CIM in MCFNet on the ORSSD, EORSSD, and ORSI4199 datasets. The best results are marked in bold.
MethodORSSDEORSSDORSI4199
S α M E ξ max F β max S α M E ξ max F β max S α M E ξ max F β max
Baseline0.92030.01120.96590.89640.91980.00970.96760.86850.84980.03520.92670.8485
w/o CIM0.93420.00930.97070.91050.92130.00750.97010.87420.86610.03360.93900.8588
w/o SAM0.94010.00810.97390.91460.92960.00690.97220.87830.86830.03140.94020.8655
MCFNet0.94720.00750.98350.92020.93740.00580.97690.88510.87680.02900.94670.8760
Table 3. Ablation analysis of branch effectiveness in the CIM. THE w/o denotes ‘WITHOUT’, and the best results are highlighted in bold.
Table 3. Ablation analysis of branch effectiveness in the CIM. THE w/o denotes ‘WITHOUT’, and the best results are highlighted in bold.
MethodORSSDEORSSDORSI4199
S α M E ξ max F β max S α M E ξ max F β max S α M E ξ max F β max
w/o CB0.93750.00880.97260.89760.92450.00700.96470.86370.86730.03220.93160.8591
w/o AB0.94020.00820.97410.90090.93090.00640.96920.86640.87090.03160.93380.8604
MCFNet0.94720.00750.97930.90700.93740.00580.97330.87120.87680.02900.93980.8695
Table 4. Performance evaluation of our MCFNet on various encoder backbones. The best results are marked in bold.
Table 4. Performance evaluation of our MCFNet on various encoder backbones. The best results are marked in bold.
MethodORSSDEORSSDORSI4199
S α M E ξ max F β max S α M E ξ max F β max S α M E ξ max F β max
MCFNet-VGG0.94340.00800.98130.91420.93280.00650.97240.88190.87050.02990.94110.8711
MCFNet-PVT0.94080.00830.98020.91080.92990.00660.97000.87930.86830.03070.93870.8684
MCFNet-ResNet0.94630.00780.98200.91930.93510.00620.97530.88420.87350.02960.94360.8734
MCFNet(Ours)0.94720.00750.98350.92020.93740.00580.97690.88510.87680.02900.94670.8760
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ding, J.; Quan, Y.; Xu, H. MCFNet: Multi-Scale Contextual Fusion Network for Salient Object Detection in Optical Remote Sensing Images. Sensors 2025, 25, 3035. https://doi.org/10.3390/s25103035

AMA Style

Ding J, Quan Y, Xu H. MCFNet: Multi-Scale Contextual Fusion Network for Salient Object Detection in Optical Remote Sensing Images. Sensors. 2025; 25(10):3035. https://doi.org/10.3390/s25103035

Chicago/Turabian Style

Ding, Jinting, Yueqian Quan, and Honghui Xu. 2025. "MCFNet: Multi-Scale Contextual Fusion Network for Salient Object Detection in Optical Remote Sensing Images" Sensors 25, no. 10: 3035. https://doi.org/10.3390/s25103035

APA Style

Ding, J., Quan, Y., & Xu, H. (2025). MCFNet: Multi-Scale Contextual Fusion Network for Salient Object Detection in Optical Remote Sensing Images. Sensors, 25(10), 3035. https://doi.org/10.3390/s25103035

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop