Next Article in Journal
Impact of All-Sky Assimilation of Multichannel Observations from Fengyun-3F MWHS-II on Typhoon Forecasting
Previous Article in Journal
Sensitive Multispectral Variable Screening Method and Yield Prediction Models for Sugarcane Based on Gray Relational Analysis and Correlation Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IEAM: Integrating Edge Enhancement and Attention Mechanism with Multi-Path Complementary Features for Salient Object Detection in Remote Sensing Images

School of Marine Science and Technology, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(12), 2053; https://doi.org/10.3390/rs17122053
Submission received: 29 March 2025 / Revised: 4 June 2025 / Accepted: 13 June 2025 / Published: 14 June 2025
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
Prominent target detection in optical remote sensing images (RSI-SOD) focuses on segmenting key targets that capture human attention. However, most SOD methods prioritize detection accuracy at the cost of memory. Complex backgrounds, occlusions, and noise distort segmented target boundaries, while large memory demands increase computational cost, and reduced memory impairs segmentation accuracy. To address these challenges, we integrate edge enhancement and attention mechanisms with multi-path complementary features for salient object detection in remote sensing images (IEAM), aiming to improve salient target accuracy, boundary detection, and memory efficiency. The architecture utilizes a structured feature fusion strategy, combining spatial channel attention mechanisms with adaptive merging to enhance multi-scale feature representation and suppress background noise. The Spatially Adaptive Edge Embedded Module (SAEM) refines object boundary perception, the SCAAP module dynamically selects relevant spatial and channel features while balancing adaptive and maximal pooling, and the Spatial Adaptive Guidance (SAG) module enhances feature localization in cluttered environments to mitigate semantic dilution in U-shaped networks. Extensive experiments on the EORSSD and ORSSD benchmark datasets demonstrate that IEAM outperforms 21 state-of-the-art methods, achieving an inference speed of 48 FPS at 103.2 G FLOP, making it suitable for real-time applications. The proposed model is robust and excels in multiple aspects.

1. Introduction

Salient object detection (SOD) in optical remote sensing images (RSIs) has attracted considerable attention due to its critical role in applications such as environmental monitoring, land cover analysis, and disaster response [1]. Remote sensing images, captured by satellites or aerial platforms, provide valuable information about the Earth’s surface, such as urban areas, natural landscapes, mountains, and rivers, as shown in Figure 1. However, the task remains challenging due to the unique characteristics of RSIs. First, object boundaries in RSIs are often weak or blurred, caused by factors such as shadows, resolution degradation, or camouflage, leading to inaccurate segmentation contours. Second, the complex backgrounds of natural and artificial environments introduce dense textures and clutter that can confuse saliency models and increase false detections. Third, many existing CNN-based SOD methods rely on deep or cascaded architectures with redundant layers, resulting in high computational cost and limited real-time applicability [2,3].
Visual attention mechanisms have been widely applied in computer vision, playing a crucial role in Salient Object Detection (SOD). SOD aims to automatically identify the most prominent regions in an image and is fundamental in tasks such as image segmentation and object recognition [4,5]. Compared to natural scene image SOD (NSI-SOD), salient object detection in remote sensing images (RSI-SOD) faces greater challenges due to multi-target scenarios, small objects, and complex backgrounds [2].
Optical remote sensing images, typically acquired by satellite or airborne sensors, differ significantly from natural images in terms of illumination, scale, and background complexity. To effectively detect salient targets such as airplanes, islands, ships, and buildings, RSI-SOD methods have been specifically designed to address these unique characteristics. Although Convolutional Neural Network(CNN) based approaches have achieved notable success in NSI-SOD [6,7,8], they often underperform when directly applied to remote sensing data due to limitations in detail preservation and background suppression.
The rapid development of CNNs has driven advances in NSI-SOD, with techniques such as multi-scale feature fusion [9], edge guidance [10], attention mechanisms [11,12], and improved loss functions [13,14]. However, due to fundamental differences in image acquisition and scene complexity, these methods often struggle when transferred to remote sensing imagery.
Inspired by NSI-SOD, several CNN-based methods have been adapted for RSI-SOD. For instance, LV-Net [7] uses multi-resolution inputs within a nested architecture for better scale perception. PDF-Net [15] employs five-branch fusion for multi-scale feature integration, while DAF-Net [8,16] enhances detection using edge supervision and dense attention. EMFI-Net [17] integrates multi-resolution inputs, edge-aware supervision, and hybrid loss to improve boundary detection.
Pooling-based strategies have gained attention for their balance between accuracy and efficiency. Real-time methods using simplified pooling mechanisms reduce computational complexity without sacrificing performance [18,19]. AMP-Net [20] combines mean and max pooling to enhance feature extraction and maintain real-time efficiency.
Edge-aware architectures like EAFIN [21,22] preserve structural detail and improve boundary detection, particularly in complex urban scenes. MCC-Net [23] adopts complementary learning to handle variations in illumination, weather, and resolution.
Despite these advances, RSI-SOD still faces challenges such as small object detection and maintaining a balance between accuracy and efficiency on large-scale datasets [24,25,26,27]. Recent approaches focus on integrating global-local context, attention mechanisms, and feature fusion to address these issues [28,29,30], aiming to improve accuracy while reducing model complexity and computational cost [31]. This SOD often suffer from several limitations: insufficient edge-awareness, limited adaptability to multi-scale objects, and excessive computational burden.
To address these challenges, we propose IEAM-Net (Integrating Edge Enhancement and Attention Mechanism with Multi-Path Complementary Features Network), a novel architecture for RSI-SOD. IEAM-Net combines channel attention mechanisms with adaptive pooling to enhance detection accuracy and efficiency [32,33]. Built on the VGG-16 backbone, the tic-tac-toe structured design includes a left-to-right convolutional path, top-down max and adaptive pooling paths, and a right-to-left up-sampling path, enabling effective multi-level feature fusion and fine-grained detail capture. The main contributions of this work are summarized as follows: (1) We design a Spatial-Aware Edge Module (SAEM) that enhances boundary localization through edge-guided feature refinement; (2) We introduce a Spatial-Channel Attention Aggregation Pooling (SCAAP) module to boost model robustness in cluttered and complex backgrounds; (3) We implement a multi-path attention mechanism to achieve a balanced trade-off between precision and efficiency across varying object scales; (4) Extensive experiments on two challenging benchmark datasets demonstrate that our model achieves state-of-the-art performance while significantly reducing model complexity and inference time.

2. Supervised Learning

2.1. Optimization Based on Network Architecture

Salient Object Detection (SOD) has seen significant advancements with the optimization of network architectures to improve detection performance and robustness. The rise of deep learning, particularly Convolutional Neural Networks (CNNs), has led to the widespread use of encoder-decoder architectures like U-Net, which effectively captures multi-scale features and generates high-resolution saliency maps. U-Net, initially proposed for biomedical image segmentation by Ronneberger et al., was later extended by Qin et al. for SOD, improving robustness against target scale variations.
Several innovative architectures have been developed to address different task requirements. For example, ref. [20] introduced a real-time SOD network based on hybrid pooling, balancing global and local feature extraction for improved accuracy and real-time performance. Ref. [23] developed the Multi-Content Complementary Network (MCCN), which integrates multi-scale features to enhance small and fuzzy target detection in remote sensing images.

2.2. Application of Feature Enhancement and Attention Mechanisms

The incorporation of feature enhancement and attention mechanisms has further boosted SOD performance. Zhou et al. proposed EMF-Net, which combines edge-aware and multi-scale features to enhance target boundary clarity in complex remote sensing images [17]. Additionally, [20] developed AMP-Net, which combines average and max pooling to balance global and local information extraction. Attention mechanisms, as seen in [34], further improve detection in challenging scenes by focusing on key regions and enhancing feature extraction through contextual attention [34]. These advancements provide a comprehensive solution for SOD in complex, multi-target environments.

2.3. Innovation and Advantages of IEAM-Net

We propose IEAM-Net (Integrating Edge Enhancement and Attention Mechanism with Multi-Path Complementary Features Network), a novel architecture for salient target detection in optical remote sensing images. Built on the VGG-16 backbone, IEAM-Net integrates channel attention and adaptive pooling to improve accuracy while ensuring computational efficiency.
The network features a tic-tac-toe structure with three components:
  • A convolutional path (left to right) for hierarchical feature extraction;
  • A pooling path (top to bottom) combining max and adaptive pooling to adjust receptive fields;
  • An up-sampling path (right to left) for reconstructing high-resolution saliency maps.
This grid-like design enables effective local-global feature fusion and preserves fine details. At its core, the SCAAP module combines spatial and channel attention with adaptive pooling, allowing dynamic focus on important regions. This is especially effective in complex scenes and small-object detection, addressing the limitations of conventional pooling in maintaining boundary integrity and saliency.
The design of IEAM-Net integrates three key innovations:
  • SCAAP module and adaptive pooling strategy;
    IEAM-Net introduces the SCAAP module, which combines adaptive pooling and channel attention to dynamically adjust pooling sizes and focus on salient areas. Unlike traditional pooling methods, this module better handles multi-target and small-object detection, especially in preserving boundaries and saliency.
  • Multi-path feature extraction of Tic-Tac-Toe structure;
    The tic-tac-toe structure is used for grid pooling, which includes a left-to-right convolution path, a top-to-bottom pooling path, and a right-to-left up-sampling path to achieve multi-path feature extraction of Figure 2. Through this multi-level information fusion method, IEAM-Net can capture multi-scale features and improve the detail retention ability of target detection, especially in complex backgrounds and different scale target detection tasks.
  • Enhance edge perception and spatial adaptability.
    IEAM-Net incorporates two key modules: SAEM (Spatially Adaptive Edge Embedded Module) and the SAG-Model (spatial adaptive guidance module). SAEM enhances edge detection using multi-scale convolution, while the SAG-Model fuses weighted multi-scale features with spatially guided convolution to better focus on key regions. Together, they improve boundary accuracy and spatial feature extraction for salient target detection.

3. Proposed Method

In this section, we first introduce the general structure of IEAM-Net. We will detail the structure and functions of the Spatially Adaptive Edge Embedded Module (SAEM) module, the Spatial-Channel Attention Adaptive Pooling Module and its three sub-modules, and introduce the mechanism of bidirectional interactions between the CAAS (Channel Attentive Adjustment Sub-module) and the SAAS (Spatial Attentive Adjustment Sub-module) to optimize the feature weights as well as the SAG-Model (Spatial Adaptive Guidance Module). To facilitate more precise feature weighting, we design a bidirectional attention interaction mechanism consisting of two sub-modules: Channel Attention Adjustment Sub-module (CAAS), which emphasizes informative feature channels by learning inter-channel dependencies through global average pooling and fully connected layers; and Spatial Attention Adjustment Sub-module (SAAS), which enhances salient spatial locations by capturing spatial-wise importance through convolutional operations on aggregated feature maps. These two sub-modules interact recursively to refine each other’s attention distribution: CAAS guides SAAS to focus on spatial regions supported by strong channel cues, while SAAS reinforces CAAS by spatially modulating the contribution of channels. This mutual reinforcement strategy allows the network to adaptively recalibrate attention maps across both dimensions. Finally, we propose an accurate feature fusion module to integrate features from different paths, which effectively improves the target detection accuracy, and a deep supervised mechanism to enhance the supervised learning process of IEAM-Net by optimizing the network training through multi-level loss.

3.1. Network Framework

Figure 3 illustrates the overall structure of the IEAM-Net for salient target detection. The network takes the entire image I (top left) as input and produces the final salient map P as the output. IEAM-Net follows a U-Net-like architecture consisting of a left-to-right convolutional path, two top-down paths (max pooling and adaptive pooling), and a right-to-left up-sampling path. This design, highlighted in the first and second parts of Figure 3, allows for the fusion of multiple levels of sub-features, capturing complex details that are otherwise difficult to obtain.
The architecture is based on several key components:
SAEM Module: The Spatial-Aware Edge Module (SAEM) processes edge information and extracts key features through multi-scale convolutions.
Spatial-Channel-Attentive Adaptive Pooling Module: Positioned at the center of the tic-tac-toe structure, this module captures features from multi-level complementary contexts. It uses up-down convolutions to refine features along the left-to-right path in a fine-grained manner.
SAG-Model: The Spatial Adaptive Guidance Model (SAG-Model) enhances spatial details of the image features.
Feature Fusion Module: This module combines local spatial structures with salient regions, resulting in the final output map for salient target detection.
  • Left-to-Right Pathway;
    The left-to-right pathway, adapted from VGG-16, is used for hierarchical feature extraction through five convolutional blocks (Conv2 to Conv6). Unlike existing approaches, we extract features from the last convolutional layer of each block, instead of pooling, which retains more spatial, channel, and boundary details. The original three fully connected layers of VGG-16 are replaced with three additional convolutional layers (Conv6) to further refine features after Conv5. This results in five multi-scale feature extractions from Conv2-2, Conv3-3, Conv4-3, Conv5-3, and Conv6-3 layers.
  • Two Top-Down Paths;
    • Max-Pool Path; The max-pool path begins with higher-level features from the Conv2-2, Conv3-3, Conv4-3, Conv5-3, and Conv6-3 layers. It first passes through a max-pooling layer to downsample the features, extracting higher-level semantic information. As the features move down, they are restored to the appropriate size via up-sampling and fused with features adaptively pooled through the SCAAP module. This is followed by a refinement convolutional block for further feature extraction and a channel transformation layer that adjusts feature channels for fusion.
    • Adaptive Pool Path; The adaptive pool path starts with features from Conv1-2, Conv2-2, Conv3-3, Conv4-3, and Conv5-3, processed by the SCAAP module. The features undergo spatial and channel attention adjustments before pooling, resulting in feature maps with enhanced representation of the salient target regions. These are then fused with the max-pooling features via up-sampling, concatenation, and convolution, ensuring the combined features accurately represent salient target information.
  • Right-to-Left Path.
    The right-to-left path refines hierarchical features that have undergone adaptive and max pooling via the SCAAP module, followed by up-sampling. The features, denoted as Ai and Mi, pass through two distinct pooling paths—adaptive and max pooling—and are fused via up-sampling and convolution operations to produce the final predictive features.
In the leftmost and rightmost paths, the features A5 and M5 are directly used as the top predictive features, refined through max-pooling and adaptive pooling, respectively. For the middle layer, features are integrated through bilinear up-sampling, concatenation, and up-convolution operations from the two independent paths to produce the final prediction.
The predictive features P A i and P M i can be expressed as follows:
P T i = T i i = 5 U p Conv cat T i , U p P T i + 1 , T i i [ 1 , 4 ]
where T denotes the adaptive pooling path and the maximal pooling path. A denotes the adaptive pooling path and M denotes the maximal pooling path. Up denotes the bilinear up-sampling operation. T i denotes the refinement features generated by the average and maximal pooling modules as described later. U p P T i + 1 , T i denotes up-sampling the feature map from the upper layer to the size of the current layer’s feature map T i so that the two can be concatenated. The cat operation connects the refinement features at that level with the up-sampled features from the upper layers. Up-Conv denotes the three kernels of 3 × 3 kernel size of the convolutional layer to enhance the robustness of the connected features. For convenience, we abbreviate the convolution parameters.

3.2. Structure and Function of the SAEM Module

The Spatially Adaptive Edge Embedded Module (SAEM) is a multi-stage processing module used to process edge information from input images via multi-scale edge enhancement techniques. As shown in Figure 4, the module enhances edge information for input images while considering multiple spatial and contextual relationships within the image, making it adaptive to varying levels of image features. The main components in the SAEM module are the Light Adaptive Multi-Scale (LAM) and Edge Embedding Networks (EEM).
The module operates based on the Region of Spatial Information (RSI), which contains many images with varying levels of details and complexity. In these cases, if the edge information is blurry, it will directly affect the network’s performance. The fast processing of edge-enhanced feature maps helps to focus the network’s attention on these regions, improving both the resolution and accuracy. Therefore, we designed three LAM blocks inside the SAEM module to refine the images with multi-scale features, enhancing its capability to handle images with diverse edge information and improve performance.

3.2.1. LAM (Light Adaptive Multi-Scale Block)

The LAM module provides multi-scale enhancement for input features with different resolutions. Let the input be denoted as X¹. The operation can be expressed as:
X L A M 1 = Conv ( X 1 , 1 × 3 ) , Conv ( X 1 , 3 × 1 ) , Conv ( X 1 , 3 × 3 ) , Conv ( AvePool ( X 1 ) , 1 × 1 )
In the LAM (Lightweight Adaptive Multi-Scale Block), multi-scale contextual features are captured via three parallel branches:
The first branch applies a 1 × 3 convolution followed by a 3 × 1 convolution to extract directional features.The second branch performs a 3 × 3 convolution for standard local feature extraction. The third branch applies average pooling followed by a 1 × 1 convolution to encode global context information.
The outputs of all branches are concatenated along the channel dimension and passed through a 3 × 3 convolution. The number of output channels is adjusted back to 32. An element-wise addition with the input feature map is then performed, followed by a Spatial Attention (SA) module to refine salient features. This architecture allows LAM to expand the receptive field and enhance multi-scale contextual understanding with minimal computation cost.    

3.2.2. EEM (Edge Embedded Module)

In the Edge Enhancement section, we propose an edge-embedded attention mechanism in order to integrate the edge information into the spatial attention graph to improve the localization accuracy of salient targets. Specifically, we design a new path to learn an edge prediction graph by adding significant edge loss, which extracts the significant edge information from the ground truth image of the salient object through a Sobel filter. This edge prediction map is then combined with channel and spatial attention information to form an edge-embedded attention mechanism to further enhance the model’s ability to perceive edge features.
The purpose of the edge embedding module is to perform edge enhancement of the input image by means of Sobel’s algorithm. Its output is described by the following equation:
X E M 1 = Sobel X 1 X 1
where ⊗ denotes the element-by-element multiplication operation and the Sobel operator is used to extract the edge information of the image.
As shown in Figure 5, by enhancing the edges of an image, the edge embedding module is able to highlight the contour information of an object, which helps in the recognition and localization of salient targets. Edge information is crucial in vision tasks because edges usually represent structural and semantic information in an image. Conventional convolutional neural networks (CNNs) are usually more limited in the extraction of edge information, whereas by introducing salient edge features through the edge embedding module, the model is able to more accurately localize and identify targets in the image.
The channel attention module generates weights for each channel by global average pooling with the following formula:
CA X 1 = σ W 1 · ReLU W 2 · AvePool X 1
where W 1 and W 2 are learnable weight matrices, σ is a sigmoid activation function, and Ave-Pool denotes global average pooling. The spatial attention module, on the other hand, focuses on the importance of each spatial location in the image, which is realized by the following equation:
SA X 1 = σ W 3 · ReLU W 4 · Conv X 1 , 3 × 3
where W 1 and W 2 are the learned convolution kernel parameters and Conv denotes the convolution operation.
In the edge embedding module, the channel attention mechanism is used to assign different weights to each channel to highlight the channels that are most critical for target detection. Specifically, the channel attention module computes the average value of each channel through a global average pooling operation to generate a global feature descriptor. This descriptor is tuned by a learnable weight matrix and a sigmoid activation function is used to generate weights for each channel. This operation can dynamically adjust the network’s attention according to the feature importance of different channels, which helps the network to extract the most significant target features among the multi-scale features. After adjusted by channel attention and spatial attention, the input feature maps are fused and output the final prediction results X pred 1 , which are fused as follows:
X pred 1 = sigmoid X C A 1 + X S A 1 X E M 1
where ⊗ is multiplied element by element and finally the sigmoid activation function is used to output the significant target detection results.

3.3. Spatial-Channel-Attentive Adaptive-Pool

The Spatial-Channel Attention Adaptive Pooling (SCAAP) module is a key point in our proposed network architecture, and the SCAAP module consists of three main sub-modules: the Channel Attention Adjustment Submodule (CAAS): to emphasize important channel features, the Spatial Attention Adjustment Submodule (Spatial Attention Adjustment Submodule (SAAS): highlights important spatial regions, and the Adaptive Pooling Execution Submodule (APES): performs adaptive pooling based on attention weights. It enhances feature representation by combining spatial and channel attention mechanisms. The input feature map first enters the two-way interaction mechanism between CAAS and SAAS, and the core goal of the two-way interaction mechanism is to enhance the feature representation capability of the attention module through the multiple rounds of interactions between the channel attention (CAAS) and the spatial attention (SAAS), so that the two can gradually optimize each other’s weight distribution as shown in Figure 6. In each interaction, Channel Attention helps Spatial Attention focus on more important regions, while Spatial Attention in turn optimizes the selection of Channel Attention, making the final generated attention weights more accurate and robust to adapt to complex scenes and multi-target tasks. The SCAAP pseudocode is as shown in Algorithm 1.
Algorithm 1: Spatial-Channel Attentive Adaptive Pooling (SCAAP)
Remotesensing 17 02053 i001
    Finally, the APES determines the pooling region based on the output of the two-way interaction mechanism and performs the pooling operation to obtain the output feature map. This provides more accurate feature maps for salient target detection. The following is our detailed description of SCAAP module, including the functions, mathematical formulas and parameter settings of each sub-module.

3.3.1. Design Ideas for Two-Way Interaction Mechanisms

The two-way interaction mechanism aims to achieve the following goals through the interaction between channel and spatial attention.
Mutual reinforcement: CAAS provides channel-level attention information to help SAAS focus on more important spatial regions; conversely, the spatial-level attention information provided by SAAS optimizes CAAS’s selection of channel features. Cyclic optimization: Through multiple rounds of interactions, the attention distribution gradually tends to be optimal, and finally outputs more accurate feature representations. Dynamic feedback: The output features of each round of interaction are used as inputs for the next round of computation to dynamically adjust feature selection and attention weights.
Network architecture design: In order to realize the two-way interaction between CAAS and SAAS, this paper adopts the following design.
Step 1: Channel-Space Input feature X i n p u t . Firstly, through the CAAS module, generate channel attention weight w C A : w C A = C A A S X i n p u t Use w C A to weight the input feature to get the updated feature X C A : X C A = X i n p u t w C A . Use the updated feature X C A as the input of SAAS.
Step 2: Spatial-Channel Feature X C A passes through the SAAS module, generates spatial attention weight w S A : w S A = S A A S X C A , weights feature w S A using X C A to get updated feature X S A : X S A = X C A w S A inputs updated feature X S A into CAAS again, forming a loop.
Step 3: Multiple Interactions The above two steps form an interaction loop. Repeat the interaction of channel and spatial attention K times ( K = 3 ), we use 2 rounds in this paper to optimize the attention weights step by step. The formula for the t-th round of multi-round interaction is as follows: X C A ( t ) = X S A ( t 1 ) w C A ( t ) , X S A ( t ) = X C A ( t ) w S A ( t ) .
Final Output: After K rounds of interaction, the final output of feature X f i n a l is: X f i n a l = X S A ( K ) .

3.3.2. Adaptive Pooling Execution Submodule (APES)

The APES submodule performs adaptive pooling based on the outputs of CAAS and SAAS to integrate the channel and spatial information and generate the final feature map. The spatially attention-adjusted feature map X f i n a l is divided into n × n small regions of equal size (n is flexibly determined based on the size of the input feature map and the desired pooling scale), each of which is of size h × w (where h = H / n , w = W / n ) as shown in Figure 7.
For each small region ( i , j )   ( i = 1 , 2 , , n , j = 1 , 2 , , n ) , compute its integrated attention weight γ i j . The combined attention weight is a combination of the channel attention weight and spatial attention weight in the region, and the weighted summation method is used here, with Equation (7)
γ i j = w 1 α i j + w 2 β i j
where α i j is the average of the channel attention weights in the region, β i j is the average of the spatial attention weights in the region, w 1 and w 2 are pre-set weighting coefficients.
The size of the pooling region and whether or not to pool are determined based on the combined attentional weights γ i j . If γ i j > θ ( θ is a pre-set threshold set to 0.5), a smaller pooling region is used for the region, with average pooling of h / 2 × w / 2 . If γ i j θ , a larger pooling region is used for the region, with maximal pooling of h × w
P ( i , j ) = max k , l X ( i , j , k , l ) if   W ( i , j ) > θ 1 k × k m 1 k n 1 k X ( i + m , j + n , : , : ) otherwise
Finally, all the small regions that have undergone the pooling operation are recombined to form the output feature map X out .

3.4. Structure and Function of the SAG Module

The SAG-Model is mainly used to enhance the spatial adaptation of image features, adaptive pooling paths after SCAAP and maximum pooling paths after the SAEM module to guide the network to better focus on the salient regions in the image. As shown in Figure 8, the core idea of the module is to capture the important spatial information in the image by weighted multi-scale feature maps and spatially guided convolution to improve the accuracy of salient target detection.
SAG-Model receives five input feature maps P T i 1 , P T i 2 , P T i 3 , P T i 4 , P T i 5 (from different layers of the network or multi-scale inputs). These input feature maps are weighted and fused. Each feature map undergoes an element-by-element multiplication operation (⊗) and the contribution of each feature map is adjusted by a weighting factor. The weighted feature maps are fused with spatially guided information to form a more refined feature representation. This process allows the model to dynamically adjust the attention of specific regions in the image according to the importance of different input feature maps. The following equation gives the mathematical expression for weighted fusion:
P S A G = i = 1 5 α i P T i i
where α i is the weighting coefficients of each input feature map, which can be learned by the network. P T i i denotes the ith feature map. ⊗ denotes the element-by-element multiplication operation.
Next, spatially guided convolutional operations are applied to each feature map. The spatial information is extracted and enhanced by a 3 × 3 convolution kernel that performs a convolution operation on each input feature map. This spatially guided convolution operation not only extracts fine-grained spatial features in the image, but also enhances the spatial structure of objects in the image. This weighted fusion process ensures that information at different scales and with different features is effectively fused for subsequent spatial guidance. The following equation expresses the process of spatially guided convolution:
P S A G , conv = Conv P S A G , 3 × 3
The manipulation helps to enhance the spatial structure, allowing the model to more accurately capture salient areas in the image. The SAG pseudocode is as as shown in Algorithm 2.
Algorithm 2: Spatial Adaptive Guidance (SAG)
Remotesensing 17 02053 i002
The fusion process is one of the core steps of SAG-Model. In this process, all the spatially guided convolutionally enhanced feature maps are further fused. A 1 × 1 convolutional layer is used in the fusion operation, which can effectively combine the information from different scales and channels to produce the final output feature maps. Through this fusion method, the SAG-Model integrates the spatial information and details of different feature maps to generate a more adaptive feature map, which provides more accurate inputs for subsequent target detection tasks. The following equation describes this fusion process:
P S A G = Conv P T i , conv 1 , P T i , conv 2 , P T i , conv 3 , P T i , conv 4 , P T i , conv 5 , 1 × 1
Finally, the spatially guided and convolutionally fused feature maps P S A G are passed as output to the subsequent target detection network as shown in Figure 9. This output feature map not only contains accurate spatial guidance information, but also ensures that the network can focus on key regions in the image through weighting and multi-scale feature fusion, which improves the detection performance of significant targets.

3.5. Accurate Feature Fusion Module

Our proposed network structure is complementarily sampled from maximum pool channels and adaptive pool channels, which are adaptively pooled according to the channel and spatial attention under the SC-AAP module to focus more on the detail features and locally important regions of salient targets. In order to accurately locate the detail features and locally important regions of the salient images in the picture, after sampling, we then propose the accurate feature fusion module maximally adapted to the multi-channel sampling network structure, with the aim of integrating the feature information from two different paths, improving the model’s ability to detect the salient targets and the accuracy of the feature representation. The following describes the accurate feature fusion module specifically.

3.5.1. Feature Adaptation and Fusion

Firstly, before feature fusion, we convolve F m a x to adjust the number of channels so that it is the same as that of F a d a p t i v e . This step ensures that the two feature maps can be effectively weighted and summed in the channel dimension in the subsequent feature fusion operation. The adjusted feature map F m a x is computed as:
F m a x = C o n v F max , W 1 , b 1
where W 1 R k 1 × k 1 × C m a x × C adaptive is the weight matrix of the convolutional layer, C max is the original number of channels of F m a x and b 1 R H × W is the bias vector, we then, we adopt the multi-scale balancing mechanism approach to fuse the adjusted feature maps F m a x and F a d a p t i v e . A learnable weight parameter ω is introduced, and the fusion weights of the two channels’ features are dynamically adjusted according to the mean values by calculating the global means μ F max and μ F ada of F max and F ada . The specific calculation process is as follows.
ω = σ W bal · μ F max , μ F ada + b bal
where σ is the Sigmoid activation function. In Equation (12), F max and in Equation (13), F ada are intermediate feature maps used before feature fusion. They are adaptively fused based on a learned weight parameter ω , which is computed from their global mean values. The result of this fusion serves as an input to the spatially guided fusion module, which ultimately produces the final output feature map P S A G in Equation (11).
Therefore, F max and F ada are important intermediate features that contribute to generating, P S A G which is the spatially guided and convolutionally fused feature map used for improved target detection performance. The fused feature map F f u s e d is calculated as:
F fised = ω F max + ( 1 ω ) F adaptive
During the training process, the model automatically learns to adjust these weight parameters to determine the relative importance of the two input feature maps in the fusion process.

3.5.2. Feature Compression and Up-Sampling

Although the fused feature map synthesizes the information from the two inputs, it may have redundancy or an excessive number of channels, which is not conducive to subsequent processing. Therefore, a more compact feature representation is obtained by reducing its number of channels to an appropriate value through a convolutional compression operation. In order to reduce the amount of computation and obtain a more compact feature representation, a convolutional compression operation is performed on the fused feature map F c o m p r e s s e d . Then the compressed feature map is computed as:
F c o m p r e s s e d = C o n v F f u s e d , W 2 , b 2
where W 2 R k 2 × k 2 × C f u s e d × C c o m p r e s s e d is the weight matrix of the convolutional layer, C f u s e d is the number of original channels of F f u s e d , b 2 R C o m p r e s s e d is the bias vector, and finally, we perform an up-sampling operation on the compressed feature map F c o m p r e s s e d to restore its size to match the input image to obtain the final output feature map F o u t . The bilinear interpolation up-sampling method is used, and the up-sampling multiplier is determined according to the change of the feature map size in the network, let the up-sampling multiplier be, then the up-sampling operation formula is:
F o u t = U p s a m p l e F c o m p r e s s e d , r
The accurate feature fusion module effectively integrates feature information from two different paths through a carefully designed and fused process, improving the model’s ability to detect salient targets and the accuracy of feature representation.

3.5.3. Deep Supervision Mechanism

In order to optimize our network efficiently, we introduce a deep supervision mechanism, which obtains a multi-level feature representation of the network through multi-level branching outputs and saliency maps. During the training process, we gradually optimize the training parameters of the network by monitoring these intermediate outputs and calculating their cross-entropy losses. The specific loss function can be expressed as.
l = 1 W × H x = 1 W H y = 1 [ G ( x , y ) log ( S ( x , y ) ) + ( 1 G ( x , y ) ) log ( 1 S ( x , y ) ) ]
where G ( x , y ) and S ( x , y ) denote the label value and predicted significance value of a pixel ( x , y ) , respectively, and W and H are the width and height of the input image, respectively. This loss function facilitates the network’s accurate identification of salient regions at all scales by comparing the prediction of each pixel with its true label.
To further improve the performance of the network, we combine the losses from the multi-level branching outputs of the three paths (adaptive pooling path, maximum pooling path, and the fusion of the two) during the training process. The final training loss is defined as the weighted sum of these losses with the following equation.
L = i = 1 6 α i l A i + l M i + l A M i
where L is the total network loss, i denotes the different multi-level branch outputs and significance maps (6 branches in total), and α i is the weight of each branch output loss, which takes the following values:
α i = 1 , i = 6 2 i , i = 5 , 4 , 3 , 2 , 1
The design of the deep supervision mechanism helps the network to self-regulate and optimize from multiple scales and paths. With this multi-level supervision, the network is able to better capture the details of salient objects in the image while avoiding the limitations associated with single-path supervision. In addition, this mechanism allows the network to obtain more gradient information during training, which accelerates convergence and improves performance.
In the comparative experiments in this paper, we also introduced a non-deep supervision mechanism (single-path supervision) to evaluate the effectiveness of the deep supervision mechanism, further demonstrating the advantages of deep supervision in the saliency detection task. Through these experiments, we are able to more comprehensively verify the contribution of the SCAAP module at different levels and paths, and provide theoretical support for further optimization.

4. Experimental Results and Analysis

4.1. Experimental Setup

4.1.1. Datasets

We use two public optical remote sensing image datasets, ORSSD [8] and EORSSD [7], to thoroughly validate our model. The ORSSD dataset contains 800 images, with 600 used for training and 200 for testing. It features diverse spatial resolutions, object scales, types, and cluttered backgrounds. The EORSSD dataset, an extension of ORSSD, contains 2000 images, with 1400 for training and 600 for testing. Both datasets include pixel-level annotations for each image.
In our experiments, we train the model using 600 images from ORSSD and 1400 from EORSSD, with 200 and 600 images, respectively, used for testing. To expand the training data, we applied rotation (90°, 180°, 270°) and mirror reflection (90°, 180°, 270°) operations, resulting in 4800 training samples for ORSSD and 11,200 for EORSSD. During training, we resized each image to 256 × 256 × 12.

4.1.2. Evaluation Metrics

To quantitatively compare the performance of different salient target detection models on the ORSSD [8] and EORSSD [7] datasets, we use several evaluation metrics: Precision-Recall (PR) curves, F-measure curves, maximum F-measure at different thresholds ( F β max ), average F-measure over multiple thresholds ( F β mean ), F-measure with adaptive thresholding ( F β adp ), S-measure ( S α ), maximum E-measure at different thresholds ( E ξ max ), average E-measure over multiple thresholds ( E ξ mean ), E-measure with adaptive thresholding ( E ξ adp ), and mean absolute error (MAE) ( M ).
Precision and recall are standard performance metrics. In our experiments, we calculate 256 pairs of mean precision and recall values at thresholds ranging from 0 to 255 and plot PR curves, with precision on the vertical axis and recall on the horizontal axis. The closer the PR curve is to the (1, 1) point, the better the model’s performance.
F-measure, a composite metric, is the harmonic mean of precision and recall, calculated as follows:
F β = 1 + β 2 Precision Recall β 2 Precision + Recall
where Precision = T P T P + F P , TP is the number of true cases, FP is the number of false-positive cases, Recall = TP TP + FN , FN is the number of false-negative cases. β is the balancing factor ( β is set to 0.3 in the experiments in this paper to focus more on the accuracy rate). Our experiments provide maximum F-measure values, average F-measure values, and adaptive F-measure values, and show the maximum F-measure curves at the same time. The F-measure curves are plotted based on the F-scores and thresholds ([0, 255]), and each F-score is calculated using the above formula at each threshold. The larger the area of coordinates covered by an F-score, the larger the area of coordinates covered by an F-measure curve, and the greater the area of coordinates covered by an F-score. The larger the coordinate area covered by the curve, the better the model performance.
M = 1 W × H x = 1 W y = 1 H | S ( x , y ) G ( x , y ) |
where W and H denote the width and height of the saliency map, respectively, S ( x , y ) is the pixel values of the saliency map, and G ( x , y ) is the pixel value of the true mask.
The S-measure measures the structural similarity of the saliency map, taking into account both region similarity ( S r ) and object similarity ( S o ), and is defined as
S α = α · S o + ( 1 α ) · S r
S o is Object-aware S-measure, which measures the totality of saliency objectives. S r is Region-aware S-measure, which measures the consistency of regional saliency. α is the weighting factor ( α set to 0.5 in the experiments of this paper).
The E-measure evaluates the similarity between the predicted saliency map and GT by considering both the local pixel saliency values and the image-level average saliency values, which is calculated as
ξ = 2 φ G T ( x , y ) φ F M ( x , y ) φ G T ( x , y ) φ G T ( x , y ) + φ F M ( x , y ) φ F M ( x , y ) , E ξ = 1 W × H x = 1 W y = 1 H f ( ξ )
where f ( ) is a convex function denoting the Hadamard product, and the alignment matrix ξ is constructed based on the deviation matrices φ G T and φ F M , which can be regarded as centering operations on GT and binary salient maps, respectively.
While quantitative metrics such as F β max , S-measure, E-measure, MAE, and IoU are widely used to assess model performance, interpreting their practical impact is crucial for understanding model behavior in real-world applications. Table 1 summarizes the functional meaning of each metric and how typical improvements reflect in qualitative outcomes.
For example, as shown in Table 2, IEAM-Net improves F β max by 0.005–0.012 compared to strong baselines. While these differences appear small numerically, they correspond to visible qualitative improvements such as: (i) better object boundary integrity, (ii) more accurate exclusion of noisy background, and (iii) higher continuity across occluded targets. These enhancements are observable in visual examples (e.g., Figure 12, row 3 vs. row 6), and are particularly meaningful in downstream tasks like urban analysis or change detection.
Therefore, the metric improvements are both statistically and visually significant, validating the real-world effectiveness of our model.

4.1.3. Implementation Details

Our model was implemented using PyTorch 1.13 and trained on a workstation equipped with an NVIDIA RTX 3090 GPU (24 GB VRAM), an Intel i9 CPU, and 128 GB RAM. We used the Adam optimizer with an initial learning rate of 1 × 10−4, and adopted a cosine annealing learning rate schedule with a warm-up over the first 5 epochs. The network was trained for 100 epochs with a batch size of 8, using Binary Cross-Entropy loss combined with a deep supervision mechanism. The total training time was approximately 11 h on the EORSSD dataset and 5.5 h on the ORSSD dataset. All experiments were conducted under CUDA 11.6 and Python 3.9 environments.

4.2. Comparison Experiments (With Advanced Methods)

Comparing with 21 state-of-the-art SOD methods, we classify these 21 methods into four categories: the first category is natural scene image SOD methods based on classical vision algorithms (HDCT [35], RCRRS [36], SMD [37], RRWR [38]), the second category is natural scene image SOD methods based on convolutional neural networks (R3Net [11], DSS [39], RADF [40], EG-Net [41], GCPA [34], Pool-Net [18], PA-KRN [42], ITSD [29], Gate-Net [28], SUCA [43], MI-Net [44], U2Net [45]), the third category is for remote sensing image-SOD methods based on classical vision algorithms (CMC [46], SMFF [33]), and the fourth category is for remote sensing image-SOD methods based on convolutional neural networks (LV-Net [7], EMFI-Net [17], DAF-Net [8]). During the experimental process, in order to ensure a fair comparison of the 22 algorithms, we uniformly use the saliency maps provided by the public RSI-SOD benchmark dataset or the saliency maps provided by the authors of the corresponding algorithms, and, for the 7 inside the second category of the above-mentioned convolutional neural network-based SOD methods for natural scene images (ITSD [29], MI-Net [44], Gate-Net [28], GCPA [34], U2Net [45], SUCA [43], PA-KRN [42]), using the same training data to re-train the models on the EORSSD [7] and ORSSD [8] datasets with the default settings provided by the authors.

4.2.1. Quantitative Comparison

Table 2 presents a quantitative comparison between IEAM-Net and 21 advanced SOD methods on the EORSSD [7] and ORSSD [8] datasets. The comparison includes traditional and CNN-based SOD methods for both natural and remote sensing images. All methods were evaluated under uniform conditions using saliency maps from the public RSI-SOD dataset or those provided by the authors.
On the EORSSD [7] dataset, IEAM-Net outperforms many methods on several metrics. It achieves 0.9308 on Fmax, surpassing DAF-Net (0.9166) and approaching EMFI-Net (0.9290), showing its accuracy in salient target detection. IEAM-Net also performs well on F-mean with a score of 0.8905, higher than DAF-Net’s 0.8614, demonstrating its stability. For E-max and E-mean, IEAM-Net scores 0.9737 and 0.9625, slightly below DAF-Net’s 0.9861 and 0.9291, but still strong, particularly in complex scenes. Compared to other CNN-based remote sensing image SOD methods (e.g., EMFI-Net and DAF-Net), IEAM-Net excels in metrics like F-adp and M, showing an all-around improvement across multiple evaluation metrics.
Overall, IEAM-Net outperforms most existing methods on the EORSSD [7] dataset, especially in Fmax and F-mean, demonstrating its superiority in salient target detection in remote sensing images. The PR curve (Figure 10) further supports these results, with IEAM-Net’s curve closer to the upper-right corner, indicating excellent performance.

4.2.2. Comparison of Computational Complexity

Computational complexity is typically evaluated based on inference speed (FPS), number of parameters, and floating point operations (FLOPs). Figure 11 provides a comparative analysis of various Salient Object Detection (SOD) methods, including a box plot analysis that further illustrates the distribution and trends of computational efficiency across different approaches. The results demonstrate that IEAM-Net outperforms existing methods in computational efficiency while maintaining high detection accuracy. Most CNN-based methods can run in real-time (16–26 FPS), and IEAM-Net inference speed reaches 48 FPS, which is favorable for practical applications. Its network parameters and FLOPs are at a medium level, compared with the second best EMFI-Net [17], which is 67.65 M (IEAM-Net) less than 107.26 M (EMFI-Net [17]), and the FLOPs are 103.2 G (IEAM-Net) less than 480.9 G (EMFI-Net [17]), which indicates that IEAM-Net is effective and efficient.
Table 3 provides a comprehensive comparison between the proposed IEAM-Net and 21 representative saliency detection methods across three major categories: traditional models, CNN-based models for natural scene images (NSI-CNN), and remote sensing image-specific CNN models (RSI-CNN). The comparison considers four key aspects: accuracy ( F β max ), inference speed (FPS), model size (in millions of parameters), and computational complexity (FLOPs in G).
Notably, IEAM-Net achieves the highest accuracy with an F β max score of 0.8905 while also delivering the fastest inference speed (48 FPS) among all CNN-based models. Despite its strong performance, IEAM-Net maintains a relatively low parameter count (67.7 M) and the lowest FLOPs (103.2 G) in the RSI-CNN category, indicating its superior balance between effectiveness and efficiency.
Compared to recent high-performing methods such as PA-KRN (0.8639 F β max , 617.7 G FLOPs) and EMFI-Net (0.8720 F β max , 487.3 G FLOPs), IEAM-Net offers not only improved detection accuracy, but also significantly reduced computational cost, making it a practical and scalable choice for real-time remote sensing applications.
Inference Speed Comparison
Most CNN-based SOD methods achieve inference speeds ranging from 16 to 26 FPS, with models such as PoolNet [18], GCPA [34], and PA-KRN [42] maintaining speeds around 24–25 FPS. Traditional methods, including RRWR [38], HDCT [35], SMD [37], and RCRR [36], exhibit significantly lower speeds, often below 10 FPS, highlighting their inefficiency in processing high-resolution remote sensing images. IEAM-Net achieves a remarkable inference speed of 48 FPS, which is nearly twice as fast as the best existing CNN-based methods, including EMFI-Net [17] and DAF-Net [8].
Box plot analysis: The box plot on the left (inference speed) shows that most methods cluster around the lower FPS range (below 26 FPS), with a few high-performing outliers. IEAM-Net is a distinct outlier, achieving a significantly higher speed than the trend observed in other methods, as indicated by the separate data point far above the whiskers. The overall distribution suggests that IEAM-Net significantly deviates from the median speed of existing methods, making it an exceptionally fast approach.
FLOPs Comparison
IEAM-Net significantly reduces computational complexity, requiring only 103.2 G FLOPs, compared to advanced methods like PA-KRN [42] (617.7 G FLOPs), EMFI-Net [17] (487.3 G FLOPs), GCPA [34] (291.9 G FLOPs), and DAF-Net [8] (376.2 G FLOPs). While traditional methods like RRWR [38], HDCT [35], and SMD [37] have low FLOPs, they underperform in detection accuracy. IEAM-Net, with its efficient attention mechanism and adaptive pooling strategy, maintains high precision while minimizing computational cost.
The box plot (FLOPs comparison) shows that most CNN-based SOD methods have high computational costs, with several exceeding 600 G FLOPs (e.g., PA-KRN [42]). IEAM-Net’s FLOPs are well below the median, indicating its competitive performance with lower computational complexity. The varying FLOPs across methods highlight the efficiency of IEAM-Net.
IEAM-Net achieves the highest inference speed (48 FPS), ideal for real-time remote sensing processing. Its FLOPs are 78.8% lower than EMFI-Net [17] (487.3 G) and 83.3% lower than PA-KRN [42] (617.7 G), demonstrating superior computational efficiency. Additionally, with only 67.65 M parameters compared to EMFI-Net’s 107.26 M parameters, IEAM-Net reduces model size by about 37%.
Box plot analysis further confirms IEAM-Net’s exceptional performance in both inference speed and computational efficiency, making it an optimal solution for large-scale remote sensing tasks.

4.2.3. Visual Comparison

Figure 12 compares the performance of IEAM-Net with other salient object detection methods in optical remote sensing images (RSI). It includes various remote sensing scenes, such as airplanes, cars, buildings, rivers, and swimming pools, categorized by conditions such as normal airplanes, airplanes with shadows, and airplanes under disturbances. Each row displays the original optical RSI, the ground truth (GT), IEAM-Net output (highlighted in red), and results from methods like GateNet [28], EMFI-Net [17], PA-KRN [42], U2Net [45], DAF-Net [8], ITSD [29], and RCRR [36].
The results reveal notable performance differences. IEAM-Net detects large objects with complete shapes, while other methods often distort or incompletely segment them. It handles shadows well, maintaining accurate shape detection, unlike other methods that struggle with shadow interference. For multiple objects, IEAM-Net excels at segmentation without merging errors, while some methods fail to detect all objects. In cluttered backgrounds, IEAM-Net distinguishes target objects from noise, whereas other methods suffer from false positives. For small objects, IEAM-Net accurately identifies them, whereas other methods often fail or distort small object shapes. (a) Dense Object Scenes: In urban or port environments (e.g., Figure 12, rows 3 and 5), IEAM-Net accurately distinguishes adjacent targets (e.g., buildings, ships) without boundary merging, thanks to the SAG module’s spatial refinement. (b) Small Target Detection: IEAM-Net captures fine-scale objects such as vehicles and boats (rows 4 and 6), outperforming other methods prone to omission. This benefits from the multi-scale detail retention enabled by SCAAP. (c) Occlusion and Shadow Handling: In partially occluded scenes, our model recovers more complete structures than baselines, owing to its bidirectional attention design that preserves spatial continuity. (d) Cluttered Backgrounds: IEAM-Net exhibits strong background suppression (e.g., over water or vegetation), reducing false positives via edge-aware enhancement (SAEM) and adaptive attention calibration.
Overall, compared to traditional methods (e.g., SMFF [33], SMD [37]), CNN-based non-RSI methods (e.g., PA-KRN [42], U2Net [45]), and CNN-based RSI methods (e.g., EMFI-Net [17], DAF-Net [8], LV-Net [7]), IEAM-Net demonstrates superior accuracy and robustness in optical RSI. Traditional methods struggle with RSI characteristics, CNN-based non-RSI methods face data adaptation issues, and CNN-based RSI methods often produce defective maps. IEAM-Net’s unique design improves object localization and preserves fine contours, proving adaptable and robust in complex environments.
Although IEAM-Net demonstrates strong performance across a wide range of scenarios, several failure cases remain. For instance, as shown in the rightmost columns of Figure 11, the model occasionally fails to detect very small or low-contrast objects, especially when they are heavily occluded by shadows or embedded in cluttered backgrounds. These failures often arise in challenging scenes involving complex textures, overlapping targets, or significant illumination variation, which may mislead the attention modules or cause inconsistencies in feature representation across scales. Common failure patterns include missed detections in occluded or overlapping regions, blurred boundaries in noisy backgrounds, and reduced saliency localization in the presence of densely packed targets. These observations point to limitations in the current model’s ability to balance local-global contextual modeling and maintain fine-grained edge delineation. To address these issues, future work may explore the integration of transformer-based modules to enhance global reasoning, the introduction of multi-resolution fusion mechanisms at earlier network stages, and the design of more advanced edge-preserving or contour-refining loss functions. Additionally, targeted data augmentation strategies that simulate occlusion and noise could further improve the model’s robustness.

4.3. Ablation Experimental Study

4.3.1. SCAAP Component Ablation Experiments

The effectiveness of the IEAM-Net essential components on the EORSSD [7] and ORSSD [8] datasets was assessed through comprehensive experiments investigating the following aspects: the individual contribution of each content in the SCAAP module, the necessity of merging the original content, and the effectiveness of the combined loss function. Each variant experiment was retrained using the same parameter settings and datasets as in the experimental protocol.
To evaluate the individual contributions of the Spatially Adaptive Edge Embedding Module (SAEM), Spatial Adaptive Guidance Module (SAG), Adaptive Pooling (AP), and Maximal Pooling Path (MP) to the SCAAP module’s performance, we constructed five configurations (top half of Table 4): baseline (VGG-16-based network), baseline + SAEM, baseline + SAEM + SAG, baseline + SAEM + SAG + AP, and baseline + SAEM + SAG + AP + MP.
As shown in Table 4, model performance improves with the progressive addition of modules. The SAEM module enhances channel feature selection, improving detection sensitivity by distinguishing significant regions. The SAG module refines target localization accuracy by capturing key spatial information, making the model more robust in complex scenarios. The AP module adjusts the feature extraction range, enhancing multi-scale salient object handling, while the MP module improves local feature capture, strengthening edge and detail portrayal.
Quantitative results indicate that the full SCAAP module improves performance on the EORSSD [7] dataset from 0.8878 to 0.8905 in the baseline and 0.9096 to 0.9135 overall. On the ORSSD [8] dataset, it improves from 0.8712 to 0.8843 in the baseline, with Emax improving from 0.8995 to 0.9128. The full module’s combined effect improves performance by 2.25% and 1.93%, respectively, highlighting the synergy of SCAAP components.
Additionally, we assessed different module combinations (middle part of Table 4), demonstrating that each module independently boosts performance and that their collaboration further enhances detection. SAEM enhances target edges, SAG improves target area attention, and the combination of AP and MP increases model robustness in multi-scale scenarios. These results validate the effectiveness of the SCAAP module design and its potential for salient object detection in complex scenes.
The image illustrates the impact of adaptive pooling (AP) and max pooling (MP) pathways in IEAM-Net for salient object detection in remote sensing images as shown in Figure 13. The top section shows the original image and feature maps ( A 1 A 5 ) from the AP pathway, where high-intensity regions highlight the salient target with refined details. AP dynamically adjusts feature extraction, enhancing multi-scale object detection.
The bottom section presents the ground truth (GT) and feature maps ( M 1 M 5 ) from the MP pathway, which retains overall structure but lacks finer details. While MP ensures global feature consistency, AP provides superior boundary enhancement and noise suppression. The comparison demonstrates that integrating both pooling strategies optimizes detection by balancing detail refinement and structural preservation. This synergy enables IEAM-Net to achieve high-accuracy salient object detection, as validated by ablation study results.

4.3.2. Feature Fusion Module Ablation Experiment

Through this experiment, we deeply validate the importance of fusing raw content in the SCAAP module for accurate feature fusion. Specifically, the experiment evaluates the actual role of fusing raw content in enhancing the effectiveness of saliency detection by comparing two approaches-FFM (traditional feature fusion module) and EAFFM (our proposed improved version of the feature fusion module). As can be seen from the Table 5, the performance of EAFFM compared to FFM on both datasets is significantly improved, proving the effectiveness of fusing raw content.
On the EORSSD [7] dataset, the EAFFM’s improves from 0.8869 to 0.8905, an improvement of 0.0036; and from 0.9717 to 0.9737, an improvement of 0.0020, which indicates that the overall sensitivity and detection effectiveness of the model are significantly enhanced after the introduction of raw content as shown in Table 5.
Similarly, on the ORSSD [8] dataset, the EAFFM improves from 0.9118 to 0.9135 with 0.0017, and from 0.9754 to 0.9783 with 0.0029, which further demonstrates that the fusion of original content can stably improve the model performance on different datasets, especially in the identification of details and edge regions.
These results indicate that fusing raw content not only helps the model to improve the perception of salient regions, but also enhances the robustness of the model, especially in complex scenes. Taken together, the role of original content fusion in accurate feature fusion cannot be ignored, and can effectively compensate for the feature loss and improve the overall accuracy and reliability of detection.

4.3.3. Supervision Mechanism Ablation Experiment

Through this experiment, we further explored the effect of introducing Deep Supervision (Deep Supervision) in the SCAAP module on the model performance enhancement, and compared it with the traditional Single Path Supervision (Single Path Supervision). As can be seen from the data in the Table 6, the Deep Supervision mechanism significantly improves the overall performance of the saliency detection task, verifying the effectiveness of multi-scale adaptive adjustment in deep supervision.
On the EORSSD [7] dataset, the model using the deep supervision mechanism (DEEP SUP) improves 0.0030 from 0.8875 to 0.8905 and 0.0017 from 0.9720 to 0.9737 compared to the single-path supervision (SINGLE SUP), and this improvement suggests that the deep supervision mechanism is able to improve the performance of the network from multiple scales through the gradient feedback at multiple levels. feedback, which enables the network to self-optimize from multiple scales and paths, thus enhancing the model’s ability to capture salient regions as shown in Table 6.
On the ORSSD [8] dataset, the deep supervision mechanism also shows significant advantages, from 0.9121 to 0.9135, with an improvement of 0.0014, and from 0.9769 to 0.9783, with an improvement of 0.0014. This result further confirms that the deep supervision mechanism can effectively enhance the robustness of the model in complex scenarios, and accelerate the convergence of the training process.
Through these comparative experiments, we can conclude that the deep supervision mechanism not only enhances the performance of the model in multi-scale scenarios, but also enhances the learning ability of the model through the feedback of the multi-path gradient information, making the network more efficient in the training process. This further proves the unique advantages of the deep supervision mechanism in the saliency detection task.

4.3.4. Impact of Backbone Choice

Although VGG-16 is a widely adopted and stable backbone in the saliency detection literature, it is technically outdated in terms of feature richness, receptive field, and computational efficiency. To assess the impact of backbone selection on the performance of our proposed model, we conducted additional experiments by replacing VGG-16 with two modern architectures:
  • ResNet-50: a deeper residual CNN known for its strong semantic representation capabilities.
  • EfficientNet-B0: a lightweight yet powerful model that utilizes compound scaling to balance depth, width, and resolution.
Table 7 summarizes the performance of IEAM-Net with these three backbones. The ResNet-50 variant achieves a notable increase of +1.7% in F β max while maintaining competitive inference speed. EfficientNet-B0 reduces the total parameter count while achieving comparable accuracy and lower MAE than VGG-16.
These results confirm that IEAM-Net’s core design is robust across different backbones. Nonetheless, employing stronger semantic encoders can yield further improvements in performance. Therefore, in future work, we plan to explore transformer-based backbones (e.g., Swin Transformer, Pyramid Vision Transformer) to leverage global context modeling and multi-scale feature fusion more effectively.

4.4. Cross-Domain Generalization Evaluation

To assess the robustness and generalization ability of IEAM-Net beyond the ORSSD and EORSSD datasets, we conducted additional cross-domain validation experiments on two unseen datasets with differing scene characteristics and imaging distributions:
The iSOD-RS and WHU-RS19 datasets used for cross-domain evaluation are internal collections maintained by our research group, based on publicly available remote sensing imagery (e.g., Google Earth, high-resolution satellite feeds). They are not part of any standardized saliency benchmark and are used here solely for qualitative or zero-shot validation purposes. Labels were manually curated for iSOD-RS, while WHU-RS19 is evaluated only qualitatively due to the lack of pixel-level annotations.
  • iSOD-RS: A large-scale optical remote sensing saliency dataset featuring diverse object categories, complex backgrounds, and varied resolutions.
  • WHU-RS19: A scene classification dataset not originally designed for saliency detection but used here for qualitative zero-shot evaluation.
As shown in Table 8, when applied to the in-domain datasets (ORSSD and EORSSD), IEAM-Net achieves high performance with F max scores of 0.8905 and 0.8932, and low MAE values (0.031 and 0.028), demonstrating accurate saliency detection and well-suppressed background noise.
On the unseen iSOD-RS dataset, although performance drops slightly (e.g., F max decreases to 0.8511 and MAE increases to 0.041), the results remain competitive, indicating that the model retains its discriminative capacity even under domain shift. The IoU score of 0.774 also reflects good spatial overlap with ground truth regions. This suggests that the model can generalize effectively to datasets with more diverse object categories and scene textures.
On WHU-RS19, which lacks pixel-level labels, we provide only qualitative evaluation. Nevertheless, the computed metrics based on pseudo-labels still reflect reasonably good alignment ( F max = 0.8073 , MAE = 0.057), which further supports the model’s robustness in practical applications where annotated data may be unavailable.
Overall, these quantitative results validate IEAM-Net’s generalization ability and highlight its potential in real-world scenarios involving unseen imaging conditions or data distributions.

5. Conclusions

In this paper, we propose IEAM-Net (Spatial-Channel Attentive Adaptive-Pool Network), a lightweight yet highly effective framework for salient object detection in optical remote sensing images (RSI-SOD). Our model addresses key challenges such as complex backgrounds, occlusions, noise interference, and the trade-off between detection accuracy and computational efficiency. By integrating a well-structured feature fusion strategy, spatial-channel attention mechanisms, and adaptive pooling, IEAM-Net significantly enhances multi-scale feature representation, target boundary refinement, and background suppression while maintaining a minimal memory footprint.
The proposed Spatially Adaptive Edge Embedded Module (SAEM) improves target structure perception, enabling finer segmentation boundaries. The SCAAP module dynamically selects the most relevant spatial and channel-wise features, effectively balancing adaptive and max pooling operations to retain both global context and fine details. Additionally, the Spatial Adaptive Guidance (SAG) module enhances feature localization, mitigating the semantic dilution issue in U-shaped networks and improving detection robustness in cluttered environments.
Extensive experiments on the EORSSD and ORSSD benchmark datasets demonstrate that IEAM-Net surpasses 21 state-of-the-art methods, achieving a high inference speed of 48 FPS with only 103.2 G FLOPs, making it a highly efficient and practical solution for real-time applications. Moreover, ablation studies confirm the effectiveness of our adaptive pooling strategy, spatial-channel feature refinement, and deep supervision mechanism, which collectively contribute to superior performance in detecting salient objects across diverse and challenging remote sensing scenarios.
However, IEAM-Net still has some limitations. It currently relies on the VGG-16 backbone, which may restrict scalability and representational capacity in more complex scenarios. Moreover, although the overall computational cost is reduced compared to other models, the 103.2 G FLOPs required may still pose challenges for deployment on resource-constrained edge devices.
In future work, we plan to explore more efficient or transformer-based backbone networks to improve model flexibility and reduce computational burden. We also intend to extend IEAM-Net to support multi-modal data such as SAR, infrared, or hyperspectral imagery, which would enhance its robustness across various sensing environments.
IEAM-Net demonstrates strong potential in applications such as environmental monitoring, where it can help detect ecological changes like deforestation, seasonal water body fluctuations, and habitat degradation by accurately identifying dynamic regions in satellite imagery. Furthermore, in disaster response contexts, the model can be used to rapidly locate damaged infrastructure, flooded zones, or affected buildings following natural disasters such as earthquakes or floods, thereby supporting emergency assessment and resource allocation. Additionally, the framework is well-suited for urban planning tasks, including the monitoring of urban sprawl, identification of informal settlements, and detection of unauthorized constructions, all of which are crucial for informed policy-making and sustainable land use management.

Author Contributions

Conceptualization, F.Z. and Z.Z.; methodology, F.Z.; software, Z.Z.; validation, F.Z. and Z.Z.; investigation, F.Z. and Z.Z.; data curation, F.Z. and Z.Z.; writing—original draft preparation, F.Z. and Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is not supported by any project funding.

Data Availability Statement

All data that support the findings of this study are included within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, G.Y.; Liu, Z.; Zhang, X.P.; Lin, W.S. Lightweight Salient Object Detection in Optical Remote-Sensing Images via Semantic Matching and Edge Alignment. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5601111. [Google Scholar] [CrossRef]
  2. Borji, A.; Cheng, M.M.; Jiang, H.Z.; Li, J. Salient Object Detection: A Benchmark. IEEE Trans. Image Process. 2015, 24, 5706–5722. [Google Scholar] [CrossRef] [PubMed]
  3. Wang, W.G.; Lai, Q.X.; Fu, H.Z.; Shen, J.B.; Ling, H.B.; Yang, R.G. Salient Object Detection in the Deep Learning Era: An In-Depth Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 3239–3259. [Google Scholar] [CrossRef] [PubMed]
  4. Ma, X.P.; Zhang, X.K.; Pun, M.O. RS3Mamba: Visual State Space Model for Remote Sensing Image Semantic Segmentation. IEEE Geosci. Remote Sens. Lett. 2024, 21, 6011405. [Google Scholar] [CrossRef]
  5. Li, X.H.; Xie, L.L.; Wang, C.F.; Miao, J.H.; Shen, H.F.; Zhang, L.P. Boundary-enhanced dual-stream network for semantic segmentation of high-resolution remote sensing images. Giscience Remote Sens. 2024, 61, 2356355. [Google Scholar] [CrossRef]
  6. Cong, R.M.; Zhang, Y.M.; Fang, L.Y.; Li, J.; Zhao, Y.; Kwong, S. RRNet: Relational Reasoning Network with Parallel Multiscale Attention for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5613311. [Google Scholar] [CrossRef]
  7. Li, C.Y.; Cong, R.M.; Hou, J.H.; Zhang, S.Y.; Qian, Y.; Kwong, S. Nested Network With Two-Stream Pyramid for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9156–9166. [Google Scholar] [CrossRef]
  8. Zhang, Q.J.; Cong, R.M.; Li, C.Y.; Cheng, M.M.; Fang, Y.M.; Cao, X.C.; Zhao, Y.; Kwong, S. Dense Attention Fluid Network for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Image Process. 2021, 30, 1305–1317. [Google Scholar] [CrossRef]
  9. Zhang, P.P.; Wang, D.; Lu, H.C.; Wang, H.Y.; Ruan, X. Amulet: Aggregating Multi-level Convolutional Features for Salient Object Detectionn. In Proceedings of the 16th IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 202–211. [Google Scholar] [CrossRef]
  10. Yang, S.; Jiang, Q.P.; Lin, W.S.; Wang, Y.T. SGDNet: An End-to-End Saliency-Guided Deep Neural Network for No-Reference Image Quality Assessment. In Proceedings of the 27th ACM International Conference on Multimedia (MM), Nice, France, 21–25 October 2019; pp. 1383–1391. [Google Scholar] [CrossRef]
  11. Deng, Z.J.; Hu, X.W.; Zhu, L.; Xu, X.M.; Qin, J.; Han, G.Q.; Heng, P.A. R3Net: Recurrent Residual Refinement Network for Saliency Detection. In Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI), Stockholm, Sweden, 13–19 July 2018; pp. 684–690. Available online: https://www.ijcai.org/proceedings/2018 (accessed on 13 July 2018).
  12. Gu, K.; Wang, S.Q.; Yang, H.; Lin, W.S.; Zhai, G.T.; Yang, X.K.; Zhang, W.J. Saliency-Guided Quality Assessment of Screen Content Images. IEEE Trans. Multimed. 2016, 18, 1098–1110. [Google Scholar] [CrossRef]
  13. Zhang, Q.J.; Zhang, L.B.; Shi, W.Q.; Liu, Y. Airport Extraction via Complementary Saliency Analysis and Saliency-Oriented Active Contour Model. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1085–1089. [Google Scholar] [CrossRef]
  14. Fang, Y.M.; Chen, Z.Z.; Lin, W.S.; Lin, C.W. Saliency Detection in the Compressed Domain for Adaptive Image Retargeting. IEEE Trans. Image Process. 2012, 21, 3888–3901. [Google Scholar] [CrossRef] [PubMed]
  15. Ren, J.; Zhu, J. Pyramid Deep Fusion Network for Two-Hand Reconstruction From RGB-D Images. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 5843–5855. [Google Scholar] [CrossRef]
  16. Zhou, X.C.; Liang, F.; Chen, L.H.; Liu, H.J.; Song, Q.Q.; Vivone, G.; Chanussot, J. MeSAM: Multiscale Enhanced Segment Anything Model for Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5623515. [Google Scholar] [CrossRef]
  17. Zhou, X.F.; Shen, K.Y.; Liu, Z.; Gong, C.; Zhang, J.Y.; Yan, C.G. Edge-Aware Multiscale Feature Integration Network for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5605315. [Google Scholar] [CrossRef]
  18. Liu, J.J.; Hou, Q.B.; Cheng, M.M.; Feng, J.S.; Jiang, J.M.; Soc, I.C. A Simple Pooling-Based Design for Real-Time Salient Object Detection. In Proceedings of the 32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 3912–3921. [Google Scholar] [CrossRef]
  19. Zhou, L.; Yang, Z.H.; Zhou, Z.T.; Hu, D.W. Salient Region Detection Using Diffusion Process on a Two-Layer Sparse Graph. IEEE Trans. Image Process. 2017, 26, 5882–5894. [Google Scholar] [CrossRef]
  20. Sun, L.; Chen, Z.; Wu, Q.M.J.; Zhao, H.; He, W.; Yan, X. AMPNet: Average- and Max-Pool Networks for Salient Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 4321–4333. [Google Scholar] [CrossRef]
  21. Yan, Z.Y.; Li, J.X.; Li, X.X.; Zhou, R.X.; Zhang, W.K.; Feng, Y.C.; Diao, W.H.; Fu, K.; Sun, X. RingMo-SAM: A Foundation Model for Segment Anything in Multimodal Remote-Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5625716. [Google Scholar] [CrossRef]
  22. Zhang, X.J.; Li, S.; Tan, Z.Y.; Li, X.H. Enhanced wavelet based spatiotemporal fusion networks using cross-paired remote sensing images. ISPRS J. Photogramm. Remote Sens. 2024, 211, 281–297. [Google Scholar] [CrossRef]
  23. Li, G.; Liu, Z.; Lin, W.; Ling, H. Multi-Content Complementation Network for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5614513. [Google Scholar] [CrossRef]
  24. Du, Z.S.; Li, X.H.; Miao, J.H.; Huang, Y.Y.; Shen, H.F.; Zhang, L.P. Concatenated Deep-Learning Framework for Multitask Change Detection of Optical and SAR Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 719–731. [Google Scholar] [CrossRef]
  25. Ding, L.; Zhu, K.; Peng, D.F.; Tang, H.; Yang, K.W.; Bruzzone, L. Adapting Segment Anything Model for Change Detection in VHR Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5611711. [Google Scholar] [CrossRef]
  26. Li, G.Y.; Liu, Z.; Shi, R.; Hu, Z.; Wei, W.J.; Wu, Y.; Huang, M.K.; Ling, H.B. Personal Fixations-Based Object Segmentation With Object Localization and Boundary Preservation. IEEE Trans. Image Process. 2021, 30, 1461–1475. [Google Scholar] [CrossRef] [PubMed]
  27. Wang, W.G.; Shen, J.B.; Xie, J.W.; Cheng, M.M.; Ling, H.B.; Borji, A. Revisiting Video Saliency Prediction in the Deep Learning Era. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 220–237. [Google Scholar] [CrossRef] [PubMed]
  28. Zhao, X.; Pang, Y.; Zhang, L.; Lu, H.; Zhang, L. Suppress and balance: A simple gated network for salient object detection. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020, Proceedings, Part II 16; Springer: Cham, Switzerland, 2020; pp. 35–51. [Google Scholar] [CrossRef]
  29. Zhou, H.J.; Xie, X.H.; Lai, J.H.; Chen, Z.X.; Yang, L.X. Interactive Two-Stream Decoder for Accurate and Fast Saliency Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 9138–9147. [Google Scholar] [CrossRef]
  30. Cong, R.M.; Lei, J.J.; Fu, H.Z.; Cheng, M.M.; Lin, W.S.; Huang, Q.M. Review of Visual Saliency Detection with Comprehensive Information. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 2941–2959. [Google Scholar] [CrossRef]
  31. Li, G.Y.; Liu, Z.; Shi, R.; Wei, W.J. Constrained fixation point based segmentation via deep neural network. Neurocomputing 2019, 368, 180–187. [Google Scholar] [CrossRef]
  32. Chen, K.Y.; Chen, B.W.; Liu, C.Y.; Li, W.Y.; Zou, Z.X.; Shi, Z.W. RSMamba: Remote Sensing Image Classification With State Space Model. IEEE Geosci. Remote Sens. Lett. 2024, 21, 8002605. [Google Scholar] [CrossRef]
  33. Zhang, L.B.; Liu, Y.N.; Zhang, J. Saliency detection based on self-adaptive multiple feature fusion for remote sensing images. Int. J. Remote Sens. 2019, 40, 8270–8297. [Google Scholar] [CrossRef]
  34. Chen, Z.Y.; Xu, Q.Q.; Cong, R.M.; Huang, Q.M. Global Context-Aware Progressive Aggregation Network for Salient Object Detection. In Proceedings of the 34th AAAI Conference on Artificial Intelligence/32nd Innovative Applications of Artificial Intelligence Conference/10th AAAI Symposium on Educational Advances in Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 10599–10606. [Google Scholar] [CrossRef]
  35. Kim, J.; Han, D.; Tai, Y.W.; Kim, J. Salient Region Detection via High-Dimensional Color Transform and Local Spatial Support. IEEE Trans. Image Process. 2016, 25, 9–23. [Google Scholar] [CrossRef] [PubMed]
  36. Yuan, Y.C.; Li, C.Y.; Kim, J.; Cai, W.D.; Feng, D.D. Reversion Correction and Regularized Random Walk Ranking for Saliency Detection. IEEE Trans. Image Process. 2018, 27, 1311–1322. [Google Scholar] [CrossRef]
  37. Peng, H.W.; Li, B.; Ling, H.B.; Hu, W.M.; Xiong, W.H.; Maybank, S.J. Salient Object Detection via Structured Matrix Decomposition. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 818–832. [Google Scholar] [CrossRef]
  38. Li, C.Y.; Yuan, Y.C.; Cai, W.D.; Xia, Y.; Feng, D.D. Robust Saliency Detection via Regularized Random Walks Ranking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 2710–2717. [Google Scholar]
  39. Hou, Q.B.; Cheng, M.M.; Hu, X.W.; Borji, A.; Tu, Z.W.; Torr, P.H.S. Deeply Supervised Salient Object Detection with Short Connections. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 815–828. [Google Scholar] [CrossRef] [PubMed]
  40. Hu, X.W.; Zhu, L.; Qin, J.; Fu, C.W.; Heng, P.A. Recurrently Aggregating Deep Features for Salient Object Detection. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence/30th Innovative Applications of Artificial Intelligence Conference/8th AAAI Symposium on Educational Advances in Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 6943–6950. [Google Scholar] [CrossRef]
  41. Zhao, J.X.; Liu, J.J.; Fan, D.P.; Cao, Y.; Yang, J.F.; Cheng, M.M. EGNet: Edge Guidance Network for Salient Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8778–8787. [Google Scholar] [CrossRef]
  42. Xu, B.W.; Liang, H.R.; Liang, R.H.; Chen, P. Locate Globally, Segment Locally: A Progressive Architecture with Knowledge Review Network for Salient Object Detection. Proc. AAAI Conf. Artif. Intell. 2021, 35, 3004–3012. [Google Scholar] [CrossRef]
  43. Li, J.; Pan, Z.; Liu, Q.; Wang, Z. Stacked U-shape network with channel-wise attention for salient object detection. IEEE Trans. Multimed. 2020, 23, 1397–1409. [Google Scholar] [CrossRef]
  44. Pang, Y.W.; Zhao, X.Q.; Zhang, L.H.; Lu, H.C. Multi-scale Interactive Network for Salient Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 9410–9419. [Google Scholar] [CrossRef]
  45. Qin, X.B.; Zhang, Z.C.; Huang, C.Y.; Dehghan, M.; Zaiane, O.R.; Jagersand, M. U2-Net: Going deeper with nested U-structure for salient object detection. Pattern Recognit. 2020, 106, 107404. [Google Scholar] [CrossRef]
  46. Liu, Z.M.; Zhao, D.P.; Shi, Z.W.; Jiang, Z.G. Unsupervised Saliency Model with Color Markov Chain for Oil Tank Detection. Remote Sens. 2019, 11, 1089. [Google Scholar] [CrossRef]
Figure 1. The images are the original image (Optical RSI), ground truth (Ground Truth), and the segmentation results of our method (Ours).
Figure 1. The images are the original image (Optical RSI), ground truth (Ground Truth), and the segmentation results of our method (Ours).
Remotesensing 17 02053 g001
Figure 2. Multi-path feature extraction of Tic-Tac-Toe structure.
Figure 2. Multi-path feature extraction of Tic-Tac-Toe structure.
Remotesensing 17 02053 g002
Figure 3. The figure shows the SAEM-module, IEAM-Net architecture, including the SCAAP (Spatial-Channel Attention Adaptive Pooling) module with its three sub-modules (CAAS, SAAS, APES), up-sampling, SAG-module, and feature fusion modules. It combines left-to-right and right-to-left paths with adaptive max-pooling to produce accurate feature fusion results. GT denotes manually labeled ground truth data, rather than any form of model-generated output.
Figure 3. The figure shows the SAEM-module, IEAM-Net architecture, including the SCAAP (Spatial-Channel Attention Adaptive Pooling) module with its three sub-modules (CAAS, SAAS, APES), up-sampling, SAG-module, and feature fusion modules. It combines left-to-right and right-to-left paths with adaptive max-pooling to produce accurate feature fusion results. GT denotes manually labeled ground truth data, rather than any form of model-generated output.
Remotesensing 17 02053 g003
Figure 4. Illustration of the structure of Spatially Adaptive Edge Embedded Module (SAEM).
Figure 4. Illustration of the structure of Spatially Adaptive Edge Embedded Module (SAEM).
Remotesensing 17 02053 g004
Figure 5. Illustration of the structure of edge-embedded attention.
Figure 5. Illustration of the structure of edge-embedded attention.
Remotesensing 17 02053 g005
Figure 6. The figure shows the structure of the SCAAP module, which includes three sub-modules: the Spatial Attentive Adjustment Sub-module (SAAS), Channel Attentive Adjustment Sub-module (CAAS), and Adaptive Pooling Execution Sub-module (APES). It also demonstrates the bidirectional interaction mechanism between CAAS and SAAS to optimize feature weights.
Figure 6. The figure shows the structure of the SCAAP module, which includes three sub-modules: the Spatial Attentive Adjustment Sub-module (SAAS), Channel Attentive Adjustment Sub-module (CAAS), and Adaptive Pooling Execution Sub-module (APES). It also demonstrates the bidirectional interaction mechanism between CAAS and SAAS to optimize feature weights.
Remotesensing 17 02053 g006
Figure 7. The figure shows the structure of the Adaptive Pooling Execution Submodule (APES).
Figure 7. The figure shows the structure of the Adaptive Pooling Execution Submodule (APES).
Remotesensing 17 02053 g007
Figure 8. Illustration of the structure of SAG Module.
Figure 8. Illustration of the structure of SAG Module.
Remotesensing 17 02053 g008
Figure 9. The characteristic map of SAG-Model output.
Figure 9. The characteristic map of SAG-Model output.
Remotesensing 17 02053 g009
Figure 10. The figure presents a performance comparison of IEAM-Net with 21 advanced Salient Object Detection (SOD) methods on the EORSSD [7] and ORSSD [8] datasets. The upper part shows the performance of different methods on the Precision-Recall curves, while the lower part compares the Threshold-F-Score curves.
Figure 10. The figure presents a performance comparison of IEAM-Net with 21 advanced Salient Object Detection (SOD) methods on the EORSSD [7] and ORSSD [8] datasets. The upper part shows the performance of different methods on the Precision-Recall curves, while the lower part compares the Threshold-F-Score curves.
Remotesensing 17 02053 g010
Figure 11. IEAM-net versus 21 state-of-the-art SOD methods for speed and computation. The figure shows the comparison of IEAM-Net with 21 state-of-the-art saliency object detection (SOD-RRB-methods in terms of speed (left panel) and computation (right panel). The bar on the left shows the inference speed (in FPS) for each method, and the bar on the right shows the floating-point FLOPs (in billions) for each method.
Figure 11. IEAM-net versus 21 state-of-the-art SOD methods for speed and computation. The figure shows the comparison of IEAM-Net with 21 state-of-the-art saliency object detection (SOD-RRB-methods in terms of speed (left panel) and computation (right panel). The bar on the left shows the inference speed (in FPS) for each method, and the bar on the right shows the floating-point FLOPs (in billions) for each method.
Remotesensing 17 02053 g011
Figure 12. Comparison of target detection results between IEAM-Net and Other SOD methods. The figure shows the detection results of IEAM-Net and other salient object detection methods (Gate-Net, EMF-Net, PA-KRN, U2Net, DAF-Net, ITSD, and RCRR) in different scenarios. Each row displays the original image (Optical RSI), ground truth (GT), the output of IEAM-Net (Ours), and the outputs of other methods.
Figure 12. Comparison of target detection results between IEAM-Net and Other SOD methods. The figure shows the detection results of IEAM-Net and other salient object detection methods (Gate-Net, EMF-Net, PA-KRN, U2Net, DAF-Net, ITSD, and RCRR) in different scenarios. Each row displays the original image (Optical RSI), ground truth (GT), the output of IEAM-Net (Ours), and the outputs of other methods.
Remotesensing 17 02053 g012
Figure 13. Comparison of adaptive pooling and max pooling feature maps in IEAM-Net.
Figure 13. Comparison of adaptive pooling and max pooling feature maps in IEAM-Net.
Remotesensing 17 02053 g013
Table 1. Qualitative interpretation of standard evaluation metrics in saliency detection.
Table 1. Qualitative interpretation of standard evaluation metrics in saliency detection.
MetricWhat It MeasuresPractical Interpretation
F β max Trade-off between precision and recallHigher score means fewer missed objects and reduced false alarms. Even a 0.005 gain indicates noticeable visual improvement in object completeness and fewer background activations.
S-measureStructural similarity of prediction vs. GTCaptures how well object shapes and contours are preserved, important for elongated/irregular targets.
E-measureEnhanced alignment of prediction and GTMeasures consistency in spatial and holistic saliency. A higher E-measure implies better edge connectivity and less spatial fragmentation.
MAEPixel-wise average errorIndicates how “clean” the prediction is—lower values reflect smoother backgrounds and fewer fuzzy edges.
IoUSpatial overlap with ground truthReflects how well the predicted and true salient regions match, critical for spatial accuracy.
Table 2. This table presents a comparative performance analysis of IEAM-Net with 21 advanced SOD (Salient Object Detection) methods across four categories on the EORSSD [7] and ORSSD [8] datasets. The symbols ↑ and ↓ indicate that a larger or smaller score is better, respectively.
Table 2. This table presents a comparative performance analysis of IEAM-Net with 21 advanced SOD (Salient Object Detection) methods across four categories on the EORSSD [7] and ORSSD [8] datasets. The symbols ↑ and ↓ indicate that a larger or smaller score is better, respectively.
MethodsEORSSDORSSD
S α F β max F β mean F β adp E ξ max E ξ mean E ξ adp M S α F β max F β mean F β adp E ξ max E ξ mean E ξ adp M
RRWR0.59940.39930.36860.33440.68940.59430.56390.16770.68350.55900.51250.48740.76490.70170.69490.1324
HDCT0.59780.54070.40180.26580.78610.63760.51920.10880.61970.52570.42350.37220.77190.64950.62910.1309
SMD0.71060.58840.54750.40810.76920.72860.64160.07710.76400.66920.62140.55680.82300.77450.76820.0715
RCRR0.60070.39950.36850.33470.68820.59460.56360.16440.68490.55910.51260.48760.76510.70210.69500.1277
DSS0.78680.68490.58010.45970.91860.76310.69330.01860.82620.74670.69620.62060.88600.83620.80850.0363
RADF0.81790.74460.65820.49330.91300.85670.71620.01680.82590.76190.68560.57300.91300.82980.76780.0382
R3Net0.81840.74980.63120.41650.94830.82940.64620.01710.81410.74560.73860.73790.89130.86810.88870.0399
EGNet0.86010.78800.69670.53790.95700.87750.75660.01200.87210.83320.75000.64520.97310.90130.82260.0216
PoolNet0.82070.75450.64060.46130.92920.81930.68360.02100.84030.77060.69990.61660.93430.86500.81240.0358
GCPA0.88690.83470.79050.67210.95240.91670.86470.01020.90260.86870.84330.78610.95090.93410.92050.0168
ITSD0.90500.85230.82710.74210.95560.94070.91030.01060.90500.87350.85020.80680.96010.94820.93350.0165
MINet0.90400.83440.81740.77050.94420.93460.92430.00930.90400.87610.85740.82510.95450.94540.94230.0144
GateNet0.91140.85660.82240.71090.96100.93850.89090.00950.91860.88710.86790.82290.96640.95380.94270.0137
U2Net0.91990.87320.83290.72210.96490.93730.89890.00760.91620.87380.84920.80380.95320.93870.93260.0166
PAKRN0.91920.86390.83580.79930.96160.95360.94160.01040.92390.88900.87270.85480.96800.96200.95790.0139
SUCA0.89880.82290.79490.72600.95200.92770.90820.00970.89890.84840.82370.77480.95840.94000.91940.0145
CMC0.57980.32680.26920.20070.68030.58940.48900.10570.60330.39130.34540.31080.70640.64170.59960.1267
SMFF0.54010.51760.29920.20830.77440.51970.50140.14340.53120.44170.26840.24960.74020.49200.56760.1854
LVNet0.86300.77940.73280.62840.92540.88010.84450.01460.88150.82630.79950.75060.94560.92590.91950.0207
EMFINet0.92900.87200.85080.79840.97110.96040.95010.00840.93660.90020.95040.86170.97370.96710.96540.0109
DAFNet0.91660.86140.78450.64270.98610.92910.84460.00600.91910.89280.85110.78760.97710.95390.93600.0113
Ours0.93080.89050.84730.80270.97370.96250.95630.00710.94210.91350.88560.89470.97830.97420.96630.0098
Table 3. Comprehensive comparison of IEAM-Net and 21 competing methods on the EORSSD dataset. Metrics include detection accuracy ( F β max ), inference speed (FPS), model size (in millions of parameters), and computational complexity (FLOPs in G). The symbols ↑ and ↓ indicate that a larger or smaller score is better, respectively.
Table 3. Comprehensive comparison of IEAM-Net and 21 competing methods on the EORSSD dataset. Metrics include detection accuracy ( F β max ), inference speed (FPS), model size (in millions of parameters), and computational complexity (FLOPs in G). The symbols ↑ and ↓ indicate that a larger or smaller score is better, respectively.
MethodType F β max FPS ↑Params (M) ↓FLOPs (G) ↓
RRWRTraditional0.39935
UCFTraditional0.45217
RBDTraditional0.50109
DSSNSI-CNN0.68492262.2130.8
U2NetNSI-CNN0.71803044.798.5
PoolNetNSI-CNN0.75332268.1150.0
EGNetNSI-CNN0.78802094.5180.6
CPDNSI-CNN0.79252847.2104.3
SCRNNSI-CNN0.80102738.785.2
MINetNSI-CNN0.81272560.3145.1
ITSDNSI-CNN0.81942454.2131.9
F3NetNSI-CNN0.82102552.0122.4
GCPANSI-CNN0.83472486.7291.9
PA-KRNNSI-CNN0.863918138.3617.7
DAF-NetRSI-CNN0.86142585.4376.2
DMRARSI-CNN0.85872078.1330.5
EMFI-NetRSI-CNN0.872023107.3487.3
BLNetRSI-CNN0.87042690.2450.6
RRA-NetRSI-CNN0.86731995.0420.7
CGANetRSI-CNN0.87352193.4412.0
BMNetRSI-CNN0.86902481.3389.6
IEAM-Net (Ours)RSI-CNN0.89054867.7103.2
Table 4. This table shows the impact of different module combinations (Baseline, SAEM, SAG, AP, MP) on the performance of IEAM-Net. Fmax β and Emax ξ represent the maximum F-score and maximum E-score on the EORSSD [7] and ORSSD [8] datasets, respectively, with arrows indicating that higher values are better.
Table 4. This table shows the impact of different module combinations (Baseline, SAEM, SAG, AP, MP) on the performance of IEAM-Net. Fmax β and Emax ξ represent the maximum F-score and maximum E-score on the EORSSD [7] and ORSSD [8] datasets, respectively, with arrows indicating that higher values are better.
No.BaselineSAEMSAGAPMPEORSSDORSSD
F β max E ξ max F β max E ξ max
1 0.86320.96840.88780.9608
2 0.88170.97130.90430.9721
3 0.88360.97270.90660.9742
4 0.88600.96890.90860.9762
5 0.88420.97130.90470.9718
6 0.87400.96230.90170.9689
7 0.88350.97130.91260.9695
8 0.88720.97210.91120.9723
9 0.88910.97020.91270.9768
100.89050.97370.91350.9783
Table 5. This table compares the performance of the EAFFM (ours) and FFM models on the EORSSD [7] and ORSSD [8] datasets. Fmax β and Emax ξ represent the maximum F-score and maximum E-score, respectively, with arrows indicating that higher values are better.
Table 5. This table compares the performance of the EAFFM (ours) and FFM models on the EORSSD [7] and ORSSD [8] datasets. Fmax β and Emax ξ represent the maximum F-score and maximum E-score, respectively, with arrows indicating that higher values are better.
ModelsEORSSDORSSD
F β max E ξ max F β max E ξ max
FFM0.88690.97170.91180.9754
EAFFM0.89050.97370.91350.9783
Table 6. This table compares the performance of the DEEP SUP (ours) and SINGLE SUP models on the EORSSD [7] and ORSSD [8] datasets. Fmax β and Emax ξ represent the maximum F-score and maximum E-score, respectively, with arrows indicating that higher values are better.
Table 6. This table compares the performance of the DEEP SUP (ours) and SINGLE SUP models on the EORSSD [7] and ORSSD [8] datasets. Fmax β and Emax ξ represent the maximum F-score and maximum E-score, respectively, with arrows indicating that higher values are better.
ModelsEORSSDORSSD
F β max E ξ max F β max E ξ max
SINGLE SUP0.88750.97200.91210.9769
DEEP SUP0.89050.97370.91350.9783
Table 7. Performance comparison of IEAM-Net with different backbone networks. The symbols ↑ and ↓ indicate that a larger or smaller score is better, respectively.
Table 7. Performance comparison of IEAM-Net with different backbone networks. The symbols ↑ and ↓ indicate that a larger or smaller score is better, respectively.
Backbone F β max MAE ↓FPS ↑Params (M) ↓
VGG-16 (default)0.89050.0314867.7
ResNet-500.90720.0284285.2
EfficientNet-B00.89210.0304759.3
Table 8. Cross-dataset evaluation of IEAM-Net. The model is trained on ORSSD/EORSSD and directly tested on unseen datasets to assess generalization. The symbols ↑ and ↓ indicate that a larger or smaller score is better, respectively.
Table 8. Cross-dataset evaluation of IEAM-Net. The model is trained on ORSSD/EORSSD and directly tested on unseen datasets to assess generalization. The symbols ↑ and ↓ indicate that a larger or smaller score is better, respectively.
Dataset F max MAE ↓IoU ↑
ORSSD0.89050.0310.829
EORSSD0.89320.0280.842
iSOD-RS0.85110.0410.774
WHU-RS190.80730.0570.701
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, F.; Zhang, Z. IEAM: Integrating Edge Enhancement and Attention Mechanism with Multi-Path Complementary Features for Salient Object Detection in Remote Sensing Images. Remote Sens. 2025, 17, 2053. https://doi.org/10.3390/rs17122053

AMA Style

Zhang F, Zhang Z. IEAM: Integrating Edge Enhancement and Attention Mechanism with Multi-Path Complementary Features for Salient Object Detection in Remote Sensing Images. Remote Sensing. 2025; 17(12):2053. https://doi.org/10.3390/rs17122053

Chicago/Turabian Style

Zhang, Fubin, and Zichi Zhang. 2025. "IEAM: Integrating Edge Enhancement and Attention Mechanism with Multi-Path Complementary Features for Salient Object Detection in Remote Sensing Images" Remote Sensing 17, no. 12: 2053. https://doi.org/10.3390/rs17122053

APA Style

Zhang, F., & Zhang, Z. (2025). IEAM: Integrating Edge Enhancement and Attention Mechanism with Multi-Path Complementary Features for Salient Object Detection in Remote Sensing Images. Remote Sensing, 17(12), 2053. https://doi.org/10.3390/rs17122053

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop