Next Article in Journal
Delving Deep into Reverse Engineering of UEFI Firmwares via Human Interface Infrastructure
Previous Article in Journal
Bluetooth 5.0 Suitability Assessment for Emergency Response within Fire Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Adaptive Multi-Content Complementary Network for Salient Object Detection

Key Laboratory of Network and Information Security, Hebei Provincial Engineering Research Center for Supply Chain Big Data Analytics and Data Security, College of Computer and Cyber Security, Hebei Normal University, Shijiazhuang 050024, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(22), 4600; https://doi.org/10.3390/electronics12224600
Submission received: 22 September 2023 / Revised: 4 November 2023 / Accepted: 8 November 2023 / Published: 10 November 2023
(This article belongs to the Section Artificial Intelligence)

Abstract

:
Deep learning methods for salient object detection (SOD) have been studied actively and promisingly. However, the existing methods mainly focus on the decoding process and ignore the differences in the contributions of different encoder blocks. To address this problem for SOD, we propose an adaptive multi-content complementary network (PASNet) for salient object detection which aims to exploit the valuable contextual information in the encoder fully. Unlike existing CNN-based methods, we adopt the pyramidal visual transformer (PVTv2) as the backbone network to learn global and local representations with its self-attention mechanism. Then, we follow the coarse-to-fine strategy and introduce two novel modules, including an advanced semantic fusion module (ASFM) and a self-refinement module (SRM). Among these, the ASFM takes local branches and adjacent branches as inputs and collects semantic and location information of salient objects from high-level features to generate an initial coarse saliency map. The coarse saliency map serves as the location guidance for low-level features, and the SRM is applied to capture detailed information disguised in low-level features. We expand the location information with high-level semantics from top to bottom across the salient region, which is effectively fused with detailed information through feature modulation. The model effectively suppresses noises in the features and significantly improves their expressive capabilities. To verify the effectiveness of our PASNet, we conducted extensive experiments on five challenging datasets, and the results show that the proposed model is superior to some of the current state-of-the-art methods under different evaluation metrics.

1. Introduction

Salient object detection aims to imitate the human visual system and rapidly identify the most attention-grabbing objects or areas [1], followed by precise segmentation. The research on SOD has made extensive progress over the past decade. Significant object detection is crucial in computer vision, serving various purposes such as object recognition [2], visual tracking [3], person re-identification [4], image retargeting [5], image retrieval, etc. In the beginning, saliency object detection methods used hand-crafted features. However, this approach often fails to capture high-level semantic information, resulting in poor performance when dealing with complex scenes or images with complex textures and semantic contexts. Additionally, traditional methods could not adapt to new scenes and objects and had poor generalization abilities. As deep learning advances, traditional methods for salient object detection based on hand-designed features are being replaced by deep detection learning methods. In particular, the application of deep learning techniques such as convolutional neural networks (CNNs) [6] has further advanced the development of salient object detection and greatly facilitated research in computer vision.
In the era of deep learning, several convolutional neural network (CNN)-based salient object detection methods [7,8,9,10,11] have emerged, which significantly improve detection accuracy. The introduction of deep learning techniques has led to significant breakthroughs in the performance and efficiency of salient object detection, providing powerful tools and methods for solving real-world vision tasks. These methods almost adopt the typical structure of fully convolutional networks (FCNs [7]) which is called “Encoder–Decoder”. The encoder is responsible for using a series of convolutions to extract the multi-level features. The decoder has to take advantage of the different characteristics of multi-level features to output the final segmentation result. The encoder–decoder structure has precipitated the classical U-shaped structure in continuous development [8,12,13]. The U-shaped structure can extract multi-level feature information through top-down paths and utilize the feature information to construct high-quality feature maps. The classical network structure is often accompanied by many ingenious strategies. For example, MINet [14] adopts a lower up/down sampling rate to integrate features from adjacent levels, yielding more efficient multi-scale features from the integrated features. GateNet [15] employs multi-level gate units to balance the contribution of individual encoder blocks to the overall network, all the while suppressing unimportant feature regions. AGNet [16] introduces an attention mechanism to provide attention-guided information to the network from global semantic and local contextual perspectives, and gradually refines the prediction results by mining object details through attention and inverse attention mechanisms. MCCNet [17] comprehensively integrates foreground, background, and global information to deal with salient objects in complex contexts and generate good saliency maps. These methods are proposed based on an in-depth study of the features of natural images in different scenes and achieve excellent performance. However, the current methods use convolutional operations with local perceptual fields for feature extraction, which may not capture global contextual relevance for some scenarios that rely on semantic information for salient object detection. The cross-layer connection in the U-shape structure connects the encoder features directly to the decoder, which does not fully consider the relationship between the features of each layer and the information provided and does not mine the information adequately. The lack of processing of interfering information between the encoder and the decoder introduces misleading contextual information into the decoder, resulting in the underutilization of genuinely useful features in each layer.
Driven by the above observation, we propose a new framework for detecting saliency objects that differs from current convolutional neural network-based methods. Inspired by [18], we use the Pyramid Vision Transformer (PVTv2) as an encoder. Due to its powerful modeling of global dependencies, global contextual information can be fully captured for more robust features. Among these features, the low-level features always have a relatively higher resolution but less semantic information than high-level features [19]. Therefore, it is necessary to face different layer features using different strategies to explore them. Furthermore, we believe that, in addition to the foreground, the background [20] and adjacent features [21] are also important for integrating global image information, considering that the top layer of the encoder has access to a larger sensory field and a powerful representation that contains rich semantic information. Therefore, the advanced semantic fusion module (ASFM) is designed for the top layer of the encoder to comprehensively integrate foreground, background, and adjacent feature information, and to generate coarse saliency maps to provide positional guidance for the low-level features. Most of the studies have come to a consensus that low-level features usually enrich our prediction models by providing some important information about the local details of the final prediction results. Similar to [16,20,22], we design the self-refinement module (SRM) at layers of the encoder other than the top layer. The attention mechanism is utilized to mine the key details of salient objectives from multiple dimensions, gradually enriching the coarse saliency map and eventually generating an accurate saliency map.
The encoder has different layers, and each contains specific information. The corresponding modules are designed based on this information, and a feature interaction strategy is introduced between adjacent modules [21]. This helps to comprehensively explore the contextual information of adjacent features, thus expanding the coverage of inter-feature interactions. The top-down progressive path transfers high-level semantic information to the pyramid features at all levels. This compensates for the U-shaped network’s weakness, which loses strength as it passes information from top to bottom. Additionally, we introduce a dense lightweight refinement decoder (DLRB [22]) to fuse various features using feature modulation. The fused feature maps are fed into the decoder to be convolved by three extensions with different expansion rates to capture information at different scales. This provides a comprehensive view of important objects from multiple scales, expands the perceptual range of the entire network, and comprehensively captures multi-scale features. The decoder has a 1 × 1 convolution after each dilated convolution, which is used to minimize the aliasing effect caused by the dilated convolution. This helps to merge the captured features and creates a robust feature representation in the decoding stage.
In summary, the contributions to this paper include the following:
(1)
Replacing the encoder network with a pure transformer backbone called Pyramid Vision Transformer (PVTv2) to provide both global and local contextual branches for more reliable and robust features.
(2)
To design the advanced semantic fusion module (ASFM) at the top layer of the encoder to gather semantic cues from previous and present branches, sharpen object details, and create coarse saliency maps for location guidance. This improves the network’s ability to obtain global information.
(3)
The self-refinement module (SRM) is proposed to acquire detailed information from different dimensions by fusing adjacent branches through the attention mechanism. The information is combined with location guidance data from ASFM using feature modulation to obtain a more accurate saliency map.
The rest of this paper is organized as follows: In Section 2, we briefly review some previous work related to SOD. In Section 3, the proposed PASNet method is described in detail. In Section 4, we perform several experiments and discuss experimental comparisons and ablation analysis to validate the proposed model. Finally, the conclusions of this paper are given in Section 5.

2. Related Work

Early salient object detection methods relied heavily on manually crafted features for saliency calculations [6,23,24,25]. Although hand-crafted features can allow traditional salient object detection methods to perform detection in real time, the drawbacks of these methods limit their ability to perform salient object detection in complex scenes. In recent years, the convolutional neural network (CNN) [26] has achieved great success in the field of salient object detection (SOD) and has been dominating SOD research due to its superior feature representation capabilities. The excellent performance and capabilities of CNNs have made them the tool of choice in saliency detection tasks. Currently, most CNN-based work uses pre-trained image classification models for saliency detection, such as VGG [27] and ResNet [28] as encoders. These methods focus on building efficient decoders by integrating multi-level features [29,30]. However, since the model structure of CNNs is characterized by modeling for local information aggregation, it is difficult to model long periods and cannot capture global contextual information well for some scenarios that rely on global semantic information for saliency object detection. To overcome the problem that the CNN model structure is limited to extracting local information while ignoring global information, many CNN-based studies have made a series of improvements to further enhance the performance. For example, Zhang et al. [31] proposed integrating multi-level features at multiple resolutions and then further integrating these features at each resolution for saliency prediction. Finally, these saliency predictions are fused to obtain the final saliency map. Cheng et al. [30] solved the salient object detection problem by extending the pooling part of the convolutional neural network. Qin et al. [32] proposed an innovative “predict–refine architecture” with a new hybrid loss function to improve the boundary detection of salient objects. This approach not only emphasizes accurate prediction but also focuses on the boundary information of the object, thus improving the accuracy and quality of saliency detection. Differently, Zhao et al. [33] extracted salient edge information from ground truth and used it to purposefully guide the neural network to learn edge features for one-to-one guidance. Yu et al. [34] proposed to the use of either a large convolution kernel or a dilated convolution (atrous convolution) to increase the receptive field. The network’s receptive field can be directly expanded by a large convolution kernel. The dilated convolution increases the receptive field by introducing voids (intervals) in the convolution operation to better capture long-period information. Furthermore, in [26], Liu and his research team investigated local and global attention mechanisms that object the pixel level for more efficient detection. Chen et al. [20] successfully incorporated background information to improve detection performance by introducing a reverse attention mechanism. Prajit Ramachandran et al. [35] proposed an independent self-attentive layer and a spatially aware independent self-attentive layer to replace the convolution. The spatially aware independent self-attentive layer adds some position-related information to the independent self-attentive layer. However, these strategies still have limitations in global dependency modeling. We recently found that the Vision Transformer [36,37,38], with a complete sequence of remote dependencies, excels in global dependency modeling in image processing, which inspired us to delve deeper into how we can utilize the Transformer to improve SOD performance.
Originally proposed by researchers such as Vaswani [36], the Transformer is originally designed to be used in the field of natural language processing, especially for solving machine translation tasks. This innovative modeling architecture debuted in 2017, generating a great deal of interest in the field of computer science. The transformer network employs a self-attention mechanism (self-attention) to model long-term dependencies by alternately superimposing layers of multi-head self-attention (MHSA), which aims to enhance feature representations. Unlike convolutional layers, MHSA layers have dynamic weights and global perceptual ranges, and thus are more flexible and efficacious. In recent years, with the great success of Transformers in NLP, researchers have introduced transformers into the field of computer vision with satisfactory results. Dosovitskiy et al. [37] proposed a Vision Transformer (ViT), which is the first pure transformer model dedicated to image classification tasks. The ViT divides the incoming raw image into image blocks (patches) at regular intervals of a certain region size and then combines the divided image blocks into sequences. The sequences are fed into the encoder part of the original Transformer model, and finally the images are classified through the fully connected layer. Since then, a large number of studies on visual transformers [33,39,40] have emerged, which outperformed the current leading CNN models in terms of performance. Mao et al. [38] adopted the Swin Transformer [39] as an encoder and also designed a simple and effective decoder based on CNNs for predicting saliency maps. After years of development, the CNN backbone in computer vision has gradually formed some general network patterns, the most classic of which is the pyramid structure. Almost all dense prediction algorithms are designed based on the feature pyramid structure. To better handle intensive prediction tasks such as semantic segmentation, several approaches [15,39,41] have also introduced the pyramid structure of convolutional neural networks (CNNs) into the design of transformer backbone networks to improve the network performance and enrich the semantic information. For example, the PVT-based model [15,41] has been successfully applied with a hierarchical transformer structure, which contains four different stages, achieving excellent performance. It indicates that the pure transformer backbone network is comparable to the conventional convolutional neural network (CNN) in terms of generalization. When applied to detection and segmentation tasks with the same number of parameters, the pure transformer backbone network shows better performance compared to CNNs.
Although pure transformer backbone networks perform well in practical applications, the feature pyramid structure is favored for its superior feature extraction capability. This structure can fully exploit the contextual information in the network. Therefore, the feature pyramid structure is still widely used and continuously improved. With the rapid development of deep learning in the field of computer vision, salient object detection methods based on feature pyramid structures [13,42,43,44] have emerged and continue to improve their performance. These methods allow models to simultaneously consider features at different scales by effectively utilizing multi-scale information. This allows the network to obtain a fuller picture of the image content and capture salient objects of various shapes and sizes. Hou et al. [45] achieved efficient feature representation by continuously mixing deep features into shallow layers. Liu et al. [26] combined global and local pixel-level contextual attention networks to enable the model to simultaneously consider the entire image as well as detailed information from different regions. The network can capture global and local contextual information in the image in a more comprehensive and fine-grained way. Zhang et al. [46] introduced a progressively improved architecture with a knowledge review aimed at enhancing the effectiveness of object detection. This architecture is realized by gradually integrating multiple levels of contextual information. With the above approach, it is easy to find that contextual information has a crucial role in saliency objective detection (SOD). This information is not only present in a single feature layer but also well-represented between adjacent feature layers. The parallel use of convolutional layers with different convolutional kernels within a single feature layer is a popular strategy for capturing both local and global content. It helps to efficiently capture salient objects of variable sizes or uncertain numbers in natural images. In this work [21], the synergy of feature interaction strategies is introduced to harmonize the three scales of the current, preceding, and following features. In this way, the previous and the latter features provide global auxiliary information for the current feature. Meanwhile, the fusion of contextual information is further enhanced by using the convolution of different convolution kernels in parallel in different feature layers. Gongyang Li et al. [22,47] enable the model to comprehensively analyze image content at different abstraction levels and scales by using multiple strategies to deeply mine the information in high-level features and low-level features. It is well known that the features extracted from a feature extraction network can be categorized into low-level features and high-level features. The low-level features mainly contain the detail and texture information of the image, while the high-level features cover the semantic and location information of the image. Considering that the two contain different information, it is necessary to object the extraction of the information contained in the features, which is especially crucial.
Inspired by the above methods, PVTv2 [41] is used as the backbone network in our proposed PASNet method to overcome the problem that CNN models are difficult to model for a long time. Network construction is performed from a global perspective, and a new adaptive multi-content complementary network framework for salient object detection is designed. To solve the problems of inaccurate localization and the loss of detailed information in salient object detection, we use global information to localize salient objects. And we utilize the attention mechanism to obtain rich detail information so that the network pays more attention to the salient regions that contain rich detail information. To cope with the problem of insufficient information extraction from the encoder, we introduce a feature interaction strategy. This strategy aims to comprehensively mine the information in the encoder and employs multiple methods to explore different levels of features.

3. Proposed Salient Region Detection Model

In this section, we deeply explore the design of the PASNet. The overall architecture of the PASNet is shown in Figure 1, and we will detail its various components and key designs step by step. As seen in Figure 1, the network proposed in this paper consists of four main components: the pyramidal visual transformer encoder block (PVTE), the advanced semantic fusion module (ASFM), the self-refinement module (SRM), and the dense lightweight refinement decoding block (DLRB). In this section, we will first provide a detailed description of the general architecture of the model and then introduce each component of the network one by one.

3.1. Network Overview

Specifically, given an input image, I R H × W × 3 , we first extracted the four pyramid features using the backbone [41] based on the PVTv2 transformer, P V T E i R H 2 i + 1 × W 2 i + 1 × C i , C i 64 , 128 , 320 , 512 , and i 1 , 2 , 3 , 4 . Multi-scale long-period dependent features are extracted from input images. This method collects semantic cues and locates salient objects by progressively aggregating high-level features. Then, the pyramid features P V T E 3 and P V T E 4 , which contain more semantic information, are transferred to the advanced semantic fusion module ( ASFM ). The ASFM further expands the sensory field by fusing different levels of high-level features to collect global contextual information. This module can capture the exact location of salient objects more accurately and generate a coarse saliency map to provide location guidance for the fusion of the underlying features. Pyramid features with more detailed P V T E 1 , P V T E 2 , and P V T E 3 information interactions are fully exploited for texture and boundary information in the features through the self-refinement module ( S R M i ). Moreover, to prevent high-level semantic information from being diluted in a top-down path, we directly aggregate ASFM high-level features into the feature mapping of each feature layer to provide multi-scale information for decoders at all levels. Finally, the detailed features and high-level semantic features of the encoder are continuously fused via feature modulation. The DLRB decoder senses the multi-scale information and refines the incoming fused features to ensure that feature maps of different scales can be seamlessly fused and output.

3.2. Pyramid Vision Transformer Encoder

The backbone network plays a central role in the whole algorithmic process. The backbone network is mainly responsible for extracting features from the input image, which includes underlying texture, edge information, and high-level semantic information. The quality and efficiency of these feature extraction processes are critical to the performance of saliency object detection. The design and selection of the backbone network directly affect the effectiveness and accuracy of the overall saliency object detection network. Several recent studies [48,49] find visual transformers [15,41] to have stronger performance and robustness against interference than CNNs [27,28]. Inspired by these results, we choose to use visual transformers as our backbone network to extract robust and comprehensive feature information of salient objects in images. This backbone network helps to provide a more reliable basis and information foundation for the next salient object detection task. The visual transformer serves as a powerful deep-learning architecture with excellent feature extraction capabilities. It is able to better capture the semantic and structural information in the image, which is expected to further improve the performance and robustness of salient object detection. Unlike [14], which uses a fixed “columnar” structure or a shifted window approach, PVT [40] uses a pyramidal structure to compute the representation by reducing spatial attention operations, thus reducing resource consumption. It is worth mentioning that the backbone network we propose to model is interchangeable, and our network framework is equally applicable to other well-known transformer backbones.
Specifically, we use an improved version of PVT, PVTv2 [41], which makes up for three shortcomings of PVT: (1) the use of convolution to extract locally continuous features; (2) the overlapping block embedding with zero padding to extract location information encoding; and (3) the attention layer with average pooling and linear complexity. Thanks to the above three improvements, the resulting PVTv2 achieves significantly better performance than PVTv1 and has a more powerful feature extraction capability. To adapt PVTv2 to the saliency object detection task, we remove its last classification layer and generate four multi-scale feature maps ( P V T E 1 P V T E 4 ) at different stages, where ( P V T E 4 ) provides high-level semantic cues.

3.3. Advanced Semantic Fusion Module

In general, high-level features usually contain richer semantic information and are not limited to a single feature level but span across adjacent levels. Therefore, we use the features extracted from the last two levels of the encoder to collect semantic information, and the advanced semantic fusion module (ASFM) is designed for this purpose. Specifically, the module fully extracts high-level semantic information from different features by fusing current and previous cross-scale features. Figure 2 shows the internal structure of the module. The ASFM module has two branches, including the current branch and the previous branch, and the current branch ( P V T E 4 ) is noted as ( f p v t 4 ). The current branch undergoes three dilated convolutions [50] with dilation rates of 1, 3, and 5 to obtain the output features ( f d c 4 , i ) with different dilation rates. The output characteristics can expand the receptive field by expanding the convolution so that the model can better understand the global structure and the relationship between the objects in the image. Secondly, dilated convolution can help the model to capture the contextual information in the image so that the model can better understand the semantics of the objects. At the same time, dilation convolution can reduce the ambiguity generated during the convolution process so that the network can accurately locate the position of salient objects. These output features are then aggregated and fused through splicing convolution operations to obtain features with rich contextual cues ( f c 4 ). The aggregated fused features are used to reduce redundant information by obtaining feature ( f c a 4 ) through the channel attention mechanism [20], which can be formulated as follows:
f d c 4 , i = D C o n v σ ( f p v t 4 ; W 3 × 3 4 , i ; r i ) , i ( 1 , 3 , 5 ) ,
f c 4 = C o n v σ ( C o n c a t ( f d c 4 , 1 , f d c 4 , 3 , f d c 4 , 5 ) ; W 3 × 3 4 , i ) ,
f c a 4 = C A ( f c 4 ) f p v t 4 ,
where D C o n v σ ( ) is the dilated convolution, W 3 × 3 t , i is the parameters with a 3 × 3 kernel, r i is the dilation rate, C o n v σ ( ) is the convolution layer, C o n c a t ( ) is the splicing operation between the channel axes, C A ( ) is the channel attention, and is the channel-wise multiplication.
Then, the foreground map is obtained through spatial attention [20], notated as f s a 4 , as the foreground map can provide an effective guide for locating salient regions. Therefore, we employ foreground maps to emphasize salient areas at the feature level to make them more prominent, which can be expressed as follows:
f s a 4 = C o n v 3 × 3 ( S A ( f c a 4 ) f c a 4 ) ,
where c o n v 3 × 3 ( ) is the 3 × 3 convolution layer, S A ( ) is the spatial attention, and is the element-by-element multiplication.
Although the foreground image obtained through the attention mechanism can accurately localize the position and general outline of the salient object, it cannot completely extract the boundary and detail information of the salient object. This missing information can be mistaken for the background, so it is essential to extract the detail information from the background. Since the generation of the background map is closely related to the foreground map, it can be noted in reverse [20] that the background map is a b 4 . Adopting the prominent non-significant regions of the background map at the feature level, the foreground and background maps are spliced and fused to further mine richer and more significant information, obtaining the spliced and fused feature notated as f s b 4 . This process can be expressed as follows:
a b 4 = 1 S A ( f c a 4 ) , f b t = C o n v 3 × 3 ( a b 4 f c a 4 ) ,
f s b t = C o n v 3 × 3 ( C o n c a t ( f s a t , f b t ) ) ,
where is the element-wise subtraction, S A ( ) is the spatial attention, is the element-by-element multiplication, c o n v 3 × 3 ( ) is a 3 × 3 convolution layer, and C o n c a t ( ) is a splicing operation between channel axes.
The previous branch, P V T E 3 , is denoted as f p v t 3 . Since the features of the adjacent branches differ in scale from the current feature, the resulting output feature after the down-sampling, inflationary convolution, and spatial attention operations of the adjacent branches can be referred to as f p p 4 . The output feature provides global auxiliary information for the current branch and aligns the saliency region. After effective coordination, we fuse the output features of these two branches. This integrated feature is subsequently activated via feature selection to generate a coarse salient map, P r e d , for positional guidance. The details are as follows:
f p p 4 = S A ( D C o n v σ ( D o w n ( f p v t 3 ) ; W 3 × 3 3 , i ; r i ) , i ( 1 , 3 , 5 ) ) f c 3 ,
f A S F M 4 = f s b 4 f p p 4 , P r e d = s i g m o i d ( c o n v 1 × 1 ( f A S F M 4 ) ) ,
where S A ( ) is the spatial attention, is the element-by-element multiplication, D o w n ( ) is the down-sampling operation, is the sum of elements, s i g m i o d ( ) is the activation function, and c o n v 1 × 1 ( ) is the 1 × 1 convolutional layer.
With the advanced semantic fusion module (ASFM), each salient region can be activated comprehensively and accurately. It effectively fuses multi-level advanced semantic information and enhances the contrast between salient and non-salient regions to obtain an accurate location guidance map.

3.4. Self-Refinement Module

Typically, the shallow features of a network carry rich information about the local details in the image, which are crucial for the final prediction results. These shallow features include not only details such as texture, color, and edges, but also some local information that may be lost or blurred in deeper features. Therefore, utilizing shallow features can provide us with more comprehensive and precise information, which helps to improve the quality and accuracy of prediction results. To obtain robust segmentation results and fully integrate the detailed information, we employ a self-refinement module (SRM) in the network. This module is designed for encoder hierarchies other than the top encoder. Its main task is to collect rich detailed semantic cues at different scales to capture the detailed information of salient objects more comprehensively. With the introduction of the self-refinement module (SRM), we are able to effectively enhance the network’s perception of detailed information at all layers, thus improving the quality and accuracy of the segmentation results. The application of this strategy allows the network to better adapt to various complex scenes and salient object characteristics. Figure 3 shows the internal structure of this module. S R M i takes adjacent features as input, and makes the features at the same scale after a sampling operation. The detailed information of the two layers of features is fused through channel splicing and convolution operations to obtain the fused feature, f S R M i . The representation is as follows:
f S R M i = C o n v 3 × 3 ( P V T E i C o n c a t ( U p ( P V T E i + 1 ) ) C o n c a t ( D o w n ( P V T E i 1 ) ) ) ,   i ( 1 , 2 , 3 )
where c o n v 3 × 3 ( ) is the 3 × 3 convolutional layer, C o n c a t ( ) is the cross-channel concatenation, D o w n ( ) is the down-sampling operation, and U p ( ) is the up-sampling operation.
The fused feature maps are convolved to obtain an attention map focusing on salient regions. The nonsignificant regions often contain some information ignored by the significant regions. To address this challenge, a reverse attention map is generated by inverting the attention map to refocus on information that the network may have missed. This approach focuses attention on regions that are usually ignored to ensure that important details in the image are not missed. In this way, the information in the image is captured more comprehensively, improving the robustness and effectiveness of feature fusion. By multiplying and fusing the two attention maps with features f S R M i at the element level, the network is guided to dig deeper into saliency information, both in salient and non-salient regions [16]. It is well known that low-level features usually contain rich detail information such as texture, color, edges, etc. When recognizing significant object details, an efficient feature extractor is needed to fully capture and utilize these rich detail features. Therefore, we add the channel attention operation [51] and spatial attention operation [52] to the SRM to capture salient object details from different dimensions of the fused feature map, f S R M i , to obtain the spatial channel attention map, f s a c a i . Finally, these three results are fused via a convolution operation to obtain the output, f o u t i , of the S R M i . This process can be represented as follows:
f a t t i = c o n v 1 × 1 ( f S R M i ) , f a t t _ r i = 1 f a t t i ,
f s a c a i = S A ( C A ( f S R M i ) f S R M i ) f S R M i ,
f o u t i = c o n v 1 × 1 ( f s a c a i + C B R ( f a t t i f S R M i ) + C B R ( f a t t _ r i f S R M i ) ) ,
where c o n v 1 × 1 ( ) is the 1 × 1 convolution layer, S A ( ) / C A ( ) is the space/channel attention, is the channel-wise multiplication, is the element-by-element multiplication, and C B R is the convolution, normalization, and activation operations.
Finally, the output detailed feature map, f o u t i , from S R M i contains rich detail information, but the localization of salient objects is not accurate and requires accurate location information for guidance. At this point, the location guidance map generated by the ASFM and the detailed feature map f o u t i is harmonized by means of feature modulation to generate the final output, f o u t _ a l l i , and deliver it to the decoder. Feature modulation is a coarse saliency map generated by the ASFM that is up-sampled to generate the same size as each layer’s feature map. Then, the coarse saliency map is augmented with the layers’ feature maps by multiplying them element by element to enhance the saliency information. Finally, the information lost during the information enhancement process is supplemented by employing residual concatenation. The formula is as follows:
f o u t _ a l l i = U p ( p r e d ) f o u t i f o u t i ,
where U p ( ) is the up-sampling operation, is the element-by-element multiplication, and is the element-wise addition.
The self-refinement module ( S R M i ), which fuses adjacent branches, captures details in low-level features from different dimensions and fuses them with high-level semantic information produced by the ASFM through feature modulation, since high-level semantic information usually contains clues about important objects in the image. By fusing it with other features through feature modulation, the representation of low-level features is enhanced and made more distinguishable. Secondly, high-level semantic information helps in understanding the relationship between objects and their surroundings. Feature modulation fuses contextual information with other features, and the network can better understand the relationship between salient objects and the background, thus improving detection accuracy. In addition, high-level semantic information is usually more stable, so fusion with other levels of features can help to suppress noise in low-level features. This helps to improve the stability of salient object detection in complex scenes. Overall, fully fusing high-level semantic information with features from other layers through feature modulation can increase the network’s understanding and perception of saliency objects. This improves the performance of salient object detection, making it more accurate, stable, and robust. Through feature modulation, the PASNet is able to obtain richer and more diverse contextual information in pictures.

3.5. Dense Lightweight Refinement Block Decoder

The network goes through an advanced semantic module to generate a location guidance map, which fuses the global location guidance information with other feature maps at different levels. However, during the feature fusion process, we have to deal with some challenging situations that occur in images from different scales as well as in natural images, such as the coexistence of multiple objects and small objects. These situations can cause the salient objects in the fused image to appear to be aliased, and also increase the difficulty of information acquisition. Therefore, we are faced with the profound problem of how to effectively reduce the aliasing effect during the fusion process while capturing and utilizing the information in the images more comprehensively. This problem deserves in-depth thinking and research to improve the effect and performance of feature fusion. For this purpose, we introduce the dense lightweight refinement decoder (DLRB) [22], which takes the fused feature maps that have been processed by feature modulation as input. Each decoder has three depth-separable convolutions ( D S c o n v s ) with dilation rates of 2, 4, and 6, three 1 × 1 convolutional layers, and one 3 × 3 convolutional layer. The dilated D S c o n v s expands the perceptual field to fully capture multi-scale features. The output features of the DLRB module are denoted as f d l r b i , and we take f o u t _ a l l 1 as its input as an example. The formula is as follows:
f d l r b 1 , 1 = c o n v 1 × 1 ( D S c o n v 2 ( f o u t _ a l l 1 ) f o u t _ a l l 1 ) ,
f d l r b 1 , 2 = c o n v 1 × 1 ( D S c o n v 4 ( f d l r b 1 , 1 ) f d r l b 1 , 1 f o u t _ a l l 1 ) ,
f d l r b 1 , 3 = c o n v 3 × 3 ( c o n v 1 × 1 ( D S c o n v 6 ( f d l r b 1 , 2 ) f d l r b 1 , 2 f d l r b 1 , 1 f d l r b 1 , 1 f o u t _ a l l 1 ) ) ,
where is the element summation, c o n v 1 × 1 ( ) is the 1 × 1 convolutional layer, D S c o n v r ( ) is a dilation-separable convolution (where r is the dilation rate), and c o n v 3 × 3 ( ) is the 3 × 3 convolutional layer.
In this way, the DLRB is able to perceive multi-scale information, further expanding the perceptual range of the entire network. It also helps to reduce the aliasing effect caused by model up-sampling, and captures detailed information more comprehensively. In the refinement phase, the DLRB can provide a robust and fine-grained feature representation of the network.

3.6. Loss Function

To improve the training effectiveness of the PASNet, we adopt a comprehensive loss function that combines the classical binary cross-entropy (BCE) loss and the IoU (intersection over union) loss. This comprehensive loss function not only helps the network to learn the locations of saliency objects, but also motivates the network to better understand the boundaries and shapes of the objects, thus improving the effectiveness and performance of the training. The flexibility of this loss function design makes the network more adaptable to the needs of various saliency detection tasks. In addition, to further improve the effectiveness of training, we employ deep supervision [45,53] in the training phase. With deep supervision, we supervised the coarse salient maps and the fine salient maps generated during network refinement. It ensures that the network receives accurate guidance and feedback at different stages, thus improving the performance and robustness of saliency object detection. We denote the total loss function, L t o t a l , as follows:
L t o t a l = l b c e ( U p ( p r e d ) , G ) + l i o u ( U p ( p r e d ) , G ) ) ,
L t o t a l = l b c e ( U p ( S 3 ) , G ) + l i o u ( U p ( S 3 ) , G ) ) ,
where G is the truth map and l b c e ( ) and l i o u ( ) are the BCE loss and IOU loss, respectively.

4. Experiments

In this section, we provide an in-depth analysis of key performance factors in the field of salient object detection. First, we conduct an extensive performance evaluation, thoroughly testing five commonly used salient object detection datasets to measure the performance of different approaches. Using five classical evaluation metrics, we comprehensively compare the performance of various models, while also considering various factors such as the characteristics of the datasets, the choice of evaluation metrics, and implementation details. We are proud that our PASNet method significantly outperforms the current leading salient object detection methods in terms of qualitative performance. Our method not only achieves excellent performance on multiple datasets but also obtains impressive results on various evaluation metrics. This demonstrates the PASNet’s superior ability to capture salient objects. In addition, we conduct a series of ablation experiments to deeply investigate the effectiveness of each module within the PASNet method. These experiments reveal the contribution of each module to the overall performance, providing key insights for further optimization of the method. Our study not only provides an efficient solution for salient object detection but also provides valuable references and inspiration for future related research.

4.1. Datasets

We train and evaluate the proposed PASNet model and fifteen recent state-of-the-art salient object detection methods in terms of five benchmark datasets.
ECSSD: The dataset [54] contains 1000 images, including many structurally complex images such as natural landscapes, city streets, buildings, and people. The images also vary in resolution. Each image in the ECSSD dataset has a pixel-level saliency annotation to indicate the location of salient objects in the image.
HKU-IS: This dataset [55] has 4447 complex scenes with multiple disconnected salient objects, some of which have overlapping borders and similar fore-/background appearances. These images have different contents and scenes, including people, animals, landscapes, streets, etc. In addition, the images in the dataset have different resolutions, covering the range from low to high resolution. This variation in resolution allows researchers to evaluate and compare the performance of image segmentation algorithms in different contexts.
PASCAL-S: This dataset [56] includes 850 images which select from the segmentation dataset of PASCAL VOC 2010. It contains 20 common object categories, which is a widely used dataset for image recognition and object detection. The images in PASCAL-S are from different categories in the PASCAL VOC 2010 dataset, covering various scenes such as indoor, outdoor, natural scenery, and people. These images are mainly labeled based on the human eye motion dataset without obvious salient objects, which makes this dataset difficult to predict.
DUT-OMRON: This dataset [57] consists of 5168 challenging images, which is a large-scale dataset for salient object detection. This image dataset includes different indoor and outdoor environments with rich content and diverse scenes. Typically, each image includes one or more foreground objects which have a complex background.
DUTS: This dataset [58], with 10,553 training images in the DUT-TR training set and 5019 test images in the DUT-TE test set, is the largest salient object detection dataset. The goal of this dataset is to bring together images from different domains and scenes to provide a comprehensive and diverse resource for salient object detection. These images contain very complex scenes and are rich in content.

4.2. Evaluation Indicators and Implementation Details

(1)
Evaluation metrics: For quantitative evaluation, we use five commonly used metrics—precision–recall (PR) curves, MaxF-measure (MaxF) scores, E-measure scores, S-measure scores, and mean absolute error (MAE)—to visually compare the significance maps with the ground truth. Each metric provides a different dimensional performance measure to fully assess the effectiveness of the model. The selection of these metrics reflects the diversity and comprehensiveness of the evaluation and ensures a comprehensive assessment of the model’s performance in different domains and tasks.
Precision–recall (PR) curve: Precision, also referred to as the check rate, indicates the percentage of samples correctly predicted as positive cases to all samples predicted as positive cases by the model. Recall, also referred to as the check rate, represents the percentage of samples correctly predicted as positive cases out of all samples that are actually positive cases (ground truth). We binarized the salient map using variable thresholds from 0 to 255, and for each threshold, we compute the precision and recall of the binarized map versus the true graph, and thus plotted the PR curve. The closer the position of the PR curve is to the upper right corner, the better the performance of the model. This means that the model is able to maintain high precision while also having high recall.
MaxF-measure (MaxF) score: The MaxF score is an important performance metric that balances the values of inspection accuracy and inspection completeness by taking a reconciled average of the two metrics. It combines two key performance metrics, inspection accuracy and inspection completeness, to fully evaluate the performance of the model at different thresholds. We use MaxF-measure, denoted as MaxF, for various thresholds varying in the range from 0 to 255, calculated as follows:
M a x F = M a x ( ( I + β 2 ) × P r e c i s i o n × R e c a l l β 2 × P r e c i s i o n + R e c a l l )
where precision and recall are the precision score and recall score, respectively, and β 2 is fixed at 0.3 according to the suggestion of [59].
E-measure: This metric is introduced to more fully assess the similarity and improve the performance between two binary images. First, E-measure removes the effects of luminance and contrast by aligning the global means of the two binary images, ensuring that the comparison is fair. Then, the similarity between the two images is measured using a local pixel-level correlation that takes into account the luminance values of each pixel and its surrounding pixels. E-measure takes values between 0 and 1, where 0 means that the two binary images are not at all similar to each other, while 1 means that they are identical to each other.
E m = 1 W × H w = 1 W h = 1 H θ ( φ ) ,
where θ ( φ ) denotes the enhanced alignment matrix and H and W are the height and width of the saliency map.
S-measure: The structural measure is an assessment tool dedicated to measuring the structural similarity between a continuous-valued saliency map (CVSM) and the ground truth. It typically ranges from 0 to 1, where higher S-measure values indicate greater structural similarity between the saliency map and the ground truth, thus reflecting the efficacy of the test method. S o represents the object-perceived structural similarity, which considers the structural similarity between an object or objects and its background. In this metric, the comparative factors of the foreground and background are summed with certain weights to reflect the structural relationship between the object and its surroundings. S r involves region-aware structural similarity, which partitions the saliency map and ground-truth into multiple local regions and computes the structural similarity metrics for each region, and then weights and sums them. Thus, S m can be defined as
S m = α × S o + ( 1 α ) × S r ,
where α is the balance parameter, which is set to 0.5, as suggested by Fan et al. [60].
Mean absolute error (MAE): Mean absolute error (MAE) is an important metric used to measure the difference between the significance map (M) and the ground truth map (T). It takes the absolute value of the difference between each pixel point of the significance plot and the true value plot, and then adds these absolute values and averages them. The smaller the value of the MAE, the lower the error between the significance plot and the true value, further reflecting that the better the test method and the more effective its performance. It can be defined as
M A E = 1 W × H w = 1 W h = 1 H | M ( w , h ) T ( w , h ) | ,
where H and W are the height and width of the image, respectively.
(2)
Implementation details: We use the DUTS-TR dataset [58] as a training dataset during our research, which contains up to 10,553 images, providing a rich data resource for our training. To implement our proposed PASNet model and conduct related experiments, we rely on a personal computer equipped with RTX 2080 Ti GPUs and use the PyTorch2.0.1 [61] repository as the basis for an end-to-end implementation. During the training phase, we normalize the size of the input training samples to 256 × 256 pixels to ensure that the model is well adapted to images of different resolutions. Our training process employs an asymptotic learning rate strategy, where we first set an initial learning rate, 1 × 10−4, and then the model automatically reduces the learning rate to one-tenth of the original rate after 30 epochs, which helps to optimize the convergence of the model. Given the GPU memory limitations, we set a relatively small batch size of 6 to ensure the stability of training. We choose Adam as the optimization strategy [62], while setting the weight decay parameter to 0.0005 to prevent model overfitting. To improve the robustness of the model, we also introduced data enhancement techniques. These tricks include random horizontal flipping, random rotation to increase the diversity of viewpoints, and random adjustments of brightness, saturation, and contrast to simulate different lighting and image conditions. These strategies make our model more robust and better able to cope with salient object detection tasks in a variety of real-world scenes.

4.3. Performance Comparison with State-of-the-Art Technology

Our proposed PASNet method is compared with 16 state-of-the-art salient object detection methods, including Amulet [31], C2SNet [63], PAGR [64], AFNet [65], BASNet [32], CPD [19], SCRN [66], PoolNet [30], EGNet [33], MINet [14], GateNet [15], DNA [67], SAMNet [68], SCWSSOD [69], UMNet [70], and Polyp-PVT [18]. Since the datasets on which the Polyp-PVT network is trained and tested are mostly medical images, for a fair comparison, we trained the Polyp-PVT network on its own training dataset and subsequently performed significance prediction. For a fair comparison, the PASCAL-S dataset in the UMNet [70] method was retrained using the original code project published by the authors. The saliency maps for all other methods were used as provided directly by their respective authors. If the results of both types of backbones (VGG and ResNet) were present in the methods, the results of the ResNet backbone training were used.
(1)
Quantitative Evaluation
To implement a comprehensive quantitative assessment, we show the scores of the four evaluation metrics, MaxF-measure, E-measure, S-measure, and MAE, and the inference speeds and FLOPs of several models in Table 1. In addition to evaluating the models from different perspectives, we show the precision–recall curves and F-measure curves in Figure 4. These comparisons enable us to comprehensively assess the performance of the model and analyze its performance in all aspects in detail. The first row in Figure 4 shows the PR curves for different datasets, while the second row shows the Fm curves for different datasets. The closer the position of the PR curve is to the upper right corner, the better the performance of the model. This means that the model is able to maintain high precision while also having high recall. The Fm curves being closer to the upper right corner indicates that the model is able to achieve highly accurate classification performance with high robustness under various thresholds. In Figure 4a, we present the PR curve and Fm curve for the DUT-OMRON dataset. The DUT-OMRON dataset is a large-scale dataset containing 5168 images. The image backgrounds in this dataset are more complex and thus have higher requirements for the accuracy of the detection models. The PR curves and Fm curves (red solid lines) of our method on this dataset are significantly ahead of the second-ranked EGNet [33] method, showing excellent accuracy. The PR and Fm curves for the DUTS-TE and ECSSD datasets are shown in Figure 4b,c, respectively. The DUTS-TE dataset covers both indoor and outdoor scenes, different lighting conditions, and diverse object types. The ECSSD dataset includes both natural objects (e.g., people, animals, plants, etc.) and artificial objects (e.g., buildings, signs, transportation, etc.). These two datasets cover a variety of real-life scenarios. The PR curves and Fm curves (red solid lines) of our method in these two datasets are still higher than other methods. This fully demonstrates the strong generalization ability of our proposed PASNet model. In Figure 4d, we show the PR curves and Fm curves of the HKU-IS dataset. The HKU-IS dataset contains 4447 images with multiple disconnected objects. Some of the images in this dataset have overlapping boundaries and the salient objects are similar to the background. It is challenging to perform saliency detection on the HKU-IS dataset. Through Figure 4d, it is obvious that our method outperforms other models in terms of both PR curves and Fm curves. This fully demonstrates the excellent performance of our method in processing complex images. The PR curves and Fm curves of the PASCAL-S dataset are exhibited in Figure 4e. The images in the PASCAL-S dataset are labeled mainly based on human eye movements, with no obvious salient objects. This makes the prediction of this dataset more difficult. Although the PR curves of our method are not optimal on this dataset, it still maintains a high performance. It is worth mentioning that the Fm curves of our method on this dataset are still ahead of other methods, which further demonstrates the competitiveness of our method. The experimental results in Figure 4 adequately demonstrate the robustness and generalizability of our model to achieve excellent performance across different datasets and scenarios. Overall, our PASNet method performs well on multiple datasets, proving its effectiveness and superior performance in the field of salient object detection. These results provide strong support for the broad potential and feasibility of our method in practical applications.
To present a clearer picture of the comparative performance of different methods, we provide quantitative measurements of the various methods in Table 1. These measurements provide us with a comprehensive perspective that helps us to evaluate and compare the effectiveness of various methods in the task of saliency detection. From Table 1, we can draw conclusions that are consistent with the PR curves and the Fm curves. Unlike other methods that fuse multi-scale features or utilize post-processing, EGNet [33] focuses on the complementarity between significant edge information and significant object information, which makes its measurements on the three datasets of DUT-OMRON, DUTS-TE, and HKU-IS. It is significantly more able than other methods to reach a higher level. Our proposed PASNet model achieves a 4.7-times-faster inference speed than the EGNet model (42.22 vs. 8.92), while the FLOPs of the PASNet model are 30 times higher than the EGNet model (9.72 vs. 291.83). In addition, our proposed method, the PASNet, obtains the best performance under all quantitative metrics on the four datasets, DUT-OMRON, ECSSD, HKU-IS, and DUTS-TE. The PASNet has a large performance gain compared to other methods. In addition, our proposed PASNet method achieves the best performance under all quantitative metrics on four datasets, DUT-OMRON, ECSSD, HKU-IS, and DUTS-TE datasets. It has a large performance gain compared with other methods. Specifically, the best performance was obtained for the challenging DUT-OMRON (0.8338 vs. 0.8181, 0.8553 vs. 0.8412, 0.8971 vs. 0.8737, and 0.0492 vs. 0.0520) and HKU-IS (0.9408 vs. 0.9352, 0.9234 vs. 0.9179, 0.9632 vs. 0.9564, and 0.0259 vs. 0.0309) datasets. The PASNet is clearly ahead of the second-ranked method in terms of the MaxF-measure, E-measure, S-measure, and MAE. This remarkable result is not only an affirmation of our method’s performance but also a clear demonstration of its strong adaptability in dealing with a wide range of complex visual scenes and tasks. In the cross-domain, multi-scenario DUTS-TE (0.8893 vs. 0.8882, 0.9257 vs. 0.9077, 0.8877 vs. 0.8865, 0.0364 vs. 0.0375) and ECSSD (0.8630 vs. 0.8591, 0.8690 vs. 0.8616, 0.8417 vs. 0.8364, and 0.1164 vs. 0.1219) datasets, our method outperforms the second-ranked method, which further demonstrates the generalization ability of the PASNet in different scenarios. To summarize, the excellence and effectiveness of this methodology are clearly observed through the experimental results, which further consolidate its potential and importance in applications.
(2)
Qualitative Evaluation
To demonstrate the merits of our proposed method in a more comprehensive and detailed manner, we conduct a series of visual comparison experiments to meticulously contrast it with six representative state-of-the-art methods in challenging natural scenarios. Such comparisons not only help to highlight the excellence of our method but also help us to gain a deeper understanding of its potential and superiority in practical applications. As shown in Figure 5, some representative images and various detection aspects are included, including complex scenes, low-contrast scenes, multiple objects, complex backgrounds, and small objects. The method proposed in this paper is advantageous and competitive in the following aspects.
(a)
Localization accuracy advantage. The method can accurately locate prominent objects and has a good ability to suppress background interference. In the first row of Figure 5, the image background is more cluttered. In the sixth row of Figure 5, the color of the salient object puppy is very similar to the color of the letter sign in the background. Some advanced SOD methods (e.g., Amulet [31] and UMNet [70]) are difficult to use to detect effective salient regions. The saliency detection algorithms that focus on edges, including BASNet [32], cannot accurately detect salient objects. In contrast, our method can detect salient objects accurately and comprehensively.
(b)
Advantages in the interference resistance. In the second and third rows of the images in Figure 5, the complexity and clutter of the background reach extremely high water. This leads to the fact that some saliency detection methods will inevitably detect some background information along with the salient objects (e.g., Amulet [31], BASNet [32], PoolNet [30], DNA [67], GateNet [15], UMNet [70]). Our method can accurately detect salient objects without background interference.
(c)
Advantages in content integrity. In the seventh row of Figure 5, the building has many fine structural details and relatively cluttered building branches. In the eighth row of Figure 5, the foreground is indistinguishable from the background and the foreground is extremely complex. Due to the increased difficulty of detection, these methods can roughly locate all significant objects, but cannot guarantee the structural integrity of the detected objects. For example, the results of the PoolNet [30] and DNA [67] methods in detecting the eighth row of Figure 5 show that the edges of the butterfly’s wings are missing. In contrast, our method achieves better performance in terms of both structural integrity and detail integrity.
(d)
Advantages of challenging solutions. Small objects in large scenes usually have lower resolution and smaller visual features, which increases the difficulty of saliency detection. Multi-object scenes, on the other hand, involve multiple salient objects and their mutual interference and overlap. In these cases, our algorithm is able to accurately recognize and segment small objects and effectively handle the complex relationships among multiple objects. This means that our method still performs well in the face of challenging visual environments and tasks, and has the ability to cope with both small-object scenes and multi-objective scenes. For example, in the last two rows of Figure 5, our method accurately detects salient objects and obtains accurate segmentation results. Other methods also fail to clearly and completely detect small objects in large scenes and even incorrectly retain background regions (e.g., BASNet [32] and DNA [67]). Similarly, in the fourth and fifth rows of images, our method accurately detects multiple salient objects, while other methods barely detect all salient objects completely. The experimental results clearly show that our PASNet method has higher accuracy in salient object detection. Compared with other state-of-the-art methods, the PASNet is able to distinguish the detailed features of salient objects in different scenes more clearly, further improving the quality and accuracy of saliency detection. At the same time, these results demonstrate the superiority of our method in providing a more accurate solution for a variety of vision applications, thus bringing significant improvements and innovations to both research and practical applications.

4.4. Ablation Experiment

(1) Quantitative Evaluation: There are three key modules in our PASNet, including the advanced semantic fusion module (ASFM), the self-refinement module (SRM), and the dense lightweight refined block decoder (DLRB). To validate the effectiveness of the PASNet methodology, we perform a large number of ablation experiments. The main purpose of these experiments is to evaluate the necessity of each module and its contribution to the overall performance. The results of these experiments are essential to ensure the reliability and practicality of our method. We use the network consisting of the PVTv2 encoder and the non-densely connected decoder as a baseline and use the same experimental parameter settings throughout the experiments. As shown in Table 2, we quantitatively analyze the contribution of each module to the whole network. Ablation experiments are performed on four classical datasets, with red data indicating optimality. It is worth noting that while our baseline model performs the worst in all three metrics, it still performs excellently compared to many of the methods in Table 1. To fully integrate the multi-scale information provided by the decoder to enrich and expand the saliency mapping, we introduce the DLRB. This improvement significantly improves the network performance. Taking the HKU-IS dataset as an example, the images in this dataset contain multiple objects, and the boundaries of these objects overlap. After replacing the decoder in the baseline with the DLRB, the MaxF-measure metric improves from 0.9362 to 0.9394, and the S-measure metric increases from 0.9195 to 0.9210, while the MAE metric decreases from 0.0283 to 0.0269. To encode detailed information in feature learning, we design SRMs for the network and bring performance improvements. For example, in the cross-domain, multi-scene ECSSD dataset, the MaxF-measure improves from 0.8586 to 0.8627, the S-measure index rises from 0.8359 to 0.8404, and the MAE metric decreases from 0.1197 to 0.1190. In addition, to accurately localize the location of significant objects and solve the problem of object scale variation in different scenes, we design the ASFM. This module enables the network to obtain more comprehensive semantic information. With the addition of the ASFM to the network, we have a complete PASNet structure. Compared to the other variants, the complete PASNet structure achieves optimal values for all metrics on the four datasets. For example, on the SOD dataset relative to the baseline model, the percentage gain of the MaxF-measure metric reaches 1.33%, the percentage gain of the MAE reaches 0.82%, and the percentage gain of the S-measure metric reaches 1.79%. This shows that our proposed framework benefits from all key components. In summary, it is evident that all the indexes are significantly improved when the PVT encoder, SRM, ASFM, and DLRB are present at the same time, resulting in the best performance and the strongest generalization ability.
To verify the generality of each module in the PASNet approach and the superior performance of the PVTv2 backbone, we replace the PASNet backbone with VGG16 and ResNet50. The degree of generalization and contribution is verified by introducing different modules one by one. As shown in Table 3, we use the network consisting of the ResNet50 [28] backbone and the non-densely connected decoders as the baseline. The DLRB, SRM, and ASFM are added to the baseline network step by step. We perform ablation experiments on four classical datasets, where the red data indicate the optimum. As shown in Table 4, we quantitatively analyze the contribution of each module in the whole network using the network consisting of the VGG16 [27] backbone and the non-densely connected decoders as the baseline. The same four classical datasets are still used for the ablation experiments, as in the previous experiments, where the red data indicate the optimum. It is evident from Table 2, Table 3 and Table 4 that the performance of the network model continues to improve with the continuous introduction of individual modules. In addition, the optimal values of three different backbone networks are compared in Table 5, where the data in red indicate the optimal. The PVTv2 [41] backbone network outperforms the other backbone networks in all metrics in the four datasets. For example, the optimal value of the PVTv2 network architecture in the MaxF-measure metric for the HKU-IS dataset is 0.9408. It is 2.69% higher than the optimal value of the ResNet50 network architecture and 3.75% higher than the optimal value of the VGG16 network architecture. The optimal value of the PVTv2 network architecture in the MAE metric for the DUTS-TE dataset is 0.0364. It is 1.6% higher than the optimal value of the ResNet50 network architecture and 2.54% higher than the optimal value of the VGG16 network architecture. The above ablation experiments demonstrate the superior performance of the PASNet architecture and the effectiveness and generalization of the modules.
(2) Qualitative Evaluation: In order to give the reader a more intuitive feeling of the contribution of different modules, a qualitative analysis of the network is shown in Figure 6. The detection effect is poor when the network is only a baseline network with a combination of the PVTv2 encoder and non-densely connected decoders. For example, in the first row of Figure 6, the background color is similar to the foreground of a single large object, and in the second row of images with multiple objects, the detection of the baseline is not good. The background in the third row of images in Figure 6 is extremely similar to the salient objects, which disrupts object detection. At this point, the baseline network fails to correctly distinguish between the foreground and the background. In the fourth row of images, where the background is more complex, the baseline network only partially detects the salient objects. After we replace the non-densely connected decoder with the DLRB, we can see from the first and second rows of Figure 6 that the network’s detection of single large objects and multiple objects is significantly improved. However, at this stage, the network is still not sufficient to mine detailed information and semantic information. With the addition of the SRM module, as seen in the fourth and fifth rows of Figure 6, the network’s effectiveness in detecting objects with more complex backgrounds and containing more details is significantly improved. It can better suppress the background information of small objects, such as the eagle in the sixth row of Figure 6. Nevertheless, it still struggles to correctly distinguish the foreground from the background when detecting images with strong background interference factors (such as the goose in the third row of Figure 6). Finally, the semantic information in the high-level encoder is fully fused by further introducing the ASFM to further suppress the background (such as the goose in the third row of Figure 6) and extract a more complete object structure (such as the butterfly in the fourth row of Figure 6). In addition, the network is suitable for single-object (such as the wine glass in the first row of Figure 6) and multi-object detection (such as the bottle in the second row of Figure 6). In summary, as each module gradually joins the network, we can more intuitively feel the contribution of each module. The synergistic cooperation between modules makes the detection effect of the whole network model significantly improved.

4.5. Failure Cases and Areas of Application

(1)
Failure Cases: We discuss some failure cases of the PASNet in Figure 7. For some very challenging examples, our approach still does not yield perfect results.
(1) It is still a challenge to effectively distinguish from the surrounding environment for salient objects that are extremely similar to the background, for example, in the first row of Figure 7, the yellow cobra in the desert and the small bird in the reeds. The salient objects are largely distracted by the texture and color information of the background, causing the detector to fail to detect the salient objects completely and accurately. To solve this problem, we need to try to develop better learning strategies to understand the abstract semantic information in the saliency detection task. A robust learning strategy can make the network more resistant to interference.
(2) Handling multiple small objects in a large cluttered context is still a challenge for our approach. For example, the mallard below the left image cannot be detected in the second row of Figure 7. In addition, the wildebeest above the image is not detected in its entirety. The three small birds on the tree trunk in the right image can only be detected as one of them accompanied by a lot of background information. The main reason for this is that we are limited by the GPU performance and only use a relatively small input size. As a result, most of the features of multiple small salient objects are lost during down-sampling, hindering the effectiveness of the detection. We can solve this problem by using larger images as inputs and utilizing more computational resources for training. However, it is more desirable to explore more efficient global context-aware attention operations and construct more robust global dependencies.
Figure 7. Some examples of failures. (a) Natural image. (b) Ground truth. (c) Significance plot of PASNet detection. (d) Natural image. (e) Ground truth. (f) Significance plot of PASNet detection.
Figure 7. Some examples of failures. (a) Natural image. (b) Ground truth. (c) Significance plot of PASNet detection. (d) Natural image. (e) Ground truth. (f) Significance plot of PASNet detection.
Electronics 12 04600 g007
(2)
Areas of Application: The improvement of the PASNet model has important applications and real-world scenario value in the field of saliency object detection, which is shown in the following aspects:
(1) Computer vision tasks: saliency object detection is a fundamental task in the field of computer vision and is of key significance for other vision tasks. For example, in tasks such as object tracking, object segmentation, automatic driving, and robot navigation, saliency object detection can be used to determine the object of attention, thus improving the accuracy and robustness of the task.
(2) Advertising and recommender systems: In advertising and content recommendation, saliency object detection can be used to identify elements that may be of interest to users, such as products, advertisements, or recommended content. This can help to improve the click-through rate of advertisements and the attractiveness of content recommendations.
(3) Medical image processing: In medical image analysis, saliency object detection helps doctors to identify and localize disease foci, abnormalities, or important structures. This is useful for medical diagnosis, treatment planning, and surgical navigation.
(4) Military and security applications: In the military field, salient object detection can be used to identify hostile objects or potential threats. In security applications, it can help surveillance systems to better detect abnormal behavior or significant objects.
(5) Intelligent transportation systems: In automated driving and intelligent traffic management, salient object detection can be used to detect important objects on the road, such as other vehicles, pedestrians, and traffic signs. This is essential for improving traffic safety and the performance of autonomous driving systems.
(6) Remote sensing image processing: In remote sensing images, saliency object detection can be used to identify important locations or objects in a specific geographic area, such as urban planning, resource management, and environmental monitoring.
In summary, the improvement of the PASNet model has a wide range of applications and practical value in saliency object detection, covering multiple domains and real-world scenarios. The improved model can increase the accuracy, efficiency, and applicability of saliency object detection. This helps in exploring its application for technology development in various fields.

5. Conclusions

In recent years, although some progress has been made in salient object detection, there are still some challenging issues and problems that need to be further researched and solved. In this paper, we find that the saliency object detection algorithm based on a convolutional neural network is modeled for local information aggregation and lacks remote interaction between pixels. Therefore, we propose an adaptive multi-content complementary network for saliency object detection called the PASNet, which focuses on the Vision Transformer-based saliency object detection algorithm. To obtain long-range dependencies between pixels, the pyramidal Vision Transformer is used as an encoder (PVTv2) to learn global and local representations. The advanced semantic fusion module (ASFM) at the top layer of the encoder can be used to acquire a larger perceptual field and stronger representation containing rich global contextual information. We design a self-refinement module (SRM), which is designed to obtain detailed information about the low-level features from multiple dimensions. The effective contextual information in the encoder is introduced into the dense lightweight refinement decoder (DLRB) to sense multi-scale information that can improve the discriminability of the whole network. In addition, we employ a feature interaction strategy to acquire the contextual information of adjacent features. Through extensive experiments on five challenging datasets, we clearly find that the proposed model significantly outperforms the current state-of-the-art methods on four evaluation metrics without any preprocessing. These experimental results solidly corroborate the excellence of our approach. Although the PASNet method proposed in this paper achieves good results on natural image datasets, there is still much room for improvement in edge detection. Next, we will further optimize the edge detection effect of our algorithm to obtain more complete and finer prediction results. The code and results of our method are available at https://github.com/GKDLX/PASNet (accessed on 21 September 2023).

Author Contributions

Conceptualization, K.G.; formal analysis, W.W.; supervision, L.H.; writing—original draft, K.G.; writing—review and editing, L.H. and W.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 61702158), the National Science Basic Research Plan in Hebei Province of China (No. F2018205137), the Educational Commission of Hebei Province of China (No. ZD2020317), and the Central Guidance on Local Science and Technology Development Fund of Hebei Province (226Z1808G, 236Z0102G).

Data Availability Statement

The datasets used and/or analyzed during the current study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cong, R.; Lei, J.; Fu, H.; Cheng, M.M.; Lin, W.; Huang, Q. Review of visual saliency detection with comprehensive information. IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 2941–2959. [Google Scholar] [CrossRef]
  2. Rutishauser, U.; Walther, D.; Koch, C.; Perona, P. Is bottom-up attention useful for object recognition? In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 27 June–2 July 2004; IEEE: Piscataway, NJ, USA, 2004; pp. 37–44. [Google Scholar]
  3. Mahadevan, V.; Vasconcelos, N. Saliency-based discriminant tracking. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 1007–1013. [Google Scholar]
  4. Zhao, R.; Ouyang, W.; Wang, X. Unsupervised salience learning for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 3586–3593. [Google Scholar]
  5. Fang, Y.; Chen, Z.; Lin, W.; Lin, C.W. Saliency detection in the compressed domain for adaptive image retargeting. IEEE Trans. Image Process. 2012, 21, 3888–3901. [Google Scholar] [CrossRef] [PubMed]
  6. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef]
  7. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  8. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015, Munich, Germany, 5–9 October 2015; Springer: Munich, Germany, 2015; pp. 234–241. [Google Scholar]
  9. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  10. Zhang, L.; Dai, J.; Lu, H.; He, Y.; Wang, G. A bi-directional message passing model for salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–23 June 2018; pp. 1741–1750. [Google Scholar]
  11. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  12. Weng, W.; Zhu, X. INet: Convolutional networks for biomedical image segmentation. IEEE Access 2021, 9, 16591–16603. [Google Scholar] [CrossRef]
  13. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Honolulu, HI, USA, 2017; pp. 2117–2125. [Google Scholar]
  14. Pang, Y.; Zhao, X.; Zhang, L.; Lu, H. Multi-scale interactive network for salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 9413–9422. [Google Scholar]
  15. Zhao, X.; Pang, Y.; Zhang, L.; Lu, H.; Zhang, L. Suppress and balance: A simple gated network for salient object detection. In Proceedings of the Computer Vision–ECCV 2020, Glasgow, UK, 23–28 August 2020; Springer: Glasgow, UK, 2020; pp. 35–51. [Google Scholar]
  16. Lin, Y.; Sun, H.; Liu, N.; Bian, Y.; Cen, J.; Zhou, H. Attention guided network for salient object detection in optical remote sensing images. In Proceedings of the International Conference on Artificial Neural Networks, Bristol, UK, 6–9 September 2022; Springer: Cham, Switzerland, 2022; pp. 25–36. [Google Scholar]
  17. Li, G.; Liu, Z.; Lin, W.; Ling, H. Multi-content complementation network for salient object detection in optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–13. [Google Scholar] [CrossRef]
  18. Dong, B.; Wang, W.; Fan, D.; Li, J.; Fu, H.; Shao, L. Polyp-pvt: Polyp segmentation with pyramid vision transformers. arXiv 2021, arXiv:210806932. [Google Scholar] [CrossRef]
  19. Wu, Z.; Su, L.; Huang, Q. Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–19 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 3907–3916. [Google Scholar]
  20. Chen, S.; Tan, X.; Wang, B.; Lu, H.; Hu, X.; Fu, Y. Reverse attention-based residual network for salient object detection. IEEE Trans. Image Process. 2020, 29, 3763–3776. [Google Scholar] [CrossRef]
  21. Li, G.; Liu, Z.; Zeng, D.; Lin, W.; Ling, H. Adjacent Context Coordination Network for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans Cybern 2023, 53, 526–538. [Google Scholar] [CrossRef]
  22. Li, G.; Liu, Z.; Bai, Z.; Lin, W.; Ling, H. Lightweight Salient Object Detection in Optical Remote Sensing Images via Feature Correlation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5617712. [Google Scholar]
  23. Klein, D.A.; Frintrop, S. Center-surround divergence of feature statistics for salient object detection. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 2214–2219. [Google Scholar]
  24. Cheng, M.-M.; Mitra, N.J.; Huang, X.; Torr, P.H.; Hu, S.M. Global contrast based salient region detection. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 569–582. [Google Scholar] [CrossRef] [PubMed]
  25. Liu, T.; Yuan, Z.; Sun, J.; Wang, J.; Zheng, N.; Tang, X.; Shum, H.Y. Learning to detect a salient object. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 353–367. [Google Scholar]
  26. Liu, N.; Han, J.; Yang, M.-H. Picanet: Learning pixel-wise contextual attention for saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–23 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 3089–3098. [Google Scholar]
  27. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:14091556. [Google Scholar]
  28. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 770–778. [Google Scholar]
  29. Qiu, Y.; Liu, Y.; Ma, X.; Liu, L.; Gao, H.; Xu, J. Revisiting multi-level feature fusion: A simple yet effective network for salient object detection. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 4010–4014. [Google Scholar]
  30. Liu, J.-J.; Hou, Q.; Cheng, M.-M.; Feng, J.; Jiang, J. A simple pooling-based design for real-time salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 3917–3926. [Google Scholar]
  31. Zhang, P.; Wang, D.; Lu, H.; Wang, H.; Ruan, X. Amulet: Aggregating multi-level convolutional features for salient object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 202–211. [Google Scholar]
  32. Qin, X.; Zhang, Z.; Huang, C.; Gao, C.; Dehghan, M.; Jagersand, M. Basnet: Boundary-aware salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision And Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 7479–7489. [Google Scholar]
  33. Zhao, J.-X.; Liu, J.-J.; Fan, D.-P.; Cao, Y.; Yang, J.; Cheng, M.M. EGNet: Edge guidance network for salient object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 8779–8788. [Google Scholar]
  34. Yu, F.; Koltun, V.; Funkhouser, T. Dilated residual networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 472–480. [Google Scholar]
  35. Ramachandran, P.; Parmar, N.; Vaswani, A.; Bello, I.; Levskaya, A.; Shlens, J. Stand-alone self-attention in vision models. Adv. Neural Inf. Process. Syst. 2019, 32, 68–80. [Google Scholar]
  36. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
  37. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16×16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:201011929. [Google Scholar]
  38. Mao, Y.; Zhang, J.; Wan, Z.; Dai, Y.; Li, A.; Lv, Y.; Tian, X.; Fan, D.P.; Barnes, N.L. Transformer Transforms Salient Object Detection and Camouflaged Object Detection. arXiv 2021, arXiv:2104.10127. [Google Scholar]
  39. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 10012–10022. [Google Scholar]
  40. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 568–578. [Google Scholar]
  41. Wang, W.H.; Xie, E.Z.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. PVT v2: Improved baselines with Pyramid Vision Transformer. Comput. Vis. Media 2022, 8, 415–424. [Google Scholar] [CrossRef]
  42. Li, Y.; Chen, Y.; Wang, N.; Zhang, Z. Scale-aware trident networks for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 6054–6063. [Google Scholar]
  43. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–23 June 2018; pp. 8759–8768. [Google Scholar]
  44. Fan, R.; Cheng, M.M.; Hou, Q.; Mu, T.J.; Wang, J.; Hu, S.M. S4Net: Single Stage Salient-Instance Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6103–6112. [Google Scholar]
  45. Hou, Q.; Cheng, M.-M.; Hu, X.; Borji, A.; Tu, Z.; Torr, P.H. Deeply supervised salient object detection with short connections. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 815–828. [Google Scholar]
  46. Xu, B.; Liang, H.; Liang, R.; Chen, P. Locate globally, segment locally: A progressive architecture with knowledge review network for salient object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; AAAI: Palo Alto, CA, USA, 2021; pp. 3004–3012. [Google Scholar]
  47. Li, G.; Liu, Z.; Bai, Z.; Lin, W. Lightweight Salient Object Detection in Optical Remote-Sensing Images via Semantic Matching and Edge Alignment. IEEE Trans. Geosci. Remote Sens. 2022, 61, 5601111. [Google Scholar]
  48. Bhojanapalli, S.; Chakrabarti, A.; Glasner, D.; Li, D.; Unterthiner, T.; Veit, A. Understanding robustness of transformers for image classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 10231–10241. [Google Scholar]
  49. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  50. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:151107122. [Google Scholar]
  51. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  52. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  53. Xie, S.; Tu, Z. Holistically-nested edge detection. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1395–1403. [Google Scholar]
  54. Yan, Q.; Xu, L.; Shi, J.; Jia, J. Hierarchical saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 1155–1162. [Google Scholar]
  55. Li, G.; Yu, Y. Visual saliency based on multiscale deep features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 5455–5463. [Google Scholar]
  56. Li, Y.; Hou, X.; Koch, C.; Rehg, J.M.; Yuille, A.L. The secrets of salient object segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 280–287. [Google Scholar]
  57. Yang, C.; Zhang, L.; Lu, H.; Ruan, X.; Yang, M.H. Saliency detection via graph-based manifold ranking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 3166–3173. [Google Scholar]
  58. Wang, L.; Lu, H.; Wang, Y.; Feng, M.; Wang, D.; Yin, B.; Ruan, X. Learning to detect salient objects with image-level supervision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 136–145. [Google Scholar]
  59. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-tuned salient region detection. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 1597–1604. [Google Scholar]
  60. Fan, D.-P.; Cheng, M.-M.; Liu, Y.; Li, T.; Borji, A. Structure-measure: A new way to evaluate foreground maps. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4548–4557. [Google Scholar]
  61. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. Adv. Neural Inf. Process. Syst. 2019, 32, 8024–8035. [Google Scholar]
  62. Wei, J.; Wang, S.; Huang, Q. F3Net: Fusion, feedback and focus for salient object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York Midtown, NY, USA, 7–12 February 2020; AAAI: Palo Alto, CA, USA, 2020; pp. 12321–12328. [Google Scholar]
  63. Li, X.; Yang, F.; Cheng, H.; Liu, W.; Shen, D. Contour knowledge transfer for salient object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; ECCV: Berlin, Germany, 2018; pp. 370–385. [Google Scholar]
  64. Zhang, X.; Wang, T.; Qi, J.; Lu, H.; Wang, G. Progressive attention guided recurrent network for salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–23 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 714–722. [Google Scholar]
  65. Feng, M.; Lu, H.; Ding, E. Attentive feedback network for boundary-aware salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1623–1632. [Google Scholar]
  66. Wu, Z.; Su, L.; Huang, Q. Stacked cross refinement network for edge-aware salient object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 7264–7273. [Google Scholar]
  67. Liu, Y.; Cheng, M.M.; Zhang, X.Y.; Nie, G.Y.; Wang, M. DNA: Deeply Supervised Nonlinear Aggregation for Salient Object Detection. IEEE Trans. Cybern. 2022, 52, 6131–6142. [Google Scholar] [CrossRef] [PubMed]
  68. Liu, Y.; Zhang, X.Y.; Bian, J.W.; Zhang, L.; Cheng, M.M. SAMNet: Stereoscopically Attentive Multi-Scale Network for Lightweight Salient Object Detection. IEEE Trans. Image Process. 2021, 30, 3804–3814. [Google Scholar] [CrossRef] [PubMed]
  69. Yu, S.; Zhang, B.; Xiao, J.; Lim, E.G. Structure-consistent weakly supervised salient object detection with local saliency coherence. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; AAAI: Palo Alto, CA, USA, 2021; pp. 3234–3242. [Google Scholar]
  70. Wang, Y.; Zhang, W.; Wang, L.; Liu, T.; Lu, H. Multi-source uncertainty mining for deep unsupervised saliency detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 11727–11736. [Google Scholar]
Figure 1. The overall architecture of the PASNet. It consists of a PVT encoder ( P V T E 1 P V T E 4 ), an advanced semantic fusion module (ASFM), three self-refinement modules ( S R M 1 S R M 3 ), and three dense lightweight refinement decoders ( D L R B 3 D L R B 1 ).
Figure 1. The overall architecture of the PASNet. It consists of a PVT encoder ( P V T E 1 P V T E 4 ), an advanced semantic fusion module (ASFM), three self-refinement modules ( S R M 1 S R M 3 ), and three dense lightweight refinement decoders ( D L R B 3 D L R B 1 ).
Electronics 12 04600 g001
Figure 2. Detailed description of the advanced semantic fusion module (ASFM). It consists of a current branch and a previous branch, where the current branch takes P V T E 4 as input and the previous branch takes P V T E 3 as input. The two branches are operated in series and elements are summed and fed into a 1 × 1 convolutional layer, and finally, feature activation is performed.
Figure 2. Detailed description of the advanced semantic fusion module (ASFM). It consists of a current branch and a previous branch, where the current branch takes P V T E 4 as input and the previous branch takes P V T E 3 as input. The two branches are operated in series and elements are summed and fed into a 1 × 1 convolutional layer, and finally, feature activation is performed.
Electronics 12 04600 g002
Figure 3. Detailed description of the self-refinement module ( S R M i ). It contains channel/space attention, inverse attention, one 3 × 3 convolution, two 1 × 1 convolutions, and two CBR operations to obtain saliency object detail information from different dimensions.
Figure 3. Detailed description of the self-refinement module ( S R M i ). It contains channel/space attention, inverse attention, one 3 × 3 convolution, two 1 × 1 convolutions, and two CBR operations to obtain saliency object detail information from different dimensions.
Electronics 12 04600 g003
Figure 4. Comparison of PR and F-measure curves of the PASNet (red curve) with existing methods on five datasets. (ae) shows the PR curves and F-measure curves of different methods on DUT-OMRON, DUT-TE, ECSSD, HKU-IS, and PASCAL-S datasets.
Figure 4. Comparison of PR and F-measure curves of the PASNet (red curve) with existing methods on five datasets. (ae) shows the PR curves and F-measure curves of different methods on DUT-OMRON, DUT-TE, ECSSD, HKU-IS, and PASCAL-S datasets.
Electronics 12 04600 g004
Figure 5. Visual comparison of different methods. Each row shows the saliency mapping of an image. Each column represents an algorithm. As is clear in the figure, our method finds salient objects more accurately than other state-of-the-art methods and clearly distinguishes the boundaries of salient objects.
Figure 5. Visual comparison of different methods. Each row shows the saliency mapping of an image. Each column represents an algorithm. As is clear in the figure, our method finds salient objects more accurately than other state-of-the-art methods and clearly distinguishes the boundaries of salient objects.
Electronics 12 04600 g005
Figure 6. Comparison of predictions between different networks. (a) Image; (b) ground truth; (c) baseline; (d) PVTE + DLRB; (e) PVTE + SRM + DLRB; (f) PVTE + SRM + ASFM + DLRB.
Figure 6. Comparison of predictions between different networks. (a) Image; (b) ground truth; (c) baseline; (d) PVTE + DLRB; (e) PVTE + SRM + DLRB; (f) PVTE + SRM + ASFM + DLRB.
Electronics 12 04600 g006
Table 1. Quantitative comparisons include Max F-measure, E-measure, S-measure, and MAE for five widely used datasets. The best performance is marked in red and the second best performance is marked in blue. The year of publication is indicated in parentheses in the first column. The second and third columns show the inference speed and the FLOPs for the different models. ↑ indicates that a larger score is better, ↓ indicates that a smaller score is better. Our method achieves state-of-the-art performance on these five widely used datasets under four evaluation metrics.
Table 1. Quantitative comparisons include Max F-measure, E-measure, S-measure, and MAE for five widely used datasets. The best performance is marked in red and the second best performance is marked in blue. The year of publication is indicated in parentheses in the first column. The second and third columns show the inference speed and the FLOPs for the different models. ↑ indicates that a larger score is better, ↓ indicates that a smaller score is better. Our method achieves state-of-the-art performance on these five widely used datasets under four evaluation metrics.
ModelSpeedFLOPsDUT-OMRONDUTS-TEECSSDHKU-ISPASCAL-S
(fps) ↑(G) ↓MaxF ↑Em ↑Sm ↑MAE ↓MaxF ↑Em ↑Sm ↑MAE ↓MaxF ↑Em ↑Sm ↑MAE ↓MaxF ↑Em ↑Sm ↑MAE ↓MaxF ↑Em ↑Sm ↑MAE ↓
Amulet (2018)0.74360.78460.78050.09740.77860.80080.80310.08520.83750.83210.81130.13830.89420.91160.88170.05270.86030.82780.81820.1045
C2SNet (2018)0.75920.83150.79920.07230.81090.85610.83070.06220.83240.84140.81160.13490.89810.93150.88790.04590.86610.85170.82330.0957
PAGR (2018)0.77090.84350.77510.07090.85470.88350.83700.05610.84100.83170.80310.14250.91770.93950.88730.04750.86700.82470.78950.1169
AFNet (2019)7.2363.120.79720.85950.82630.05730.86290.89490.86710.04580.84800.85370.82510.12730.92260.94750.90540.03580.88300.86810.83060.0874
BASNet (2019)1.51127.560.80530.86910.83620.05650.85950.88420.86570.04760.85530.83500.83050.12190.92840.94580.90890.03220.87570.83510.81940.0924
CPD (2019)37.2159.380.79660.87260.82480.05600.86550.90370.86900.04340.85260.86160.83080.12290.92500.95030.90550.03420.88410.86760.82770.0890
SCRN (2019)18.4115.160.81150.86880.83650.05600.88760.90150.88490.03970.85910.85370.83640.12400.93420.95280.91590.03370.89800.86440.84250.0847
PoolNet (2019)15.21123.450.80550.86340.83230.05610.87640.89660.87880.04180.85330.85560.82930.12670.93050.95190.91180.03330.89260.86650.84040.0858
EGNet (2019)8.92291.830.81540.87370.84120.05280.88820.90770.88650.03950.85780.85820.83450.12480.93520.95640.91790.03090.89070.86260.83460.0899
MINet (2020)25.83162.380.80990.87340.83290.05550.88350.91790.88340.03750.84220.83040.82000.13000.92630.94580.91330.03540.89120.88900.85020.0774
GateNet (2020)13.41161.190.81810.86820.83810.05490.88720.90360.88440.04050.85580.85510.83120.12570.93350.95340.91540.03310.88940.86040.83100.0897
DNA (2021)37.25132.120.80310.83360.81810.06270.87280.86670.86180.04630.85130.84750.82590.12780.92840.93730.90480.03540.88460.85620.82650.0927
SAMNet (2021)23.280.570.80260.84660.82990.06520.83610.86520.84800.05820.84480.84650.82280.13200.91480.93710.89840.04490.85680.82990.80270.1113
SCWSSOD (2021)35.8144.510.78270.86980.81190.06020.84490.90520.83920.04920.83480.84950.80210.13030.90860.94280.88200.03750.86600.85530.79360.1000
UMNet (2022)70.5688.840.78700.85960.80470.06310.79880.86240.80110.06740.82790.82560.78970.14240.90760.93940.88660.04120.85650.83260.81320.0952
Polyp-PVT (2022)48.2225.100.79730.86050.82100.06220.86160.90450.86580.04480.85390.86500.83050.12170.91640.94270.91350.03610.88020.86150.81680.0933
OURS42.229.720.83390.89710.85530.04920.88930.92570.88770.03640.86300.86900.84170.11640.94080.96320.92340.02590.89330.87080.83430.0841
Table 2. Experimental analysis of ablation on four datasets. ↑ indicates that a larger score is better; ↓ indicates that a smaller score is better.
Table 2. Experimental analysis of ablation on four datasets. ↑ indicates that a larger score is better; ↓ indicates that a smaller score is better.
MethodMetricBaselinePVTE
+ DLRB
PVTE + DLRB
+ SRM
PVTE + DLRB + SRM
+ ASFM
HKU-ISMaxF ↑0.93620.93940.94060.9408
Sm ↑0.91950.92100.92300.9234
MAE ↓0.02830.02690.02610.0259
DUTS-TEMaxF ↑0.88120.88630.88770.8893
Sm ↑0.88100.88540.88700.8877
MAE ↓0.03880.03710.03660.0364
ECSSDMaxF ↑0.85860.85900.86270.8629
Sm ↑0.83590.83790.84040.8417
MAE ↓0.11970.11900.11810.1164
SODMaxF ↑0.86450.86520.86770.8778
Sm ↑0.78930.79120.80180.8072
MAE ↓0.10330.10140.09760.0951
Table 3. Experimental analysis of ablation for four datasets with ResNet50 as the backbone network. ↑ indicates that a larger score is better; ↓ indicates that a smaller score is better.
Table 3. Experimental analysis of ablation for four datasets with ResNet50 as the backbone network. ↑ indicates that a larger score is better; ↓ indicates that a smaller score is better.
MethodMetricBaselineResNet
+ DLRB
ResNet + DLRB
+ SRM
ResNet + DLRB + SRM
+ ASFM
HKU-ISMaxF ↑0.90850.90880.90950.9139
Sm ↑0.87860.87920.88300.8935
MAE ↓0.04280.04080.03980.0389
DUTS-TEMaxF ↑0.83390.83530.83870.8441
Sm ↑0.82440.83950.84220.8437
MAE ↓0.05620.05540.05350.0524
ECSSDMaxF ↑0.83750.83960.84240.8431
Sm ↑0.80220.80920.81270.8186
MAE ↓0.13260.13140.12940.1275
SODMaxF ↑0.80310.81620.82820.8383
Sm ↑0.72150.72830.74690.7606
MAE ↓0.13120.13020.12540.1139
Table 4. Experimental analysis of ablation for four datasets with VGG16 as the backbone network. ↑ indicates that a larger score is better; ↓ indicates that a smaller score is better.
Table 4. Experimental analysis of ablation for four datasets with VGG16 as the backbone network. ↑ indicates that a larger score is better; ↓ indicates that a smaller score is better.
MethodMetricBaselineVGG
+ DLRB
VGG + DLRB
+ SRM
VGG + DLRB + SRM
+ ASFM
HKU-ISMaxF ↑0.89150.89390.89960.9033
Sm ↑0.86240.87260.87840.8827
MAE ↓0.04580.04430.04320.0405
DUTS-TEMaxF ↑0.79570.80120.81770.8360
Sm ↑0.79630.80360.82700.8392
MAE ↓0.06710.06360.06230.0618
ECSSDMaxF ↑0.81720.82360.82870.8332
Sm ↑0.78310.78560.79630.8171
MAE ↓0.14620.14230.14010.1324
SODMaxF ↑0.79830.79950.81240.8273
Sm ↑0.71940.72610.73850.7562
MAE ↓0.14460.13820.13530.1318
Table 5. Comparison of optimal values for three different backbone networks. ↑ indicates that a larger score is better; ↓ indicates that a smaller score is better.
Table 5. Comparison of optimal values for three different backbone networks. ↑ indicates that a larger score is better; ↓ indicates that a smaller score is better.
MethodMetricVGG + DLRB + SRM
+ ASFM
ResNet + DLRB + SRM+ ASFMPVTE + DLRB + SRM
+ ASFM
HKU-ISMaxF ↑0.90330.91390.9408
Sm ↑0.88270.89350.9234
MAE ↓0.04050.03890.0259
DUTS-TEMaxF ↑0.83600.84410.8893
Sm ↑0.83920.84370.8877
MAE ↓0.06180.05240.0364
ECSSDMaxF ↑0.83320.84310.8629
Sm ↑0.81710.81860.8417
MAE ↓0.13240.12750.1164
SODMaxF ↑0.82730.83830.8778
Sm ↑0.75620.76060.8072
MAE ↓0.13180.11390.0951
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huo, L.; Guo, K.; Wang, W. An Adaptive Multi-Content Complementary Network for Salient Object Detection. Electronics 2023, 12, 4600. https://doi.org/10.3390/electronics12224600

AMA Style

Huo L, Guo K, Wang W. An Adaptive Multi-Content Complementary Network for Salient Object Detection. Electronics. 2023; 12(22):4600. https://doi.org/10.3390/electronics12224600

Chicago/Turabian Style

Huo, Lina, Kaidi Guo, and Wei Wang. 2023. "An Adaptive Multi-Content Complementary Network for Salient Object Detection" Electronics 12, no. 22: 4600. https://doi.org/10.3390/electronics12224600

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop