Next Article in Journal
Borehole Radar Experiment in a 7500 m Deep Well
Previous Article in Journal
Forest Fire Detection Algorithm Based on Improved YOLOv11n
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Modal Weakly Supervised RGB-D Salient Object Detection with a Focus on Filamentary Structures

by
Yifan Ding
,
Weiwei Chen
,
Guomin Zhang
,
Zhaoming Feng
and
Xuan Li
*
School of Electrical and Information Engineering, Wuhan Institute of Technology, Wuhan 430205, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 2990; https://doi.org/10.3390/s25102990
Submission received: 31 March 2025 / Revised: 6 May 2025 / Accepted: 7 May 2025 / Published: 9 May 2025
(This article belongs to the Section Optical Sensors)

Abstract

:
Current weakly supervised salient object detection (SOD) methods for RGB-D images mostly rely on image-level labels and sparse annotations, which makes it difficult to completely contour object boundaries in complex scenes, especially when detecting objects with filamentary structures. To address the aforementioned issues, we propose a novel cross-modal weakly supervised SOD framework. The framework can adequately exploit the advantages of cross-modal weak labels to generate high-quality pseudo-labels, and it can fully couple the multi-scale features of RGB and depth images for precise saliency prediction. The framework mainly consists of a cross-modal pseudo-label generation network (CPGN) and an asymmetric salient-region prediction network (ASPN). Among them, the CPGN is proposed to sufficiently leverage the precise pixel-level guidance provided by point labels and the enhanced semantic supervision provided by text labels to generate high-quality pseudo-labels, which are used to supervise the subsequent training of the ASPN. To better capture the contextual information and geometric features from RGB and depth images, the ASPN, an asymmetrically progressive network, is proposed to gradually extract multi-scale features from RGB and depth images by using the Swin-Transformer and CNN encoders, respectively. This significantly enhances the model’s ability to perceive detailed structures. Additionally, an edge constraint module (ECM) is designed to sharpen the edges of the predicted salient regions. The experimental results demonstrate that the method shows better performance in depicting salient objects, especially the filamentary structures, than other weakly supervised SOD methods.

1. Introduction

Salient object detection (SOD) aims to emulate human visual attention by capturing the most salient objects and regions in an image, which offers crucial cues for higher-level visual tasks such as image retrieval [1,2,3], object tracking [4], image quality assessment [5], scene analysis [6,7], object classification [8], and autonomous driving [9]. Fully supervised SOD tasks require extensive pixel-level annotation, which is time-consuming and labor-intensive [10,11,12,13,14]. The emergence of weakly supervised learning methods effectively addresses the data labeling issue and enhances the transferability of the model. However, weakly supervised methods [15,16,17,18,19] usually rely on image-level labels or limited annotation information, and it is difficult for weakly supervised methods to provide sufficient annotation information as guidance during the model’s training.
To enrich the information of annotation, researchers propose multi-source weakly supervised methods. However, obtaining high-quality pseudo-labels is still difficult. Low-quality pseudo-labels are inadequate in enabling the model to recognize complete edges, particularly filamentary structures. Since semantic information provides feature representations of the object and visual information provides its positional information, this work selects complementary cross-modal weak labels to generate supervisory information that is both rich and accurate. Moreover, as depth images provide rich geometric information [20,21,22,23,24], an increasing number of studies have utilized depth images as a crucial 3D spatial cue for SOD tasks. Depth images help the model distinguish salient objects from the background by perceiving the depth structure of the scene, which is particularly useful in complex scenes. However, there are significant differences among different modalities. For example, RGB images contain rich contextual information, while depth images primarily provide geometric information. This difference makes it difficult for traditional fusion methods such as channel concatenation or element-wise summation to effectively utilize the complementary information of the two modalities. Therefore, when fusing different modalities [8,25,26,27], researchers must design complex network structures and loss functions. At the same time, the number of existing RGB-D datasets is limited, and some datasets have low annotation quality, which results in the inability to meet the training requirements of models. Therefore, in the context of weakly supervised learning, it remains a challenge to utilize multi-modal and multi-dimensional data to address the issue of missing salient regions in complex scenarios, especially the detection of filamentary structures.
To address the aforementioned issue, this approach explores the complementarity of cross-modal weak labels by combining the semantic information of text labels with the visual information of point labels, thereby enriching the dimensionality of the supervisory information. Meanwhile, it enhances the utilization of the advantageous properties of RGB and depth images, providing the saliency detection model with a more comprehensive understanding of the scene. Specifically, a cross-modal pseudo-label generation network (CPGN) is designed, which leverages the semantic information contained in text labels as prior knowledge and integrates the powerful image–text matching capability of CLIP to generate initial pseudo-labels. The annotations and positional information provided by point labels are then employed to hierarchically refine the initial pseudo-labels, providing high-quality pseudo-labels aligned with human perception for salient object detection. To fully leverage the contextual and geometric features, an asymmetric salient-region prediction network (ASPN) is proposed to capture the advantageous features of different images. Various attention mechanisms are embedded to explore the correlation between RGB and depth features, enabling efficient coupling of multi-dimensional information. Additionally, an edge constraint module (ECM) is integrated into the ASPN to sharpen the edges of the saliency results. By efficiently combining multi-modal and multi-dimensional information, the proposed approach achieves salient object detection in complex backgrounds under a weakly supervised learning framework and demonstrates outstanding performance in filamentary structure detection.
The main contributions of this study are as follows:
  • We propose a cross-modal weakly supervised method for the RGB-D SOD task, in which the text and point annotations are used to provide rich semantic information and position information, respectively. By fully integrating the advantages of weak labels from different modalities, the method effectively enhances the detection performance of salient objects in complex scenes, especially in highlighting filamentary structures.
  • A cross-modal pseudo-label generation network (CPGN) is designed to provide stronger supervision for model training, which first leverages the multimodal alignment capability of CLIP to activate semantic priors from text labels, obtaining coarse annotations, and then point labels progressively enhance the richness and accuracy of these annotations to generate high-quality pseudo-labels.
  • An asymmetric salient-region prediction network (ASPN) is proposed to extract and integrate the contextual information of RGB images and the geometric information of depth images. In ASPN, an edge constraint module is introduced to sharpen the edges of salient objects. Meanwhile, we construct a cross-modal weakly supervised saliency detection dataset (CWS) for the weakly supervised RGB-D SOD task.

2. Related Work

2.1. Weakly Supervised Salient Object Detection

Salient object detection (SOD) has garnered increasing attention, and although current methods have achieved excellent results, they rely on time-consuming and labor-intensive pixel-level manual annotations. As a result, weakly supervised strategies have become mainstream. These strategies use image-level labels [18,28,29], scribbles [30,31,32], bounding boxes [33], and points [34] to supervise model training. For example, Li et al. [28] and Yu et al. [31] perform self-training by using predicted saliency maps as part of the pseudo-labels. Zeng et al. [35] employs multi-task learning, including multi-label classification and caption generation, to assist in generating pseudo-labels. Liu et al. [33] iteratively refine pseudo-labels by learning a multi-task mapping refinement network with salient bounding boxes. Point-based methods, such as the PSOD proposed by Piao et al. [34], integrate point labels and edges to refine pseudo-labels.
Our research focuses on leveraging the powerful capabilities of CLIP, using it to convert image–text similarity into pixel-text matching for precise segmentation, while also utilizing the strong foreground–background differentiation ability of point labels to create complete pseudo-labels.

2.2. Weakly Supervised RGB-D Salient Object Detection

With the advancement of depth imaging technology, depth maps have been introduced as a crucial 3D cue in the SOD task. Depth maps enhance the ability of SOD models to detect salient objects in complex scenes, leading to the rapid development of RGB-D salient object detection. Meanwhile, in order to reduce the cost of pixel-level annotations, weakly supervised RGB-D SOD solutions need to be further explored. Xu et al. [17] addressed the sparsity issue of scribbles through a secondary labeling method. Li et al. [15] modeled the pixels in RGB images to supplement the missing pixel-level annotations in scribbles.
The method leverages the attribute advantages of cross-modal weak labels. Meanwhile, it utilizes the Swin-Transformer and a CNN network to create a dual-branch architecture for feature extraction, and it proposes an edge constraint module to refine the edge details of the salient object.

2.3. Contrastive Language–Image Pre-Training

CLIP (Contrastive Language–Image Pre-Training) is a visual–language pre-training model introduced by OpenAI [36,37]. Its core idea is to establish a correlation between images and text, enabling better mutual understanding through measuring the similarity between images and text. CLIP is implemented by combining and matching a large number of images with their corresponding textual information to build a large-scale dataset. CLIP can pair input images with text and compute similarity scores between them. In practice, for an input image or text, the model outputs similar images or text, facilitating more intelligent information processing.
Most existing weakly supervised SOD methods use CAMs as pseudo-labels. However, standard CAMs are often noisy, which can significantly degrade model performance. This method uses CLIP to generate class activation maps (CAMs), delving into the relationship between text and objects in images to create a robust foundation for obtaining results that meet our requirements.

3. Proposed Method

3.1. Overview

We solve some existing issues in the weakly supervised salient object detection (SOD) task for RGB-D images, e.g., the insufficient supervision ability of weak labels, the incomplete prediction of object regions and the failure to detect filamentary structures. The overall framework we proposed is shown in Figure 1. Among them, it can be divided into a cross-modal pseudo-label generation network (CPGN) and an asymmetric salient-region prediction network (ASPN). Firstly, we designed the CPGN, which takes as input D = { R i , T i , P i } i = 1 , 2 , 3 , , n , where i is the index of dataset samples and { R i , T i , P i } represents the paired RGB image ( R i ), text labels ( T i ), and point labels ( P i ). The network outputs high-quality pseudo-labels ( P h ) to guide the ASPN to predict results with complete regions, especially for the detection of filamentary structures. Secondly, in view of the attribute differences between RGB images and depth images, the ASPN employs an asymmetric encoder structure based on the Swin-Transformer and a CNN to progressively extract the features of RGB-D images. The structure is designed to enable the ASPN to fully extract contextual information from RGB images while making the ASPN more sensitive to the spatial structure information in depth maps. Finally, we designed an edge constraint module (ECM), which utilizes shallow features and deep features to generate edge features to sharpen the edges of the salient objects by the ASPN. In conclusion, this method capitalizes on the synergy of the two networks to tackle the issues of insufficient supervision of weak labels, incomplete salient regions, and the absence of filamentous structures in current weakly supervised SOD tasks for RGB-D images.

3.2. Cross-Modal Pseudo-Label Generation Network

3.2.1. CLIP-Based Grad-CAM

Class activation mapping (CAM) has been widely used in weakly supervised SOD tasks. However, the pseudo-labels generated only by using CAM often contain noise, and the separation between the foreground and the background is not satisfactory, which makes it difficult to directly use them to guide the model training. To achieve better results, we divide the text label activation task into two stages: pseudo-label generation and refinement.
The text labels are obtained by calibrating the Caps dataset in [38]. The text labels consist of 122 foreground object descriptions (such as cat, dog, and people riding on bicycles) and 23 common background classes (such as sky, ground, and beach). As shown in Figure 1, the RGB image and the text labels (foreground label and background set) of the image are fed into the CPGN simultaneously.
We need to match the information from the text labels with the object regions in the image, and the CLIP model is well suited for this task due to its strong zero-shot capability. Therefore, our text labels include both foreground information of the object and a common background set. We use the softmax function to ensure mutual exclusivity between different classes, thereby suppressing the background and achieving good segmentation performance in the presentation of the matching results. Specifically, the scores after applying softmax are calculated as follows:
s c = e x p y c c = 1 C e x p y c
where s c represents the score for the c - th object text, C is the total number of object texts, and y c is the cosine similarity score for the c - th object.
We use Grad-CAM as the visualization method for the matching results. Grad-CAM is a visualization tool for deep learning models that allows us to intuitively understand the regions of the image that CLIP focuses on when given an input object text. It then visualizes this result and generates CAMs M c . The specific formula is as follows:
α k c = 1 Z i j y c A i j k s c 1 s c + 1 Z i j c c y c A i j k s c s c
M c = R e L U k α k c A i j k
where α k c is the weight of the k - th channel in the last feature layer A of the CLIP model for the c - th class, Z is the total number of pixels, A ij k is the pixel coordinate i , j in the k - th channel of the last feature layer A of the CLIP model, and y c A ij k is the gradient information of class c backpropagated through the k - th channel of the last feature layer A in the CLIP model.
Finally, the result refinement is achieved by combining the CAMs M c of each class c with the attention weights W a from the CLIP image encoder. Specifically, to make W a pay more attention to the object, we threshold the CAMs to obtain masks and then cover the regions related to category c in the masks with boxes to obtain class masks B c . After that, we perform sinkhorn normalization and transpose sum processing on W a , respectively, to enhance the object weights. By multiplying B c by W a , we obtain the class attention map. We use the semantic similarity among pixels in the class attention map to refine the CAM, thereby covering more incomplete areas of the object in M c . The generation of our initial pseudo-labels M init is shown in the following equation:
M i n i t = B c W A + W A T 2 · v e c M c
W A = S i n k h o r n ( W a )
where B c is the class mask, S i n k h o r n ( · ) is the sinkhorn normalization, and v e c · denotes matrix vectorization. Since the class activation map is essentially the confidence of each pixel in the image regarding class c, a hyperparameter λ (i.e., the filtering parameter) is set to ignore pixels with low confidence, completing the text label activation process, as represented by:
P i n i t = 1 , M i , j i n i t > λ 0 , M i , j i n i t < λ
where P init is the initial pseudo-label and i , j represents the coordinates of a pixel in the class activation map M init .

3.2.2. Point Label Activation and Pseudo-Label Improvement

Due to the semantic ambiguity of text labels, the masks generated from them contain a lot of noise, which causes the model to learn incorrect information. To address this issue, we introduce point labels into the pseudo-label generation process. The method leverages the advantages of multiple modalities, refining the initial pseudo-labels from both the foreground and background to obtain the final pseudo-label with sufficient supervisory information. We fully leverage the advantages of point labels to improve the quality of pseudo-labels. We divide point labels into foreground points S f and background points S b . The flood fill algorithm is a method used to fill a connected region with a specified color or value, starting from a given seed point. Using the flood fill algorithm, we fill the edge E detected by the edge detector [39] with the foreground point labels to obtain the foreground activation map P p . In this method, the foreground point is set as the seed point, with a radius of one-fifth of the given image size, and the Low Difference and High Difference are set to 20 and 50, respectively. The specific operation is defined as:
P p = F F ( S f , e )
where F F · represents the flood fill algorithm. Then, we add the foreground activation map pixel-wise with the initial pseudo-labels to complete the foreground refinement process, followed by using the background points as seed points to flood fill the entire image to remove non-object noise and obtain the final pseudo-label P h . The specific operation is defined as:
P h = F F ( S f S b , P i n i t P p )
We use P h to guide the training of the ASPN.

3.3. Asymmetric Salient-Region Prediction Network

3.3.1. Asymmetric Encoder

Weakly supervised SOD tasks for RGB-D images typically employ encoder structures based on ResNet, VGG, or Transformer backbones. These fundamental encoders are commonly pre-trained on RGB images to initialize the model. However, recent studies have found attribute differences between RGB and depth images. The extraction of depth image features using these encoders may introduce interference in the network. Depth images contain rich spatial positional information, and CNN-based encoders are more sensitive to spatial structures. Therefore, we use CNN-based encoders to extract features from depth images. Moreover, encoders based on the Swin-Transformer are better equipped to capture contextual information. Therefore, we use the Swin-Transformer encoder to extract features from RGB images.
We extract hierarchical features from the RGB features. Specifically, the Swin-Transformer first divides the input RGB image I into non-overlapping P × P patches, which are then flattened and represented as:
X = L i n e a r P a t c h I
Here, P a t c h · refers to dividing the RGB image into P × P patches, and L i n e a r · is the operation that transforms each patch into a higher-dimensional embedding feature space. The Swin-Transformer updates features layer by layer through hierarchical feature extraction. Each layer l consists of two core operations:
(a)
Performing multi-head self-attention operations on each patch:
Z l = W M S A X l + X l
Here, W M S A · refers to the window-based multi-head self-attention operation, X l is the input feature at layer l , and Z l is the feature after the window-based multi-head self-attention operation.
(b)
To better capture the contextual relationships, multi-head self-attention operations are performed between windows:
X l + 1 = S W M S A Z l + Z l
Here, S W M S A · refers to the cross-window multi-head self-attention operation. By stacking multiple layers of such operations, contextual features are gradually extracted. The final output is the contextual features { F rgb i | i = 1 , 2 , 3 , 4 } after hierarchical feature extraction. Depth features are typically used to supplement the spatial positional information of the RGB features. Networks trained from scratch tend to achieve better performance in tasks related to depth maps. Therefore, we adopted a CNN-based backbone to extract depth features in the simplest way possible. As shown in Figure 2, the depth map is converted into a higher-dimensional embedding feature space through convolution layers with different receptive fields. Then, the learned depth features are transformed into multi-scale features with three different resolutions via three identical CNN-based blocks.
Finally, for the fourth stage, we use a 1 × 1 convolution layer to obtain the final output. In this way, we acquire the depth features, represented as { F d i | i = 1 , 2 , 3 , 4 } . We use a serial structure in the depth branch, which is capable of extracting deeper information and helps reduce structural distortion during the feature extraction process of depth maps.

3.3.2. Mixed Attention Module (MAM)

As shown in Figure 3, we feed the RGB features { F rgb i | i = 1 , 2 , 3 , 4 } and the depth features { F d i | i = 1 , 2 , 3 , 4 } extracted from the asymmetric encoder into the Transformer basic blocks. The cross-modal information is output by mixing their respective attention weights in the mixed self-attention layer. Specifically, before the mixed self-attention layer, we extract the query Q , key K , and value V from both F rgb i and F d i . as Q R ,   K R ,   V R and Q D ,   K D ,   V D , respectively. We use Q R along with K D and V D as the head in the RGB feature stream, while Q D is used together with K R and V R as the head in the depth feature stream. This process is defined as:
H e a d r g b = M H S A Q R ,   K D ,   V D
H e a d d = M H S A Q D ,   K R ,   V R
After mixing the two through mixed self-attention, the output information is summed to obtain the attention weights:
M H S A r g b = C a t ( n = 1 n H e a d r g b j ) W 1 + F R G B
M H S A d = C a t ( n = 1 n H e a d d j ) W 1 + F D
where W 1 and W 2 are linear projections. After obtaining the feature maps { F j i | i = 1 , 2 , 3 , 4 , j = rgb , d } mixed through the Transformer basic blocks, we further utilize spatial attention and channel attention to better fuse the features of different modalities. This helps generate the final multi-scale features at each scale, { F i | i = 1 , 2 , 3 , 4 } , with the following operations:
F i = F j i F j i SA F j i CA F j i
Here, CA · and SA · represent channel attention and spatial attention, respectively.

3.3.3. Edge Constraint Module (ECM)

Complete edges can enhance the performance of saliency detection models. Therefore, we propose an edge constraint module (ECM), which progressively refines the output of edge features. Low-level features contain rich texture information, while high-level features contain rich semantic information. The ECM combines multi-scale features to sharpen the edges of the salient object. Figure 4 illustrates the schematic diagram of our edge constraint module. The module is divided into two parts: the generation of edge features F e and the generation of the learning target for edge features E gt . Specifically, F 1 and F 4 are fed into two separate 1 × 1 convolution layers and upsampled. Then, the obtained features are concatenated to generate a new feature. Next, the feature is fed into two consecutive C U 3 and C U 1 layers, and the edge features are output through a sigmoid function for learning the object boundaries, which can be expressed as follows:
F e d g e t = C a t C U 1 F 1 , C U 1 F 4
F e = s i g m o i d ( C U 1 ( C U 3 ( F e d g e t ) )
where C U 1 refers to a 1 × 1 convolution layer followed by an upsampling operation, C U 3 refers to a 3 × 3 convolution layer followed by an upsampling operation, and s i g m o i d ( · ) denotes the sigmoid activation function. In the generation phase of E gt , we obtain the edge map of the salient object in the final output of each iteration. This edge map guides the ECM in learning the edges of the prediction results. This process can be expressed as:
E g t = s o b e l S 1
where S 1 represents the prediction results of the ASPN in each iteration and s o b e l ( · ) denotes the edge extraction using the sobel operator.

3.3.4. Progressive Decoder

To obtain the prediction result S o , we use a progressive decoder. We first feed F 3 and F 4 into the TB layer. The features from the TB layer are then summed element-wise to generate enhanced features. These refined features are subsequently split into two branches: one branch performs upsampling before being fed back into the TB layer, while the other branch combines with F 2 in the same manner as the previous step. After incorporating F 1 into the decoding process, the edge features are added by element-wise multiplication during the decoding process, as shown in Figure 5. The decoder equation is as follows:
S f = δ T B δ T B F i T B F i 1
D t = C o n B R S f S f S i g F e
where D t | t = 1 , 2 , 3 represents the intermediate features, C o n B R is a 3 × 3 convolution layer followed by BN and ReLU layers, TB consists of three C o n B R blocks stacked together, F i denotes the i - th level feature, δ represents the upsampling operation, and Sig · is the sigmoid activation function.
Finally, a 1 × 1 convolution layer is used as the projection head for the salient features, outputting the salient object image S o , expressed as:
S o = C o n v 1 D t
Here, S o | o = 1 , 2 , 3 is the predicted result and C o n v 1 · represents the convolution operation with a 1 × 1 convolution kernel. We take S 1 as the final saliency map.

3.4. Hybrid Loss Function

Since the pseudo-labels generated by the CPGN provide rich supervisory information and suppress the noise introduced by weak labels, we employ a hybrid loss function that includes pixel-level contrast loss L PCL , enhanced hybrid loss L HEL , and edge loss L E to train the ASPN. L PCL treats the pseudo-labels and the pixels with the same predictions as positive sample pairs, while the rest are considered negative sample pairs. It trains the ASPN by increasing the similarity of positive pairs and decreasing the similarity of negative pairs. The total loss function is given by the following expression:
L = λ 1 × L P C L + λ 2 × L H E L + λ 3 × L E
We fix the value of λ 1 to λ 3 in the order of 0.9, 0.1, 1, in all our experiments, where L PCL is defined as:
L P C L = β 1 × L P C L 1 D 1 , P h + β 2 × L P C L 2 D 2 , P h + β 3 × L P C L 3 D 3 , P h
where D 1 , D 2 , D 3 D t is the output before the projection head of the final layer of the progressive decoder. L H E L is defined as:
L H E L = β 1 × L H E L 1 S 1 , P h + β 2 × L H E L 2 S 2 , P h + β 3 × L H E L 3 S 3 , P h
L HEL is proposed in [40], where S 1 , S 2 , S 3 S o is the output after the projection head of the final layer of the progressive decoder. L HEL optimizes the predicted results from both the edge and region perspectives. We fix the value of β 1 to β 3 in the order of 1, 0.5, 0.5, in all our experiments.
Since the foreground in the edge image contains only a small number of pixels, it may be dominated by the abundance of background pixels. Therefore, we use L Dice and ω L BCE to optimize the edge features. L Dice is robust to the imbalance between positive and negative samples. It focuses more on the overlapping part of the two sets in the predicted results and pseudo-labels, rather than simply considering the number of elements in each set. This allows L Dice to better balance the contributions of foreground and background pixels when the pixel distribution is uneven (e.g., when foreground pixels are far fewer than background pixels). ω L BCE is adjusted by weighting to balance, and ω is calculated based on the ratio of foreground pixels to background pixels in E gt . We use a combination of L Dice and ω L BCE to optimize the edge features. L E is defined as:
L E = L D i c e F e , E g t + ω L B C E F e , E g t

4. Experiment and Result Analysis

To maximize the effectiveness of our proposed model, we followed the traditional training setup and performed re-annotation. We selected 1485 images from the NLPR [41] dataset and 700 images from the NJU2K [42] dataset to construct a new cross-modal dataset, CWS. Furthermore, we tested our model on the public datasets LFSD [43], DES [44], SIP [45], DUT-RGBD [46], NLPR [41], NJU2K [42], and STERE [47].

4.1. Evaluation Metrics

We evaluate the performance of the models on four golden evaluation metrics, i.e., S-measure, F-measure, E-measure, and mean absolute error (MAE).

4.2. Implementation Details

During the construction of the training set, we utilized the CLIP pre-trained ViT-B-16 [48] model for CAM generation. The feature map used to generate the CAM was extracted from the final feature map before the last self-attention layer of ViT. Our model was trained on a single NVIDIA GTX 3090 GPU with 24 GB of memory. The input RGB and depth images were resized to 256 × 256. To avoid overfitting during training, all images were augmented with data augmentation techniques such as random flipping, random cropping, random rotation, and color enhancement. With the backbone network for the RGB stream employed by the Swin-Transformer model pre-trained on ImageNet, we used the AdamW optimizer with β s = ( 0.9 , 0.999 ) , ε = 1 × 10 8 , and a weight decay of 1 × 10 4 to optimize our model. For the proposed PCL, we set the temperature hyperparameter τ = 0.3 . The initial learning rate was set to 5 × 10 5 , and we applied a polynomial decay strategy with the formula: l r = i n i t l r × ( 1 ( c u r r l r / m a x i t e r ) p o w e r ) where p o w e r = 0.9 . The model was trained for 300 epochs with a batch size of 6.

4.3. Performance Comparison with the State of the Art

We compare our proposed method with both weakly supervised and fully supervised models, including MIRV [15], SSSD [32], DENet [17], and SCWS [31], which are based on scribble supervision; PSOD [34], which is based on point supervision; and MSOD [49], JSM [38], and MSW [35], which are based on multiple weak labels. For pixel-level supervision, we compare CPNet [50], HFIL [51], TPCL [28], CATNet [11], and C2DFNet [52]. To ensure a fair comparison, we either used the salient object maps provided by the authors or ran the models released by the authors to predict on the RGB-D datasets. For models designed for RGB images, we only input the RGB images from the RGB-D datasets for comparison.

4.3.1. Quantitative Evaluation

Our method was compared against state-of-the-art single-label weakly supervised, multiple-label weakly supervised, and fully supervised SOD models for RGB-D images. We evaluated the performance across all datasets listed in Table 1 using the F-measure ( F β ), S-measure ( S α ), E-measure ( E m ), and MAE ( M ). Among the compared methods, MSOD, JSM, and MSW employ two or more weak labels as supervision sources, similar to our approach, while MIRV, Denet, SCWS, and POSD rely on a single supervision source. CPNet, HFIL, TPCL, CATNet, and C2DFNet, on the other hand, are fully supervised models using pixel-level labels.
As shown by the results, our model achieves state-of-the-art performance on datasets with rich filamentary structures (SIP) and those with complex scenes (DUT-RGBD). Although our method lags behind on the DES dataset, further investigation revealed that this is due to the dataset’s limited size (only 135 images) and high homogeneity, with simple structures that diminish the advantages of our method in handling complex structural information.

4.3.2. Qualitative Evaluation

To further demonstrate the effectiveness of our method, Figure 6 shows a comparison with state-of-the-art methods in various scenes. It can be seen that compared to single weakly supervised models (such as MIRV, SSSD, DENet, SCWS, and PSOD), our model exhibits a stronger ability to localize salient objects and segment complete regions. When compared to multiple weakly supervised models (such as MSOD, JSM, and MSW), our model produces smoother salient object edges. Notably, our model shows an overwhelming advantage in representing fine filamentary structures (e.g., rows 1, 2, 4, 6, and 7), even demonstrating competitive performance when compared to fully supervised tasks. For occluded objects (e.g., row 3), our model is also able to accurately identify and segment salient objects, exhibiting a superior understanding of complex spatial structures—surpassing even fully supervised models like CATNet and C2DFNet, which aligns with the original intent of the SOD task for RGB-D images. Furthermore, our model excels in recognizing multiple objects (e.g., row 5), where it accurately segments the structures of multiple targets without introducing background noise. Overall, our model surpasses the limitations of weak labels in object completeness recognition and remains competitive even among fully supervised models.

4.3.3. Complexity Comparisons

We calculate the FLOPs and parameters for weakly supervised methods based on RGB-D images, and the results are shown in Table 2 below. The FLOPs and parameters of our model are relatively high. However, compared with other methods, our method achieves higher detection performance for filamentary structures. Additionally, it is worth noting that the primary computational cost of our model lies in the Swin-Transformer backbone, which accounts for 87 M parameters.

4.4. Ablation Studies

We conducted various ablation studies to validate the effectiveness of each key component of the model on the RGB-D dataset. This includes an examination of the effectiveness of weakly supervised pseudo-labels in relation to the model architecture. We quantitatively present the benefits of each component in a table. Our model achieved the best performance when all components were utilized.
We selected four datasets as the test datasets for the ablation experiments. The NJU2K and NLPR are two representative datasets for the SOD task on RGB-D images. The depth maps have better quality in DUT-RGBD, while DES represents a case where our method performs relatively poorly. The performance on these four datasets is sufficient to represent the effectiveness of our method.

4.4.1. Effectiveness of Cross-Modal Weak Labels

As shown in Table 3, we disassembled our cross-modal weak labels to train the model with the pseudo-labels formed by the text labels and the point labels, respectively. We found that the performance of the model was significantly reduced when only single weak labels were used, which was sufficient to prove that a single source was not enough to meet the requirements of SOD tasks for RGB-D images. Figure 7 shows the comparison of prediction results using cross-modal weak labels and using single-modal weak labels. It can be seen from the figure that when we only use text labels, the background is noisy, and when only point labels are used, the predicted results are incomplete. These are consistent with the defects of single-modal weak labels, respectively, while the prediction results obtained by the combination of cross-modal weak labels proposed by us have abundant filamentous structures and a clean background.

4.4.2. Effectiveness of Asymmetric Encoder and ECM

We added two sets of comparative experiments using symmetric encoders to verify the effectiveness of the asymmetric architecture. Moreover, we validated the effectiveness of the ECM in each set of experiments. The qualitative evaluation is shown in Figure 8, where it can be clearly observed that our asymmetric structure better captures the filamentary structure of the salient object. The ECM also sharpens the edges of the salient region. The experimental results are shown in Table 4. From the quantitative results, we observed that the asymmetric structures based on the Swin-Transformer and CNN encoders achieved better performance in the experiments without the ECM. After incorporating the ECM into the network, the performance improved significantly, which strongly demonstrates the effectiveness of the proposed asymmetric encoder combined with the ECM.

5. Discussion

Weakly supervised salient object detection holds great potential for applications in medical imaging, defect detection, and autonomous driving. The reduced annotation cost and accurate localization of salient objects provide a solid technical foundation for practical applications. The introduction of depth information can enhance the ability of the predictive model to recognize objects in low-contrast or complex background scenarios. In this study, we propose a cross-modal weak supervision framework that integrates text and point labels to provide both semantic and pixel-level guidance. This approach helps the model to learn a more complete object contour, especially in capturing filamentary structures. It also introduces depth information as guidance to overcome adverse environmental factors and to clearly delineate the geometric information of the object. These advantages contribute to the competitiveness of our approach in real-world applications. Our method significantly improves the detection of complete salient objects by aggregating multi-modal and multi-dimensional information. In future work, how to use a small amount of supervision information to complete the detection of complex structures is expected to become a development direction for weakly supervised salient object detection, and this is also the main issue we need to explore.

6. Conclusions

We propose a high-performance cross-modal weakly supervised salient object detection (SOD) framework for RGB-D images that consists of two main components: pseudo-label generation and salient object detection. Since the issue of weak labels is insufficient supervision, we propose the CPGN, which can overcome the limitations of different weak labels by combining the advantages of text labels and point labels, thus generating high-quality pseudo-labels to train the ASPN and providing sufficient and accurate saliency information. Additionally, considering the attribute differences between RGB and depth images, we utilize an asymmetric encoder structure, where RGB-D image features are extracted using a Swin-Transformer and CNN-based encoders, respectively. To further enhance the model’s perception of edge information, we propose the ECM, which constrains the edges of salient objects through multi-scale feature integration. Based on experimental results from seven datasets, our proposed method outperforms other weakly supervised approaches and demonstrates competitive performance even when compared to fully supervised methods in fine filamentary structure segmentation.

Author Contributions

Conceptualization, Y.D. and X.L.; methodology, Y.D.; software, Y.D.; validation, Y.D., G.Z. and Z.F.; formal analysis, X.L.; investigation, W.C.; resources, W.C.; data curation, Y.D.; writing—original draft preparation, Y.D.; writing—review and editing, X.L.; visualization, Y.D.; supervision, X.L.; project administration, W.C.; funding acquisition, W.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China under Grant No. 62401411, and it is also supported by Wuhan Municipal Natural Science Foundation under Grant No. 2024040801020323.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ma, L.; Luo, X.; Shi, Y.; Meng, F.; Wu, Q.; Hong, H. Optimal Transport Quantization Based on Cross-X Semantic Hypergraph Learning for Fine-grained Image Retrieval. IEEE Trans. Circuits Syst. Video Technol. 2025. [Google Scholar] [CrossRef]
  2. Ma, L.; Hong, H.; Meng, F.; Wu, Q.; Wu, J. Deep progressive asymmetric quantization based on causal intervention for fine-grained image retrieval. IEEE Trans. Multimed. 2023, 26, 1306–1318. [Google Scholar] [CrossRef]
  3. Ma, L.; Luo, X.; Hong, H.; Meng, F.; Wu, Q. Logit variated product quantization based on parts interaction and metric learning with knowledge distillation for fine-grained image retrieval. IEEE Trans. Multimed. 2024, 26, 10406–10419. [Google Scholar] [CrossRef]
  4. Chen, J.; Lin, J.; Zhong, G.; Yao, Y.; Li, Z. Multi-granularity Localization Transformer with Collaborative Understanding for Referring Multi-Object Tracking. IEEE Trans. Instrum. Meas. 2025, 74, 5004613. [Google Scholar]
  5. Zhang, Y.; Wan, L.; Liu, D.; Zhou, X.; An, P.; Shan, C. Saliency-Guided No-Reference Omnidirectional Image Quality Assessment via Scene Content Perceiving. IEEE Trans. Instrum. Meas. 2024, 73, 5039115. [Google Scholar] [CrossRef]
  6. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 20, 1254–1259. [Google Scholar] [CrossRef]
  7. Li, X.; Huang, Z.; Ma, L.; Xu, Y.; Cheng, L.; Yang, Z. Reliable metrics-based linear regression model for multilevel privacy measurement of face instances. IET Image Process. 2022, 16, 1935–1948. [Google Scholar] [CrossRef]
  8. Li, J.; Huang, B.; Pan, L. SMCNet: State-Space Model for Enhanced Corruption Robustness in 3D Classification. Sensors 2024, 24, 7861. [Google Scholar] [CrossRef]
  9. Zunair, H.; Khan, S.; Hamza, A.B. RSUD20K: A dataset for road scene understanding in autonomous driving. In Proceedings of the 2024 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 27–30 October 2024; pp. 708–714. [Google Scholar]
  10. Chen, H.; Shen, F.; Ding, D.; Deng, Y.; Li, C. Disentangled cross-modal transformer for RGB-D salient object detection and beyond. IEEE Trans. Image Process. 2024, 33, 1699–1709. [Google Scholar] [CrossRef]
  11. Sun, F.; Ren, P.; Yin, B.; Wang, F.; Li, H. CATNet: A cascaded and aggregated transformer network for RGB-D salient object detection. IEEE Trans. Multimed. 2024, 26, 2249–2262. [Google Scholar] [CrossRef]
  12. Xiao, F.; Pu, Z.; Chen, J.; Gao, X. DGFNet: Depth-guided cross-modality fusion network for RGB-D salient object detection. IEEE Trans. Multimed. 2023, 26, 2648–2658. [Google Scholar] [CrossRef]
  13. Wu, J.; Hao, F.; Liang, W.; Xu, J. Transformer fusion and pixel-level contrastive learning for RGB-D salient object detection. IEEE Trans. Multimed. 2023, 26, 1011–1026. [Google Scholar] [CrossRef]
  14. Ma, L.; Zhao, F.; Hong, H.; Wang, L.; Zhu, Y. Complementary parts contrastive learning for fine-grained weakly supervised object co-localization. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 6635–6648. [Google Scholar] [CrossRef]
  15. Li, A.; Mao, Y.; Zhang, J.; Dai, Y. Mutual information regularization for weakly-supervised RGB-D salient object detection. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 397–410. [Google Scholar] [CrossRef]
  16. Liu, Z.; Hayat, M.; Yang, H.; Peng, D.; Lei, Y. Deep hypersphere feature regularization for weakly supervised RGB-D salient object detection. IEEE Trans. Image Process. 2023, 32, 5423–5437. [Google Scholar] [CrossRef]
  17. Xu, Y.; Yu, X.; Zhang, J.; Zhu, L.; Wang, D. Weakly supervised RGB-D salient object detection with prediction consistency training and active scribble boosting. IEEE Trans. Image Process. 2022, 31, 2148–2161. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, L.; Lu, H.; Wang, Y.; Feng, M.; Wang, D.; Yin, B.; Ruan, X. Learning to detect salient objects with image-level supervision. In Proceedings of the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 136–145. [Google Scholar]
  19. Li, X.; Xu, Y.; Ma, L.; Huang, Z.; Yuan, H. Progressive attention-based feature recovery with scribble supervision for saliency detection in optical remote sensing image. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5631212. [Google Scholar] [CrossRef]
  20. Pei, R.; Deng, S.; Zhou, L.; Qin, H.; Liang, Q. MCS-ResNet: A Generative Robot Grasping Network Based on RGB-D Fusion. IEEE Trans. Instrum. Meas. 2024, 74, 3504012. [Google Scholar] [CrossRef]
  21. Wang, Y.; Tian, Y.; Chen, J.; Chen, C.; Xu, K.; Ding, X. MSSD-SLAM: Multi-feature Semantic RGB-D Inertial SLAM with Structural Regularity for Dynamic Environments. IEEE Trans. Instrum. Meas. 2024, 74, 5003517. [Google Scholar] [CrossRef]
  22. Yu, X.; Zhang, X.; Zeng, J.; Zhang, Y.; Zhao, H.; Tao, J.; Zhu, Z.; Xu, J.; Xie, S.; Peng, Q. Depth of Interaction in PET Detector Design: Performance Optimization with Light-Sharing Window. IEEE Trans. Instrum. Meas. 2024, 74, 4000910. [Google Scholar] [CrossRef]
  23. Luo, Y.; Shao, F.; Xie, Z.; Wang, H.; Chen, H.; Mu, B.; Jiang, Q. HFMDNet: Hierarchical fusion and multi-level decoder network for RGB-D salient object detection. IEEE Trans. Instrum. Meas. 2024, 73, 5012115. [Google Scholar] [CrossRef]
  24. Zhao, J.X.; Liu, J.J.; Fan, D.P.; Cao, Y.; Yang, J.; Cheng, M.M. EGNet: Edge guidance network for salient object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8779–8788. [Google Scholar]
  25. Chen, H.; Deng, L.; Chen, Z.; Liu, C.; Zhu, L.; Dong, M.; Lu, X.; Guo, C. SFCFusion: Spatial-frequency collaborative infrared and visible image fusion. IEEE Trans. Instrum. Meas. 2024, 73, 5011615. [Google Scholar] [CrossRef]
  26. Li, X.; Chen, R.; Wang, J.; Chen, W.; Zhou, H.; Ma, J. CASPFuse: An Infrared and Visible Image Fusion Method based on Dual-cycle Crosswise Awareness and Global Structure-tensor Preservation. IEEE Trans. Instrum. Meas. 2024, 74, 5002515. [Google Scholar] [CrossRef]
  27. Li, X.; Zhang, C.; Wang, J.; Chen, R.; Cheng, L. Bidirectional feedback network for high-level task-directed infrared and visible image fusion. Infrared Phys. Technol. 2025, 147, 105751. [Google Scholar] [CrossRef]
  28. Li, G.; Xie, Y.; Lin, L. Weakly supervised salient object detection using image labels. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  29. Hsu, K.J.; Lin, Y.Y.; Chuang, Y.Y. Weakly supervised salient object detection by learning a classifier-driven map generator. IEEE Trans. Image Process. 2019, 28, 5435–5449. [Google Scholar] [CrossRef] [PubMed]
  30. Zhang, J.; Yu, X.; Li, A.; Song, P.; Liu, B.; Dai, Y. Weakly-supervised salient object detection via scribble annotations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12546–12555. [Google Scholar]
  31. Yu, S.; Zhang, B.; Xiao, J.; Lim, E.G. Structure-consistent weakly supervised salient object detection with local saliency coherence. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 19–21 May 2021; Volume 35, pp. 3234–3242. [Google Scholar]
  32. Li, L.; Han, J.; Liu, N.; Khan, S.; Cholakkal, H.; Anwer, R.M.; Khan, F.S. Robust perception and precise segmentation for scribble-supervised RGB-D saliency detection. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 46, 479–496. [Google Scholar] [CrossRef] [PubMed]
  33. Liu, Y.; Wang, P.; Cao, Y.; Liang, Z.; Lau, R.W. Weakly-supervised salient object detection with saliency bounding boxes. IEEE Trans. Image Process. 2021, 30, 4423–4435. [Google Scholar] [CrossRef]
  34. Gao, S.; Zhang, W.; Wang, Y.; Guo, Q.; Zhang, C.; He, Y.; Zhang, W. Weakly-Supervised Salient Object Detection Using Point Supervision. In Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022. [Google Scholar]
  35. Zeng, Y.; Zhuge, Y.; Lu, H.; Zhang, L.; Qian, M.; Yu, Y. Multi-source weak supervision for saliency detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6074–6083. [Google Scholar]
  36. Gao, Z.; Chen, X.; Xu, J.; Yu, R.; Zhang, H.; Yang, J. Semantically-Enhanced Feature Extraction with CLIP and Transformer Networks for Driver Fatigue Detection. Sensors 2024, 24, 7948. [Google Scholar] [CrossRef]
  37. Asperti, A.; Naldi, L.; Fiorilla, S. An Investigation of the Domain Gap in CLIP-Based Person Re-Identification. Sensors 2025, 25, 363. [Google Scholar] [CrossRef]
  38. Li, J.; Ji, W.; Bi, Q.; Yan, C.; Zhang, M.; Piao, Y.; Lu, H. Joint semantic mining for weakly supervised RGB-D salient object detection. Adv. Neural Inf. Process. Syst. 2021, 34, 11945–11959. [Google Scholar]
  39. Liu, Y.; Cheng, M.M.; Hu, X.; Wang, K.; Bai, X. Richer convolutional features for edge detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3000–3009. [Google Scholar]
  40. Pang, Y.; Zhang, L.; Zhao, X.; Lu, H. Hierarchical dynamic filtering network for RGB-D salient object detection. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXV 16; Springer: Berlin/Heidelberg, Germany, 2020; pp. 235–252. [Google Scholar]
  41. Peng, H.; Li, B.; Xiong, W.; Hu, W.; Ji, R. RGBD salient object detection: A benchmark and algorithms. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part III 13; Springer: Berlin/Heidelberg, Germany, 2014; pp. 92–109. [Google Scholar]
  42. Ju, R.; Ge, L.; Geng, W.; Ren, T.; Wu, G. Depth saliency based on anisotropic center-surround difference. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 1115–1119. [Google Scholar]
  43. Li, N.; Ye, J.; Ji, Y.; Ling, H.; Yu, J. Saliency detection on light field. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2806–2813. [Google Scholar]
  44. Cheng, Y.; Fu, H.; Wei, X.; Xiao, J.; Cao, X. Depth enhanced saliency detection method. In Proceedings of the International Conference on Internet Multimedia Computing and Service, Xiamen, China, 10–12 July 2014; pp. 23–27. [Google Scholar]
  45. Fan, D.P.; Lin, Z.; Zhang, Z.; Zhu, M.; Cheng, M.M. Rethinking RGB-D salient object detection: Models, data sets, and large-scale benchmarks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 2075–2089. [Google Scholar] [CrossRef] [PubMed]
  46. Piao, Y.; Ji, W.; Li, J.; Zhang, M.; Lu, H. Depth-induced multi-scale recurrent attention network for saliency detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7254–7263. [Google Scholar]
  47. Niu, Y.; Geng, Y.; Li, X.; Liu, F. Leveraging stereopsis for saliency analysis. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 454–461. [Google Scholar]
  48. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021. [Google Scholar]
  49. Li, X.; Xu, Y.; Ma, L.; Yang, Z.; Huang, Z.; Hong, H.; Tian, J. Multi-source weakly supervised salient object detection via boosting weak-annotation source and constraining object structure. Digit. Signal Process. 2022, 126, 103461. [Google Scholar] [CrossRef]
  50. Hu, X.; Sun, F.; Sun, J.; Wang, F.; Li, H. Cross-modal fusion and progressive decoding network for RGB-D salient object detection. Int. J. Comput. Vis. 2024, 132, 3067–3085. [Google Scholar] [CrossRef]
  51. Gao, H.; Su, Y.; Wang, F.; Li, H. Heterogeneous fusion and integrity learning network for RGB-D salient object detection. ACM Trans. Multimed. Comput. Commun. Appl. 2024, 20, 1–24. [Google Scholar] [CrossRef]
  52. Zhang, M.; Yao, S.; Hu, B.; Piao, Y.; Ji, W. C2DFNet: Criss-Cross Dynamic Filter Network for RGB-D Salient Object Detection. IEEE Trans. Multimed. 2022, 25, 5142–5154. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of the proposed model. A CPGN is used for generating high-quality pseudo-labels, while an ASPN is used for salient object detection.
Figure 1. The overall architecture of the proposed model. A CPGN is used for generating high-quality pseudo-labels, while an ASPN is used for salient object detection.
Sensors 25 02990 g001
Figure 2. The architecture of the CNN-based encoder. It is used to progressively extract features from depth images.
Figure 2. The architecture of the CNN-based encoder. It is used to progressively extract features from depth images.
Sensors 25 02990 g002
Figure 3. The architecture of the mixed attention module.
Figure 3. The architecture of the mixed attention module.
Sensors 25 02990 g003
Figure 4. The architecture of the edge constraint module. It provides edge information for the ASPN.
Figure 4. The architecture of the edge constraint module. It provides edge information for the ASPN.
Sensors 25 02990 g004
Figure 5. Architecture of progressive decoder.
Figure 5. Architecture of progressive decoder.
Sensors 25 02990 g005
Figure 6. The results of our model are compared with those of previous models. Our method is indicated in the red box. Columns 4 to 8 are weakly supervised models for RGB-D images, columns 9 to 12 are weakly supervised models for RGB images, and columns 13 to 17 are fully supervised models for RGB-D images. We denote full-supervision methods by adding an asterisk “*” after the method names.
Figure 6. The results of our model are compared with those of previous models. Our method is indicated in the red box. Columns 4 to 8 are weakly supervised models for RGB-D images, columns 9 to 12 are weakly supervised models for RGB images, and columns 13 to 17 are fully supervised models for RGB-D images. We denote full-supervision methods by adding an asterisk “*” after the method names.
Sensors 25 02990 g006
Figure 7. Ablation study results comparison of cross-modal weak labels.
Figure 7. Ablation study results comparison of cross-modal weak labels.
Sensors 25 02990 g007
Figure 8. The effectiveness of the asymmetric encoder and the ECM. “Swin” refers to the symmetric encoder structure based on the Swin-Transformer, and “CNN” refers to the symmetric encoder structure based on a CNN.
Figure 8. The effectiveness of the asymmetric encoder and the ECM. “Swin” refers to the symmetric encoder structure based on the Swin-Transformer, and “CNN” refers to the symmetric encoder structure based on a CNN.
Sensors 25 02990 g008
Table 1. The benchmark test results of the SOD model. ↑ and ↓ denote that the larger and the smaller are better. Among them, sup. represents the supervision method used by the model, which is divided into weak supervision and full supervision. We denote full-supervision methods by adding an asterisk “*” after the method names. We separately highlight the best performance indicators of weak supervision and full supervision in bold.
Table 1. The benchmark test results of the SOD model. ↑ and ↓ denote that the larger and the smaller are better. Among them, sup. represents the supervision method used by the model, which is divided into weak supervision and full supervision. We denote full-supervision methods by adding an asterisk “*” after the method names. We separately highlight the best performance indicators of weak supervision and full supervision in bold.
Sup.Weakly Sup.Fully Sup.
Multiple Weak Labels Sup. Single Weak Label Sup.
DATASET Mtric Our MSOD JSM MSW MIRV SSSD Denet SCWS PSOD CPNet * HFIL * TPCL * CATNet * C2DFNet *
[49] [38] [35] [15] [32] [17] [31] [34] [50] [51] [28] [11] [52]
DUT-RGBD [46] S α 0.8870.8280.7910.8250.8760.8770.8450.8570.8870.9510.9500.9350.9530.930
F β 0.8810.8240.8020.7770.8810.8760.8300.8650.8850.9560.9530.9400.9510.934
E ξ 0.9370.8850.8700.8770.9190.9220.8940.9020.9370.9760.9750.9660.9710.958
M 0.0490.0780.0930.1040.0540.0670.0710.0600.0540.0190.0190.0240.0200.025
SIP [45] S α 0.8700.7350.7070.7810.8760.8760.8520.8330.8700.9070.9080.9000.9130.872
F β 0.8930.7080.6910.7090.8720.8720.8400.8300.8650.9250.9230.9140.9180.867
E ξ 0.9260.8120.7870.8360.9250.9250.9040.9000.9260.9440.9460.9410.9440.915
M 0.0490.1250.1410.1270.0490.0600.0630.0650.0500.0350.0340.0370.0340.054
NJU2K [42] S α 0.8910.8020.7230.7840.8900.9020.8830.8530.8840.9340.9360.9250.9370.912
F β 0.8910.8050.7440.7400.8880.9040.8700.8700.8820.9360.9360.9240.9290.912
E ξ 0.9390.8560.7930.8400.9290.9500.9150.9000.9210.9600.9590.9550.9330.919
M 0.0400.0910.1290.1190.0450.0480.0500.0590.0480.0240.0250.0280.0250.038
DES [44] S α 0.9090.8590.8260.8350.9280.9180.9000.8540.8950.9490.9460.9350.9450.914
F β 0.8780.8310.8270.7770.9270.9090.8900.8770.8970.9370.9350.9230.9140.896
E ξ 0.9570.9270.8900.8900.9720.9670.9570.8860.9440.9840.9800.9710.9790.955
M 0.0220.0420.0560.0650.0180.0280.0280.0370.0280.0120.0130.0160.0160.021
NLPR [41] S α 0.9030.8440.8100.8250.9130.8990.9020.8670.8850.9400.9420.9350.9390.928
F β 0.8660.7920.7890.7280.9020.8840.8740.8420.8570.9290.9270.9200.9160.899
E ξ 0.9480.8790.8900.8400.9540.9470.9430.9130.9210.9720.9720.9680.9680.958
M 0.0230.0500.0580.0750.0250.0350.0310.0390.0380.0160.0160.0170.0180.021
LFSD [43] S α 0.8620.8030.7660.8080.8540.8350.8320.8060.8480.8920.8850.8850.8980.863
F β 0.8740.8110.7990.7800.8610.8370.8270.8250.8450.8970.8870.8830.9000.863
E ξ 0.9050.8590.8230.8400.9010.8790.8680.8490.9030.9250.9170.9180.9330.883
M 0.0690.1020.1280.1290.0700.0950.0890.0980.0870.0490.0580.0580.0400.065
STERE [47] S α 0.8950.8420.8580.8380.8900.8810.8790.8760.8990.9200.9220.9160.9250.902
F β 0.8770.8310.7960.7850.8800.8700.8550.8850.8920.9090.9100.9020.9020.892
E ξ 0.9430.8920.8580.8800.9360.9290.9210.9300.9400.9540.9530.9510.9350.927
M 0.0350.0650.0920.0890.0410.0590.0510.0440.0390.0290.0300.0310.0300.038
Table 2. Comparison of FLOPs and parameters among different methods.
Table 2. Comparison of FLOPs and parameters among different methods.
MethodFLOPs (G)Params (M)
Ours274.19112.66
SSSD [32]44.28130.78
JSM [38]17.9447.85
DeNet [17]179.7118.57
MIRV [15]37.1463.58
Table 3. Effectiveness of our cross-modal weak labels strategy.
Table 3. Effectiveness of our cross-modal weak labels strategy.
NJU2KNLPR
Sup. S α F β E m M S α F β E m M
TXT0.8140.8070.8750.0810.8750.8860.9210.037
Point0.6720.7470.7440.1340.7240.7850.8040.067
Ours0.8910.8910.9390.0400.9030.8660.9480.023
DUT-RGBDDES
Sup. S α F β E m M S α F β E m M
TXT0.8180.8110.8890.0880.8790.8350.9200.038
Point0.7050.7930.7770.1100.6840.7490.7500.072
Ours0.8870.8810.9370.0490.9090.8780.9570.022
The best results are highlighted in bold. ↑ and ↓ denote that the larger and the smaller are better.
Table 4. The effectiveness of the asymmetric encoder and the ECM. “w/o” refers to “without”.
Table 4. The effectiveness of the asymmetric encoder and the ECM. “w/o” refers to “without”.
NJU2KNLPRDUT-RGBDDES
Model S α F β E m M S α F β E m M S α F β E m M S α F β E m M
CNN + CNN w/o ECM0.8070.8210.8460.0970.7960.8140.8330.0690.7830.7940.8350.1100.8320.7480.8700.054
CNN + CNN0.8320.8410.8660.0890.8280.8250.8640.0600.7910.8120.8440.0980.8910.7870.8910.045
Swin-Transformer + Swin-Transformer w/o ECM0.8470.8560.8860.0680.8690.8300.9030.0430.8570.8300.9030.0650.8690.8410.9140.033
Swin-Transformer + Swin-Transformer0.8580.8820.9040.0590.8790.8330.9190.0330.8690.8450.9280.0540.8880.8480.9360.029
Ours w/o ECM0.8530.8600.8960.0630.8730.8200.9140.0360.860.8490.9190.0660.8790.8350.9250.032
Ours0.8910.8910.9390.0400.9030.8660.9480.0230.8870.8810.9370.0490.9090.8780.9570.022
The best results are highlighted in bold. ↑ and ↓ denote that the larger and the smaller are better.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ding, Y.; Chen, W.; Zhang, G.; Feng, Z.; Li, X. Cross-Modal Weakly Supervised RGB-D Salient Object Detection with a Focus on Filamentary Structures. Sensors 2025, 25, 2990. https://doi.org/10.3390/s25102990

AMA Style

Ding Y, Chen W, Zhang G, Feng Z, Li X. Cross-Modal Weakly Supervised RGB-D Salient Object Detection with a Focus on Filamentary Structures. Sensors. 2025; 25(10):2990. https://doi.org/10.3390/s25102990

Chicago/Turabian Style

Ding, Yifan, Weiwei Chen, Guomin Zhang, Zhaoming Feng, and Xuan Li. 2025. "Cross-Modal Weakly Supervised RGB-D Salient Object Detection with a Focus on Filamentary Structures" Sensors 25, no. 10: 2990. https://doi.org/10.3390/s25102990

APA Style

Ding, Y., Chen, W., Zhang, G., Feng, Z., & Li, X. (2025). Cross-Modal Weakly Supervised RGB-D Salient Object Detection with a Focus on Filamentary Structures. Sensors, 25(10), 2990. https://doi.org/10.3390/s25102990

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop