You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

10 July 2024

MLBSNet: Mutual Learning and Boosting Segmentation Network for RGB-D Salient Object Detection

,
and
1
The First Affiliated Hospital of Anhui University of Science and Technology (Huainan First People’s Hospital), Huainan 232002, China
2
College of Computer Science and Engineering, Anhui University of Science and Technology, Huainan 232001, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Advances in Machine Learning, Volume II

Abstract

RGB-D saliency object detection (SOD) primarily segments the most salient objects from a given scene by fusing RGB images and depth maps. Due to the inherent noise in the original depth map, fusion failures may occur, leading to performance bottlenecks. To address this issue, this paper proposes a mutual learning and boosting segmentation network (MLBSNet) for RGB-D saliency object detection, which consists of a deep optimization module (DOM), a semantic alignment module (SAM), a cross-modal integration (CMI) module, and a separate reconstruct decoder (SRD). Specifically, the deep optimization module aims to obtain optimal depth information by learning the similarity between the original and predicted depth maps. To eliminate the uncertainty of single-modal neighboring features and capture the complementary features of multiple modalities, a semantic alignment module and a cross-modal integration module are introduced. Finally, a separate reconstruct decoder based on a multi-source feature integration mechanism is constructed to overcome the accuracy loss caused by segmentation. Through comparative experiments, our method outperforms 13 existing methods on five RGB-D datasets and achieves excellent performance on four evaluation metrics.

1. Introduction

Salient object detection (SOD) follows the human visual system to find the most visually appealing objects in a given scene. It has been widely used in many computer vision domains, such as visual tracking [1], image or video segmentation [2,3], face recognition [4], and person re-identification [5].
In recent years, SOD has relied on RGB images to achieve impressive results, but its performance is still limited by the use of single-modal information in certain extreme scenarios (such as low foreground–background contrast, extreme lighting conditions, background clutter, etc.). Unlike the color information provided by RGB, the spatial geometric cues provided by depth maps can help the network better segment the objects. So far, many approaches [6,7,8,9] have introduced depth information to SOD with encouraging results. However, current research methods often treat depth as the correct information to introduce into the network, often ignoring the effect of the inherent noise of the depth map.
First, some processing schemes made for depth maps have been proposed in several works of literature [10,11,12]. Among them, Fan et al. [13] proposed a filtering scheme to screen the depth maps. Ji et al. [11] avoided the effects of low-quality depth maps by giving them lower weights. To improve the quality of low-quality depth maps, Chen et al. [14] supplemented the original depth information by estimating additional high-quality depth information to improve detection performance. At the same time, all these methods take complex preprocessing operations for depth maps. Given the small amount of data, discarding low-quality depth maps or giving lower weights to low-quality depth maps may not always achieve the desired purpose. In [15], utilizing the predicted depth maps in RGB after subsequent fusion greatly improved the performance of the model. Thus, the predicted depth maps can provide useful clues. Inspired by this, we plan a mutual information minimization depth map scheme to alleviate the effects of unstable depth maps. In this work, we employ an end-to-end network to obtain predicted depth maps. To explore the relationship between the predicted depth map and the original depth map, we first combine the two to obtain maximum information and then learn from each other with the original and predicted depth map to obtain more stable depth maps. At the same time, to maximize the accuracy of the predicted depth maps, we constrain them using similarity loss. As shown in Figure 1, our proposed model can work well for those unreliable depth images compared to other models for depth map correction.
Figure 1. Visualization of significance maps in the case of unstable depth maps: (a) RGB, (b) depth, (c) GT, (d) ours, (e) DCF [11], (f) D3Net [13], and (g) DSANet [15].
Secondly, we further consider the issue of modal alignment before fusing the features. In previous works, many approaches are more biased towards considering cross layers of fused feature alignment in the decoder part. However, the network’s feature extraction process makes features at neighboring levels differ in detail while being extremely correlated in terms of abstract semantic information. While existing methods [16,17] directly knead and sum the features of each layer and then put them into the subsequent decoder for feature extraction. As a result, this approach tends to ignore intermodal matching incompatibilities. To overcome the above limitations, we propose a semantic alignment program to reduce the variability and enhance the correlation between neighboring features by employing a graph attention mechanism. Then, to integrate complementary fusion information between RGB images and depth maps, we propose a self-attention fusion scheme to learn more common representations of both.
Finally, we believe that the segmentation structure of the network very much affects the performance of the model. Most models [18,19,20] use a U-shaped network structure to integrate fused features and segment target objects. While this top-down approach is powerful in integrating multilevel features and context awareness, the incremental stacking of features may lead to redundant information and not utilize the full potential of fused features. To this end, we propose a novel framework for the separation–reconstruction decoder. Firstly, we separate the fusion features equally and then reintegrate the separated fusion features, different levels of RGB features, and depth features in the multi-source information integration module for reconstruction. On the one hand, we retain the original fusion information and avoid information loss caused by up-sampling. On the other hand, the learning of information between different input sources allows the network to learn more discriminative features.
In summary, our main contributions can be summarized as follows:
  • We propose a mutual learning and boost segmentation network (MLBSNet) that completes the deep optimization, feature fusion, and information inference process through the deep prediction and saliency goal inference stages. The five RGB-D SOD datasets yield excellent experimental results and outperform the state-of-the-art models.
  • We introduce a deep optimization module (DOM) to obtain more accurate depth information through similarity learning, which provides more effective complementary features for fusion.
  • We designed the semantic alignment module (SAM) with a graph attention mechanism to eliminate the variability between the neighboring features of a single modality. The output features are then fed into a cross-modal integration (CMI) module to collect cross-modal complementary information.
  • We build a simple and effective separate–reconstruct decoder (SRD) that not only avoids information loss caused by up-sample but also integrates rich multi-source features.

3. Proposed Methods

3.1. Architecture Overview

Figure 2 illustrates the overview architecture of our proposed MLBSNet model. Based on an end-to-end architecture, it contains two branch networks: depth prediction and RGB-D SOD network. In RGB-D SOD, the main components include depth optimization, a semantic alignment module, cross-modal integration, and a separate–reconstruct decoder. Specifically, two parallel ResNet50 are used to extract multilevel features f i r and f i d from RGB and depth data, where i represents the number of layers of the encoder. After that, a U-shaped decoder is employed to complete the depth prediction and output the feature M p . To obtain more accurate depth maps, we propose a depth optimization module (DOM) to balance the error caused by unstable depth maps. A semantic alignment module (SAM) is designed to strengthen the links between neighboring feature layers and reduce fusion variability. Next, a cross-modal integration module is implemented to enable the fusion of RGB and depth features. Finally, a separation–reconstruction decoder is designed to segment out saliency objects.
Figure 2. The overall architecture of the proposed MLBSNet.

3.2. Depth Optimization Module (DOM)

As mentioned above, unstable depth maps may provide inaccurate depth information, which may pose challenges for cross-modal fusion. To solve this problem, as shown in Figure 3, we propose a depth optimization module that learns more effective depth information between the predicted depth and the original depth. Specifically, we first predict pseudo-depth maps from RGB images utilizing a depth prediction network. In addition, the predicted depth features that are used to perform the optimization maintain the D-dimensional. Empirically, D is set to 128, which maintains a wealth of depth information. With multilayer perceptron (MLP) operations, we input the output feature into the cosine function to compute the similarity between M o and M f , and between M f and M p , respectively. The above process is expressed as follows:
M f = C C ( M o , M p ) , S 1 = c o s ( M L P ( M o , M f ) ) , S 2 = c o s ( M L P ( M f , M p ) ) ,
where C C denotes concatenation and convolutions, M p denotes the predicted depth features, M L P denotes the multilayer perceptron, c o s denotes the cosine function, and S 1 and S 2 denote the similarity values of M o , M f , and M f , M p respectively.
Figure 3. Architecture of the proposed DOM.

3.3. Semantic Alignment Module (SAM)

Considering the matching problem between the neighboring hierarchical features of a single modality, SAM is proposed to balance the variability between the neighboring features in terms of details and enhance its semantic relevance. Inspired by [36], graph convolutional network operations can be used to explore semantic correlations between different pixels to learn more discriminative information. Next, concatenation and convolution are employed to extract features f i r / d and f i + 1 r / d ( i 2 , 3 , 4 ) , followed by a graph convolutional network and a pooling operation to learn more subtle features. The above process is expressed as follows:
f i j r s a / d s a = C C ( f i d / r , f i + 1 d / r ) P o o l i n g ( G C N ( C C ( f i d / r , f i + 1 d / r ) ) ) ,
where C C denotes concatenate and convolution, ⊗ denotes the multiplication operation, G C N denotes the graph convolutional network, P o o l i n g denotes the adaptive averaging pooling operation, and f i j r s a / d s a denotes the output features of SAM, where j = i + 1 .

3.4. Cross-Modal Integration (CMI) Module

To extract cross-modal features, a cross-modal integration module is proposed to learn more common representations between RGB images and depth maps. As shown in Figure 2, we integrate the feature output from SAM through concatenation and convolution operations. We use two 1 × 1 convolutions separately for the feature extraction of the features being integrated. After the reshaping and chunking operations, we obtain the features Q, K, and V. After that, we obtain the weight map Q K T by matrix multiplication and softmax operation. The last multiplication operation is used to enhance the feature V. The above process is represented as follows:
Q = R e s h a p e ( c o n v 1 × 1 ( C C ( f i j r s a , f i j d s a ) ) ) , K = R e s h a p e ( C h u n k ( c o n v 1 × 1 ( C C ( f i j r s a , f i j d s a ) ) ) ) , V = R e s h a p e ( C h u n k ( c o n v 1 × 1 ( C C ( f i j r s a , f i j d s a ) ) ) ) , f i j f = s o f t m a x ( Q K T ) V
where C C denotes concatenate and convolution, ⊗ denotes the multiplication operation, and C h u n k denotes the channel split operation.

3.5. Separate–Reconstruct Decoder (SRD)

We propose the SRD module in response to the question of how to make the best use of multiple sources of information. As shown in Figure 4, the module first integrates the fused features from all the outputs in the encoder. Considering the interdependence of features between channels, we performed a channel shuffling operation before separating the features, which allowed the network to learn different features, thus increasing the robustness of the model. Afterward, the separated features are combined with RGB and depth maps from different levels in the encoder to learn the saliency information provided in different modalities. To fully utilize features from different sources, we also propose a multi-source feature (MSF) integration module to learn more discriminative features to highlight saliency objects. The output features are then multiplied with RGB and depth maps to obtain enhanced features. It is worth noting that we use top-down reconstruction to gradually refine the segmentation target, which preserves more delicate saliency target details. The above process is represented as follows:
X 1 , X 2 , , X n = s p l i t ( s h u f f l e ( C C ( f 45 f , f 34 f , f 23 f ) ) ) , F i f = f i r C C ( X i , F i 1 f ) + f i d C C ( X i , F i 1 f ) ,
where C C denotes concatenate and convolution, ⊗ denotes the multiplication operation, s p l i t indicates the channel split operation, and s h u f f l e denotes the channel shuffle operation.
Figure 4. Architecture of the proposed SRD.

3.6. Supervision

Our network is an end-to-end training process. The overall loss function is described as follows:
L t o t a l = L D + L S + L O ,
We follow previous work [37] in the depth prediction phase, using SSIM and L1 as loss functions to predict the depth map. Following [38], we introduce a difference minimization loss L M to reduce the difference between the predicted depth map and the depth map.
L D = L S S I M ( M o , M p ) + L L 1 + L M ( M o , M p ) , L M = L c e ( M p , M o ) + L c e ( ( M o , M p ) + K L ( M p , M o ) + K L ( M o , M p ) ,
where L c e denotes the binary cross-entropy loss, K L denotes Kullback–Leibler divergence, M p denotes the predicted depth map, and M o denotes the original depth map. This K L loss is used to measure the similarity between the two modalities.
In the depth optimization phase, we use the cosine function as a similarity loss to predict more useful depth features. In the saliency prediction stage, we use the commonly used binary cross-entropy loss and IoU loss to optimize the network.
L O = c o s ( M c , M p ) + c o s ( M c , M o ) ,
L S = i = 1 4 ( L b c e ( S , G ) + L i o u ( S , G ) ) ,
where S denotes the saliency map, G denotes the ground truth map, and M c denotes the optimized depth map.

4. Experiments

4.1. Experimental Protocol

Datasets. For a fair comparison, we followed previous works [11,39] and selected 1485 images from the NJU2K [40] dataset and 700 images from the NLPR [41] dataset for training. We then tested the performance of our model and existing RGB-D saliency detection models on the NJU2K, NLPR, LFSD [42], DES [21], SIP [13], and STERE [43] test sets.
Evaluation metrics. We used four widely used metrics in our quantitative comparisons: mean absolute error ( M A E ) [44], S-measure ( S m ) [45], maximum F-measure ( F β ) [46], and maximum E-measure ( E m ) [47]. M A E was used to express the pixel-by-pixel mean absolute error between the predicted saliency map and the ground truth. S-measure is used to measure the structural integrity of the predictions. F-measure is used to balance precision and recall. E-measure is used to obtain image-level statistics and local pixel information.
Implementation details. Our model was implemented based on Pytorch and trained 120 times on an A4000 GPU, taking 6 h. We chose ResNet50 [48] as the backbone network for our model, and the network parameters were initialized by a pre-trained model on ImageNet [49]. The input RGB and depth map sizes were resized to 352 × 352 during both the training and testing phases. In addition, we utilized the Adam algorithm to optimize the network. The learning rate was set to 0.0001, and the batch size was set to 10. To avoid overfitting, we augmented the data with random flipping, rotating, and clipping.

4.2. Comparison with State-of-the-Art Models

We compare with 13 state-of-the-art models, including DGFNet [50], MRVI [51], SPSN [6], EMANet [30], SMAC [52], AirSOD [53], HINet [54], DCF [11], BBSNet [39], BiANet [55], JLDCF [56], D3Net [13], and CoNet [57]. In this study, we test the performance results using the authors’ prediction maps and the code generation result maps.
As shown in Figure 5, it can be observed through the PR curves that MLBSNet achieves significantly improved performance on the five datasets. Table 1 shows the quantitative comparison results of our model on the five datasets. Our model achieves excellent performance on all four evaluation metrics. In particular, the model outperforms all state-of-the-art models on the NJU2K and STERE datasets. This reflects the good generalization ability of the model. Regarding M A E metrics, our method achieves a minimum percentage gain of 5.8 % for DES, 4.6 % for SIP, 10.2 % for STERE, and 9.3 % for NJU2K compared to the second-best method. In terms of F β metrics, our method reaches a minimum percentage gain of 0.8 % in NJU2K and 0.7 % . Figure 6 visualizes the results of the qualitative comparison between MLBSNet and other state-of-the-art methods. Our method achieves better detection results in some complex scenes. For example, in the first and second rows, for detecting plants with complex edges, our proposed method can segment the edges of the plants more accurately compared to other methods. In the third row, the salient object in the RGB image is in a situation where the background is cluttered and the depth information is missing, but our proposed method still segments the salient object completely. Some incorrect depth maps may provide noise for fusion, and in the fourth row, our proposed method obtains accurate results. Benefiting from the proposed CMI module makes the network more robust. In the fifth row, our proposed method still extracts the salient objects completely in the face of the foreground–background similarity in RGB images and the continuous distribution of depth values of salient and non-salient objects in the depth map. These results are more dependent on the proposed DOM module, which can effectively optimize the imperfect depth map and provide effective fusion information for fusion. In the sixth, seventh, and eighth rows, our proposed method can detect multiple salient objects completely without losing the salient object information during the detection of multiple objects, while other methods failed to detect the salient information during the detection process.
Figure 5. The PR curves on NJU2K [58], NLPR [41], SIP [13], DES [21], and STERE [43].
Table 1. The quantitative comparison of our proposed model with 13 state-of-the-art RGB-D saliency models on five benchmark datasets. Red color indicates the best result, and blue color indicates the second-best result. The ↑/↓ means that the larger/smaller value is better.
Figure 6. Visual comparison of different methods.

4.3. Ablation Study

To verify the usefulness of the different modules in our proposed model, we demonstrate the validity of the added modules by removing them from the complete model.
Effectiveness of the depth optimization module. To demonstrate the effectiveness of the added depth optimization module, we removed the module from the network and labeled it as w/o DOM. As seen in Table 2, compared to w/o DOM, our proposed method resulted in a percentage gain of 3.4 % on the M A E of NJU2K and a percentage gain of 12.7 % on the M A E of SIP. In addition, our method increases the value of F β from 0.904 to 0.910 on the SIP dataset. Our proposed depth optimization module can largely improve the quality of the depth map to enhance the performance of the model. As seen from column (e) in Figure 7, removing the DOM module makes the network produce a lot of interfering information. In addition, to demonstrate the effect of the proposed module on the depth map, we visualized the depth map at different stages, as shown in Figure 8. To further illustrate the impact of the proposed supervised learning on the model performance, we remove L O . As can be seen in Table 2, all metrics are greatly improved on the DES, NJU2K, and SIP datasets. Therefore, similarity learning can learn deep information more reliably. To demonstrate the effectiveness of the proposed L M even further, we delete L M and only keep L O . As shown in Table 3, the model’s performance is greatly improved.
Table 2. Quantitative comparison of ablation experiments on NJU2K and SIP datasets, with ↑/↓ denoting that the larger/smaller value is better.
Figure 7. Comparison of ablation experiment visualizations: (a) RGB; (b) depth; (c) GT; (d) saliency maps obtained by MLBSNet; (e) saliency maps obtained by w/o DOM; (f) saliency maps obtained by w/o SAM; (g) saliency maps obtained by w/o CIM; (h) saliency maps obtained by w/o SAM.
Figure 8. Visualisation of depth maps at different stages:(a) RGB; (b) depth; (c) predicted depth maps; (d) the depth maps obtained by the DOM module in layer 2; (e) the depth maps obtained by the DOM module in layer 3; (f) the depth maps obtained by the DOM module in layer 4; (g) the depth maps obtained by the DOM module in layer 5; (h) saliency maps obtained by MLBSNet; (i) GT.
Table 3. Ablation experiments on supervise learning, with ↑/↓ denoting that the larger/smaller value is better.
Effectiveness of the semantic alignment module. To prove the effectiveness of the added semantic alignment module, we delete the module in the network and only carry out the channel splicing operation on the neighboring features and mark it as w/o SAM. As shown in Table 2, the added semantic alignment module can better learn the features of the neighboring levels and reduce the differences between the different levels of modalities, thus enhancing the relevance of the neighboring features and improving the model’s performance. Our proposed method has a gain of 6.4 % on the M A E of NJU2K and 8.8 % on the M A E of SIP. In addition, our method increases S m from a value of 0.889 to 0.895 on the SIP dataset, and E m from 0.924 to 0.930. The semantic alignment module can obtain more significant information for subsequent fusion, which in turn improves the model performance.
Effectiveness of the cross-modal integration module. To demonstrate the importance of the added cross-modal integration module, we removed the module from the network and only performed simple addition operations on the RGB and depth maps and labeled it as w/o CIM. Table 2 shows that the added cross-modal integration module facilitates performance improvement. On both the NJU2K and SIP datasets, all four metrics were greatly improved. Especially on the S m metric, S m increases from 0.922 to 0.925 on the NJU2K dataset, and S m improves from 0.923 to 0.930 on the SIP dataset. These results demonstrate that the cross-modal integration module can efficiently fuse the RGB image and depth information and promote the expression of saliency objects.
Effectiveness of separate–reconstruct decoder. To demonstrate the contribution of the designed decoder to the model’s performance, we used a U-shaped network for the substitution and labeled it as w/o SRD. As demonstrated in Table 2 and Figure 7, the separate–reconstruct decoder helps to improve the model performance and maintain the details of the saliency objects. The percentage gain on M A E is 16.1 % and 7.3 % respectively. S m metrics on NJU2K is 1.1 % . The F β metric has a percentage gain of 1.3 % on NJU2K. These results show that our proposed decoder can learn saliency features and separate saliency objects from multi-source information compared to the U-decoder.
Efficiency evaluation. To demonstrate the efficiency of the proposed model even further, we compare the inference time (fps), model complexity (FLOPs), and number of parameters of the proposed model with other models. For a fair comparison, all models are computed on 3090 GPUs. As shown in Table 4, our model has relatively lower FLOPs and faster fps compared to other approaches.
Table 4. Parameters, FLOPs, and fps of different RGB-D SOD methods.

5. Conclusions

We propose a mutual learning and boosting segmentation network that completes the depth optimization, feature fusion, and information inference processes through the depth prediction and saliency object inference stages. Satisfactory results are achieved on five RGB-D saliency object detection datasets, and they outperform state-of-the-art models. First, we introduced a depth optimization module to obtain more accurate depth information through mutual learning. Then, a semantic alignment module is designed to align modalities between neighboring features. Then, the output features are fed into a cross-modal integration module to obtain complementary cross-modal information. Finally, we design a simple and effective separation–reconstruction decoder for segmenting saliency objects. In addition, predicted depth maps are not always accurate, and inaccurate depth maps may lead to optimization failures. Depth prediction networks can be added to RGB-D SOD in future works to improve the quality of depth maps and improve performance.

Author Contributions

Data curation, methodology, and writing—original draft, J.W. and C.X.; Writing—review and editing, C.X. and B.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Medical Special Cultivation Project of Anhui University of Science and Technology (YZ2023H2B003), National Natural Science Foundation of China (62102003), Anhui Postdoctoral Science Foundation (2022B623), Natural Science Foundation of Anhui Province (2108085QF258), Huainan City Science and Technology Plan Project (2023A316), the University Synergy Innovation Program of Anhui Province (GXXT-2021-006, GXXT-2022-038), university-level general projects of Anhui University of Science and Technology (xjyb2020-04), and central guiding local technology development special funds (202107d06020001).

Data Availability Statement

All datasets utilized in this article are open source and publicly available for researchers to use. Interested individuals can obtain the datasets using the following link: http://mmcheng.net/socbenchmark (accessed on 13 June 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yang, G.; Li, M.; Zhang, J.; Lin, X.; Ji, H.; Chang, S.F. Video event extraction via tracking visual states of arguments. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 3136–3144. [Google Scholar]
  2. Athar, A.; Hermans, A.; Luiten, J.; Ramanan, D.; Leibe, B. Tarvis: A unified approach for target-based video segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 18738–18748. [Google Scholar]
  3. Bai, Y.; Chen, D.; Li, Q.; Shen, W.; Wang, Y. Bidirectional Copy-Paste for Semi-Supervised Medical Image Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 11514–11524. [Google Scholar]
  4. Chai, J.C.L.; Ng, T.S.; Low, C.Y.; Park, J.; Teoh, A.B.J. Recognizability Embedding Enhancement for Very Low-Resolution Face Recognition and Quality Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 9957–9967. [Google Scholar]
  5. Chen, C.; Ye, M.; Jiang, D. Towards Modality-Agnostic Person Re-Identification with Descriptive Query. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 15128–15137. [Google Scholar]
  6. Lee, M.; Park, C.; Cho, S.; Lee, S. Spsn: Superpixel prototype sampling network for rgb-d salient object detection. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 630–647. [Google Scholar]
  7. Zhou, J.; Wang, L.; Lu, H.; Huang, K.; Shi, X.; Liu, B. Mvsalnet: Multi-view augmentation for rgb-d salient object detection. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 270–287. [Google Scholar]
  8. Wu, Z.; Allibert, G.; Meriaudeau, F.; Ma, C.; Demonceaux, C. HiDAnet: RGB-D Salient Object Detection via Hierarchical Depth Awareness. arXiv 2023, arXiv:2301.07405. [Google Scholar] [CrossRef] [PubMed]
  9. Wu, Z.; Wang, J.; Zhou, Z.; An, Z.; Jiang, Q.; Demonceaux, C.; Sun, G.; Timofte, R. Object Segmentation by Mining Cross-Modal Semantics. arXiv 2023, arXiv:2305.10469. [Google Scholar]
  10. Zhang, W.; Ji, G.P.; Wang, Z.; Fu, K.; Zhao, Q. Depth quality-inspired feature manipulation for efficient RGB-D salient object detection. In Proceedings of the 29th ACM International Conference on Multimedia, Virtual Event, 20–24 October 2021; pp. 731–740. [Google Scholar]
  11. Ji, W.; Li, J.; Yu, S.; Zhang, M.; Piao, Y.; Yao, S.; Bi, Q.; Ma, K.; Zheng, Y.; Lu, H.; et al. Calibrated RGB-D salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9471–9481. [Google Scholar]
  12. Zhang, Q.; Qin, Q.; Yang, Y.; Jiao, Q.; Han, J. Feature Calibrating and Fusing Network for RGB-D Salient Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 1493–1507. [Google Scholar] [CrossRef]
  13. Fan, D.P.; Lin, Z.; Zhang, Z.; Zhu, M.; Cheng, M.M. Rethinking RGB-D salient object detection: Models, data sets, and large-scale benchmarks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 2075–2089. [Google Scholar] [CrossRef] [PubMed]
  14. Chen, C.; Wei, J.; Peng, C.; Zhang, W.; Qin, H. Improved saliency detection in RGB-D images using two-phase depth estimation and selective deep fusion. IEEE Trans. Image Process. 2020, 29, 4296–4307. [Google Scholar] [CrossRef]
  15. Zhao, J.; Zhao, Y.; Li, J.; Chen, X. Is depth really necessary for salient object detection? In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 1745–1754. [Google Scholar]
  16. Wu, Z.; Gobichettipalayam, S.; Tamadazte, B.; Allibert, G.; Paudel, D.P.; Demonceaux, C. Robust rgb-d fusion for saliency detection. In Proceedings of the 2022 International Conference on 3D Vision (3DV), Prague, Czechia, 12–15 September 2022; pp. 403–413. [Google Scholar]
  17. Song, M.; Song, W.; Yang, G.; Chen, C. Improving RGB-D salient object detection via modality-aware decoder. IEEE Trans. Image Process. 2022, 31, 6124–6138. [Google Scholar] [CrossRef] [PubMed]
  18. Sun, F.; Ren, P.; Yin, B.; Wang, F.; Li, H. CATNet: A Cascaded and Aggregated Transformer Network For RGB-D Salient Object Detection. IEEE Trans. Multimed. 2023, 26, 2249–2262. [Google Scholar] [CrossRef]
  19. Wen, H.; Yan, C.; Zhou, X.; Cong, R.; Sun, Y.; Zheng, B.; Zhang, J.; Bao, Y.; Ding, G. Dynamic selective network for RGB-D salient object detection. IEEE Trans. Image Process. 2021, 30, 9179–9192. [Google Scholar] [CrossRef]
  20. Ji, W.; Yan, G.; Li, J.; Piao, Y.; Yao, S.; Zhang, M.; Cheng, L.; Lu, H. DMRA: Depth-induced multi-scale recurrent attention network for RGB-D saliency detection. IEEE Trans. Image Process. 2022, 31, 2321–2336. [Google Scholar] [CrossRef]
  21. Cheng, Y.; Fu, H.; Wei, X.; Xiao, J.; Cao, X. Depth enhanced saliency detection method. In Proceedings of the International Conference on Internet Multimedia Computing and Service, Xiamen, China, 10–12 July 2014; pp. 23–27. [Google Scholar]
  22. Ciptadi, A.; Hermans, T.; Rehg, J.M. An In Depth View of Saliency. In Proceedings of the BMVC, Bristol, UK, 9–13 September 2013; pp. 1–11. [Google Scholar]
  23. Piao, Y.; Rong, Z.; Zhang, M.; Ren, W.; Lu, H. A2dele: Adaptive and attentive depth distiller for efficient RGB-D salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9060–9069. [Google Scholar]
  24. Zhang, C.; Cong, R.; Lin, Q.; Ma, L.; Li, F.; Zhao, Y.; Kwong, S. Cross-modality discrepant interaction network for RGB-D salient object detection. In Proceedings of the ACM International Conference on Multimedia, Virtual Event, 20–24 October 2021; pp. 2094–2102. [Google Scholar]
  25. Liao, G.; Gao, W.; Jiang, Q.; Wang, R.; Li, G. Mmnet: Multi-stage and multi-scale fusion network for rgb-d salient object detection. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2436–2444. [Google Scholar]
  26. Qu, L.; He, S.; Zhang, J.; Tian, J.; Tang, Y.; Yang, Q. RGBD salient object detection via deep fusion. IEEE Trans. Image Process. 2017, 26, 2274–2285. [Google Scholar] [CrossRef]
  27. Zhang, J.; Fan, D.P.; Dai, Y.; Anwar, S.; Saleh, F.S.; Zhang, T.; Barnes, N. UC-Net: Uncertainty inspired RGB-D saliency detection via conditional variational autoencoders. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8582–8591. [Google Scholar]
  28. Han, J.; Chen, H.; Liu, N.; Yan, C.; Li, X. CNNs-based RGB-D saliency detection via cross-view transfer and multiview fusion. IEEE Trans. Cybern. 2017, 48, 3171–3183. [Google Scholar] [CrossRef] [PubMed]
  29. Wang, Y.; Wang, F.; Wang, C.; Sun, F.; He, J. Learning Saliency-Aware Correlation Filters for Visual Tracking. Comput. J. 2022, 65, 1846–1859. [Google Scholar] [CrossRef]
  30. Feng, G.; Meng, J.; Zhang, L.; Lu, H. Encoder deep interleaved network with multi-scale aggregation for RGB-D salient object detection. Pattern Recognit. 2022, 128, 108666. [Google Scholar] [CrossRef]
  31. Sun, P.; Zhang, W.; Wang, H.; Li, S.; Li, X. Deep RGB-D saliency detection with depth-sensitive attention and automatic multi-modal fusion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1407–1417. [Google Scholar]
  32. Cong, R.; Lin, Q.; Zhang, C.; Li, C.; Cao, X.; Huang, Q.; Zhao, Y. CIR-Net: Cross-modality interaction and refinement for RGB-D salient object detection. IEEE Trans. Image Process. 2022, 31, 6800–6815. [Google Scholar] [CrossRef] [PubMed]
  33. Zhou, T.; Fu, H.; Chen, G.; Zhou, Y.; Fan, D.P.; Shao, L. Specificity-preserving RGB-D saliency detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 4681–4691. [Google Scholar]
  34. Zhou, W.; Zhu, Y.; Lei, J.; Wan, J.; Yu, L. CCAFNet: Crossflow and cross-scale adaptive fusion network for detecting salient objects in RGB-D images. IEEE Trans. Multimed. 2021, 24, 2192–2204. [Google Scholar] [CrossRef]
  35. Jin, W.D.; Xu, J.; Han, Q.; Zhang, Y.; Cheng, M.M. CDNet: Complementary depth network for RGB-D salient object detection. IEEE Trans. Image Process. 2021, 30, 3376–3390. [Google Scholar] [CrossRef] [PubMed]
  36. Te, G.; Liu, Y.; Hu, W.; Shi, H.; Mei, T. Edge-aware graph representation learning and reasoning for face parsing. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 258–274. [Google Scholar]
  37. Zhao, X.; Pang, Y.; Zhang, L.; Lu, H. Joint learning of salient object detection, depth estimation and contour extraction. IEEE Trans. Image Process. 2022, 31, 7350–7362. [Google Scholar] [CrossRef] [PubMed]
  38. Zhang, J.; Fan, D.P.; Dai, Y.; Yu, X.; Zhong, Y.; Barnes, N.; Shao, L. RGB-D saliency detection via cascaded mutual information minimization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 4338–4347. [Google Scholar]
  39. Fan, D.P.; Zhai, Y.; Borji, A.; Yang, J.; Shao, L. BBS-Net: RGB-D salient object detection with a bifurcated backbone strategy network. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 275–292. [Google Scholar]
  40. Ju, R.; Ge, L.; Geng, W.; Ren, T.; Wu, G. Depth saliency based on anisotropic center-surround difference. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 1115–1119. [Google Scholar]
  41. Peng, H.; Li, B.; Xiong, W.; Hu, W.; Ji, R. RGBD salient object detection: A benchmark and algorithms. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part III 13. Springer: Berlin/Heidelberg, Germany, 2014; pp. 92–109. [Google Scholar]
  42. Li, N.; Ye, J.; Ji, Y.; Ling, H.; Yu, J. Saliency detection on light field. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2806–2813. [Google Scholar]
  43. Niu, Y.; Geng, Y.; Li, X.; Liu, F. Leveraging stereopsis for saliency analysis. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 454–461. [Google Scholar]
  44. Perazzi, F.; Krahenbuhl, P.; Pritch, Y.; Hornung, A. Saliency filters: Contrast based filtering for salient region detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 733–740. [Google Scholar]
  45. Fan, D.P.; Cheng, M.M.; Liu, Y.; Li, T.; Borji, A. Structure-measure: A new way to evaluate foreground maps. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4548–4557. [Google Scholar]
  46. Kulshreshtha, A.; Deshpande, A.; Meher, S.K. Time-frequency-tuned salient region detection and segmentation. In Proceedings of the IEEE International Advance Computing Conference, Ghaziabad, India, 22–23 February 2013; pp. 1080–1085. [Google Scholar]
  47. Fan, D.P.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.M.; Borji, A. Enhanced-alignment measure for binary foreground map evaluation. arXiv 2018, arXiv:1805.10421. [Google Scholar]
  48. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  49. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  50. Xiao, F.; Pu, Z.; Chen, J.; Gao, X. DGFNet: Depth-guided cross-modality fusion network for RGB-D salient object detection. IEEE Trans. Multimed. 2023, 26, 2648–2658. [Google Scholar] [CrossRef]
  51. Li, A.; Mao, Y.; Zhang, J.; Dai, Y. Mutual information regularization for weakly-supervised RGB-D salient object detection. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 397–410. [Google Scholar] [CrossRef]
  52. Liu, N.; Zhang, N.; Shao, L.; Han, J. Learning selective mutual attention and contrast for RGB-D saliency detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 9026–9042. [Google Scholar] [CrossRef]
  53. Zeng, Z.; Liu, H.; Chen, F.; Tan, X. AirSOD: A Lightweight Network for RGB-D Salient Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 1656–1669. [Google Scholar] [CrossRef]
  54. Bi, H.; Zhang, J.; Wu, R.; Tong, Y.; Jin, W. Cross-modal refined adjacent-guided network for RGB-D salient object detection. Multimed. Tools Appl. 2023, 82, 37453–37478. [Google Scholar] [CrossRef]
  55. Zhang, Z.; Lin, Z.; Xu, J.; Jin, W.D.; Lu, S.P.; Fan, D.P. Bilateral attention network for RGB-D salient object detection. IEEE Trans. Image Process. 2021, 30, 1949–1961. [Google Scholar] [CrossRef] [PubMed]
  56. Fu, K.; Fan, D.P.; Ji, G.P.; Zhao, Q.; Shen, J.; Zhu, C. Siamese network for RGB-D salient object detection and beyond. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5541–5559. [Google Scholar] [CrossRef] [PubMed]
  57. Ji, W.; Li, J.; Zhang, M.; Piao, Y.; Lu, H. Accurate RGB-D salient object detection via collaborative learning. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 52–69. [Google Scholar]
  58. Ju, R.; Liu, Y.; Ren, T.; Ge, L.; Wu, G. Depth-aware salient object detection using anisotropic center-surround difference. Signal Process. Image Commun. 2015, 38, 115–126. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.