Next Article in Journal
Unified Object Detector for Different Modalities Based on Vision Transformers
Next Article in Special Issue
Unified Scaling-Based Pure-Integer Quantization for Low-Power Accelerator of Complex CNNs
Previous Article in Journal
An Extended Reality System for Situation Awareness in Flood Management and Media Production Planning
Previous Article in Special Issue
Instance Segmentation of Irregular Deformable Objects for Power Operation Monitoring Based on Multi-Instance Relation Weighting Module
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Camouflaged Object Detection with a Feature Lateral Connection Network

1
HDU-ITMO Joint Institute, Hangzhou Dianzi University, Hangzhou 310018, China
2
College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310014, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(12), 2570; https://doi.org/10.3390/electronics12122570
Submission received: 26 April 2023 / Revised: 23 May 2023 / Accepted: 2 June 2023 / Published: 7 June 2023

Abstract

:
We propose a new framework for camouflaged object detection (COD) named FLCNet, which comprises three modules: an underlying feature mining module (UFM), a texture-enhanced module (TEM), and a neighborhood feature fusion module (NFFM). Existing models overlook the analysis of underlying features, which results in extracted low-level feature texture information that is not prominent enough and contains more interference due to the slight difference between the foreground and background of the camouflaged object. To address this issue, we created a UFM using convolution with various expansion rates, max-pooling, and avg-pooling to deeply mine the textural information of underlying features and eliminate interference. Motivated by the traits passed down through biological evolution, we created an NFFM, which primarily consists of element multiplication and concatenation followed by an addition operation. To obtain precise prediction maps, our model employs the top-down strategy to gradually combine high-level and low-level information. Using four benchmark COD datasets, our proposed framework outperforms 21 deep-learning-based models in terms of seven frequently used indices, demonstrating the effectiveness of our methodology.

1. Introduction

Camouflage is a common biological phenomenon observed in nature that allows organisms to blend into their surroundings, thus avoiding predators [1,2,3]. As shown in Figure 1, we present a camouflaged object and the ground truth (GT). Camouflaged objects include naturally camouflaged and artificially camouflaged objects. Natural camouflage primarily refers to animals that blend in with their surroundings, while artificial camouflage is increasingly used in various fields (such as art, combat, etc.). In contrast to salient object detection (SOD) [4], camouflaged object detection (COD) is more challenging due to the high similarity between the object and the background in terms of texture, color, and shape, resulting in low visual recognition of its boundary and surrounding environment. In addition to its scientific value, studying COD has significant engineering applications (such as surface defect detection, search, rescue, etc.) [5,6].
Research on camouflage can be traced back to 1998. Researchers have proposed various camouflaged object detection methods based on direct visual features (such as color, texture, optical flow, etc.) [7,8]. However, traditional models fail to detect in cases where there is shallow contrast between the foreground and background [9,10]. To overcome this limitation, researchers have recently introduced deep learning into COD and proposed various deep-learning-based models (such as SINet_V2 [11], UGTR [12], C2FNet [13], MGL-R [14], etc.). These models have strong feature extraction abilities and self-learning capacities, which improve the accuracy of camouflaged object detection and boost the model’s generalization.
Similarly, we propose a new deep-learning-based model for camouflaged object detection named feature lateral connection networks (FLCNet), as shown in Figure 2. FLCNet primarily consists of three modules: the underlying feature mining module (UFM), the texture-enhanced module (TEM) [15], and the neighborhood feature fusion module (NFFM). Low-level features and high-level features contain different information, but existing models focus more on high-level features and overlook the analysis of low-level features, only using C o n v 3 to compress the number of channels on low-level features, leading to the loss of a lot of texture information. We designed the UFM to thoroughly mine the texture information of low-level features and eliminate unnecessary data. The UFM obtains information under different receptive fields via convolution with different rates of expansion, removes redundant information via methods such as cross-multiplication and addition, and excavates texture information via max-pooling and avg-pooling. In the feature fusion process, most existing methods consider skip-connection fusion, and sometimes even fuse features that are completely different. Inspired by the similarity between the biological evolution process and feature fusion, we designed the NFFM, mainly composed of element multiplication and concatenation followed by element addition. The NFFM fuses adjacent features layer by layer, generating descendants with similar features while suppressing unsuitable ones. Finally, we employ a top-down strategy, starting from high-level features and gradually refining and integrating until reaching the lowest level. The purpose of this method is to ensure that each level is closely connected to the next, in order to ensure the effectiveness and stability of the integration. During training, we use the common loss functions L I O U w and L B C E w for object segmentation to focus on the pixel and map levels.
In conclusion, we can summarize our main contributions as follows:
  • We propose a new COD model which incorporates three modules: an underlying feature mining module, a texture-enhanced module [15], and a neighborhood feature fusion module. We conduct a series of experiments to validate the effectiveness of our model.
  • To fully mine the spatial texture information of low-level features, we design an underlying feature mining module. Drawing inspiration from biological evolution, we created the neighborhood feature fusion module. To improve the accuracy of our prediction map, we utilized a top-down strategy to gradually integrate high-level and low-level features.
  • To demonstrate the superiority of our model, we compare it with 21 other models on four benchmark datasets (CAMO [16], CHAMELEON [17], COD10K [11], and NC4K [1]). Based on seven public evaluation metrics, our model shows distinct advantages.
Overall, our proposed framework outperforms existing methods and demonstrates the effectiveness of our methodology.

2. Related Works

In this section, we will provide a brief overview of the work related to our model.

2.1. Salient Object Detection (SOD)

In 1998, Ltti et al. [18] proposed a visual attention mechanism, enabling computer vision researchers to explore salient object detection. Nowadays, researchers regard salience detection as an image segmentation problem, which involves segmenting the salient object area of an image from the background as a guide for image description [19]. This process is called salient object detection (SOD).
In recent years, a large number of deep-learning-based methods have emerged in the field of SOD, and their performance is significantly better than previous methods. Liu et al. [20] added global and local context decoding blocks between the sub-modules of the U-Net [21] decoding part to selectively construct context information for each pixel. The corresponding attention weights are then assigned based on context correlation, resulting in accurate and uniform detection results. Zhang et al. [22] connected the output of the last sub-module in the coding phase step by step in the form of feedback. High-level global semantic data are transferred to the shallower convolution layers by providing the input of each sub-module in the encoder. This improves the network’s capacity for feature learning. Wang et al. [23] demonstrated that high-level semantic information could be directly integrated into all low-level features by merging deeper features for each convolution layer and updating them regularly. This prevents long-term dependencies that occur due to mixing adjacent layer features. The network model is gradually optimized in a repeated procedure.

2.2. Camouflaged Object Detection (COD)

The purpose of COD is to find and distinguish objects from complex backgrounds. As an emerging field, COD has received increasingly more attention in recent years.
Fan et al. [24] used a search module (SM) and a partial decoder component (PDC) [15] to refine a rough area. Sun et al. [13] employed attention direction for feature fusion in their study. They created C2F-Net using multi-scale channel attention guidance to aggregate hierarchical features, paying attention to both local and global information simultaneously. This improves the multi-scale degree of object detection performance. Mei et al. [2] proposed PFNet, which multiplies the higher level and inverted prediction maps with the properties of the current layer and inputs them into the context exploration block. This helps to find false positive and false negative predictions and uses element-by-element subtraction to suppress two interferences. Li et al. [25] proposed JCSOD to account for the uncertainties introduced by fully labeling camouflaged objects. They use the full convolution discriminator to estimate the confidence in the predicted results, and the adversarial training strategy is used to explicitly model confidence estimation.

3. The Proposed Method

In this section, before delving into the specifics of each module, we outline the structure of FLCNet. Furthermore, we will discuss the proposed model’s training loss function.

3.1. Overall Architecture

The overall architecture of FLCNet is depicted in Figure 2, and it basically comprises the underlying feature mining module (UFM), the texture-enhanced module (TEM) [15], and the neighborhood feature fusion module (NFFM) that enables the model to detect camouflaged objects. Specifically, we use Res2Net-50 [26] as the backbone to extract five different levels of features, denoted as F i , i 1 , 2 , 3 , 4 , 5 . F 1 and F 2 contain low-level details, including spatial details (such as edges, textures, etc.) as well as interference (such as noise, etc.). We create an underlying feature mining module for low-level features to strengthen details, filter out noise, and obtain the features F 1 m and F 2 m . F 3 , F 4 , and F 5 contain high-level features, including specific details (such as semantic information, position, etc.). To combine the information more effectively, we employ the TEM [15] with a higher expansion rate to extend the sense of sensation, obtaining the features F 3 m , F 4 m , and F 5 m . All features F i m , i 1 , 2 , 3 , 4 , 5 have 64 channels. Then, we employ the neighborhood feature fusion module to gradually combine high-level and low-level features using a top-down strategy to improve the COD aims. Finally, we compute the loss and update the parameters using an optimizer. The algorithm of FLCNet is described in Algorithm 1. Then, we will detail the operation of each module, which will enable the readers to understand our method more effectively.
Algorithm 1 Camouflaged Object Detection with a Feature Lateral Connection Network
Input: Training datasets D. Maximal number of learning epochs E.
Output: Parameters α for Res2Net-50 [26], β for UFM, θ for TEM, γ for NFFM.
 1:
Initialize α , and β , θ , γ randomly.
 2:
for t 1 to E do
 3:
     Obtain encoding features F i , i 1 , 2 , 3 , 4 , 5 from Res2Net-50.
 4:
     Obtain low-level features F j m , j 1 , 2 from UFM.
 5:
     Obtain high-level features F j m , j 3 , 4 , 5 from TEM.
 6:
     Generate supervised map l k , k 2 , 3 , 4 , 5 , 6 and predicted map l 1 .
 7:
     Compute loss according loss function and update α , β , θ , γ .
 8:
end for

3.2. Underlying Feature Mining Module (UFM)

Camouflaged objects blend into the background by altering properties (such as color, texture, shape, etc.), resulting in low-level features with more interference information when using backbone extraction. To address this issue, we create a UFM with interference reduction and extensive texture information mining. As shown in Figure 2, we use this module to handle F 1 and F 2 and to obtain F 1 m and F 2 m .
Specifically, each UFM includes two parallel branches, as shown in Figure 2. The blue box represents the first branch, consisting of a C o n v 3 + B N + Re L U layer with a dilation rate of one, while the purple box represents the second branch, consisting of a C o n v 3 + B N + Re L U layer with a dilation rate of three. In each branch, the convolution operation extracts features at different dilation rates without reducing the channel count. After processing through both branches, we obtain two separate features H 1 C and H 2 C . Then, we extract shared features from H 1 C and H 2 C through element-wise multiplication and enhance them by adding the shared features back to H 1 C and H 2 C through element-wise addition. After these operations, we obtain R 1 and R 2 .
H 1 C = C 1 B R ( H ) H 2 C = C 3 B R ( H ) R 1 = H 1 C H 2 C + H 1 C R 2 = H 1 C H 2 C + H 2 C
where C 1 B R · represents the “ C o n v 3 + B N + R e L U , dilate = 1” operation and C 3 B R · represents the “ C o n v 3 + B N + R e L U , dilate = 3” operation. ⊗ represents element-wise multiplication. + represents element-wise addition.
Moving on, we concatenate features R 1 and R 2 in dim = 1 to obtain R C , which we then subject to the max-pooling operation to retain texture features, resulting in R M . Additionally, we use the avg-pooling operation to preserve the overall data quality while better highlighting the background information. Finally, we combine features R M . and R A . to obtain feature R , as shown below:
R C = C o n c a t R 1 , R 2 R M = M a x R C R A = A v g R C R = R M + R A
where C o n c a t · represents the concatenating operation in dim = 1. M a x · represents the max-pooling operation and A v g · represents the avg-pooling operation. + represents element-wise addition.

3.3. Neighborhood Feature Fusion Module (NFFM)

The limitations of natural selection force organisms to evolve in a way that is compatible with their immediate environment. Only creatures with the same traits can reproduce in this process. Natural selection simultaneously promotes genes that are adapted to the environment and represses ones that are unsuited to it [27]. Taking into account the relationship between the process of biological evolution and feature fusion, we design the neighborhood feature fusion module (NFFM), which mainly consists of element-wise multiplication and concatenation, followed by element-wise addition.
As shown in Figure 2, the NFFM has two inputs, which we denote as P 1 and P 2 . First, we use element-wise multiplication on features P 1 and P 2 to improve related features and filter out unrelated features. At the same time, we concatenate P 1 and P 2 in dim = 1 to extract more highly abstract features and use C o n v 3 + B N + Re L U to reduce the number of channels to 64. Finally, to create feature K with a wealth of similar features, we combine S 1 and S 2 by element-wise addition. The neighborhood feature fusion module can be described as follows:
S 1 = P 1 P 2 S 2 = C o n c a t P 1 , P 2 K = S 1 + S 2
where C o n c a t · represents concatenating in dim = 1. ⊗ represents element-wise multiplication. + represents element-wise addition.
To obtain accurate detection maps of camouflaged objects, we use a top-down strategy with the neighborhood feature fusion module to gradually fuse similar features of high-level and low-level features. To ensure that the neighborhood features have the same size, we added the operation of double sampling in the middle of this step.

3.4. Loss Function

We employed a combined loss function inspired by [28,29], which consists of binary cross-entropy (BCE) [30] loss and intersection over union (IoU) loss. The loss function is defined as:
L = L I O U w + L B C E w
where L B C E w is the weighted binary cross-entropy loss, which measures the pixel-wise difference between the predicted mask and the ground truth mask. This loss function is commonly used in image segmentation tasks. The formula is expressed as:
L B C E w = i , j H , W [ w × g ( i , j ) log ( p ( i , j ) ) + ( 1 g ( i , j ) ) log ( 1 p ( i , j ) ) ]
where g ( i , j ) represents the GT, p ( i , j ) represents the predicted map, and w represents the weight of different categories. W and H stand for the width and height of the map.
L I O U w measures the similarity between the predicted map and the GT at a global level and is commonly used in object detection and segmentation tasks. The formula is expressed as:
L I O U w = 1 i = 1 H j = 1 W p ( i , j ) g ( i , j ) i = 1 H j = 1 W [ p ( i , j ) + g ( i , j ) p ( i , j ) g ( i , j ) ]
where g ( i , j ) represents the GT and p ( i , j ) represents the predicted map. W and H stand for the width and height of the map.
By combining these two loss functions, we are able to capture both the pixel-level differences and the global structure of the predicted map. Furthermore, we use weighted loss functions to focus on hard samples, which can help to improve the accuracy of the model [29,31].
The model has six outputs, all of which are carefully supervised. The locations of these outputs are depicted in Figure 2. The total loss can be obtained using the following formula:
L all = L l 1 + L l 2 + L l 3 + L l 4 + L l 5 + L l 6
where L l 1 , L l 2 , L l 3 , L l 4 , L l 5 , and L l 6 refer the loss calculated between the output and the GT.

4. Experimental Results

In this section, we will delve deeper into the benchmark datasets in the field of COD, the evaluation metrics employed, the experimental setup, and the ablation study.

4.1. Datasets and Implementation

We conducted extensive comparisons on four publicly available datasets (CAMO [16], CHAMELEON [17], COD10K [11], and NC4K [1]) for COD to thoroughly verify our method.
The CAMO [16] dataset includes one camouflaged dataset (CAMO) and another non-camouflaged dataset (MS-Coco). The dataset has 1000 images for training and the remaining 250 are for testing.
The CHAMELEON [17] dataset consists of only 76 images collected from the internet using the keyword “camouflaged animal”. It primarily focuses on creatures in nature.
The COD10K [11] dataset contains 10K images divided into 5 super and 69 subclasses, including naturally camouflaged land, ocean, flying, and amphibious animals. It has 6000 images for training and 4000 images for testing.
The NC4K [1] dataset is currently the largest COD testing dataset, comprising 4121 images downloaded from the internet. Most of the camouflaged object categories are natural, with some artificial camouflage.
Implementation Details: We built our model using PyTorch and ran it on a PC with an NVIDIA GTX 2080Ti GPU. We optimized the network using the Adam algorithm [32], with the initial hyperparameters of a learning rate set at 10 4 , a batch size of 14, and a maximum epoch number of 100. We initialized some parameters with Res2Net-50 [26], while other values were randomly initialized. To ensure fair comparisons, we trained our model using the same training dataset as [11], which comprised 4040 photos from the COD10K and CAMO datasets. For the testing phase, we used the remaining images. During the training phase, we changed the size of each training image to 352 × 352 and supplemented the training dataset by randomly flipping photos to perform data preprocessing.

4.2. Evaluation Metrics

To conduct a quantitative comparison of different models on COD datasets, we utilize seven standard metrics, including the precision–recall ( PR ) curve, S-measure [33] ( S m ), weighted F-measure [34] ( F β w ), F-measure [35] ( F β ), E-measure [36] ( E m ), and mean absolute error ( MAE ). By evaluating these metrics, we can assess the performance of different models in camouflaged object detection.
Precision and recall are popular metrics used to assess the efficacy of a model by computing the pertinent precision and recall scores.
S-measure [33] is used to evaluate the structural similarity between the predicted map and the GT. It is defined as:
S = α × S o + ( 1 α ) × S r
where S o represents the structural similarity measurement based on the object level and S r represents the region-based similarity. As suggested in [33], α is set to 0.5 .
F-measure [35] is used to calculate the relationship between the precision and recall and to calculate and display the average harmonic measurement value between P and R.The formula is expressed as:
F β = 1 + β 2 P R β 2 P + R
where P represents precision and R represents recall. β 2 is set to 0.3 to emphasize precision, as suggested in [37]. Similar to [34], we determined the weights of recall and precision with the following formula:
F β w = 1 + β 2 P w R w β 2 P w + R w
where the parameters are the same as F β and w represents the weighted harmonic mean of the precision and recall.
The MAE is used to calculate the mean absolute error for each pixel between the model’s output and the input’s GT. The formula is as follows:
MAE = 1 W × H i = 1 W × H S ( i ) G ( i ) ,
where S ( i ) represents the predicted map. G ( i ) represents the GT. W and H stand for the width and height of the map, respectively.
E-measure [36] is used to evaluate the overall and local accuracy of the camouflaged object detection results by comparing the difference between the prediction map and the GT. The formula is as follows:
E = 1 W × H x = 1 W y = 1 H f ( i ) ,
where the enhanced alignment term f(·) is used to record statistics between the predicted map and the GT at the pixel and image levels. W and H stand for the width and height of the map, respectively.

4.3. Comparison with State-of-the-Art Methods

We refer to our model FLCNet as “Ours”. To validate its efficacy, we compared it to 21 other models. The models we evaluated are EGNet [38], F3Net [31], SCRN [39], PoolNet [40], CSNet [41], SSAL [42], UCNet [43], MINet [4], ITSD [44], PraNet [45], ANet-SRM [16], MirrorNet [46], PFNet [2], UJSC [25], SLSR [1], SINet [24], MGL-R [14], C2FNet [13], UGTR [12], SINet_V2 [11], and FAPNet [3]. We derived the results of all these methods from publicly available data, from data created by the model, or by retraining the model using the author’s code.

4.3.1. Quantitative Comparison

As shown in Figure 3, we present PR and F-measure curves for the COD datasets to demonstrate the clear advantages of our model. The curves clearly show that our model outperforms the alternatives.
Additionally, as listed in Table 1, our model outperformed other models on four public COD datasets, particularly on the COD10K dataset, and achieved better results in five camouflaged map quality evaluation metrics. Our model achieved impressive MAEs of 0.071, 0.034, and 0.046 in the CAMO, COD10K, and NC4K datasets, respectively. This is a decline of 6.58%, 5.56%, and 2.13% over FAPNet. Moreover, for the four superclasses in COD10K, our model’s M A E scores significantly improved, declining by 12.50%, 2.04%, and 4.00% in amphibian, aquatic, and flying animals, respectively (as demonstrated in Table 2).
Overall, the excellence and efficiency of our model were significantly improved compared to others, as evidenced by the results presented in Figure 3 and Table 1 and Table 2.

4.3.2. Qualitative Comparison

Our model’s results are more similar to GT, as shown in Figure 4. This figure summarizes the qualitative comparisons we conducted for all models using various visual contrast experiments. In other words, our results are more comprehensive and accurate compared to other models. Overall, our model has two significant advantages. Our model is capable of improving the edge information of large objects, as shown in the first, second, third, fifth, and sixth rows of Figure 4. This is primarily due to the UFM module, which effectively filters interference and distinguishes between the foreground and background by thoroughly mining low-level information. Our model can accurately segment small camouflaged objects, as shown in the fourth, seventh, eighth, ninth, and tenth rows of Figure 4. This is because we gradually fuse low-level and high-level features using the NFFM and top-down strategy, resulting in accurate prediction maps.
Based on the aforementioned comparison, we can conclusively demonstrate the superiority of FLCNet. Whether identifying small objects or edge features, our model outperforms existing methods in COD.

4.4. Comparisons of Inference

We also compare our model with other methods in terms of parameters (Params), floating point operations (FLOPs), and frames per second (FPS), as shown in Table 3. While our model produced positive results, the inference structure’s architecture still contains some redundant elements. We want to emphasize that by including UFM and NFFM modules, our method outperformed the comparison methods, striking a compromise between model complexity and performance. This shows that the primary goal of this work has been achieved. For future work, we will explore and improve our approach by considering integrated solutions for accuracy and efficiency, as well as addressing the increased inference cost consumption.

4.5. Ablation Studies

The effectiveness of the proposed UFM and NFFM is demonstrated in this section through our ablation studies on three COD datasets. Quantitative and qualitative comparisons are shown in Table 4 and Figure 5.
Table 4 illustrates how each module improves the model’s performance, with the best results achieved when all suggested modules are combined. Adding UFM or NFFM to the basic structure results in gradual increases in the evaluation scores across all three datasets, particularly in terms of the M A E . Our model’s M A E scores increased by 8.97%, 15.00%, and 11.54% compared to the basic model, respectively. In addition, the 2nd and 3rd rows in Table 4 show the two suggested modules significantly enhance the model’s performance compared to the basic model. The NFFM module has the most significant impact on MAE performance, increasing it by 7.69%, 12.50%, and 11.54% in the three datasets, compared to the basic module.
We also provide the prediction maps for each ablation experiment, demonstrating the effectiveness of our suggested modules. Figure 5 shows that UFM distinguishes between the foreground and background by mining low-level information, thus improving the resolution of edge information. Meanwhile, the NFFM and top-down strategy gradually fuse low-level features with high-level features to accurately locate the target location and generate an accurate predicted map.
Based on both qualitative comparisons and quantitative comparisons of the ablation investigation, it is clear that our model fully conforms to anticipated design standards and demonstrates a superior performance.

4.6. Failure Cases and Analysis

Based on the previous findings, our model still has some drawbacks, even though it performs better overall than other models. As shown in Figure 6, we display some scenarios where it fails. Detecting highly intricate artificial camouflage was not particularly successful using our methodology. As seen in first, second, and third rows of figures, the background and objects combined seamlessly, making it impossible for our algorithm to infer boundaries accurately. This issue could be due to the insufficient depth of mining and poor usage of edge detail information. Meanwhile, when an image contains both salient objects and camouflaged objects, our model can become confused and mistakenly segment the salient objects as camouflaged objects. As seen in the third and fourth rows, containing objects of two categories, the model has identified the objects. However, salient objects generally stand out significantly from the background, whereas a camouflaged object differs only slightly. This suggests that feature extraction and refinement alone will not enable the model to discern the differences between the two. Meanwhile, in our model, ReLU was mainly used as the activation function. In future research, we will consider using adaptive activation functions, which have a better learning ability than fixed activations and can greatly improve the convergence speed and increase model accuracy [47,48,49,50]. This also provides new directions for future improvements.

5. Conclusions

We introduce a new model named FLCNet, which comprises a UFM, a TEM, and an NFFM. The proposed UFM addresses the problem of existing models ignoring the deep exploration of low-level features, while the proposed NFFM avoids the problem of low efficiency in cross-level feature fusion. Finally, we use a top-down strategy to gradually integrate high-level and low-level features to generate an accurately predicted map. Through a series of comparison and ablation experiments, we have demonstrated the effectiveness and superiority of our model. However, our model still has some limitations, as it is not particularly successful at detecting highly complex artificial camouflage. At the same time, our model cannot distinguish salient objects from camouflaged objects. In the following research, we will further optimize this issue. Additionally, we hope that our model can be applied to more fields, such as polyp detection, steel defect detection, etc.

Author Contributions

Methodology, T.W.; software, R.W.; supervision, J.W.; writing—original draft, T.W.; writing—review and editing, T.W. and J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data will be available at https://github.com/Tao-Wang-CV/FLCNet.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lv, Y.; Zhang, J.; Dai, Y.; Li, A.; Liu, B.; Barnes, N.; Fan, D.P. Simultaneously localize, segment and rank the camouflaged objects. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 18–22 June 2021; pp. 11591–11601. [Google Scholar]
  2. Mei, H.; Ji, G.P.; Wei, Z.; Yang, X.; Wei, X.; Fan, D.P. Camouflaged object segmentation with distraction mining. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 18–22 June 2021; pp. 8772–8781. [Google Scholar]
  3. Zhou, T.; Zhou, Y.; Gong, C.; Yang, J.; Zhang, Y. Feature Aggregation and Propagation Network for Camouflaged Object Detection. IEEE Trans. Image Process. 2022, 31, 7036–7047. [Google Scholar] [CrossRef]
  4. Pang, Y.; Zhao, X.; Zhang, L.; Lu, H. Multi-scale interactive network for salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9413–9422. [Google Scholar]
  5. Le, X.; Mei, J.; Zhang, H.; Zhou, B.; Xi, J. A learning-based approach for surface defect detection using small image datasets. Neurocomputing 2020, 408, 112–120. [Google Scholar] [CrossRef]
  6. Lidbetter, T. Search and rescue in the face of uncertain threats. Eur. J. Oper. Res. 2020, 285, 1153–1160. [Google Scholar] [CrossRef] [Green Version]
  7. Zhang, X.; Zhu, C.; Wang, S.; Liu, Y.; Ye, M. A Bayesian approach to camouflaged moving object detection. IEEE Trans. Circuits Syst. Video Technol. 2016, 27, 2001–2013. [Google Scholar] [CrossRef]
  8. Feng, X.; Guoying, C.; Richang, H.; Jing, G. Camouflage texture evaluation using a saliency map. Multimed. Syst. 2015, 21, 169–175. [Google Scholar] [CrossRef]
  9. Tankus, A.; Yeshurun, Y. Detection of regions of interest and camouflage breaking by direct convexity estimation. In Proceedings of the 1998 IEEE Workshop on Visual Surveillance, Bombay, India, 2 January 1998; pp. 42–48. [Google Scholar]
  10. Guo, H.; Dou, Y.; Tian, T.; Zhou, J.; Yu, S. A robust foreground segmentation method by temporal averaging multiple video frames. In Proceedings of the 2008 International Conference on Audio, Language and Image Processing, Shanghai, China, 7–9 July 2008; pp. 878–882. [Google Scholar]
  11. Fan, D.P.; Ji, G.P.; Cheng, M.M.; Shao, L. Concealed object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 6024. [Google Scholar] [CrossRef]
  12. Yang, F.; Zhai, Q.; Li, X.; Huang, R.; Luo, A.; Cheng, H.; Fan, D.P. Uncertainty-guided transformer reasoning for camouflaged object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Nashville, TN, USA, 18–22 June 2021; pp. 4146–4155. [Google Scholar]
  13. Sun, Y.; Chen, G.; Zhou, T.; Zhang, Y.; Liu, N. Context-aware cross-level fusion network for camouflaged object detection. arXiv 2021, arXiv:2105.12555. [Google Scholar]
  14. Zhai, Q.; Li, X.; Yang, F.; Chen, C.; Cheng, H.; Fan, D.P. Mutual graph learning for camouflaged object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 18–22 June 2021; pp. 12997–13007. [Google Scholar]
  15. Wu, Z.; Su, L.; Huang, Q. Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3907–3916. [Google Scholar]
  16. Le, T.N.; Nguyen, T.V.; Nie, Z.; Tran, M.T.; Sugimoto, A. Anabranch network for camouflaged object segmentation. Comput. Vis. Image Underst. 2019, 184, 45–56. [Google Scholar] [CrossRef]
  17. Skurowski, P.; Abdulameer, H.; Błaszczyk, J.; Depta, T.; Kornacki, A.; Kozieł, P. Animal camouflage analysis: Chameleon database. Unpubl. Manuscr. 2018, 2, 7. [Google Scholar]
  18. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef] [Green Version]
  19. Fang, H.; Gupta, S.; Iandola, F.; Srivastava, R.K.; Deng, L.; Dollár, P.; Gao, J.; He, X.; Mitchell, M.; Platt, J.C.; et al. From captions to visual concepts and back. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1473–1482. [Google Scholar]
  20. Liu, N.; Han, J.; Yang, M.H. Picanet: Learning pixel-wise contextual attention for saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3089–3098. [Google Scholar]
  21. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  22. Zhang, X.; Wang, T.; Qi, J.; Lu, H.; Wang, G. Progressive attention guided recurrent network for salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 714–722. [Google Scholar]
  23. Wang, B.; Chen, Q.; Zhou, M.; Zhang, Z.; Jin, X.; Gai, K. Progressive feature polishing network for salient object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12128–12135. [Google Scholar]
  24. Fan, D.P.; Ji, G.P.; Sun, G.; Cheng, M.M.; Shen, J.; Shao, L. Camouflaged object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2777–2787. [Google Scholar]
  25. Li, A.; Zhang, J.; Lv, Y.; Liu, B.; Zhang, T.; Dai, Y. Uncertainty-aware joint salient object and camouflaged object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 18–22 June 2021; pp. 10071–10081. [Google Scholar]
  26. Gao, S.H.; Cheng, M.M.; Zhao, K.; Zhang, X.Y.; Yang, M.H.; Torr, P. Res2net: A new multi-scale backbone architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 652–662. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Dobzhansky, T. Nothing in biology makes sense except in the light of evolution. Am. Biol. Teach. 2013, 75, 87–91. [Google Scholar] [CrossRef]
  28. Fan, D.P.; Ji, G.P.; Qin, X.; Cheng, M.M. Cognitive vision inspired object segmentation metric and loss function. Sci. Sin. Inform. 2021, 6, 6. [Google Scholar]
  29. Qin, X.; Zhang, Z.; Huang, C.; Gao, C.; Dehghan, M.; Jagersand, M. Basnet: Boundary-aware salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7479–7489. [Google Scholar]
  30. De Boer, P.T.; Kroese, D.P.; Mannor, S.; Rubinstein, R.Y. A tutorial on the cross-entropy method. Ann. Oper. Res. 2005, 134, 19–67. [Google Scholar] [CrossRef]
  31. Wei, J.; Wang, S.; Huang, Q. F3Net: Fusion, feedback and focus for salient object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12321–12328. [Google Scholar]
  32. Da, K. A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  33. Fan, D.P.; Cheng, M.M.; Liu, Y.; Li, T.; Borji, A. Structure-measure: A new way to evaluate foreground maps. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4548–4557. [Google Scholar]
  34. Margolin, R.; Zelnik-Manor, L.; Tal, A. How to evaluate foreground maps? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 248–255. [Google Scholar]
  35. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-tuned salient region detection. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1597–1604. [Google Scholar]
  36. Fan, D.P.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.M.; Borji, A. Enhanced-alignment measure for binary foreground map evaluation. arXiv 2018, arXiv:1805.10421. [Google Scholar]
  37. Zhang, D.; Han, J.; Li, C.; Wang, J.; Li, X. Detection of co-salient objects by looking deep and wide. Int. J. Comput. Vis. 2016, 120, 215–232. [Google Scholar] [CrossRef]
  38. Zhao, J.X.; Liu, J.J.; Fan, D.P.; Cao, Y.; Yang, J.; Cheng, M.M. EGNet: Edge guidance network for salient object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8779–8788. [Google Scholar]
  39. Wu, Z.; Su, L.; Huang, Q. Stacked cross refinement network for edge-aware salient object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7264–7273. [Google Scholar]
  40. Liu, J.J.; Hou, Q.; Cheng, M.M.; Feng, J.; Jiang, J. A simple pooling-based design for real-time salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3917–3926. [Google Scholar]
  41. Gao, S.H.; Tan, Y.Q.; Cheng, M.M.; Lu, C.; Chen, Y.; Yan, S. Highly efficient salient object detection with 100k parameters. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 8–14 September 2020; pp. 702–721. [Google Scholar]
  42. Zhang, J.; Yu, X.; Li, A.; Song, P.; Liu, B.; Dai, Y. Weakly-supervised salient object detection via scribble annotations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12546–12555. [Google Scholar]
  43. Zhang, J.; Fan, D.P.; Dai, Y.; Anwar, S.; Saleh, F.S.; Zhang, T.; Barnes, N. UC-Net: Uncertainty inspired RGB-D saliency detection via conditional variational autoencoders. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8582–8591. [Google Scholar]
  44. Zhou, H.; Xie, X.; Lai, J.H.; Chen, Z.; Yang, L. Interactive two-stream decoder for accurate and fast saliency detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9141–9150. [Google Scholar]
  45. Fan, D.P.; Ji, G.P.; Zhou, T.; Chen, G.; Fu, H.; Shen, J.; Shao, L. Pranet: Parallel reverse attention network for polyp segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; pp. 263–273. [Google Scholar]
  46. Yan, J.; Le, T.N.; Nguyen, K.D.; Tran, M.T.; Do, T.T.; Nguyen, T.V. Mirrornet: Bio-inspired camouflaged object segmentation. IEEE Access 2021, 9, 43290–43300. [Google Scholar] [CrossRef]
  47. Jagtap, A.D.; Kawaguchi, K.; Karniadakis, G.E. Adaptive activation functions accelerate convergence in deep and physics-informed neural networks. J. Comput. Phys. 2020, 404, 109136. [Google Scholar] [CrossRef] [Green Version]
  48. Jagtap, A.D.; Kawaguchi, K.; Em Karniadakis, G. Locally adaptive activation functions with slope recovery for deep and physics-informed neural networks. Proc. R. Soc. A 2020, 476, 20200334. [Google Scholar] [CrossRef]
  49. Jagtap, A.D.; Karniadakis, G.E. How important are activation functions in regression and classification? A survey, performance comparison, and future directions. J. Mach. Learn. Model. Comput. 2023, 4, 21–75. [Google Scholar] [CrossRef]
  50. Jagtap, A.D.; Shin, Y.; Kawaguchi, K.; Karniadakis, G.E. Deep Kronecker neural networks: A general framework for neural networks with adaptive activation functions. Neurocomputing 2022, 468, 165–180. [Google Scholar] [CrossRef]
Figure 1. Examples of camouflaged objects. (a) Camouflaged object, (b) GT.
Figure 1. Examples of camouflaged objects. (a) Camouflaged object, (b) GT.
Electronics 12 02570 g001
Figure 2. The overall architecture of the proposed FLCNet, which includes three key components, an underlying feature mining module (UFM), a texture-enhanced module (TEM), and a neighborhood feature fusion module (NFFM). The input is I , and the result is a prediction map M .
Figure 2. The overall architecture of the proposed FLCNet, which includes three key components, an underlying feature mining module (UFM), a texture-enhanced module (TEM), and a neighborhood feature fusion module (NFFM). The input is I , and the result is a prediction map M .
Electronics 12 02570 g002
Figure 3. Quantitative evaluation of different models. (a) shows CHAMELEONs Recall, (b) shows CHAMELEONs threshold.
Figure 3. Quantitative evaluation of different models. (a) shows CHAMELEONs Recall, (b) shows CHAMELEONs threshold.
Electronics 12 02570 g003
Figure 4. Visual comparison of our and other models on four COD testing datasets. (a) Input, (b) GT, (c) Ours, (d) FAPNet [3] (e) SINet_V2 [11], (f) UGTR [12], (g) C2FNet [13], (h) MGL-R [14], (i) SLSR [1], (j) UJSC [25], (k) PFNet [2], (l) SINet [24], (m) PraNet [45], (n) ITSD [44], (o) MINet [4], (p) UCNet [43].
Figure 4. Visual comparison of our and other models on four COD testing datasets. (a) Input, (b) GT, (c) Ours, (d) FAPNet [3] (e) SINet_V2 [11], (f) UGTR [12], (g) C2FNet [13], (h) MGL-R [14], (i) SLSR [1], (j) UJSC [25], (k) PFNet [2], (l) SINet [24], (m) PraNet [45], (n) ITSD [44], (o) MINet [4], (p) UCNet [43].
Electronics 12 02570 g004
Figure 5. Qualitative comparisons of our model. (a) Input, (b) GT, (c) Ours, (d) Basic, (e) Basic+UFM, (f) Basic+NFFM.
Figure 5. Qualitative comparisons of our model. (a) Input, (b) GT, (c) Ours, (d) Basic, (e) Basic+UFM, (f) Basic+NFFM.
Electronics 12 02570 g005
Figure 6. Some failed examples. (a) Input, (b) GT, (c) Ours.
Figure 6. Some failed examples. (a) Input, (b) GT, (c) Ours.
Electronics 12 02570 g006
Table 1. Quantitative comparison of different methods on four COD testing datasets. Here, “↑” (“↓”) means that the larger (smaller) the better. The best three results in each column are marked in red, green, and blue.
Table 1. Quantitative comparison of different methods on four COD testing datasets. Here, “↑” (“↓”) means that the larger (smaller) the better. The best three results in each column are marked in red, green, and blue.
CAMO DatasetCHAMELEON DatasetCOD10K DatasetNC4K Dataset
S m F β w F β E m MAE S m F β w F β E m MAE S m F β w F β E m MAE S m F β w F β E m MAE
EGNet [38]0.7320.6040.6700.8000.1090.7970.6490.7020.8600.0650.7360.5170.5820.8100.0610.7770.6390.6960.8410.075
PoolNet [40]0.7300.5750.6430.7470.1050.8450.6910.7490.8640.0540.7400.5060.5760.7770.0560.7850.6350.6990.8140.073
F3Net [31]0.7110.5640.6160.7410.1090.8480.7440.7700.8940.0470.7390.5440.5930.7950.0510.7800.6560.7050.8240.070
SCRN [39]0.7790.6430.7050.7970.0900.8760.7410.7870.8890.0420.7890.5750.6510.8170.0470.8300.6980.7570.8540.059
CSNet [41]0.7710.6420.7050.7950.0920.8560.7180.7660.8690.0470.7780.5690.6350.8100.0470.7500.6030.6550.7730.088
SSAL [42]0.6440.4930.5790.7210.1260.7570.6390.7020.8490.0710.6680.4540.5270.7680.0660.6990.5610.6440.7800.093
UCNet [43]0.7390.6400.7000.7870.0940.8800.8170.8360.9300.0360.7760.6330.6810.8570.0420.8110.7290.7750.8710.055
MINet [4]0.7480.6370.6910.7920.0900.8550.7710.8020.9140.0360.7700.6080.6570.8320.0420.8120.7200.7640.8620.056
ITSD [44]0.7500.6100.6630.7800.1020.8140.6620.7050.8440.0570.7670.5570.6150.8080.0510.8110.6800.7290.8450.064
PraNet [45]0.7690.6630.7100.8240.0940.8600.7630.7890.9070.0440.7890.6290.6710.8610.0450.8220.7240.7620.8760.059
ANet [16]0.6820.4840.5410.6850.126---------------
MirrorNet [46]0.7850.7190.7540.8480.077---------------
SINet [24]0.7450.6440.7020.8040.0920.8720.8060.8270.9360.0340.7760.6310.6790.8640.0430.8080.7230.7690.8710.058
PFNet [2]0.7820.6950.7460.8420.0850.8820.8100.8280.9310.0330.8000.6600.7010.8770.0400.8290.7450.7840.8880.053
UJSC [25]0.8000.7280.7720.8590.0730.8910.8330.8470.9450.0300.8090.6840.7210.8840.0350.8420.7710.8060.8980.047
SLSR [1]0.7870.6960.7440.8380.0800.8900.8220.8410.9350.0300.8040.6730.7150.8800.0370.8400.7660.8040.8950.048
MGL-R [14]0.7750.6730.7260.8120.0880.8930.8130.8340.9180.0300.8140.6660.7110.8520.0350.8330.7400.7820.8670.052
C2FNet [13]0.7960.7190.7620.8540.0800.8880.8280.8440.9350.0320.8130.6860.7230.8900.0360.8380.7620.7950.8970.049
UGTR [12]0.7840.6840.7360.8220.0860.8870.7940.8200.9100.0310.8170.6660.7110.8530.0360.8390.7470.7870.8750.052
SINet_V2 [11]0.8200.7430.7820.8820.0700.8880.8160.8350.9420.0300.8150.6800.7180.8870.0370.8470.7700.8050.9030.048
FAPNet [3]0.8150.7340.7760.8650.0760.8930.8250.8420.9400.0280.8220.6940.7310.8880.0360.8510.7750.8100.8990.047
Ours0.8080.7410.7820.8730.0710.8910.8370.8510.9480.0280.8180.7000.7340.8930.0340.8450.7800.8130.9050.046
Table 2. Quantitative comparison of different methods on four COD10K testing dataset categories. Here, “↑” (“↓”) means that the larger (smaller) the better. The best three results in each column are marked in red, green, and blue.
Table 2. Quantitative comparison of different methods on four COD10K testing dataset categories. Here, “↑” (“↓”) means that the larger (smaller) the better. The best three results in each column are marked in red, green, and blue.
COD10K-AmphibianCOD10K-AquaticCOD10K-FlyingCOD10K-Terrestrial
S m F β w F β E m MAE S m F β w F β E m MAE S m F β w F β E m MAE S m F β w F β E m MAE
EGNet [38]0.7760.5880.6500.8430.0560.7120.5150.5840.7840.0910.7690.5580.6210.8380.0460.7130.4670.5310.7940.056
PoolNet [40]0.7810.5840.6440.8230.0500.7370.5340.6070.7820.0780.7670.5390.6100.7970.0450.7070.4410.5080.7450.054
F3Net [31]0.8080.6570.7000.8460.0390.7280.5540.6110.7880.0760.7600.5710.6180.8180.0400.7120.4900.5380.7700.048
SCRN [39]0.8390.6650.7290.8670.0410.7800.6000.6740.8180.0640.8170.6080.6830.8400.0360.7580.5090.5880.7840.048
CSNet [41]0.8280.6490.7110.8570.0410.7680.5870.6560.8080.0670.8090.6100.6760.8380.0360.7440.5010.5660.7760.047
SSAL [42]0.7290.5600.6370.8170.0570.6320.4280.5090.7370.1010.7020.5040.5760.7950.0500.6470.4050.4710.7560.060
UCNet [43]0.8270.7170.7560.8970.0340.7670.6490.7030.8430.0600.8060.6750.7180.8860.0300.7420.5660.6170.8300.042
MINet [4]0.8230.6950.7320.8810.0350.7670.6320.6840.8310.0580.7990.6500.6970.8560.0310.7320.5360.5840.8020.043
ITSD [44]0.8100.6280.6790.8520.0440.7620.5840.6480.8110.0700.7930.5880.6450.8310.0400.7360.4960.5520.7770.051
PraNet [45]0.8420.7170.7500.9050.0350.7810.6430.6920.8480.0650.8190.6690.7070.8880.0330.7560.5650.6070.8350.046
ANet [16]--------------------
MirrorNet [46]--------------------
SINet [24]0.8200.7140.7560.8910.0340.7660.6430.6980.8540.0630.8030.6630.7070.8870.0310.7490.5770.6250.8450.042
PFNet [2]0.8480.7400.7750.9110.0310.7930.6750.7220.8680.0550.8240.6910.7290.9030.0300.7730.6060.6470.8550.040
UJSC [25]0.8410.7420.7690.9050.0310.8050.7050.7470.8790.0490.8360.7190.7520.9060.0260.7780.6240.6640.8630.037
SLSR [1]0.8450.7510.7830.9060.0300.8030.6940.7400.8750.0520.8300.7070.7450.9060.0260.7720.6110.6550.8550.038
MGL-R [14]0.8540.7340.7700.8860.0280.8070.6880.7360.8550.0510.8390.7010.7430.8730.0260.7850.6060.6510.8230.036
C2FNet [13]0.8490.7520.7790.8990.0300.8070.7000.7410.8820.0520.8400.7240.7590.9140.0260.7830.6270.6640.8720.037
UGTR [12]0.8570.7380.7740.8960.0290.8100.6860.7340.8550.0500.8430.6990.7440.8730.0260.7890.6060.6530.8230.036
SINet_V2 [11]0.8580.7560.7880.9160.0300.8110.6960.7380.8830.0510.8390.7130.7490.9080.0270.7870.6230.6620.8660.039
FAPNet [3]0.8540.7520.7830.9140.0320.8210.7170.7570.8870.0490.8450.7250.7600.9060.0250.7950.6390.6780.8680.037
Ours0.8630.7760.8030.9250.0280.8170.7250.7640.8950.0480.8440.7330.7630.9130.0240.7850.6380.6750.8690.037
Table 3. Comparisons of the number of parameters, FLOPs, and FPS corresponding to COD methods. All evaluations follow the inference settings in the corresponding papers.
Table 3. Comparisons of the number of parameters, FLOPs, and FPS corresponding to COD methods. All evaluations follow the inference settings in the corresponding papers.
MethodOursFAPNet [3]SINet_V2 [11]UGTR [12]C2FNet [13]MGL-R [14]SINet [24]SLSR [1]UJSC [25]PFNet [2]
Params.31.554 M29.524 M26.976 M48.868 M28.411 M63.595 M48.947 M50.935 M217.982 M46.498 M
FLOPs43.435 G59.101 G24.481 G1.007 T26.167 G553.939 G38.757 G66.625 G112.341 G53.222 G
FPS34.01328.47638.94815.44636.94112.79334.08332.54718.24629.175
Table 4. Ablation studies on three testing datasets.
Table 4. Ablation studies on three testing datasets.
MethodCAMO DatasetCOD10K DatasetNC4K Dataset
S m F β w F β E m MAE S m F β w F β E m MAE S m F β w F β E m MAE
Basic0.7970.7120.7570.8610.0780.8000.6600.6980.8780.0400.8320.7490.7850.8930.052
Basic+UFM0.8040.7280.7740.8650.0760.8140.6850.7250.8830.0360.8410.7650.8030.8940.049
Basic+NFFM0.8060.7380.7810.8730.0720.8170.7000.7340.8920.0350.8430.7770.8100.9030.046
Ours0.8080.7410.7820.8730.0710.8180.7000.7340.8930.0340.8450.7800.8130.9050.046
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, T.; Wang, J.; Wang, R. Camouflaged Object Detection with a Feature Lateral Connection Network. Electronics 2023, 12, 2570. https://doi.org/10.3390/electronics12122570

AMA Style

Wang T, Wang J, Wang R. Camouflaged Object Detection with a Feature Lateral Connection Network. Electronics. 2023; 12(12):2570. https://doi.org/10.3390/electronics12122570

Chicago/Turabian Style

Wang, Tao, Jian Wang, and Ruihao Wang. 2023. "Camouflaged Object Detection with a Feature Lateral Connection Network" Electronics 12, no. 12: 2570. https://doi.org/10.3390/electronics12122570

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop