Next Article in Journal
Individual Contrast Preferences in Natural Images
Previous Article in Journal
Fully Self-Supervised Out-of-Domain Few-Shot Learning with Masked Autoencoders
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Features Split and Aggregation Network for Camouflaged Object Detection

1
HDU-ITMO Joint Institute, Hangzhou Dianzi University, Hangzhou 310018, China
2
School of Automation, Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
J. Imaging 2024, 10(1), 24; https://doi.org/10.3390/jimaging10010024
Submission received: 10 December 2023 / Revised: 8 January 2024 / Accepted: 15 January 2024 / Published: 18 January 2024

Abstract

:
Higher standards have been proposed for detection systems since camouflaged objects are not distinct enough, making it possible to ignore the difference between their background and foreground. In this paper, we present a new framework for Camouflaged Object Detection (COD) named FSANet, which consists mainly of three operations: spatial detail mining (SDM), cross-scale feature combination (CFC), and hierarchical feature aggregation decoder (HFAD). The framework simulates the three-stage detection process of the human visual mechanism when observing a camouflaged scene. Specifically, we have extracted five feature layers using the backbone and divided them into two parts with the second layer as the boundary. The SDM module simulates the human cursory inspection of the camouflaged objects to gather spatial details (such as edge, texture, etc.) and fuses the features to create a cursory impression. The CFC module is used to observe high-level features from various viewing angles and extracts the same features by thoroughly filtering features of various levels. We also design side-join multiplication in the CFC module to avoid detail distortion and use feature element-wise multiplication to filter out noise. Finally, we construct an HFAD module to deeply mine effective features from these two stages, direct the fusion of low-level features using high-level semantic knowledge, and improve the camouflage map using hierarchical cascade technology. Compared to the nineteen deep-learning-based methods in terms of seven widely used metrics, our proposed framework has clear advantages on four public COD datasets, demonstrating the effectiveness and superiority of our model.

1. Introduction

When viewing an image or encountering a scene, it can be challenging to notice the object at first glance if the difference between the foreground and background is minimal [1,2,3,4,5,6]. The attempt by one or more objects to modify their traits (such as color, texture, etc.) to blend into their environment to avoid discovery is known as camouflage. There are two types of camouflage, as depicted in Figure 1. Natural camouflaged objects [7,8] are animals that use their inherent advantages to blend in with their surroundings and protect themselves. For instance, chameleons or other animals can change their color and physical appearance to match the hues and patterns of their surroundings. Additionally, artificial camouflaged objects were initially employed in battle, where soldiers and military gear employ camouflage to blend into their surroundings. In daily life, we can also see artificial camouflage, such as body art. Because of the characteristics of the camouflaged objects, the study of COD has not only scientific value but also significant engineering applications, such as surface defect detection [9], polyp segmentation [10], pest control, search, and rescue [11,12], and other applications.
The research on camouflage can be traced back to 1998. In recent years, COD has attracted more and more attention from researchers. Traditional models are mainly based on hand-crafted features (color, texture, optical flow, etc.) to describe the unified features of the object [13,14,15,16]. However, limited by hand-crafted features, the traditional model cannot work well when the background environment changes [17,18]. To address this, deep-learning-based techniques for COD have been developed, which utilize deep features automatically learned by the network from extensive training images. These features are more generic and effective than hand-crafted features. For example, Fan et al. [19] designed the first deep-learning-based model, SINet, which simulates the human visual mechanism, especially used for COD. Zheng et al. [20] successfully predicted the camouflage map using the short connection of the frame. However, there are still some shortcomings in the existing models. Specifically, (1) they cannot deeply explore high-level features, leading to imprecisely locating small objects. (2) There is no particularly effective method for integrating high-level and low-level features, even directly discarding low-level features, resulting in suboptimal performance in handling object edges.
Inspired by the description above, we propose a new model for COD named features split and aggregation network (FSANet) to solve the above problems, as shown in Figure 2, which primarily consists of three modules. Taking into account how low-level features and high-level features contribute differently to the creation of camouflage maps [19,20], we use the backbone to extract five feature layers and divide them into two parts with the second layer as the boundary. The first stage consists of the backbone’s first two layers, simulating a human’s cursory examination of the scene to gather spatial details (such as the edge, texture, etc.). The SDM module is used in this phase to fuse the features to create a cursory impression. The second stage consists of the backbone’s last three layers, simulating a person’s additional observation and reworking of imperfect scenes. Specifically, both the same and different information is observed for the same feature across different viewing angles. We employ an ordinary convolution layer and TEM [1] module to mimic the human evaluation of features from various angles. After that, we fuse these features to enhance similar features. To avoid detail distortion and filter out noise by feature multiplication, we simultaneously utilize side-join multiplication (SJM), as shown by the red line in Figure 2. These operations constitute the CFC module. Finally, to obtain more thorough detection results, we build the HFAD module to thoroughly mine effective information from the two stages. We guide the fusion of low-level features by using high-level semantic information, and we enhance the camouflage map generated in the earlier stage by using a hierarchical cascade technique.
Overall, we can summarize our main contributions as follows:
  • We simulate the human observation camouflage scenes to propose a new COD method that includes the spatial detail mining module, the cross-scale feature combination module, and the hierarchical feature aggregation decoder. We rigorously test our model against nineteen others using four public datasets (CAMO [21], CHAMELEON [22], COD10K [1], and NC4K [2]) and evaluate it across seven metrics, where it demonstrates clear advantages.
  • To fully mine spatial detail information, we design a spatial detail mining module that interacts with first-level feature information, simulating the human’s cursory examination. To effectively mine information in high-level features, we designed a cross-scale feature combination module to strengthen high-level semantic information by combining features from adjacent scales, simulating humans’ evaluation of features from various angles. Furthermore, we build a hierarchical feature aggregation module to fully integrate multi-level deep features, simulating humans’ aggregation and processing of information.

2. Related Works

This section discusses COD based on deep learning and context-aware deep learning, both of which are related to our model.

2.1. Camouflaged Object Detection (COD)

Camouflaged Object Detection (COD) has become an important area of research for identifying objects that are blended in with their surroundings. In this emerging field, significant contributions have been made.
Fan et al. [19] utilized a search module (SM) alongside a partial decoder component (PDC) [23] to enhance the accuracy of initial detection zones, fine-tuning the identification of camouflaged objects by focusing on salient features within the rough areas. Sun et al. [24] developed C2F-Net, which employs multi-scale channel attention to guide the fusion of features across different levels. Their approach ensures that both local nuances and global context are considered, thus improving the detection of objects across various scales. Mei et al. [3] introduced PFNet, which cleverly combines high-level feature maps with inverted predictions. By integrating these with the current layer’s attributes, and processing them through a context exploration block, the network is able to effectively reduce false positive and negative detections by strategically employing subtraction techniques. Li et al. [25] proposed JCOSOD, which considers the uncertainties inherent in fully labeling camouflaged objects. They used a full convolutional discriminator to gauge the confidence in predictions, and an adversarial training strategy was applied to refine the model’s ability to estimate prediction confidence.
Together, these advancements reflect a growing sophistication in COD, showing a trend towards more nuanced algorithms capable of distinguishing objects that are naturally or artificially designed to be hard to detect.

2.2. Context-Aware Deep Learning

Contextual information is important in object segmentation tasks as it has the ability to improve feature representation and, in turn, improve performance. Efforts have been made to improve contextual information. Stars et al. [26] proposed a salience detection algorithm based on four psychological principles. The model defines the algorithm using local low-level considerations, global considerations, and visual organization rules. High-level factors are used for post-processing and are helpful in producing compact, attractive, and rich information. Chen et al. [27] created ASPP, which collects contextual data using various dilated convolutions. They proposed an approach that, in the end, produces accurate semantic segmentation results based on the DCNN’s capability to detect objects and the fcCRF’s capability to localize objects with fine detail. To improve the features of the local context, Tan et al. [28] employed LCANet to merge the local area context and the global scene context in a coarse-to-fine framework.

3. The Proposed Method

In this section, we present the overall architecture of FSANet before delving into the specifics of each module. Finally, we discuss the training loss function of the proposed model.

3.1. Overall Architecture

The overall architecture of FSANet can be seen in Figure 2, which consists primarily of the spatial detail mining module, the cross-scale feature combination module, and the hierarchical feature aggregation decoder to endow the model with the ability to detect camouflaged objects. Specifically, for the input image I R W × H × 3 , we use Res2Net-50 [29] as the backbone to extract five different levels of information, denoted as F i , i 1 , 2 , 3 , 4 , 5 . The resolutions of each layer are H k , W k , k = 4 , 4 , 8 , 16 , 32 . We divide the backbone into two parts, with the second layer as a boundary. The first two levels of features are low-level fine-detail features, including spatial details (such as edge, texture, etc.), while the last three layers F i , i 3 , 4 , 5 are high-level semantic features that include specific details (such as semantic information, position, etc.). We obtain the spatial details from low-level features by using the spatial detail mining module and combine the features to provide a superficial impression, denoted by P 1 . However, it contains more redundant information. For high-level features, we design a cross-scale feature combination module to obtain three layers of high-level features denoted as P i , i 2 , 3 , 4 , each with different specific semantic information. Finally, we employ the hierarchical feature aggregation decoder, which utilizes the high-level features layer to refine and fuse the low-level features layer by layer, yielding the prediction map of COD. Below are detailed descriptions of each key component.

3.2. Spatial Detail Mining (SDM)

Because the camouflaged object is very similar to the background, the extracted low-level features contain rich spatial detail and adjacent features have great similarity, but they also contain more noise information. Therefore, we use this module to find similar features and eliminate noise, simulating a human’s cursory examination of the scene to gather spatial details. To retain the fine-detail information in the low-level features while discarding the noise information and unsuitable features, we combine the adjacent features of the first two layers F i C , i 1 , 2 , with element-wise multiplication on these two features to extract shared features, followed by element-wise addition. We obtain R 1 and R 2 after these operactions. Then, we concatenate R 1 and R 2 to obtain R C . Global average pooling is applied to R C to weight the features, and these are further enhanced with local information through element-wise multiplication with the original feature R C . Finally, the channels of feature are reduced to 32 through convolution to obtain feature P 1 with the size H k , W k , 32 , k = 4 . The spatial detail mining process can be described as follows:
R 1 = F 1 C F 2 C F 1 C R 2 = R 1 F 2 C R C = C o n c a t C B R R 1 , C B R R 2 P 1 = C B R R C C B R G R C
where C B R · represents the C o n v + B N + R e L U operation and C o n c a t · represents concatenate operation in dim = 1, G · represents global average pooling operation, which is used to establish relationships between feature maps and categories.

3.3. Cross-Scale Feature Combination (CFC)

Different types of camouflaged objects have varying colors, physical traits, and camouflage techniques. Similarly, objects of the same type that are camouflaged can have different camouflage methods and sizes in various environments, making it more challenging to locate them. In studies on biological vision, researchers have been discussing challenges of perspective. Viewpoint-invariant theories and viewpoint-dependent theories have been proposed [30,31,32]. Viewpoint-invariant theories assert that a particular object can be recognized from diverse viewing angles while maintaining its properties. On the other hand, viewpoint-dependent theories suggest that object recognition from different viewing angles may be effective. Separately, Tarr et al. [33] proposed a multi-view model in which objects can be represented by a series of images of familiar viewpoints, with each view describing a different view-specific object characterization.
Inspired by this, we realize that using various receptive fields with reduced-channel operations can provide additional feature information about the objects. Thus, we design the cross-scale feature combination module, which processes features and adjacent features differently to obtain different viewpoints and finally fuses them to obtain advanced features, simulating the person’s additional observation and reworking of imperfect scenes.
Specifically, we utilize the C o n v 3 to handle the features F i , i 2 , 3 , 4 , 5 to preserve object boundaries and enhance local context information, and we use the TEM [1] to handle the features F i , i 3 , 4 , 5 to capture multi-scale information further. This enables us to obtain F i C , i 2 , 3 , 4 , 5 and F i R , i 3 , 4 , 5 . All features’ channels are adjusted to 32. After that, we use NCD*, which selective removes upsampling from NCD [1] to ensure dimensional consistency, fine-tune, and effectively combine features from different viewpoints; the inputs are F i 1 C , F i C , and F i R with output F i N .
However, since camouflaged objects are relatively blurred, using NCD* may result in detail and other useful information loss while enhancing similar features. Therefore, we use side-join multiplication to re-add details to the output features and filter out noise through multiplication to further enhance the object’s features; obtain P i , i 3 , 4 , 5 with the size H k , W k , 32 , k = 4 , 4 , 8 . This operation is depicted by the red line in Figure 2. The cross-scale feature combination process can be described as follows:
F i N = N CBR F i 1 , C B R F i , T F i i = 3 F i N = N R F i 1 , C B R F i , T F i i = 4 , 5 P i 1 = F i N R F i i = 3 , 4 , 5
where C B R · represents C o n v + B N + R e L U operation, N · represents neighbor connection decoder (NCD*), T · represents texture-enhanced module (TEM).

3.4. Hierarchical Feature Aggregation Decoder (HFAD)

We obtain improved features { P 1 , P 2 , P 3 , P 4 } using the method described above. The next crucial problem is how to successfully bridge the context and fuse these features. To address this, our model employs hierarchical cascade technology, which gradually guides the fusion of low-level features using high-level semantics, simulating human processing of aggregated information obtained from different sources. We regard the process of fusing rich features as a decoder. Formally, the hierarchical feature aggregation decoder contains four inputs, as shown in Figure 2. The module’s general structure is an inverted triangle hierarchical structure, primarily consisting of C o n v 3 + B N + R e L U layers and element-wise multiplication to extract similarities between various features. To ensure that cascade processes may be completed, we resize the features to an appropriate size in the process by using an upsampling operation.
Specifically, we first apply an upsampling operation for P 4 to make it the same shape as P 3 , and we multiply them to obtain S 3 . Then, we upsample P 3 and P 4 , respectively, to match the size of P 2 , and multiply them to obtain S 2 . The same operations are applied to obtain S 1 . This progressive method is characterized as follows:
P 3 U = C B R ( δ 2 P 3 ) S 3 = C B R ( δ 2 P 4 ) P 3 S 2 = C B R ( δ 4 P 4 ) P 3 U P 2 S 1 = C B R ( δ 4 P 4 ) P 3 U C B R P 2 P 1
where C B R · represents C o n v + B N + R e L U operation. To ensure that the candidate features have the same size as each other, we use upsampling operation before element-wise multiplication; δ 2 · means 2 × upsampling operation by executing the bilinear interpolation, and δ 4 · 4 × upsampling operation.
After performing the above operation, we obtain three refined features, denoted by { S 1 , S 2 , S 3 } . Then, we use the concatenation operation with C o n v 3 + B N + R e L U layers to enhance the feature step by step, obtaining S 3 C a t , S 2 C a t , and S 1 C a t . Finally, we use a convolution layer to reduce the channels and obtain the final prediction map M R W × H × 1 . The following formulas express this process:
S 3 C a t = C o n c a t { S 3 , C B R ( δ 2 P 4 ) } S 2 C a t = C o n c a t { S 2 , C B R ( δ 2 ( S 3 C a t ) ) } S 1 C a t = C o n c a t { S 1 , C B R S 1 C a t } M = C o n v C B R S 1 C a t
where C B R · represents C o n v 3 + B N + R e L U operation. δ 2 · means 2 × upsampling operation by executing the bilinear interpolation. C o n c a t · represents concatenating operation in dim = 1. C o n v means 1 × 1 convolutional layer. Following these operations, we obtain the prediction map.

3.5. Loss Function

The binary cross-entropy (BCE) [34] loss, which highlights pixel-level differences, disregards discrepancies between neighboring pixels, and equally weights foreground and background pixels, is often employed in the binary classification problem. For object detection and segmentation, the IoU is a commonly used assessment metric that emphasizes global structure. Inspired by [35,36], we adopt weighted BCE loss and weighted IOU loss as the combined loss. Weighted BCE and weighted IoU losses place a greater focus on hard samples compared to regular BCE and IoU losses. The following formula shows how we define our loss:
L = L I O U w + L B C E w
where L I O U w and L B C E w denote BCE loss and IoU loss, respectively. It has been proven successful to apply the same parameter definition and setup as [36,37].
The four supervision maps in this model are all closely supervised, and their locations are illustrated in Figure 2. Here, each map is enlarged through upsampling to align its dimensions with the GT. The total loss can be calculated using the formula below:
L all = i = 1 4 L l i , G
The G represents the GT, and L l i , G represent the loss calculation between each output and the ground truth, respectively.

4. Experimental Results

In this section, we will delve into greater detail about the benchmark datasets in the COD field, evaluation measures, experimental setup, and ablation study.

4.1. Datasets and Implementation

We conducted extensive comparisons on four publicly available COD datasets (CAMO [21], CHAMELEON [22], COD10K [1], and NC4K [2]) to fully validate our method.
CAMO [21] dataset includes 1250 photos and was suggested in 2019. It contains two types of scenes: indoor scenes (artworks) and outdoor scenes (disguised humans/animals). The dataset also includes a few images that are not camouflaged.
CHAMELEON [22] dataset includes 76 natural images, each of which is matched with an instance-level annotation. This dataset collection primarily focuses on creatures that are disguised in complicated backgrounds, making it challenging for humans to identify them from the environment.
COD10K [1] dataset includes 10K images, which are classified as 5066 camouflaged, 3000 background, and 1934 non-camouflaged images. The dataset is divided into five main categories and sixty-nine subcategories, including images of land, sea, air, and amphibians in camouflaged scenes as well as images of non-camouflaged environments.
NC4K [2] dataset includes 4121 images, which is the largest existing COD testing dataset. The dataset’s camouflaged scenes can be generally classified into two categories: natural camouflaged and artificial camouflaged, and the majority of the visual scenes in this collection are also naturally hidden.
Implementation Details: In this instance, we train our model using the same training dataset as stated in [1], which consists of 4040 images from the COD10K and CAMO datasets. The remaining images are used as testing datasets. Additionally, the training dataset is strengthened by randomly flipping images to increase the sufficiency of the network training, and each training image’s size is changed to 352 × 352 in the training phase. Our model is built with PyTorch and run on a PC with an NVIDIA GTX 2080Ti GPU. Parts of the parameters are initialized with Res2Net-50 [29] during the training process, while the remaining parameters are randomly initialized. The network is optimized using the Adam algorithm [38], with the initial learning rate, batch size, and maximum epoch number set to 10 4 , 16, and 100.

4.2. Evaluation Metrics

We use seven common metrics to evaluate and conduct a quantitative comparison of different models on COD datasets, including precision–recall ( PR ) curve, S-measure [39] ( S m ), F-measure [40] ( F β ), weighted F-measure [41] ( F β w ), E-measure [42] ( E m ), and mean absolute error ( MAE ). Please refer to the evaluation code for details in https://github.com/DengPingFan/CODToolbox (accessed on 16 January 2024).
Precision and recall are common metrics used to evaluate how well the model works. The recall value is used as the horizontal coordinate and the precision value as the vertical coordinate to create a coordinate system. We can calculate the associated precision and recall scores to evaluate the effectiveness of the models.
S-measure [39] is used to determine the structural similarities between the prediction map and the related ground truth, which is defined as
S = α × S o + ( 1 α ) × S r
where S o represents the structural similarity measurement based on the object level and S r represents the region-based similarity. According to [39], the α is set to 0.5 .
F-measure [40] is used to calculate the weighted summation average of the precision and recall under non-negative weights. It is often used to compare the similarity of two images. The formula can be expressed as
F β = 1 + β 2 P R β 2 P + R
where P represents precision and R represents recall. We set β 2 to 0.3, as suggested in [43], to emphasize precision. To improve the accuracy and completeness metrics, we determine the weights of recall and precision, as similarly conducted in [41]. The following is the formula:
F β w = 1 + β 2 P w R w β 2 P w + R w
The parameters are the same as F β , and w represents the weighted harmonic mean of the precision and recall.
E-measure [42] assesses the similarity between the prediction map and the ground truth by using the pixel significance value and the average significance value. The formula is as follows:
E = 1 W × H x = 1 W y = 1 H f ( i ) ,
where f ( · ) = S ( x , y ) G ( x , y ) stands for the enhanced alignment term, which is used to record statistics at the image level and pixel level. The image’s width and height are denoted by W and H.
MAE is used to quantify the average absolute difference between the model’s output and the input’s ground truth, which is the pixel-level error evaluation index. The formula can be written as follows:
MAE = 1 W × H i = 1 W × H S ( i ) G ( i ) ,
where S ( i ) represents the predicted map. G ( i ) represents the GT. W and H denote the image’s width and height.

4.3. Comparison with the State-of-the-Art Methods

In this part, we denote our model FSANet as “ours” and include some models from salient object detection and medical image segmentation to compare. We evaluate a total of ten models in salient object recognition and medical image segmentation, with nine models in COD, including EGNet [44], F3Net [37], SCRN [45], PoolNet [46], CSNet [47], SSAL [48], UCNet [49], MINet [50], ITSD [51], PraNet [10], PFNet [3], UJSC [25], SLSR [2], SINet [19], MGL-R [52], C2FNet [24], UGTR [53], SINet_V2 [1], and FAPNet [8]. The results of all these methods were obtained from publicly available data, created by the model, or retrained using the author’s code.

4.3.1. Quantitative Comparison

For COD datasets, we first present PR curves and F-measure curves for quantitative comparison. As shown in Figure 3, we observe that our model outperforms the other models in terms of the PR curve and F β curves. This is due to the feature fusion approach we use (see Section 3.4 for details).
Moreover, as listed in Table 1, our model obtains superior scores on four COD datasets under five public camouflaged map quality evaluation metrics. For instance, our model outperforms all the advanced models in five evaluation metrics for the CHAMELEON and COD10K datasets, achieving M A E of 0.026 and 0.034, respectively, which is 7.14% and 5.56% lower than FAPNet. Similarly, compared to FAPNet, the F β also improves by 2.06% and 1.66% on the CAMO and CHAMELEON datasets, respectively. Although our model’s F β w scores rank second among the available models for the NC4K dataset, their scores only decrease by 0.26% . Furthermore, as shown in Table 2, our model achieves great results in categories within the COD10K dataset. For instance, in COD10K-Amphibian, compared with FAPNet, our model’s MAE decreases by 15.63% .
Overall, through Figure 3 and Table 1 and Table 2, the excellence and efficiency of our model, which has attained SOTA performance, are readily apparent.

4.3.2. Qualitative Comparison

We carry out several visual contrast experiments and provide corresponding images to make a qualitative comparison for all models. As shown in Figure 4, our model’s detection results are more comparable to the GT, indicating that our results are more complete and precise than those of the other models. In general, our model has two major advantages:
(a) Object placement accuracy: In the first, second, seventh, and eighth rows of Figure 4, we can see that our model’s outcomes closely resemble the GT. In contrast, other deep-learning-based models, (e.g., (d) FAPNet [8], (e) SINet_V2 [1], (l) SINet [19], etc.), shown in Figure 4, find the object but mistake a portion of the background for the object in the process.
(b) Advantages of edge details for optimization: In the third, fourth, ninth, and tenth rows of Figure 4, our model is capable of precisely locating the object and properly identifying microscopic details. For other models, (e.g., (d) FAPNet [8], (e) SINet_V2 [1], (g) C2FNet [24], (l) SINet [19], etc.), although they may detect the object’s major portion, the object’s boundary is unclear, tailing is a serious occurrence, and the edge details are not readily apparent.
Based on the above comparisons, we can indisputably establish the efficacy and superiority of the FSANet that we present. When it comes to identifying camouflaged objects, whether they are inside the object or on its edge, our model performs better than the other models.

4.4. Ablation Studies

In this section, we conduct a thorough experiment on two COD datasets to demonstrate the efficacy of each model component. Table 3 displays the quantitative comparison; Figure 5A–E display the qualitative comparisons. We conduct experiments on the SDM, TEM, SJM, and HFAD modules to validate their effectiveness. The following are the details of the implementation of the experiment.
Table 3 demonstrates how various operations can be used to further enhance the model’s performance. When all the proposed modules are combined, our model performs the best, particularly when applied to the CAMO dataset, where our model performs better than any other stage. With relation to the model without many-to-many side-join multiplication, a one-to-many side-join multiplication (No.#5) is used, and the S m and M A E of ours in CAMO are improved by 3.67% and 5.56% , respectively. When we remove TEM from the CFC module and use C o n v 3 instead (No.#1), each of the two datasets’ metrics are noticeably worse; especially, S m and M A E show the most obvious decline. Experiment No.#3 verifies the effectiveness of HFAD; if we remove the HFAD, while the F β in the COD10K dataset improves marginally compared to ours, other indicators of our model significantly decrease; in particular, E m in the CAMO dataset declines 19.25% .
We also provide the prediction map of five ablation settings to visually demonstrate the effectiveness of our strategy. When we do not use TEM to enlarge the receptive field in the CFC module (No.#1), according to Figure 5A, the camouflaged objects can roughly be resolved, but the edge details are not smooth enough. As shown in Figure 5B,C, we introduce the SDM module and CFC module (No.#2, #3) to address the issue that the prediction map is void because the high-level semantic characteristics do not contain image spatial details and other information. Furthermore, we independently confirm the many-to-many and one-to-many side-join multiplication for the CFC module (No.#4, #5), as shown in Figure 5D,E; we improve the detection accuracy by re-adding the information that NCD* overlooked to the prediction feature using the many-to-many side-join multiplication technique that we devised.
It is demonstrated that our model fully complies with the anticipated design standards based on the qualitative analysis and quantitative analysis of the aforementioned ablation study.

4.5. Failure Cases and Analysis

As shown in Table 4, we evaluate the inference speed of our model in comparison to other models. The findings demonstrate that, despite our model’s successful utilization of the SDM, CFC, and HFAD modules and achievement of the primary design goals, a significant amount of duplication still exists in our model. Thus, the model will be further developed from the efficiency standpoint. On the other hand, the first and second rows of Figure 6 depict certain failed scenarios where numerous camouflaged objects are present but only one can be detected by our model. This could be because the CFC module is being used, which focuses more on scenarios where there is only one camouflaged object and filters out other objects as background data. In our subsequent research, we will further explore methods for multi-object detection, such as instance segmentation [54]. Furthermore, the object’s edge processing is sloppy, and the background is wrongly identified as the foreground when using artificial camouflage, as demonstrated in the third and fourth rows. This could be as a result of the SDM module’s limited ability to effectively filter out interference data. The aforementioned results offer fresh perspectives for our upcoming model design.

5. Conclusions

In this paper, we propose a new model named features split and aggregation network (FSANet) to detect camouflaged objects, which can be divided into three modules to simulate the three-stage detection process of the human visual mechanism when viewing a camouflaged scene. To begin, we divide the backbone into two stages. The SDM module is used in the first stage to perform information interaction of first-level features to fully mine the spatial details (such as edge, texture, etc.) and fuse the features to create a cursory impression. In parallel, high-level semantic information from several sensory areas is mined by using the CFC module. Furthermore, we apply side-join multiplication in CFC to prevent detail distortion and reduce noise. Finally, we configure HFAD to completely fuse the effective information between the two stages to acquire more thorough detection results. Through in-depth experiments on four public camouflaged datasets, we observe that both quantitative and qualitative results verify the effectiveness of our methodology. These results prove the validity and superiority of our model. However, our model still has some limitations. When there are numerous camouflaged objects, our model can only detect one. Additionally, for artificially camouflaged objects, our model fails to perform fine-grained edge processing. The above results provide new directions for our upcoming model design. Furthermore, we aspire for our model to be adaptable across a broader range of applications, including but not limited to industrial defect detection and medical image segmentation and detection.

Author Contributions

Methodology, Z.Z. and T.W.; Software, T.W.; Supervision, J.W.; Writing—Original draft, Z.Z. and T.W.; Writing—Review and editing, J.W. and Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fan, D.P.; Ji, G.P.; Cheng, M.M.; Shao, L. Concealed object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 6024–6042. [Google Scholar] [CrossRef]
  2. Lv, Y.; Zhang, J.; Dai, Y.; Li, A.; Liu, B.; Barnes, N.; Fan, D.P. Simultaneously localize, segment and rank the camouflaged objects. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11591–11601. [Google Scholar]
  3. Mei, H.; Ji, G.P.; Wei, Z.; Yang, X.; Wei, X.; Fan, D.P. Camouflaged object segmentation with distraction mining. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8772–8781. [Google Scholar]
  4. Ren, J.; Hu, X.; Zhu, L.; Xu, X.; Xu, Y.; Wang, W.; Deng, Z.; Heng, P.A. Deep texture-aware features for camouflaged object detection. IEEE Trans. Circuits Syst. Video Technol. 2021, 33, 1157–1167. [Google Scholar] [CrossRef]
  5. Jiang, X.; Cai, W.; Zhang, Z.; Jiang, B.; Yang, Z.; Wang, X. MAGNet: A camouflaged object detection network simulating the observation effect of a magnifier. Entropy 2022, 24, 1804. [Google Scholar] [CrossRef] [PubMed]
  6. Zhuge, M.; Fan, D.P.; Liu, N.; Zhang, D.; Xu, D.; Shao, L. Salient object detection via integrity learning. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 3738–3752. [Google Scholar] [CrossRef]
  7. Merilaita, S.; Scott-Samuel, N.E.; Cuthill, I.C. How camouflage works. Philos. Trans. R. Soc. Biol. Sci. 2017, 372, 20160341. [Google Scholar] [CrossRef] [PubMed]
  8. Zhou, T.; Zhou, Y.; Gong, C.; Yang, J.; Zhang, Y. Feature Aggregation and Propagation Network for Camouflaged Object Detection. IEEE Trans. Image Process. 2022, 31, 7036–7047. [Google Scholar] [CrossRef] [PubMed]
  9. Le, X.; Mei, J.; Zhang, H.; Zhou, B.; Xi, J. A learning-based approach for surface defect detection using small image datasets. Neurocomputing 2020, 408, 112–120. [Google Scholar] [CrossRef]
  10. Fan, D.P.; Ji, G.P.; Zhou, T.; Chen, G.; Fu, H.; Shen, J.; Shao, L. Pranet: Parallel reverse attention network for polyp segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2020; pp. 263–273. [Google Scholar]
  11. Lidbetter, T. Search and rescue in the face of uncertain threats. Eur. J. Oper. Res. 2020, 285, 1153–1160. [Google Scholar] [CrossRef]
  12. Li, G.; Liu, Z.; Zhang, X.; Lin, W. Lightweight salient object detection in optical remote-sensing images via semantic matching and edge alignment. IEEE Trans. Geosci. Remote. Sens. 2023, 61, 1–11. [Google Scholar] [CrossRef]
  13. Zhang, X.; Zhu, C.; Wang, S.; Liu, Y.; Ye, M. A Bayesian approach to camouflaged moving object detection. IEEE Trans. Circuits Syst. Video Technol. 2016, 27, 2001–2013. [Google Scholar] [CrossRef]
  14. Feng, X.; Guoying, C.; Richang, H.; Jing, G. Camouflage texture evaluation using a saliency map. Multimed. Syst. 2015, 21, 169–175. [Google Scholar] [CrossRef]
  15. Hou, J.Y.Y.H.W.; Li, J. Detection of the mobile object with camouflage color under dynamic background based on optical flow. Procedia Eng. 2011, 15, 2201–2205. [Google Scholar]
  16. Bi, H.; Zhang, C.; Wang, K.; Tong, J.; Zheng, F. Rethinking camouflaged object detection: Models and datasets. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 5708–5724. [Google Scholar] [CrossRef]
  17. Tankus, A.; Yeshurun, Y. Detection of regions of interest and camouflage breaking by direct convexity estimation. In Proceedings of the Proceedings 1998 IEEE Workshop on Visual Surveillance, Bombay, India, 2 January 1998; pp. 42–48. [Google Scholar]
  18. Guo, H.; Dou, Y.; Tian, T.; Zhou, J.; Yu, S. A robust foreground segmentation method by temporal averaging multiple video frames. In Proceedings of the 2008 International Conference on Audio, Language and Image Processing, Shanghai, China, 7–9 July 2008; pp. 878–882. [Google Scholar]
  19. Fan, D.P.; Ji, G.P.; Sun, G.; Cheng, M.M.; Shen, J.; Shao, L. Camouflaged object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2777–2787. [Google Scholar]
  20. Zheng, Y.; Zhang, X.; Wang, F.; Cao, T.; Sun, M.; Wang, X. Detection of people with camouflage pattern via dense deconvolution network. IEEE Signal Process. Lett. 2018, 26, 29–33. [Google Scholar] [CrossRef]
  21. Le, T.N.; Nguyen, T.V.; Nie, Z.; Tran, M.T.; Sugimoto, A. Anabranch network for camouflaged object segmentation. Comput. Vis. Image Underst. 2019, 184, 45–56. [Google Scholar] [CrossRef]
  22. Skurowski, P.; Abdulameer, H.; Błaszczyk, J.; Depta, T.; Kornacki, A.; Kozieł, P. Animal camouflage analysis: Chameleon database. Unpubl. Manuscr. 2018, 2, 7. [Google Scholar]
  23. Wu, Z.; Su, L.; Huang, Q. Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3907–3916. [Google Scholar]
  24. Sun, Y.; Chen, G.; Zhou, T.; Zhang, Y.; Liu, N. Context-aware cross-level fusion network for camouflaged object detection. arXiv 2021, arXiv:2105.12555. [Google Scholar]
  25. Li, A.; Zhang, J.; Lv, Y.; Liu, B.; Zhang, T.; Dai, Y. Uncertainty-aware joint salient object and camouflaged object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10071–10081. [Google Scholar]
  26. Goferman, S.; Zelnik-Manor, L.; Tal, A. Context-aware saliency detection. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 1915–1926. [Google Scholar] [CrossRef]
  27. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  28. Tan, J.; Xiong, P.; Lv, Z.; Xiao, K.; He, Y. Local context attention for salient object segmentation. In Proceedings of the Asian Conference on Computer Vision, Seattle, WA, USA, 19 June 2020. [Google Scholar]
  29. Gao, S.H.; Cheng, M.M.; Zhao, K.; Zhang, X.Y.; Yang, M.H.; Torr, P. Res2net: A new multi-scale backbone architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 652–662. [Google Scholar] [CrossRef]
  30. Wilson, K.D.; Farah, M.J. When does the visual system use viewpoint-invariant representations during recognition? Cogn. Brain Res. 2003, 16, 399–415. [Google Scholar] [CrossRef]
  31. Burgund, E.D.; Marsolek, C.J. invariant and viewpoint-dependent object recognition in dissociable neural subsystems. Psychon. Bull. Rev. 2000, 7, 480–489. [Google Scholar] [CrossRef] [PubMed]
  32. Li, Y.; Pizlo, Z.; Steinman, R.M. A computational model that recovers the 3D shape of an object from a single 2D retinal representation. Vis. Res. 2009, 49, 979–991. [Google Scholar] [CrossRef]
  33. Tarr, M.J.; Pinker, S. Mental rotation and orientation-dependence in shape recognition. Cogn. Psychol. 1989, 21, 233–282. [Google Scholar] [CrossRef] [PubMed]
  34. De Boer, P.T.; Kroese, D.P.; Mannor, S.; Rubinstein, R.Y. A tutorial on the cross-entropy method. Ann. Oper. Res. 2005, 134, 19–67. [Google Scholar] [CrossRef]
  35. Fan, D.P.; Ji, G.P.; Qin, X.; Cheng, M.M. Cognitive vision inspired object segmentation metric and loss function. Sci. Sin. Informationis 2021, 6, 6. [Google Scholar]
  36. Qin, X.; Zhang, Z.; Huang, C.; Gao, C.; Dehghan, M.; Jagersand, M. Basnet: Boundary-aware salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7479–7489. [Google Scholar]
  37. Wei, J.; Wang, S.; Huang, Q. F3Net: Fusion, feedback and focus for salient object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Seattle, WA, USA, 19 June 2020; Volume 34, pp. 12321–12328. [Google Scholar]
  38. Da, K. A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  39. Fan, D.P.; Cheng, M.M.; Liu, Y.; Li, T.; Borji, A. Structure-measure: A new way to evaluate foreground maps. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4548–4557. [Google Scholar]
  40. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-tuned salient region detection. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1597–1604. [Google Scholar]
  41. Margolin, R.; Zelnik-Manor, L.; Tal, A. How to evaluate foreground maps? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 248–255. [Google Scholar]
  42. Fan, D.P.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.M.; Borji, A. Enhanced-alignment measure for binary foreground map evaluation. arXiv 2018, arXiv:1805.10421. [Google Scholar]
  43. Zhang, D.; Han, J.; Li, C.; Wang, J.; Li, X. Detection of co-salient objects by looking deep and wide. Int. J. Comput. Vis. 2016, 120, 215–232. [Google Scholar] [CrossRef]
  44. Zhao, J.X.; Liu, J.J.; Fan, D.P.; Cao, Y.; Yang, J.; Cheng, M.M. EGNet: Edge guidance network for salient object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8779–8788. [Google Scholar]
  45. Wu, Z.; Su, L.; Huang, Q. Stacked cross refinement network for edge-aware salient object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7264–7273. [Google Scholar]
  46. Liu, J.J.; Hou, Q.; Cheng, M.M.; Feng, J.; Jiang, J. A simple pooling-based design for real-time salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3917–3926. [Google Scholar]
  47. Gao, S.H.; Tan, Y.Q.; Cheng, M.M.; Lu, C.; Chen, Y.; Yan, S. Highly efficient salient object detection with 100 k parameters. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 702–721. [Google Scholar]
  48. Zhang, J.; Yu, X.; Li, A.; Song, P.; Liu, B.; Dai, Y. Weakly-supervised salient object detection via scribble annotations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 19 June 2020; pp. 12546–12555. [Google Scholar]
  49. Zhang, J.; Fan, D.P.; Dai, Y.; Anwar, S.; Saleh, F.S.; Zhang, T.; Barnes, N. UC-Net: Uncertainty inspired RGB-D saliency detection via conditional variational autoencoders. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 19 June 2020; pp. 8582–8591. [Google Scholar]
  50. Pang, Y.; Zhao, X.; Zhang, L.; Lu, H. Multi-scale interactive network for salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 19 June 2020; pp. 9413–9422. [Google Scholar]
  51. Zhou, H.; Xie, X.; Lai, J.H.; Chen, Z.; Yang, L. Interactive two-stream decoder for accurate and fast saliency detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 19 June 2020; pp. 9141–9150. [Google Scholar]
  52. Zhai, Q.; Li, X.; Yang, F.; Chen, C.; Cheng, H.; Fan, D.P. Mutual graph learning for camouflaged object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12997–13007. [Google Scholar]
  53. Yang, F.; Zhai, Q.; Li, X.; Huang, R.; Luo, A.; Cheng, H.; Fan, D.P. Uncertainty-guided transformer reasoning for camouflaged object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 4146–4155. [Google Scholar]
  54. Pei, J.; Cheng, T.; Fan, D.P.; Tang, H.; Chen, C.; Van Gool, L. Osformer: One-stage camouflaged instance segmentation with transformers. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2022; pp. 19–37. [Google Scholar]
Figure 1. Examples of camouflaged objects; from left to right are natural camouflaged objects and artificial camouflaged objects.
Figure 1. Examples of camouflaged objects; from left to right are natural camouflaged objects and artificial camouflaged objects.
Jimaging 10 00024 g001
Figure 2. The overall architecture of the proposed FSANet, which can be divided into three key components; they are spatial detail mining module, cross-scale feature combination module, and hierarchical feature aggregation decoder. The input is camouflaged object I , and the result is prediction map M .
Figure 2. The overall architecture of the proposed FSANet, which can be divided into three key components; they are spatial detail mining module, cross-scale feature combination module, and hierarchical feature aggregation decoder. The input is camouflaged object I , and the result is prediction map M .
Jimaging 10 00024 g002
Figure 3. Quantitative evaluation of different models. The first row shows PR curves; the second row shows F-measure curves; (a,b) display the results for CAMO dataset; (c,d) display the results for CHAMELEON dataset.
Figure 3. Quantitative evaluation of different models. The first row shows PR curves; the second row shows F-measure curves; (a,b) display the results for CAMO dataset; (c,d) display the results for CHAMELEON dataset.
Jimaging 10 00024 g003
Figure 4. Visual comparison of our model and others on four COD testing datasets. (a) Input, (b) GT, (c) ours, (d) FAPNet [8], (e) SINet_V2 [1], (f) UGTR [53], (g) C2FNet [24], (h) MGL-R [52], (i) SLSR [2], (j) UJSC [25], (k) PFNet [3], (l) SINet [19], (m) PraNet [10], (n) ITSD [51], (o) MINet [50], (p) UCNet [49].
Figure 4. Visual comparison of our model and others on four COD testing datasets. (a) Input, (b) GT, (c) ours, (d) FAPNet [8], (e) SINet_V2 [1], (f) UGTR [53], (g) C2FNet [24], (h) MGL-R [52], (i) SLSR [2], (j) UJSC [25], (k) PFNet [3], (l) SINet [19], (m) PraNet [10], (n) ITSD [51], (o) MINet [50], (p) UCNet [49].
Jimaging 10 00024 g004
Figure 5. Qualitative comparisons of five experiments (AE) No.#1–No.#5.
Figure 5. Qualitative comparisons of five experiments (AE) No.#1–No.#5.
Jimaging 10 00024 g005
Figure 6. Failed cases: (a) input, (b) GT, (c) ours.
Figure 6. Failed cases: (a) input, (b) GT, (c) ours.
Jimaging 10 00024 g006
Table 1. Quantitative comparison of different methods on four COD testing datasets, which contain S-measure ( S m ), weighted F-measure ( F β w ), F-measure ( F β ), E-measure ( E m ), and mean absolute error ( MAE ). Here, “↑” (“↓”) means that the larger (smaller) the better. The best three results in each column are marked in red, green, and blue.
Table 1. Quantitative comparison of different methods on four COD testing datasets, which contain S-measure ( S m ), weighted F-measure ( F β w ), F-measure ( F β ), E-measure ( E m ), and mean absolute error ( MAE ). Here, “↑” (“↓”) means that the larger (smaller) the better. The best three results in each column are marked in red, green, and blue.
CAMO DatasetCHAMELEON DatasetCOD10K DatasetNC4K Dataset
S m F β w F β E m MAE S m F β w F β E m MAE S m F β w F β E m MAE S m F β w F β E m MAE
EGNet [44]0.7320.6040.6700.8000.1090.7970.6490.7020.8600.0650.7360.5170.5820.8100.0610.7770.6390.6960.8410.075
PoolNet [46]0.7300.5750.6430.7470.1050.8450.6910.7490.8640.0540.7400.5060.5760.7770.0560.7850.6350.6990.8140.073
F3Net [37]0.7110.5640.6160.7410.1090.8480.7440.7700.8940.0470.7390.5440.5930.7950.0510.7800.6560.7050.8240.070
SCRN [45]0.7790.6430.7050.7970.0900.8760.7410.7870.8890.0420.7890.5750.6510.8170.0470.8300.6980.7570.8540.059
CSNet [47]0.7710.6420.7050.7950.0920.8560.7180.7660.8690.0470.7780.5690.6350.8100.0470.7500.6030.6550.7730.088
SSAL [48]0.6440.4930.5790.7210.1260.7570.6390.7020.8490.0710.6680.4540.5270.7680.0660.6990.5610.6440.7800.093
UCNet [49]0.7390.6400.7000.7870.0940.8800.8170.8360.9300.0360.7760.6330.6810.8570.0420.8110.7290.7750.8710.055
MINet [50]0.7480.6370.6910.7920.0900.8550.7710.8020.9140.0360.7700.6080.6570.8320.0420.8120.7200.7640.8620.056
ITSD [51]0.7500.6100.6630.7800.1020.8140.6620.7050.8440.0570.7670.5570.6150.8080.0510.8110.6800.7290.8450.064
PraNet [10]0.7690.6630.7100.8240.0940.8600.7630.7890.9070.0440.7890.6290.6710.8610.0450.8220.7240.7620.8760.059
SINet [19]0.7450.6440.7020.8040.0920.8720.8060.8270.9360.0340.7760.6310.6790.8640.0430.8080.7230.7690.8710.058
PFNet [3]0.7820.6950.7460.8420.0850.8820.8100.8280.9310.0330.8000.6600.7010.8770.0400.8290.7450.7840.8880.053
UJSC [25]0.8000.7280.7720.8590.0730.8910.8330.8470.9450.0300.8090.6840.7210.8840.0350.8420.7710.8060.8980.047
SLSR [2]0.7870.6960.7440.8380.0800.8900.8220.8410.9350.0300.8040.6730.7150.8800.0370.8400.7660.8040.8950.048
MGL-R [52]0.7750.6730.7260.8120.0880.8930.8130.8340.9180.0300.8140.6660.7110.8520.0350.8330.7400.7820.8670.052
C2FNet [24]0.7960.7190.7620.8540.0800.8880.8280.8440.9350.0320.8130.6860.7230.8900.0360.8380.7620.7950.8970.049
UGTR [53]0.7840.6840.7360.8220.0860.8870.7940.8200.9100.0310.8170.6660.7110.8530.0360.8390.7470.7870.8750.052
SINet_V2 [1]0.8200.7430.7820.8820.0700.8880.8160.8350.9420.0300.8150.6800.7180.8870.0370.8470.7700.8050.9030.048
FAPNet [8]0.8150.7340.7760.8650.0760.8930.8250.8420.9400.0280.8220.6940.7310.8880.0360.8510.7750.8100.8990.047
Ours0.8210.7520.7920.8830.0680.8970.8410.8560.9520.0260.8220.6990.7340.8900.0340.8460.7730.8080.8990.047
Table 2. Quantitative comparison of different methods on four COD10K testing dataset categories, which contain S-measure ( S m ), weighted F-measure ( F β w ), F-measure ( F β ), E-measure ( E m ), and mean absolute error ( MAE ). Here, “↑” (“↓”) means that the larger (smaller) the better. The best three results in each column are marked in red, green, and blue.
Table 2. Quantitative comparison of different methods on four COD10K testing dataset categories, which contain S-measure ( S m ), weighted F-measure ( F β w ), F-measure ( F β ), E-measure ( E m ), and mean absolute error ( MAE ). Here, “↑” (“↓”) means that the larger (smaller) the better. The best three results in each column are marked in red, green, and blue.
COD10K-AmphibianCOD10K-AquaticCOD10K-FlyingCOD10K-Terrestrial
S m F β w F β E m MAE S m F β w F β E m MAE S m F β w F β E m MAE S m F β w F β E m MAE
EGNet [44]0.7760.5880.6500.8430.0560.7120.5150.5840.7840.0910.7690.5580.6210.8380.0460.7130.4670.5310.7940.056
PoolNet [46]0.7810.5840.6440.8230.0500.7370.5340.6070.7820.0780.7670.5390.6100.7970.0450.7070.4410.5080.7450.054
F3Net [37]0.8080.6570.7000.8460.0390.7280.5540.6110.7880.0760.7600.5710.6180.8180.0400.7120.4900.5380.7700.048
SCRN [45]0.8390.6650.7290.8670.0410.7800.6000.6740.8180.0640.8170.6080.6830.8400.0360.7580.5090.5880.7840.048
CSNet [47]0.8280.6490.7110.8570.0410.7680.5870.6560.8080.0670.8090.6100.6760.8380.0360.7440.5010.5660.7760.047
SSAL [48]0.7290.5600.6370.8170.0570.6320.4280.5090.7370.1010.7020.5040.5760.7950.0500.6470.4050.4710.7560.060
UCNet [49]0.8270.7170.7560.8970.0340.7670.6490.7030.8430.0600.8060.6750.7180.8860.0300.7420.5660.6170.8300.042
MINet [50]0.8230.6950.7320.8810.0350.7670.6320.6840.8310.0580.7990.6500.6970.8560.0310.7320.5360.5840.8020.043
ITSD [51]0.8100.6280.6790.8520.0440.7620.5840.6480.8110.0700.7930.5880.6450.8310.0400.7360.4960.5520.7770.051
PraNet [10]0.8420.7170.7500.9050.0350.7810.6430.6920.8480.0650.8190.6690.7070.8880.0330.7560.5650.6070.8350.046
SINet [19]0.8200.7140.7560.8910.0340.7660.6430.6980.8540.0630.8030.6630.7070.8870.0310.7490.5770.6250.8450.042
PFNet [3]0.8480.7400.7750.9110.0310.7930.6750.7220.8680.0550.8240.6910.7290.9030.0300.7730.6060.6470.8550.040
UJSC [25]0.8410.7420.7690.9050.0310.8050.7050.7470.8790.0490.8360.7190.7520.9060.0260.7780.6240.6640.8630.037
SLSR [2]0.8450.7510.7830.9060.0300.8030.6940.7400.8750.0520.8300.7070.7450.9060.0260.7720.6110.6550.8550.038
MGL-R [52]0.8540.7340.7700.8860.0280.8070.6880.7360.8550.0510.8390.7010.7430.8730.0260.7850.6060.6510.8230.036
C2FNet [24]0.8490.7520.7790.8990.0300.8070.7000.7410.8820.0520.8400.7240.7590.9140.0260.7830.6270.6640.8720.037
UGTR [53]0.8570.7380.7740.8960.0290.8100.6860.7340.8550.0500.8430.6990.7440.8730.0260.7890.6060.6530.8230.036
SINet_V2 [1]0.8580.7560.7880.9160.0300.8110.6960.7380.8830.0510.8390.7130.7490.9080.0270.7870.6230.6620.8660.039
FAPNet [8]0.8540.7520.7830.9140.0320.8210.7170.7570.8870.0490.8450.7250.7600.9060.0250.7950.6390.6780.8680.037
Ours0.8620.7670.7950.9240.0270.8210.7200.7580.8930.0480.8510.7410.7740.9160.0230.7870.6320.6690.8590.038
Table 3. Ablation studies on two testing datasets. Here, m-m SJM represents many-to-many side-join multiplication (as shown by the red line in Figure 2); o-m SJM represents one-to-many side-join multiplication. Here, “↑” (“↓”) means that the larger (smaller) the better.
Table 3. Ablation studies on two testing datasets. Here, m-m SJM represents many-to-many side-join multiplication (as shown by the red line in Figure 2); o-m SJM represents one-to-many side-join multiplication. Here, “↑” (“↓”) means that the larger (smaller) the better.
No.SDMCFCDecoderCAMO DatasetCOD10K Dataset
SDM TEM m-m SJM o-m SJM PD HFAD S m F β E m MAE S m F β E m MAE
#1 0.8120.7770.8700.0710.8180.7340.8860.035
#2 0.8120.7830.8700.0720.8190.7320.8890.034
#3 0.8120.7780.8660.0720.8210.7400.8880.034
#4 0.8150.7840.8720.0710.8200.7350.8870.035
#5 0.8180.7840.8740.0720.8210.7300.8850.035
Ours 0.8210.7920.8830.0680.8220.7340.8900.034
Table 4. Comparisons of the number of parameters, FLOPs, and FPS corresponding to recent COD methods. All evaluations follow the inference settings in the corresponding papers.
Table 4. Comparisons of the number of parameters, FLOPs, and FPS corresponding to recent COD methods. All evaluations follow the inference settings in the corresponding papers.
MethodOursFAPNet [8]SINet_V2 [1]UGTR [53]C2FNet [24]MGL-R [52]SINet [19]SLSR [2]UJSC [25]PFNet [3]
Params.66.550 M29.524 M26.976 M48.868 M28.411 M63.595 M48.947 M50.935 M217.982 M46.498 M
FLOPs40.733 G59.101 G24.481 G1.007 T26.167 G553.939 G38.757 G66.625 G112.341 G53.222 G
FPS29.41728.47638.94815.44636.94112.79334.08332.54718.24629.175
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Z.; Wang, T.; Wang, J.; Sun, Y. Features Split and Aggregation Network for Camouflaged Object Detection. J. Imaging 2024, 10, 24. https://doi.org/10.3390/jimaging10010024

AMA Style

Zhang Z, Wang T, Wang J, Sun Y. Features Split and Aggregation Network for Camouflaged Object Detection. Journal of Imaging. 2024; 10(1):24. https://doi.org/10.3390/jimaging10010024

Chicago/Turabian Style

Zhang, Zejin, Tao Wang, Jian Wang, and Yao Sun. 2024. "Features Split and Aggregation Network for Camouflaged Object Detection" Journal of Imaging 10, no. 1: 24. https://doi.org/10.3390/jimaging10010024

APA Style

Zhang, Z., Wang, T., Wang, J., & Sun, Y. (2024). Features Split and Aggregation Network for Camouflaged Object Detection. Journal of Imaging, 10(1), 24. https://doi.org/10.3390/jimaging10010024

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop