Next Article in Journal
Compton Camera X-Ray Fluorescence Imaging Design and Image Reconstruction Algorithm Optimization
Previous Article in Journal
From Brain Lobes to Neurons: Navigating the Brain Using Advanced 3D Modeling and Visualization Tools
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid-Recursive-Refinement Network for Camouflaged Object Detection

School of Computer Science and Technology, Harbin University of Science and Technology, Harbin 150080, China
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(9), 299; https://doi.org/10.3390/jimaging11090299
Submission received: 27 July 2025 / Revised: 17 August 2025 / Accepted: 30 August 2025 / Published: 2 September 2025
(This article belongs to the Section Computer Vision and Pattern Recognition)

Abstract

Camouflaged object detection (COD) seeks to precisely detect and delineate objects that are concealed within complex and ambiguous backgrounds. However, due to subtle texture variations and semantic ambiguity, it remains a highly challenging task. Existing methods that rely solely on either convolutional neural network (CNN) or Transformer architectures often suffer from incomplete feature representations and the loss of boundary details. To address the aforementioned challenges, we propose an innovative hybrid architecture that synergistically leverages the strengths of CNNs and Transformers. In particular, we devise a Hybrid Feature Fusion Module (HFFM) that harmonizes hierarchical features extracted from CNN and Transformer pathways, ultimately boosting the representational quality of the combined features. Furthermore, we design a Combined Recursive Decoder (CRD) that adaptively aggregates hierarchical features through recursive pooling/upsampling operators and stage-wise mask-guided refinement, enabling precise structural detail capture across multiple scales. In addition, we propose a Foreground–Background Selection (FBS) module, which alternates attention between foreground objects and background boundary regions, progressively refining object contours while suppressing background interference. Evaluations on four widely used public COD datasets, CHAMELEON, CAMO, COD10K, and NC4K, demonstrate that our method achieves state-of-the-art performance.

Graphical Abstract

1. Introduction

Camouflage, whether occurring naturally or through artificial means, serves as a strategy to reduce object visibility by imitating environmental attributes such as color, texture, and morphology [1]. COD addresses the intricate problem of detecting and localizing objects that blend seamlessly into cluttered or visually similar backgrounds. This task holds considerable importance in various real-world scenarios, including medical image analysis for lesion segmentation [2,3], surface defect inspection in industrial settings [4], and automated pest detection systems in precision agriculture [5].
Early COD methods based on handcrafted features [6,7] exhibited poor generalization and were highly susceptible to background interference. The advancement of CNN architectures has been instrumental in driving progress in camouflaged object detection, primarily due to their proficiency in extracting discriminative visual features [8,9,10]. Nevertheless, the locality constraint of CNN receptive fields limits their capacity to model global context, often resulting in incomplete segmentation or inaccurate localization. To address these shortcomings, Transformer-based models have recently gained traction in visual tasks, leveraging self-attention mechanisms to effectively capture long-range dependencies. Their powerful global representation capabilities have opened new avenues for tackling the challenges inherent in COD [11,12,13]. Nevertheless, directly aggregating features with large semantic discrepancies often leads to the loss of subtle but crucial local information, making it difficult to precisely delineate camouflaged object boundaries. While the aforementioned approaches have achieved varying degrees of success, they still face inherent limitations in complex and diverse camouflaged scenarios.
Some recent studies have explored hybrid architectures that fuse CNN and Transformer encoders, showing promising results in vision tasks [14,15]. However, effectively integrating their complementary advantages remains a key challenge. Additionally, the features extracted by the encoders are inherently hierarchical: shallow high-frequency features are rich in spatial details but prone to noise, whereas deep low-frequency features contain more semantic information but often lose fine spatial cues. Therefore, designing a decoder that can effectively process and integrate multi-level features—fusing semantic and spatial information while suppressing noise—is essential [16]. Moreover, the substantial visual similarity between camouflaged objects and their surrounding backgrounds poses a critical challenge in COD, often leading to foreground regions being mistakenly identified as background. Ignoring background modeling can lead to incomplete object detection or inaccurate boundary localization. To address this, researchers have explored solutions such as texture-aware modeling [17,18] and edge-guided features [19,20]. As illustrated in Figure 1, both CNN-based methods like SINet-v2 [21] and Transformer-based methods like CamoFormer [22] still suffer from limitations such as incomplete object detection and poor edge localization.
Motivated by the aforementioned insights, we introduce a Hybrid Recursive Refinement Network (HRRNet), which comprises three cohesively designed components. Initially, multi-scale features extracted from the CNN and Transformer encoders are processed by an HFFM, which aligns and integrates the dual-branch representations through channel-wise interaction and fusion, effectively bridging the semantic gap and promoting robust feature integration. Second, to better decode the hierarchical features extracted by the encoders, we propose a CRD, which comprises three submodules: Recursive Pooling, Recursive Upsampling, and Recursive Aggregation. These modules iteratively refine and fuse multi-scale information to generate high-quality decoding features. Finally, an FBS module is proposed to refine object boundaries and effectively reduce background distractions. This module alternately attends to foreground and background regions, emphasizing ambiguous regions that are prone to being confused. By adaptively refining and re-integrating misclassified features, FBS produces more complete and accurately bounded predictions. Our contributions are as follows:
  • We introduce a new hybrid architecture, termed HRRNet, which integrates features extracted by CNN and Transformer encoders through an HFFM. This module performs decoupled channel-wise fusion to exploit the complementary strengths of both feature extractors effectively.
  • We design a CRD that adaptively adjusts the feature fusion strategy based on the feature hierarchy. By fusing low-level detail with high-level semantics, the decoder effectively estimates the probable locations of camouflaged targets.
  • We present an FBS module, which selectively focuses on either foreground or background regions to iteratively refine segmentation results, enabling accurate identification of camouflaged object boundaries.
The remainder of this paper is organized as follows. Section 2 reviews related work on camouflaged object detection and complementary information strategy. Section 3 presents the proposed framework and its key components. Section 4 first describes the experimental setup, datasets, and evaluation metrics. It then analyzes the experimental results, including comparisons with state-of-the-art methods as well as ablation studies. Finally, Section 5 concludes the study and outlines possible directions for future research.

2. Related Work

2.1. Camouflaged Object Detection

The emergence of large-scale benchmark datasets, including CAMO [23], COD10K [24], and NC4K [25], has markedly accelerated progress in deep learning approaches for COD. CNNs have been widely adopted as the foundational architecture for visual recognition and segmentation tasks, owing to their capacity for processing inputs of varying dimensions and extracting multi-level feature representations. As a result, most current COD approaches are built upon CNN-based models. For instance, some biologically inspired methods simulate predator perception mechanisms (e.g., SINet-v2 [21]) or leverage scene-level transformations (e.g., ZoomNet [26]) to capture camouflaged targets. Recent studies have proposed multi-task learning frameworks that improve prediction accuracy by introducing auxiliary supervision signals, including edge detection [27], scribble-based annotations [28], and edge reconstruction strategies such as FEDER [8]. Despite the notable performance of CNN-based models, their inherently constrained receptive fields limit the ability to capture long-range dependencies across the entire image, often resulting in suboptimal delineation of camouflaged objects with incomplete structural recovery. Given the strength of Transformer-based models in capturing long-range dependencies, Transformers have recently demonstrated excellent performance in COD tasks. For example, FSPNet [29] employs ViT [30] as the encoder to capture global contextual information. FPNet [31] and FSNet [32] utilize PVT [33] and Swin Transformer [34], respectively, to extract globally representative features from input images for camouflaged object detection. Despite their strength in capturing global dependencies, Transformer-based architectures exhibit inherent limitations in modeling local patterns and preserving fine-grained details. The reliance on long-range interactions often comes at the expense of spatial precision, which is particularly detrimental for the accurate localization of camouflaged objects. To address these challenges, this study introduces a hybrid framework that integrates CNN and Transformer encoders, coupled with a novel multi-scale feature fusion mechanism. This design aims to alleviate the semantic discrepancy between heterogeneous features, enhance the quality of initial representations, and leverage the complementary advantages of both architectures to achieve more robust camouflaged object detection.

2.2. Complementary Information Strategy

Due to the inherent scale and information diversity of features extracted by encoders, as well as the inevitable encoding of noise and blurred boundaries, the practical performance of COD is often suboptimal. To enhance feature representation, various methods have attempted to integrate complementary low- and high-level features using structures [35] such as skip connections [36], short-range links [37], and neighboring pathways [21]. Despite their effectiveness, these techniques are often biased toward transferring information from lower-resolution stages to higher ones, which may restrict semantic expressiveness and neglect the dependencies across features. To address this, Song et al. [32] proposed a Cross-layer Connection Decoder (CCD), which selectively integrates intermediate-level features via element-wise multiplication and cross-layer sampling to enable computationally efficient multi-scale interaction, generating an initial prediction map focused on key regions. Similarly, Sun et al. [16] introduced a Cascade Correction Decoder (CCD), which performs iterative multi-scale feature optimization through bottom-up aggregation and top-down correction via dual-path collaborative learning. Such complementary fusion of multi-level features effectively enhances model robustness and discrimination capability in COD. In addition, several works have demonstrated that emphasizing the disparity between foreground and background representations contributes to more precise target localization. For example, Zhang et al. [38] introduced the Bilateral Attention Module (BAM), which utilizes a dual-branch attention mechanism to improve the representation of discriminative features in boundary regions. The foreground branch suppresses background interference and reinforces the target object, while the background branch mitigates foreground artifacts, significantly improving the integrity of fine details. Mei et al. [39] introduced PFNet (Positioning and Focus Network), where a higher-level prediction map and its inverse are multiplied with current-layer features to generate foreground- and background-aware features, which are then fed into a context exploration block to detect heterogeneous interference in either region. However, these approaches primarily aim at reducing interference from the foreground or background independently, lacking an explicit mechanism to decouple and regulate the interaction between the two.

3. Method

3.1. Overall Architecture

As shown in Figure 2, HRRNet consists of two stages. Stage 1 utilizes an encoder–decoder paradigm to extract meaningful representations and reconstruct camouflaged object structures, enhancing localization precision. Specifically, the RGB image I R H × W × 3 is simultaneously processed through both CNN-based and Transformer-based encoders to extract multi-scale features, represented as C i and T i , respectively, where i = 1 , 2 , , 4 . To enhance global perception, an HFFM fuses the extracted CNN and Transformer features C i and T i through cross-level concatenation, producing the representations F i ( i = 1 , 2 , , 4 ). The resulting features are subsequently passed to the CRD to integrate multi-scale information and output a coarse prediction. In stage 2, an FBS module is introduced. Guided by the coarse prediction, this module alternately focuses on foreground interiors/background boundary region to refine the object structure and boundary, enabling precise segmentation of complex camouflaged targets. To improve segmentation accuracy in complex backgrounds, the network jointly leverages fine-level texture cues and holistic semantic context, which facilitates precise boundary localization and effective object–background separation. We adopt a unified architecture and do not instantiate any background-specific branches or parameters.

3.2. Hybrid Feature Fusion Module

In this study, features from CNN and Transformer streams are fused via a parallel integration scheme. Across the encoder’s four stages, feature maps undergo downsampling by factors of 4, 8, 16, and 32, with the number of channels set to 64, 128, 256, and 512, accordingly. The proposed HFFM comprises three key steps: (1) alternately stacking feature maps from the two branches to achieve cross-branch fusion, (2) applying L1-norm pooling to compute global features followed by channel-wise normalization, and (3) enhancing inter-channel dependencies through 1D convolution-based channel interaction. An overview of the internal design and components of the HFFM module is shown in Figure 3.
To facilitate the integration of channel-level cues from the CNN and Transformer branches, the module utilizes cross-connections to alternately arrange feature maps X C and X T , thereby embedding pixel-wise semantics within the channel dimension. The feature map X C is placed at the even indices of X T with a stride of 2, while X T is inserted at the odd indices with the same stride. The specific formulation is as follows:
X H [ 0 : : 2 , : , : ] = X C ,
X H [ 1 : : 2 , : , : ] = X T ,
where X C and X T represent the feature maps from the CNN and Transformer branches, respectively, and X H is the fused feature map. Here, X C , X T R C × H × W , X H R 2 C × H × W . Subsequently, L 1 norm average pooling is applied to extract global contextual information from the fused feature map. Specifically, this pooling operation averages each channel across all spatial locations, yielding a fixed-size global feature vector e R 2 C :
e c = 1 H · W i = 1 H j = 1 W X H [ c , i , j ] ,
Based on this, motivated by ref. [40], we use channel-wise normalization to ensure numerical stability. The normalized vector e ^ is computed as
e ^ c = 2 C · e c e 2 = 2 C · e c c = 1 2 C e c 2 + ε 1 / 2
Here, e c denotes the global descriptor for the c-th channel, while X H [ c , i , j ] represents the activation at spatial location ( i , j ) within the c-th channel of the fused feature map X H R 2 C × H × W . H and W correspond to the spatial dimensions, and C is the number of channels for each individual branch, either X C (from the CNN) or X T (from the Transformer). The normalized global context vector e ^ lies in R 2 C × 1 × 1 , where | e | 2 denotes the L 2 -norm of the vector e, and ε is a small constant introduced to ensure numerical stability and prevent division by zero.
Finally, a 1D convolution is performed on the normalized context vector e ^ to model local interactions across channels, generating a new channel-wise attention vector u:
u = Conv 1 D ( e ^ )
This attention vector u is used to recalibrate the original fused feature map X H through element-wise multiplication, resulting in the final output feature map y:
y = X H · u
Here, u R 2 C × 1 × 1 denotes the learned channel-wise attention vector, while y R 2 C × H × W corresponds to the output feature map obtained after applying channel-specific modulation. X H is the intermediate feature representation obtained through cross-connection fusion of the CNN and Transformer branches.

3.3. Combined Recursive Decoder

In previous studies, cascaded sub-decoders in decoders typically adopted the same strategy to sequentially fuse features from different levels. This repetitive approach overlooks the inherent differences among multi-scale features. To improve decoding performance, this paper proposes a sub-decoder that accounts for these discrepancies. By adopting a recursive method, it dynamically adjusts the feature fusion strategy, effectively addressing the variations between features at different levels and enhancing detection accuracy. The CRD module includes three recursive-based sub-decoders, as shown in Figure 4, to achieve the complete fusion of features F i ( i = 1 , 2 , , 4 ).
The low-level features F 1 and F 2 are characterized by rich spatial resolution, which facilitates the extraction of detailed object boundaries. Conversely, the high-level features F 3 and F 4 encode stronger semantics, enabling effective discrimination between foreground and background. The CRD sub-decoders are designed to tailor their fusion mechanisms to the input feature scale, promoting stage-wise integration of hierarchical representations. Through the comprehensive multi-scale fusion of complementary low-level and high-level information, excellent camouflaged object detection results can be obtained. Specifically, each sub-decoder in the CRD module is tasked with feature fusion at a given scale. During feature fusion, the sub-decoder normalizes all scale information to ensure that the resulting feature maps remain reliable throughout the decoding process. Additionally, to ensure better detection results, the output from the preceding sub-decoder is passed to the subsequent sub-decoder for feature refinement. The entire decoding process is defined by the following formula:
m a s k m = Des m ( F 1 , F 2 , F 3 , F 4 , m a s k m 1 ) ,
where m a s k m corresponds to the output produced by the m-th sub-decoder, and Des m ( · ) represents the operation performed by the m-th sub-decoder, with m = 1 , 2 , 3 . Within each sub-decoder, the Recursive Pooling (RP) module pools the input features, and then convolution and activation functions are applied to the pooled features. This process helps capture more spatial detail information at lower layers. The Recursive Upsampling (RU) module increases the feature map size through upsampling to recover detail information, followed by convolution and activation functions to further optimize these features. Each sub-decoder is fed with pooled features, spatially upsampled maps, weighted contextual references, and benchmark cues to guide feature alignment and refinement.
The pooled features and upsampled features are sequentially normalized using RP and RU, resulting in features F 3 m mid and F 5 m mid , which match the shape of the benchmark features F 4 m . To construct the concatenated feature F m nor , the benchmark features F 4 m are combined with the normalized features F 3 m mid and F 5 m mid along the channel axis. Subsequently, the upsampled weighted reference features produced by the Upsampling module are employed to refine the concatenated representation. The refined features are then sequentially processed by the Channel Attention (CA) module and the Conv1 module to produce the final output of the m-th sub-decoder, denoted as mask m . At every stage, CRD forwards the upsampled prediction from the previous stage U ( mask m 1 ) as a mask guidance signal, which localizes object boundaries early and propagates that cue across scales. The entire process is defined as
F k + 1 mid = Cat D ( F k mid ) , F k + 1 , k [ 1 , 2 m ]
F j mid = Cat U ( F j + 1 mid ) , F j , j [ 5 m , 3 ]
F m nor = Cat ( D ( F 3 m mid ) , F 4 m , U ( F 5 m mid ) ) ,
mask m = CC m ( F m nor U ( mask m 1 ) ) ,
coarse map = Conv ( mask 3 ) ,
where D ( · ) and U ( · ) represent downsampling and upsampling operations, ⊙ denotes element-wise multiplication, and Cat ( · ) represents channel-wise concatenation. The convolution operation CC m ( · ) consists of CA and Conv1 operations, performed by the m-th sub-decoder. Conv2 denotes the convolutional layer applied in the final stage to generate the coarse prediction map, which serves as the ultimate output of the CRD module. For the initialization, F 1 mid is set to 0, and for each sub-decoder, if 1 > 2 m , then k is an empty set. Similarly, if 5 m < 3 , j is similarly empty.
This mask-guided recursion detects and preserves boundary details of complex objects before producing the coarse map, and provides clean contours for subsequent FBS refinement.

3.4. Foreground–Background Selection Module

Camouflaged objects often reside in complex, texture-similar backgrounds. To mitigate the performance drop caused by subtle differences between the object and its surroundings, we introduce the FBS module. Rather than performing global background removal, FBS operates only in the ambiguous neighborhood of the object–background boundary, where confusion most often occurs. It alternates its focus between foreground (FFBS) interiors and the background boundary (EFBS) region, progressively refining the segmentation. By selectively refining informative areas, FBS enhances the model’s ability to recover the complete contour of camouflaged objects.
As shown in Figure 5, the FBS module takes two inputs: the current-level feature f i R C × H × W and the coarse prediction map p i R 1 × H × W . It operates in two modes: FFBS emphasizes the foreground interiors by directly weighting the feature with p i , so the network strengthens confident foreground regions and stabilizes the interior representation at the current level. EFBS emphasizes the background boundary belt around the object: we dilate the prediction and subtract the original prediction to highlight the narrow ring of pixels just outside the contour, and then use this ring to weight the feature. In effect, EFBS concentrates computation on the edge area, suppresses background leakage in texture-similar surroundings, and helps recover thin, high-curvature details (e.g., spikes, antennae, tapering tails). FFBS and EFBS are applied at different stages so that interiors are consolidated while boundaries are progressively refined. This process is summarized as
M i = D i ( p i ) p i ,
F i = f i p i , M i ,
where D i (·) denotes the dilation operation (kernel size k = 7, iterations = 1), which is a dilation algorithm in computer vision used to emphasize regions in the background that are highly similar to the foreground boundary.
Next, a sequence of operations, including three 3 × 3 convolution layers, batch normalization, and ReLU activation functions, is applied to refine the fused features, producing the output f i R C × H × W . This refined output is then processed through a 7 × 7 convolution layer to generate p i R 1 × H × W . The entire process is expressed as
f i = F 3 C 3 B R ( F i ) , p i = F C 7 ( f i ) ,
where F 3 C 3 B R represents the sequence of three 3 × 3 convolution layers, batch normalization, and ReLU activations, and F C 7 is a 7 × 7 convolution layer.

3.5. Loss Function

This paper utilizes binary cross-entropy (BCE) loss [41], L BCE , and intersection over union (IoU) loss [42], L IoU , to supervise the model’s predictions for camouflaged objects. This strategy differs from traditional loss formulations by assigning pixel-wise weights based on local contrast, which enhances the model’s focus on hard-to-segment regions. This mechanism is especially beneficial for COD tasks, where accurate boundary detection is essential. The formulas for these two losses are as follows:
L BCE = 1 N i = 1 N y i log ( y ^ i ) + ( 1 y i ) log ( 1 y ^ i ) ,
L IoU = 1 i = 1 N y ^ i y i i = 1 N y ^ i + y i y ^ i y i ,
where y i and y ^ i denote the ground truth label and predicted probability for the i-th pixel, respectively, while N represents the total number of pixels in the image. Therefore, the combined loss function, which incorporates the advantages of both, is defined as
L total = λ 1 L BCE + λ 2 L IoU ,
where λ 1 and λ 2 are coefficients that control the relative weights of the two loss functions, and are typically set to 1 by default, based on empirical observations. Through this loss function, we ensure not only pixel-level classification accuracy but also improve the structural consistency and spatial alignment of the predicted masks, thus more effectively addressing the challenges in camouflage object detection tasks.

4. Experiments and Results

4.1. Experiment Settings

4.1.1. Datasets

HRRNet is evaluated on four standard COD benchmarks: CHAMELEON [43], CAMO [23], COD10K [24], and NC4K [25]. CHAMELEON consists of 76 labeled images, all used for testing. CAMO contains 1250 samples, with a training/testing split of 1000/250. COD10K offers a broad range of camouflaged object types with 3040 training and 2026 testing images. NC4K contributes 4121 test images for further performance validation. In this study, the proposed HRRNet is trained on the 1000 images from the CAMO-Train subset and the 3040 images from the COD10K-Train subset.
The datasets used in this work include several representative object categories whose boundary geometries differ markedly, covering various geometric challenges. For example, insects (filamentary limbs/antennae; high aspect ratio and multi-branch structures), fishes (serrated/spiny contours with high-curvature segments), reptiles (elongated, tapering tails and long, narrow parts), mammals (smooth closed silhouettes under texture-similar backgrounds and touching boundaries), and seahorses (slender bodies with curled tails and multi-instance proximity).

4.1.2. Implementation Details

The training of HRRNet is conducted using PyTorch 2.3.0 on an NVIDIA GTX 4090 GPU (24 GB). Four different backbones are employed: ResNet50, Res2Net50, Swin Transformer, and PVT. Optimization is performed with the Adam optimizer initialized at a learning rate of 5 × 10 5 , which is gradually reduced using a polynomial decay schedule with an exponent of 0.9. Input images are uniformly resized to 384 × 384 prior to training. In accordance with previous works such as SINet [24] and FSPNet [29], data augmentation techniques, including horizontal flipping, random cropping, and color enhancement, are applied. The model is trained over 50 epochs with a batch size of 12.

4.1.3. Evaluation Metrics

For performance evaluation, we use four standard metrics commonly employed in camouflaged object detection (COD) tasks: the structure measure ( S α ) [44], enhanced-alignment measure ( E ϕ ) [45], weighted F-measure ( F β w ) [46], and mean absolute error (MAE) [47]. S α measures the structural similarity between the prediction and the ground truth by combining region-aware and object-aware components, thereby reflecting overall shape consistency and indirectly the integrity of object contours. E ϕ (enhanced-alignment measure) evaluates the alignment between the prediction and ground truth by jointly considering local pixel agreement and global image-level statistics, capturing both global and local coherence. F β w is a spatially weighted F-measure that re-weights false positives/negatives according to their distance to the ground-truth mask, placing larger penalties on confident errors far from the object while still accounting for boundary discrepancies; this yields a balanced assessment of interiors and boundary regions. MAE is the mean absolute error between continuous maps, penalizing deviations across the entire image. In general, higher S α , E ϕ , F β w , and lower MAE indicate superior COD performance.

4.2. Comparison with State of the Art

We evaluate the proposed HRRNet by benchmarking it against recent state-of-the-art approaches on four representative datasets to demonstrate its effectiveness. These methods include MGL [47], C2FNet [48], SINet-v2 [21], ZoomNet [26], FEDER [8], DGNet [18], FSPNet [29], Camouflageator [9], VSSNet [11], VSCode [12], FocusDiffuser [13], CamoFormer [22], EPFDNet [10], SegMar [49], LSR [25], FAPNet [50], SENet [51], and MCRNet [52]. To ensure consistency in evaluation, we distinguish between CNN- and Transformer-based backbones, and utilize prediction maps released by the original authors or obtained from official open-source implementations.

4.2.1. Quantitative Evaluation

As summarized in Table 1, the proposed HRRNet achieves consistent improvements over all the compared methods across four standard metrics. To thoroughly assess its effectiveness, we experiment with different configurations involving four widely adopted backbone encoders. Particularly on the CHAMELEON dataset, HRRNet outperforms all the baselines on every metric, with its S α surpassing the runner-up by more than 1.0%. On the CAMO dataset, HRRNet attains an S α score of 0.883, which is 1.0% higher than VSSNet (based on Swin Transformer) and 5.4% higher than Camouflageator (based on ResNet backbone). Despite the inherent challenges posed by the COD10K and NC4K datasets—known for their complex scenes and difficult object boundaries—HRRNet exhibits strong performance, validating the proposed complementary fusion and refinement designs. Furthermore, the PR and F-measure curves in Figure 6 and Figure 7 demonstrate that our method (red ‘Ours’ curves) consistently outperforms the competing approaches across all confidence thresholds.
Beyond the four standard COD metrics ( S α , E ϕ , F β w , MAE), we also compare the number of parameters (Params, M), computational complexity (FLOPs, G), and runtime throughput (FPS, images per second) to provide a more comprehensive assessment. Params reflects the model size/capacity and memory footprint, FLOPs quantifies the hardware-agnostic computation per image (lower is more efficient), and FPS captures the end-to-end inference speed. As shown in Table 2, our HRRNet variants maintain moderate Params and FLOPs while achieving real-time FPS, yielding a favorable accuracy–efficiency trade-off compared with recent CNN/Transformer baselines.

4.2.2. Qualitative Evaluation

A visual comparison is illustrated in Figure 8, showcasing the performance of HRRNet against seven leading SOTA methods. The examples, selected from all four test datasets, represent a variety of complex environments such as underwater, land-based, and forested scenes. As observed from the prediction results, our method consistently yields more precise object boundaries and more complete object structures across various scenes. Specifically, for objects with fine-grained details and complex contours (e.g., the scorpionfish in the third row and the beetle in the fourth row of Figure 8), HRRNet is capable of delineating the object edges more accurately, preserving intricate structures such as antennae and spines. In contrast, competing methods such as ZoomNet and SINet-v2 often fail to capture these subtle details, exhibiting issues like missing parts and false predictions. In cases where the objects share high color and texture similarity with the background (e.g., the marine fish in the first row and the cheetah in the fifth row), HRRNet demonstrates superior discrimination, successfully segmenting the object with more complete contours while effectively filtering out occluding background elements. Methods solely based on Transformer or CNN backbones, such as FSPNet and LSR, tend to produce blurred segmentation results and suffer from background–object confusion. Moreover, in complex scenes containing multiple closely packed objects (e.g., the seahorses in the second row), our method is able to accurately reveal the spatial relationships and biological shapes of individual instances, outperforming other models in terms of structural integrity and object separation.

4.2.3. Effectiveness Comparison of the HFFM and Dual-Encoder Strategy

To verify the impact of both the HFFM and the dual-branch encoder architecture on COD performance, a set of controlled ablation studies was conducted, with the detailed results shown in Table 3. Specifically, we compare three different model configurations: (1) a single-perception model using dual CNN encoders, (2) a global modeling model using dual Transformer encoders, and (3) a hybrid encoder model retaining both the CNN and Transformer encoders but with the HFFM module removed. These configurations are intended to assess the design rationale and performance contributions of our proposed method, focusing on network modeling capabilities, feature fusion strategies, and overall synergy. The experimental results reveal that both the dual-CNN and dual-Transformer encoder models exhibit performance limitations in handling camouflaged objects. The former lacks sufficient global perception capability, making it difficult to capture semantic consistency across large-scale camouflaged regions. In contrast, the latter, while capable of modeling long-range dependencies, falls short in representing local texture details, resulting in blurred boundaries and loss of object details. Furthermore, when the HFFM module is replaced by simple convolutional layers or naive feature concatenation, the model performance degrades significantly. This indicates that merely combining features from different encoders is insufficient for effective complementary learning. The HFFM module mitigates semantic discrepancies through channel alignment, normalization, and a strong interaction mechanism, thereby enhancing the consistency and discriminability of the fused features and ensuring the full exploitation of multi-scale and structural information. Therefore, both the dual-encoder design and the HFFM module are indispensable to the overall performance improvements of the model. Their synergy forms a solid foundation for subsequent multi-scale decoding and boundary refinement stages.

4.2.4. Effectiveness Comparison of Combined Recursive Decoder

To comprehensively assess the contribution of the CRD, we carried out ablation comparisons, the results of which are detailed in Table 4. In this experiment, the CRD module was removed and replaced with a conventional decoder that does not employ recursive pooling, recursive upsampling, or recursive aggregation. We then conducted a thorough comparison with the complete model. It is evident from the results that excluding the CRD module significantly compromises segmentation performance, largely because standard decoders lack the capacity to effectively integrate multi-scale features with varying semantic and spatial characteristics. As a result, the fusion process fails to effectively integrate spatial details with semantic information. In contrast, the CRD module incorporates a recursive feature refinement strategy that effectively captures and enhances the spatial structure of target regions. The recursive design of the CRD module enables dynamic adjustment of multi-scale feature weights, leading to improved structural coherence and more precise boundary extraction. This makes the CRD an essential component for enhancing the overall reliability and detection accuracy of the proposed model. Its contribution to multi-scale integration and the preservation of structural details is essential for achieving high-quality segmentation, especially in challenging scenarios.

4.2.5. Effectiveness Comparison of Foreground–Background Selection Module

To further investigate the contributions of FFBS and EFBS within the proposed FBS module, we conducted additional ablation experiments, as summarized in Table 5. To assess the individual contribution of each submodule, we conducted ablation experiments by independently removing the FFBS and EFBS components and analyzing the resulting performance changes. The results reveal that excluding the FFBS branch substantially impairs the model’s capacity to capture fine-grained details of the camouflaged object’s core structure. This underscores the effectiveness of the foreground-oriented strategy in enhancing salient object representation while suppressing background noise. Similarly, excluding the EFBS branch leads to a noticeable drop in boundary accuracy and detail preservation, indicating that attention to background edge regions is equally essential for mitigating foreground–background confusion and improving contour precision. Therefore, FFBS and EFBS together constitute a complete and complementary feature optimization framework. Accurate delineation of subtle structures is achieved through the dual-selection mechanism, which markedly enhances the model’s resilience and generalization performance in complex camouflaged environments.

4.2.6. Qualitative Ablation Comparison

Beyond the quantitative gains reported in Table 4 and Table 5, Figure 9 provides side-by-side visual comparisons under two ablation settings. Removing the CRD leads to broken or over-smoothed shapes and missing thin parts (top: stick insect legs; bottom: lizard tail), indicating that recursive pooling/upsampling and aggregation are crucial to reconcile multi-scale semantics with spatial details. In contrast, removing the FBS degrades boundary localization and increases foreground–background confusion (e.g., edge fraying around the insect skeleton and leakage along the lizard silhouette), confirming the benefit of alternating foreground/background selection. The complete HRRNet preserves slender structures and sharp, clean contours across diverse scenes, aligning with the trends in S α / E ϕ / F β w /MAE.

5. Conclusions and Future Work

This work addresses a central challenge in camouflaged object detection (COD): integrating global semantics while preserving fine local structure. We propose HRRNet, a hybrid architecture that combines CNNs and Transformers. HRRNet is organized around three modules. The HFFM aligns and fuses heterogeneous CNN/Transformer features via cross-stacked fusion and channel-wise interaction, strengthening the joint representation. The CRD reconciles multi-scale features and reinforces object shape integrity during decoding. The FBS alternates attention between foreground structures and background edges, sharpening boundaries and reducing uncertainty in fine-grained predictions. Extensive experiments on CHAMELEON, CAMO, COD10K, and NC4K show state-of-the-art performance. On COD10K, HRRNet achieves an S-measure of 0.869 and MAE of 0.021, improving upon the Transformer baseline CamoFormer by 0.7% and the CNN baseline SINet-v2 by up to 1.5%, while reducing the error by 0.006. Ablation studies corroborate the effectiveness of each module and their complementary benefits when combined. This study has limitations. (i) Training and evaluation focus on static images; video and real-time COD remain to be validated. (ii) The recursive decoding and multi-level fusion that aid detail recovery introduce additional computation and latency compared with the lightest models, which may constrain edge deployment. (iii) Performance can degrade under extreme conditions (very low illumination, heavy occlusion, highly cluttered backgrounds). (iv) Reliance on pixel-level ground truth may limit generalization in domains with scarce or noisy annotations. Future work will explore lightweight module designs and real-time variants, robustness under extreme conditions, and self/weakly supervised learning. We also plan to investigate generalization to few-shot settings and video-based camouflage detection.

Author Contributions

Methodology, H.C. and X.W.; software, X.W.; supervision, H.C. and H.J.; writing—original draft, H.C. and X.W.; writing—review and editing, H.C. and H.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number No. 62071157.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in https://github.com/DengPingFan/SINet/ (accessed on 25 July 2025), reference [24].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Xiao, F.; Hu, S.; Shen, Y.; Fang, C.; Huang, J.; Tang, L.; Yang, Z.; Li, X.; He, C. A Survey of Camouflaged Object Detection and Beyond. CAAI Artif. Intell. Res. 2024, 3, 9150044. [Google Scholar] [CrossRef]
  2. Fan, D.P.; Ji, G.P.; Zhou, T.; Chen, G.; Fu, H.; Shen, J.; Shao, L. Pranet: Parallel reverse attention network for polyp segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2020; pp. 263–273. [Google Scholar]
  3. Wu, Y.H.; Gao, S.H.; Mei, J.; Xu, J.; Fan, D.P.; Zhang, R.G.; Cheng, M.M. Jcs: An explainable covid-19 diagnosis system by joint classification and segmentation. IEEE Trans. Image Process. 2021, 30, 3113–3126. [Google Scholar] [CrossRef]
  4. Xiong, Y.J.; Gao, Y.B.; Wu, H.; Yao, Y. Attention u-net with feature fusion module for robust defect detection. J. Circuits Syst. Comput. 2021, 30, 2150272. [Google Scholar] [CrossRef]
  5. Wang, L.; Yang, J.; Zhang, Y.; Wang, F.; Zheng, F. Depth-aware concealed crop detection in dense agricultural scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–18 June 2024; pp. 17201–17211. [Google Scholar]
  6. Sengottuvelan, P.; Wahi, A.; Shanmugam, A. Performance of decamouflaging through exploratory image analysis. In Proceedings of the 2008 First International Conference on Emerging Trends in Engineering and Technology, Nagpur, India, 16–18 July 2008; pp. 6–10. [Google Scholar]
  7. Kavitha, C.; Rao, B.P.; Govardhan, A. An efficient content based image retrieval using color and texture of image sub blocks. Int. J. Eng. Sci. Technol. 2011, 3, 1060–1068. [Google Scholar]
  8. He, C.; Li, K.; Zhang, Y.; Tang, L.; Zhang, Y.; Guo, Z.; Li, X. Camouflaged object detection with feature decomposition and edge reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22046–22055. [Google Scholar]
  9. He, C.; Li, K.; Zhang, Y.; Zhang, Y.; You, C.; Guo, Z.; Li, X.; Danelljan, M.; Yu, F. Strategic Preys Make Acute Predators: Enhancing Camouflaged Object Detectors by Generating Camouflaged Objects. arXiv 2023, arXiv:2308.03166. [Google Scholar]
  10. Fang, X.; Chen, J.; Wang, Y.; Jiang, M.; Ma, J.; Wang, X. EPFDNet: Camouflaged object detection with edge perception in frequency domain. Image Vis. Comput. 2025, 154, 105358. [Google Scholar] [CrossRef]
  11. Yin, C.; Yang, K.; Li, J.; Li, X.; Wu, Y. Camouflaged object detection via complementary information-selected network based on visual and semantic separation. IEEE Trans. Ind. Inform. 2024, 20, 12871–12881. [Google Scholar] [CrossRef]
  12. Luo, Z.; Liu, N.; Zhao, W.; Yang, X.; Zhang, D.; Fan, D.P.; Khan, F.; Han, J. Vscode: General visual salient and camouflaged object detection with 2d prompt learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 17169–17180. [Google Scholar]
  13. Zhao, J.; Li, X.; Yang, F.; Zhai, Q.; Luo, A.; Jiao, Z.; Cheng, H. Focusdiffuser: Perceiving local disparities for camouflaged object detection. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2024; pp. 181–198. [Google Scholar]
  14. Wang, Y.; Feng, L.; Sun, W.; Wang, L.; Yang, G.; Chen, B. A lightweight CNN-Transformer network for pixel-based crop mapping using time-series Sentinel-2 imagery. Comput. Electron. Agric. 2024, 226, 109370. [Google Scholar] [CrossRef]
  15. Wang, Y.; Zhang, T.; Zhao, L.; Hu, L.; Wang, Z.; Niu, Z.; Cheng, P.; Chen, K.; Zeng, X.; Wang, Z.; et al. Ringmo-lite: A remote sensing lightweight network with cnn-transformer hybrid framework. IEEE Trans. Geosci. Remote. Sens. 2024, 62, 1–20. [Google Scholar] [CrossRef]
  16. Sun, F.; Ren, P.; Yin, B.; Wang, F.; Li, H. CATNet: A cascaded and aggregated transformer network for RGB-D salient object detection. IEEE Trans. Multimed. 2023, 26, 2249–2262. [Google Scholar] [CrossRef]
  17. Ren, J.; Hu, X.; Zhu, L.; Xu, X.; Xu, Y.; Wang, W.; Deng, Z.; Heng, P.A. Deep texture-aware features for camouflaged object detection. IEEE Trans. Circuits Syst. Video Technol. 2021, 33, 1157–1167. [Google Scholar] [CrossRef]
  18. Ji, G.P.; Fan, D.P.; Chou, Y.C.; Dai, D.; Liniger, A.; Van Gool, L. Deep gradient learning for efficient camouflaged object detection. Mach. Intell. Res. 2023, 20, 92–108. [Google Scholar] [CrossRef]
  19. Xiao, J.; Chen, T.; Hu, X.; Zhang, G.; Wang, S. Boundary-guided context-aware network for camouflaged object detection. Neural Comput. Appl. 2023, 35, 15075–15093. [Google Scholar] [CrossRef]
  20. Liu, Z.; Deng, X.; Jiang, P.; Lv, C.; Min, G.; Wang, X. Edge perception camouflaged object detection under frequency domain reconstruction. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 10194–10207. [Google Scholar] [CrossRef]
  21. Fan, D.P.; Ji, G.P.; Cheng, M.M.; Shao, L. Concealed object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 6024–6042. [Google Scholar] [CrossRef]
  22. Yin, B.; Zhang, X.; Fan, D.P.; Jiao, S.; Cheng, M.M.; Van Gool, L.; Hou, Q. Camoformer: Masked separable attention for camouflaged object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 10362–10374. [Google Scholar] [CrossRef]
  23. Le, T.N.; Nguyen, T.V.; Nie, Z.; Tran, M.T.; Sugimoto, A. Anabranch network for camouflaged object segmentation. Comput. Vis. Image Underst. 2019, 184, 45–56. [Google Scholar] [CrossRef]
  24. Fan, D.P.; Ji, G.P.; Sun, G.; Cheng, M.M.; Shen, J.; Shao, L. Camouflaged object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 2777–2787. [Google Scholar]
  25. Lv, Y.; Zhang, J.; Dai, Y.; Li, A.; Liu, B.; Barnes, N.; Fan, D.P. Simultaneously localize, segment and rank the camouflaged objects. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 11591–11601. [Google Scholar]
  26. Pang, Y.; Zhao, X.; Xiang, T.Z.; Zhang, L.; Lu, H. Zoom in and out: A mixed-scale triplet network for camouflaged object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 2160–2170. [Google Scholar]
  27. Sun, Y.; Wang, S.; Chen, C.; Xiang, T.Z. Boundary-guided camouflaged object detection. arXiv 2022, arXiv:2207.00794. [Google Scholar] [CrossRef]
  28. He, R.; Dong, Q.; Lin, J.; Lau, R.W. Weakly-supervised camouflaged object detection with scribble annotations. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 781–789. [Google Scholar]
  29. Huang, Z.; Dai, H.; Xiang, T.Z.; Wang, S.; Chen, H.X.; Qin, J.; Xiong, H. Feature shrinkage pyramid for camouflaged object detection with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 5557–5566. [Google Scholar]
  30. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  31. Cong, R.; Sun, M.; Zhang, S.; Zhou, X.; Zhang, W.; Zhao, Y. Frequency perception network for camouflaged object detection. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 1179–1189. [Google Scholar]
  32. Song, Z.; Kang, X.; Wei, X.; Liu, H.; Dian, R.; Li, S. Fsnet: Focus scanning network for camouflaged object detection. IEEE Trans. Image Process. 2023, 32, 2267–2278. [Google Scholar] [CrossRef]
  33. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 10–17 October 2021; pp. 568–578. [Google Scholar]
  34. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  35. Wei, J.; Wang, S.; Huang, Q. F3Net: Fusion, feedback and focus for salient object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12321–12328. [Google Scholar]
  36. Wu, Z.; Su, L.; Huang, Q. Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3907–3916. [Google Scholar]
  37. Hou, Q.; Cheng, M.M.; Hu, X.; Borji, A.; Tu, Z.; Torr, P.H. Deeply supervised salient object detection with short connections. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3203–3212. [Google Scholar]
  38. Zhang, Z.; Lin, Z.; Xu, J.; Jin, W.D.; Lu, S.P.; Fan, D.P. Bilateral attention network for RGB-D salient object detection. IEEE Trans. Image Process. 2021, 30, 1949–1961. [Google Scholar] [CrossRef]
  39. Mei, H.; Ji, G.P.; Wei, Z.; Yang, X.; Wei, X.; Fan, D.P. Camouflaged object segmentation with distraction mining. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8772–8781. [Google Scholar]
  40. Yang, Z.; Zhu, L.; Wu, Y.; Yang, Y. Gated channel transformation for visual recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11794–11803. [Google Scholar]
  41. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  42. Yu, J.; Jiang, Y.; Wang, Z.; Cao, Z.; Huang, T. Unitbox: An advanced object detection network. In Proceedings of the 24th ACM International Conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 516–520. [Google Scholar]
  43. Skurowski, P.; Abdulameer, H.; Błaszczyk, J.; Depta, T.; Kornacki, A.; Kozieł, P. Animal camouflage analysis: Chameleon database. 2018; 2, 7, unpublished manuscript. [Google Scholar]
  44. Fan, D.P.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.M.; Borji, A. Enhanced-alignment measure for binary foreground map evaluation. arXiv 2018, arXiv:1805.10421. [Google Scholar] [CrossRef]
  45. Margolin, R.; Zelnik-Manor, L.; Tal, A. How to evaluate foreground maps? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 248–255. [Google Scholar]
  46. Perazzi, F.; Krähenbühl, P.; Pritch, Y.; Hornung, A. Saliency filters: Contrast based filtering for salient region detection. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 733–740. [Google Scholar]
  47. Zhai, Q.; Li, X.; Yang, F.; Chen, C.; Cheng, H.; Fan, D.P. Mutual graph learning for camouflaged object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 12997–13007. [Google Scholar]
  48. Sun, Y.; Chen, G.; Zhou, T.; Zhang, Y.; Liu, N. Context-aware cross-level fusion network for camouflaged object detection. arXiv 2021, arXiv:2105.12555. [Google Scholar]
  49. Jia, Q.; Yao, S.; Liu, Y.; Fan, X.; Liu, R.; Luo, Z. Segment, magnify and reiterate: Detecting camouflaged objects the hard way. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4713–4722. [Google Scholar]
  50. Zhou, T.; Zhou, Y.; Gong, C.; Yang, J.; Zhang, Y. Feature aggregation and propagation network for camouflaged object detection. IEEE Trans. Image Process. 2022, 31, 7036–7047. [Google Scholar] [CrossRef] [PubMed]
  51. Hao, C.; Yu, Z.; Liu, X.; Xu, J.; Yue, H.; Yang, J. A simple yet effective network based on vision transformer for camouflaged object and salient object detection. IEEE Trans. Image Process. 2025, 34, 608–622. [Google Scholar] [CrossRef] [PubMed]
  52. Zhang, D.; Cheng, L.; Liu, Y.; Wang, X.; Han, J. Mamba capsule routing towards part-whole relational camouflaged object detection. Int. J. Comput. Vis. 2025, 1–21. [Google Scholar] [CrossRef]
Figure 1. Comparison with other methods. It can be observed that both the CNN-based SINet-v2 and the Transformer-based CamoFormer exhibit challenges in achieving complete target detection and accurately capturing edge details.
Figure 1. Comparison with other methods. It can be observed that both the CNN-based SINet-v2 and the Transformer-based CamoFormer exhibit challenges in achieving complete target detection and accurately capturing edge details.
Jimaging 11 00299 g001
Figure 2. An overview of the proposed HRRNet: The input image is initially passed through a CNN-Transformer encoder to extract feature representations. These features are then fused using HFFM and passed through the CRD for progressive decoding. Finally, the boundaries are refined through the FBS module to achieve accurate prediction of camouflaged objects.
Figure 2. An overview of the proposed HRRNet: The input image is initially passed through a CNN-Transformer encoder to extract feature representations. These features are then fused using HFFM and passed through the CRD for progressive decoding. Finally, the boundaries are refined through the FBS module to achieve accurate prediction of camouflaged objects.
Jimaging 11 00299 g002
Figure 3. Schematic diagram of the HFFM. This module fuses CNN- and Transformer-encoded features through cross-stacking, channel normalization, and 1D convolution, enhancing semantic consistency and discriminative power.
Figure 3. Schematic diagram of the HFFM. This module fuses CNN- and Transformer-encoded features through cross-stacking, channel normalization, and 1D convolution, enhancing semantic consistency and discriminative power.
Jimaging 11 00299 g003
Figure 4. Schematic diagram of the CRD.
Figure 4. Schematic diagram of the CRD.
Jimaging 11 00299 g004
Figure 5. Schematic diagram of the FBS.
Figure 5. Schematic diagram of the FBS.
Jimaging 11 00299 g005
Figure 6. Precision–recall curves on three common COD datasets. Best viewed by zooming in for details.
Figure 6. Precision–recall curves on three common COD datasets. Best viewed by zooming in for details.
Jimaging 11 00299 g006
Figure 7. F β curves on three common COD datasets. Best viewed by zooming in for details.
Figure 7. F β curves on three common COD datasets. Best viewed by zooming in for details.
Jimaging 11 00299 g007
Figure 8. Visual comparisons of the proposed HRRNet and other seven state-of-the-art methods.
Figure 8. Visual comparisons of the proposed HRRNet and other seven state-of-the-art methods.
Jimaging 11 00299 g008
Figure 9. Qualitative ablation comparison. Columns (leftright): (a) Image, (b) GT, (c) w/o CRD, (d) w/o FBS, (e) HRRNet.
Figure 9. Qualitative ablation comparison. Columns (leftright): (a) Image, (b) GT, (c) w/o CRD, (d) w/o FBS, (e) HRRNet.
Jimaging 11 00299 g009
Table 1. Comparisons of state-of-the-art methods on COD datasets is presented, with the top three results highlighted in red, green, and blue, respectively, where ↑ denotes that a higher score indicates better performance, and ↓ denotes that a lower score indicates better performance.
Table 1. Comparisons of state-of-the-art methods on COD datasets is presented, with the top three results highlighted in red, green, and blue, respectively, where ↑ denotes that a higher score indicates better performance, and ↓ denotes that a lower score indicates better performance.
MethodYearBackboneCAMO-TestCOD10K-TestNC4KCHAMELEON
S α E ϕ F β ω M↓ S α E ϕ F β ω M↓ S α E ϕ F β ω M↓ S α E ϕ F β ω M↓
MGL [47]2021-CVPRResNet500.7750.8120.7360.0880.8140.8520.7110.0350.8330.8670.7820.0520.8930.9170.8330.031
C 2 FNet [48]2021-IJCAIRes2Net500.7960.8540.7620.0800.8130.8900.7230.0360.8380.8970.7650.0490.8880.9350.8280.032
SINet-v2 [21]2022-TPAMIRes2Net500.8200.8820.7820.0700.8150.8870.7180.0370.8470.9030.8050.0480.8880.9420.8350.030
ZoomNet [26]2022-CVPRResNet500.8200.8770.7940.0660.8380.8880.7660.0290.8530.8960.8180.0430.9020.9430.8640.023
FEDER [8]2023-CVPRResNet500.8070.8730.7850.0690.8230.9000.7400.0320.8460.9050.8170.0450.8940.9470.8550.028
DGNet [18]2023-MIREfficientNet0.8390.9010.8060.0570.8220.8960.7280.0330.8570.9110.8140.0420.8900.9380.8340.029
FSPNet [29]2023-CVPRViT0.8570.8990.8300.0500.8510.8950.7690.0260.8790.9150.8430.0350.9080.9430.8670.023
Camouflageator [9]2024-ICLRResNet500.8290.8910.8050.0660.8430.9200.7630.0280.8690.9220.8350.0410.9030.9520.8630.026
VSSNet [11]2024-TIIPVTv20.8730.9390.8440.0430.8730.9410.8050.0210.8890.9420.8550.030----
VSCode [12]2024-CVPRSwin-S0.8730.9260.8610.0430.8730.9350.8180.0210.8890.9360.8700.030----
FocusDiffuser [13]2024-ECCVViT0.8690.9310.8420.0430.8630.9340.7850.0240.8820.9330.8400.032----
CamoFormer [22]2024-TPAMISwin-B0.8760.9350.8320.0430.8620.9320.7720.0240.8880.9410.8400.0310.8910.9530.8290.026
EPFDNet [10]2025-IVCRes2Net500.8170.8860.7570.0680.8150.8940.7000.0330.8440.9060.7800.044----
SENet [51]2025-TIPViT----0.8650.9250.7800.0240.8890.9330.8430.032----
MCRNet [52]2025-IJCVSwin0.8540.9150.8470.0540.8540.9240.8070.0260.8750.9300.8570.036----
HRRNetOursSwin-Res2Net0.8760.9280.8430.0430.8670.9320.7850.0230.8880.9330.8450.0320.9110.9480.8630.023
HRRNetOursSwin-ResNet0.8740.9260.8420.0450.8660.9310.7830.0230.8860.9310.8470.0320.9100.9480.8620.024
HRRNetOursPVT-Res2Net0.8830.9350.8530.0410.8690.9410.7970.0210.8910.9420.8550.0300.9180.9590.8690.021
HRRNetOursPVT-ResNet0.8820.9320.8490.0420.8630.9370.7950.0210.8890.9410.8570.0310.9140.9560.8670.023
“-” denotes that the author does not provide corresponding results.
Table 2. Efficiency analysis of our HRRNet and multiple COD methods.
Table 2. Efficiency analysis of our HRRNet and multiple COD methods.
MGL [47] C 2 FNet [48]ZoomNet [26]FEDER [8]SegMar [29]FSPNet [49]HRRNet
(Swin-R2N)
HRRNet
(Swin-R50)
HRRNet
(PVT-R2N)
HRRNet
(PVT-R50)
Parameters63.6028.4132.3837.3755.62273.7963.1662.5360.2459.62
FLOPs (G)553.9426.17203.5023.9833.65288.3136.6735.8035.0634.28
FPS5.18109.6714.10119.6885.2910.1378.4280.1782.1583.92
Table 3. Ablation studies on the HFFM module and dual-encoder strategy, where ↑ denotes that a higher score indicates better performance, and ↓ denotes that a lower score indicates better performance.
Table 3. Ablation studies on the HFFM module and dual-encoder strategy, where ↑ denotes that a higher score indicates better performance, and ↓ denotes that a lower score indicates better performance.
MethodCAMO-TestCOD10K-TestNC4KCHAMELEON
S α E ϕ F β ω M S α E ϕ F β ω M S α E ϕ F β ω M S α E ϕ F β ω M
CNN-CNN0.8250.8970.7980.0560.8230.9070.7600.0300.8610.9130.8250.0410.9010.9380.8550.028
Trans-Trans0.8590.9180.8360.0430.8440.9230.7740.0260.8790.9240.8360.0360.9110.9460.8590.027
w/o HFFM0.8760.9280.8410.0460.8610.9270.7880.0250.8830.9300.8410.0330.9120.9520.8600.024
HRRNet0.8830.9350.8530.0410.8690.9410.7970.0210.8910.9420.8550.0300.9180.9590.8690.021
Table 4. Ablation studies on Combined Recursive Decoder, where ↑ denotes that a higher score indicates better performance, and ↓ denotes that a lower score indicates better performance.
Table 4. Ablation studies on Combined Recursive Decoder, where ↑ denotes that a higher score indicates better performance, and ↓ denotes that a lower score indicates better performance.
MethodCAMO-TestCOD10K-TestNC4KCHAMELEON
S α E ϕ F β ω M S α E ϕ F β ω M S α E ϕ F β ω M S α E ϕ F β ω M
w/o CRD0.8230.8940.7850.0600.8150.9030.7420.0330.8550.9160.7900.0430.8860.9220.8130.034
HRRNet0.8830.9350.8530.0410.8690.9410.7970.0210.8910.9420.8550.0300.9180.9590.8690.021
Table 5. Ablation studies on Foreground–Background Selection module, where ↑ denotes that a higher score indicates better performance, and ↓ denotes that a lower score indicates better performance.
Table 5. Ablation studies on Foreground–Background Selection module, where ↑ denotes that a higher score indicates better performance, and ↓ denotes that a lower score indicates better performance.
MethodCAMO-TestCOD10K-TestNC4KCHAMELEON
S α E ϕ F β ω M S α E ϕ F β ω M S α E ϕ F β ω M S α E ϕ F β ω M
w/o FFBS0.8790.9310.8460.0450.8640.9330.7920.0230.8870.9350.8460.0330.9150.9540.8630.023
w/o EFBS0.8810.9330.8500.0430.8660.9380.7950.0230.8900.9390.8520.0310.9160.9580.8670.021
HRRNet0.8830.9350.8530.0410.8690.9410.7970.0210.8910.9420.8550.0300.9180.9590.8690.021
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, H.; Wang, X.; Jin, H. Hybrid-Recursive-Refinement Network for Camouflaged Object Detection. J. Imaging 2025, 11, 299. https://doi.org/10.3390/jimaging11090299

AMA Style

Chen H, Wang X, Jin H. Hybrid-Recursive-Refinement Network for Camouflaged Object Detection. Journal of Imaging. 2025; 11(9):299. https://doi.org/10.3390/jimaging11090299

Chicago/Turabian Style

Chen, Hailong, Xinyi Wang, and Haipeng Jin. 2025. "Hybrid-Recursive-Refinement Network for Camouflaged Object Detection" Journal of Imaging 11, no. 9: 299. https://doi.org/10.3390/jimaging11090299

APA Style

Chen, H., Wang, X., & Jin, H. (2025). Hybrid-Recursive-Refinement Network for Camouflaged Object Detection. Journal of Imaging, 11(9), 299. https://doi.org/10.3390/jimaging11090299

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop