You are currently viewing a new version of our website. To view the old version click .
Engineering Proceedings
  • Proceeding Paper
  • Open Access

13 December 2023

A Linear Differentiation Scheme for Camouflaged Target Detection using Convolution Neural Networks †

,
,
and
1
Department of ECE, KPR Institute of Engineering and Technology, Coimbatore 641407, Tamilnadu, India
2
Department of CSE, Dr.N.G.P Institute of Technology, Coimbatore 641048, Tamilnadu, India
3
Department of Mechatronics Engineering, Paavai Engineering College, Namakkal 637018, Tamilnadu, India
4
Department of CSE, Bharati Vidyapeeth’s College of Engineering, Delhi 110063, India
This article belongs to the Proceedings Eng. Proc., 2023, RAiSE-2023

Abstract

Camouflaged objects are masked within an existing image or video under similar patterns. This makes it tedious to detect target objects post classification. The pattern distributions are monotonous due to similar pixels and non-contrast regions. In this paper, a distribution-differentiated target detection scheme (DDTDS) is proposed for segregating and identifying camouflaged objects. First, the image is segmented using textural pixel patterns for which the linear differentiation is performed. Convolutional neural learning is used for training the regions across pixel distribution and pattern formations. The neural network employs two layers for linear training and pattern differentiation. The differentiated region is trained for its positive rate in identifying the region around the target. Non-uniform patterns are used for training the second layer of the neural network. The proposed scheme pursues a recurrent iteration until the maximum segmentation is achieved. The metrics of positive rate, detection time, and false negatives are used for assessing the proposed scheme’s performance.

1. Introduction

Camouflaged object detection (COD) is a process that detects the objects that are presented in videos and images. Important patterns and parameters are detected from the images using COD. A rethinking COD method is commonly used to detect the objects from the images []. The COD method detects the visual effects, patterns, color, texture, and motion features of the images. The COD method uses a feature-based structure to analyze the depth and content of the objects []. The COD method also identifies the disguised targets that improve the functional performance range of the systems. The neighbor connection-based COD method is mainly used to detect the hidden similarities in the background in the images []. Hierarchical information transfer (HIT) is used in a method that identifies the neighboring layers in the images. The HIT-based COD method enhances the effectiveness level of the object detection process []. A semi-supervised search identification network-based COD approach is also used in the detection process. Both spatial and temporal features are detected in the images. The textual and statistical parameters are also detected, which produce optimal information for the COD process [].
Machine learning (ML) algorithms and techniques are used for the COD process. ML is mainly used to increase the accuracy level of the object detection process. A polarization imaging and deep learning (DL)-based COD method is used for the applications. A self-supervised DL algorithm is used here to detect the camouflaged targets from the given outputs. The polarization imaging detects the important parameters using a segmentation algorithm. The DL algorithm improves the effectiveness and performance range in target detection processes [,]. A deep gradient learning network (DGNet) is used for COD. The DGNet analyzes the context and texture features from the images. The DGNet detects the transparent and hidden objects which are presented in the images. The DGNet provides effective datasets for detection that reduce the latency in the computation process. The COD method also identifies the similarities among the objects that reduce the computational cost ratio of the systems []. The contributions of this study are summarized as follows:
  • Designing a target object detection scheme in camouflaged images using segmentation and variation detection with high positive rates;
  • Incorporating the convolution neural network process with modified layer operations for differentiation detection and training;
  • Performing experimental and comparative analysis using external inputs and metrics for assessing the proposed scheme’s performance.

3. Distribution-Differentiated Target Detection Scheme

This proposed distribution-differentiated target detection scheme (DDTDS) helps in conveying the challenge of determining camouflaged objects within images or videos, where the target objects are masked by similar patterns and distributions, making it crucial to identify them precisely after classification. By manipulating textural pixel patterns, convolutional neural learning, and an iterative training process, this process efficaciously differentiates and determines camouflaged objects with enhanced positive rates and reduced false negatives. Furthermore, the use of non-uniform patterns in the training procedure helps in enhancing the process’s ability to detect diverse and complex camouflaged objects. A schematic illustration of the proposal is presented in Figure 1.
Figure 1. Schematic illustration of the DDTD scheme.
The camouflaged images are analyzed for the further segmentation procedure, and to extract patterns from the segmentation. This proposed DDTDS helps in determining the region that contains the potential camouflaged objects. This input analysis process helps in scrutinizing their linear differentiation to extract the valuable characteristics of the images and their distinctions between the target images and then the background. After determining the camouflage images, the segmentation process is performed according to it. The segmentation operation is employed to separate the camouflaged object from the surrounding background. This involves using the information which is obtained from the linear differentiation of the textural pixel patterns to describe the bounds of the potential target regions. The process of determining the camouflage image via the segmentation process is explained by the following equation given below:
X = x ¯ x ¯ = x ¯ i j + i y ¯ i j = x ¯ i j a , x ¯ i j a + i y ¯ i j a , y ¯ i j a w h e r e     1 i ,       j n ,     0 a 1 , A ¯ = ( a ¯ i j ) A ¯ = ( a ¯ i j ) = a ¯ i j a , a ¯ i j a + i b ¯ i j a , b ¯ i j a
where x denotes the input image, y represents the textural patterns, A denotes the segmentation operation in detecting the camouflaged object, i represents the characteristics of an image, a n d   j denotes the target regions. After the segmentation process, the pattern determination procedure is performed to extract the formation and regions of the target input images. This process involves training the neural network to distinguish between uniform and non-uniform patterns. Uniform textural patterns represent the background regions, while non-uniform patterns denote the presence of the potential camouflage object in the target images. The convolutional neural network helps in analyzing the pattern formations and distribution of the pixels within each segmented region. It extracts the patterns for every segmented region. The process of determining the patterns of segmented regions is explained by the following equation given below:
X ¯ = M ¯ + i N ¯ Y ¯ = a ¯ + i b ¯ X ¯ + Y ¯ = a ¯ + b ¯ + i ( M ¯ + N ¯ ) X ¯ Y ¯ = a ¯ b ¯ + i ( M ¯ N ¯ ) a X ¯ = a Y ¯ + i a M ¯ ,           M N X ¯ × Y ¯ = M ¯ × N ¯ a ¯ × b ¯ + i a ¯ × b ¯ + M ¯ × N ¯ X = M ¯ a , M ¯ a + i N ¯ a , N ¯ a = ( [ a i j a , a ¯ i j a ] w h e r e     0 a 1
where M represents the patterns of the segmented regions, and N denotes the training operation. Now, the formation and regions of the camouflaged images are determined for further processes. This process ensures that the determined images are formed precisely according to the input images. They are identified for further training processes and also to determine the outside target regions. This ensures that the identified formation accurately corresponds to the potential camouflaged objects. The process of determining the formation of the images is explained by the following equation given below:
K X a B = C a K Y a B = D a K = X + X X X + X a = M ¯ ( a ) M ¯ ( a ) B a = X ¯ ( a ) X ¯ ( a ) Y a = N ¯ ( a ) N ¯ ( a ) C a = Z ¯ ( a ) Z ¯ ( a )
where K represents the formation of the image, and C represents the precise information of its origination. The image formation using the patterns is illustrated in Figure 2.
Figure 2. Image formation using patterns.
The input camouflaged image is differentiated using y for multiple pixel regions. As pixels meet x , y = m a x , the i is used for pattern checking across multiple regions. The CNN’s first layer is responsible for verifying the obtained patterns across various j and M regions through N . If the N is non-deviating (incremental), then K is obtained from i C . The segments are thus classified from the patterns obtained; the patterns are used for image formation (Figure 2). Now, the regions of the images are determined in this process. They are identified for the further verification operation of the targeted objects. These are the procedures conducted to reduce false positives and to improve the detection accuracy of the process. They help in achieving image recognition depending on the training of the network, and thus this process is explained by the following equation given below:
A ( n ) = 1 2 A + A + A + + A B ( n ) = 1 2 B + B + B + + B S = x ¯ a y ¯ a x ¯ a y ¯ a B X = M ¯ a N ¯ ( a ) N ¯ a M ¯ ( a ) A
where S represents the region of the camouflage images. Now, the CNN first layer helps in checking the pattern of the target object. The training process is performed for the neural network to extract the formation and region of the target objects. Then, it is compared with the camouflaged image’s formation and region. After the training process, the neural network helps in extracting the distinctive features and formations of the region surrounding the target. In this CNN first-layer process, a procedure is performed to check whether the camouflaged formation and region are the same as the formation and region of the camouflaged image. The differentiated region is trained for its positive rate in identifying the region around the target. The process of the first layer of CNN is explained by the following equation given below:
S S + = A + A A A + a b b a = I 0 0 I A + a A b = I A + a A b = 0 a + b = A + A a b = A + + A A + A A A + P ¯ x Q ¯ x P ¯ ( x ) Q ¯ ( x ) B = M ¯ ( a ) N ¯ ( a ) M ¯ ( a ) N ¯ ( a ) A X a P = P ( a ) A Y a Q = Q ( a ) P = B + B B B +
where P represents the operation of the CNN layer 1 in this process. Now the pattern differentiation is determined in the second layer of the CNN. The non-uniform patterns are used for training the second layer of the neural network. These non-uniform patterns denote the difficult and precise characteristics present in the camouflaged images. By providing efficient training to the neural network with such patterns, it identifies the variations in the target object. After determining the pattern differentiation, the outside target region is extracted. Hence, the CNN second-layer process is explained by the following equation given below:
V ¯ a = A X ¯ a A X ¯ a V ¯ a = A Y ¯ b A Y ¯ b B = 1 2 B + B + B + + B B ( A ) = = 1 2 B + B B + + B S X a B = C a S Y a B = D a S = A + A A A + A x = X ¯ a X ¯ a B y = Y ¯ a Y ¯ a
where V represents the pattern differentiation in the second layer of CNN. Now, the outside target region is determined after the two-layer process of the CNN. With both layers’ outcomes, this CNN helps in precisely delineating the outside of the target region. The second-layer process of the CNN is illustrated in Figure 3.
Figure 3. Second-layer process of the CNN.
The CNN’s second layer is responsible for differentiating V from the input image. The mixed M and C are used for concurrent layer-based analysis through N in the first layer. In this process, either f ( x ) or B ( y ) is used for validating V throughout C . If i C is observed anywhere, the variation boundaries are identified using the pixels present (Refer to Figure 3). The outside regions of the targeted images are consolidated as the background region. This procedure helps in ensuring the CNN [,,] aims at the target area, establishing precise and efficacious target detection within the camouflaged images. The process of extracting the outside target region is explained by the following equation given below:
T = X Y A ( B ) = 1 2 B + + B + B + B B + + B B + B B + + B B + B B + + B + B + B M ¯ ( a ) M ¯ ( a ) N ¯ ( a ) N ¯ ( a ) = A x ¯ ( a ) x ¯ ( a ) y ¯ ( a ) y ¯ ( a ) B + B B B +
where T represents the outside target region. This proposed system helps in detecting the outside target region by assessing the positive rate, detection time, and false negatives. This helps to reduce unnecessary false information and the time taken for the detection of the outside target region. The patterns of the camouflaged image are detected and compared with the targeted objects to extract the pattern differentiation. Additionally, based on this output, the outside target region is determined.

4. Experimental Analysis

In this sub-section, the images from [] are used for validating the functions of the proposed scheme. This reference provides 455 camouflage images for testing and training. which are related to underwater life. From the given input set, we consider a few images for describing the proposed scheme’s process, as presented in Table 1.
Table 1. Experimental output illustration.

5. Results and Discussion

The results and discussion in this section are presented using the positive rate, detection time, and false negatives observed in the experimental analysis. The number of patterns and regions varied up to 24 and 9, respectively in this analysis. The methods BGN-COD [], and ERR-Net [] are used along the proposed scheme in this comparative analysis.

5.1. Positive Rate

The metric positive rate defines the maximum differentiation score between y and S for T detection. This is computed as p o s i t i v e r a t e = y + s A + M i C . The positive rate is efficacious in this method, with the precise camouflaged input image and then its textural patterns. The segmentation process is performed according to the input image, and thus it helps in the further formation and region determination operation. This process efficaciously differentiates and determines camouflaged objects with enhanced positive rates and reduced false negatives. By using CNN, the pattern differentiation is identified for the extraction of the outside target region (Figure 4).
Figure 4. Positive rate analysis.

5.2. Detection Time

The time taken for the detection is lesser in this process, with the help of the two layers of the CNN. The pattern determination procedure is performed to extract the formation and regions of the target input images. This process involves training the neural network to distinguish between uniform and non-uniform patterns. The differentiated region is trained for its positive rate in identifying the region around the target. By consolidating the mentioned characteristics, the detection time is reduced (Figure 5).
Figure 5. Detection time analysis.

5.3. False Negatives

The false negatives are those V that fail to identify the target regardless of y M . This means the patterns are not identified during the intersection, and hence it is computed as M y M   i C . The false negatives are lesser in this process due to the enhanced precision of the obtained patterns. By manipulating textural pixel patterns, convolutional neural learning and its training process efficaciously differentiate and determine camouflaged objects with enhanced positive rates and reduced false negatives. These are procedures conducted to reduce false positives and to improve the detection accuracy of the process (Figure 6).
Figure 6. False negatives’ analysis.

6. Conclusions

The DDTDS commences by separating the input image using textural pixel patterns, and then it performs linear differentiation on these patterns. This process helps to differentiate the regions that contain prospective camouflaged objects from the background. To attain this, a CNN is engaged in training the regions based on pixel distribution and pattern formations. The neural network consists of two layers, and its primary aim is to distinguish the patterns linearly. In the first layer, the differentiated region is trained to enhance its positive rate in determining the area around the target. The second layer of the neural network is trained using non-uniform patterns. Therefore, this proposed scheme engages are peated iteration process to continuously clarify the segmentation, mainly aiming for the maximum segmentation precision. This iterative approach allows the network to continuously enhance its capability to differentiate between the target objects and background elements, leading to better detection results. From the comparative analysis, it is seen that the proposed scheme improves the positive rate by 9.3%, and reduces detection time and false negatives by 9.39% and 13.2% for the different patterns observed.

Author Contributions

Conceptualization, J.S. and G.B.; methodology, R.; software; validation and writing—original draft preparation, J.S., G.B., R. and M.T.; supervision, M.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data can be obtained from the corresponding author on request.

Acknowledgments

The authors are grateful to the anonymous referees for their comments and suggestions. The authors are also grateful to the management and staff.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bi, H.; Zhang, C.; Wang, K.; Tong, J.; Zheng, F. Rethinking camouflaged object detection: Models and datasets. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 5708–5724. [Google Scholar] [CrossRef]
  2. Zhang, C.; Wang, K.; Bi, H.; Liu, Z.; Yang, L. Camouflaged object detection via neighbor connection and hierarchical information transfer. Comput. Vis. Image Underst. 2022, 221, 103450. [Google Scholar] [CrossRef]
  3. Mondal, A. Camouflage design, assessment and breaking techniques: A survey. Multimed. Syst. 2022, 28, 141–160. [Google Scholar] [CrossRef]
  4. PullaRao, C.; Guruva Reddy, A.; Rama Rao, C.B. Camouflaged object detection for machine vision applications. Int. J. Speech Technol. 2020, 23, 327–335. [Google Scholar] [CrossRef]
  5. Shen, Y.; Lin, W.; Wang, Z.; Li, J.; Sun, X.; Wu, X.; Huang, F. Rapid detection of camouflaged artificial target based on polarization imaging and deep learning. IEEE Photonics J. 2021, 13, 7800309. [Google Scholar] [CrossRef]
  6. Ji, G.P.; Fan, D.P.; Chou, Y.C.; Dai, D.; Liniger, A.; Van Gool, L. Deep gradient learning for efficient camouflaged object detection. Mach. Intell. Res. 2023, 20, 92–108. [Google Scholar] [CrossRef]
  7. Liu, Y.; Wang, C.Q.; Zhou, Y.J. Camouflaged people detection based on a semi-supervised search identification network. Def. Technol. 2023, 21, 176–183. [Google Scholar] [CrossRef]
  8. Xu, X.; Chen, S.; Lv, X.; Wang, J.; Hu, X. Guided multi-scale refinement network for camouflaged object detection. Multimed. Tools Appl. 2023, 82, 5785–5801. [Google Scholar] [CrossRef] [PubMed]
  9. Yuanying, G.; Chuntong, L.; Hongcai, L.; Zhongye, L. A camouflage target detection method based on local minimum difference constraints. J. Syst. Eng. Electron. 2023, 34, 696–705. [Google Scholar]
  10. Zhai, Q.; Li, X.; Yang, F.; Jiao, Z.; Luo, P.; Cheng, H.; Liu, Z. Mgl: Mutual graph learning for camouflaged object detection. IEEE Trans. Image Process. 2022, 32, 1897–1910. [Google Scholar] [CrossRef] [PubMed]
  11. Xu, X.; Zhu, M.; Yu, J.; Chen, S.; Hu, X.; Yang, Y. Boundary guidance network for camouflage object detection. Image Vis. Comput. 2021, 114, 104283. [Google Scholar] [CrossRef]
  12. Liu, Y.; Zhang, K.; Zhao, Y.; Chen, H.; Liu, Q. Bi-RRNet: Bi-level recurrent refinement network for camouflaged object detection. Pattern Recognit. 2023, 139, 109514. [Google Scholar] [CrossRef]
  13. Ji, G.P.; Zhu, L.; Zhuge, M.; Fu, K. Fast camouflaged object detection via edge-based reversible re-calibration network. Pattern Recognit. 2022, 123, 108414. [Google Scholar] [CrossRef]
  14. Bi, H.; Tong, J.; Zhang, C.; Mo, D.; Wang, X. Camouflaged objects detection network via contradiction detection and feature aggregation. Multimed. Tools. Appl. 2023, 1–19. [Google Scholar] [CrossRef]
  15. Zhang, Q.; Ge, Y.; Zhang, C.; Bi, H. TPRNet: Camouflaged object detection via transformer-induced progressive refinement network. Vis. Comput. 2022, 39, 4593–4607. [Google Scholar] [CrossRef]
  16. Xiao, J.; Chen, T.; Hu, X.; Zhang, G.; Wang, S. Boundary-guided context-aware network for camouflaged object detection. Neural. Comput. Appl. 2023, 35, 15075–15093. [Google Scholar] [CrossRef]
  17. Xu, W.; Wang, Z.; You, X.; Zhang, C. Efficient fast convolution architectures for convolutional neural network. In Proceedings of the 2017 IEEE 12th International Conference on ASIC (ASICON), Guiyang, China, 25–28 October 2017; pp. 904–907. [Google Scholar]
  18. Vijayan, R.; Mohler, G. Forecasting Retweet Count during Elections Using Graph Convolution Neural Networks. In Proceedings of the 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA), Turin, Italy, 1–3 October 2018; pp. 256–262. [Google Scholar]
  19. Amrutha, B.R.; Bindu, K.R. Detecting Hate Speech in Tweets Using Different Deep Neural Network Architectures. In Proceedings of the 2019 International Conference on Intelligent Computing and Control Systems (ICCS), Madurai, India, 15–17 May 2019; pp. 923–926. [Google Scholar]
  20. Available online: https://www.kaggle.com/datasets/mikurules/aqua-test (accessed on 13 October 2023).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.