Next Article in Journal
DPO-ESRGAN: Perceptually Enhanced Super-Resolution Using Direct Preference Optimization
Previous Article in Journal
Design and Measurements of an Electrothermal Filter Using CMOS Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

CNN–Transformer-Based Model for Maritime Blurred Target Recognition

1
School of Electronic Information, Wuhan University of Science and Technology, WuHan 430081, China
2
School of Information Engineering, Hubei University of Economics, Wuhan 430205, China
3
College of Computer Science and Electronic Engineering, Hunan University, Changsha 410082, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(17), 3354; https://doi.org/10.3390/electronics14173354 (registering DOI)
Submission received: 26 July 2025 / Revised: 21 August 2025 / Accepted: 22 August 2025 / Published: 23 August 2025

Abstract

In maritime blurred image recognition, ship collision accidents frequently result from three primary blur types: (1) motion blur from vessel movement in complex sea conditions, (2) defocus blur due to water vapor refraction, and (3) scattering blur caused by sea fog interference. This paper proposes a dual-branch recognition method specifically designed for motion blur, which represents the most prevalent blur type in maritime scenarios. Conventional approaches exhibit constrained computational efficiency and limited adaptability across different modalities. To overcome these limitations, we propose a hybrid CNN–Transformer architecture: the CNN branch captures local blur characteristics, while the enhanced Transformer module models long-range dependencies via attention mechanisms. The CNN branch employs a lightweight ResNet variant, in which conventional residual blocks are substituted with Multi-Scale Gradient-Aware Residual Block (MSG-ARB). This architecture employs learnable gradient convolution for explicit local gradient feature extraction and utilizes gradient content gating to strengthen blur-sensitive region representation, significantly improving computational efficiency compared to conventional CNNs. The Transformer branch incorporates a Hierarchical Swin Transformer (HST) framework with Shifted Window-based Multi-head Self-Attention for global context modeling. The proposed method incorporates blur invariant Positional Encoding (PE) to enhance blur spectrum modeling capability, while employing DyT (Dynamic Tanh) module with learnable α parameters to replace traditional normalization layers. This architecture achieves a significant reduction in computational costs while preserving feature representation quality. Moreover, it efficiently computes long-range image dependencies using a compact 16 × 16 window configuration. The proposed feature fusion module synergistically integrates CNN-based local feature extraction with Transformer-enabled global representation learning, achieving comprehensive feature modeling across different scales. To evaluate the model’s performance and generalization ability, we conducted comprehensive experiments on four benchmark datasets: VAIS, GoPro, Mini-ImageNet, and Open Images V4. Experimental results show that our method achieves superior classification accuracy compared to state-of-the-art approaches, while simultaneously enhancing inference speed and reducing GPU memory consumption. Ablation studies confirm that the DyT module effectively suppresses outliers and improves computational efficiency, particularly when processing low-quality input data.
Keywords: dual-branch; normalization-free transformer; blurred image recognition; Dynamic Tanh; hybrid architecture dual-branch; normalization-free transformer; blurred image recognition; Dynamic Tanh; hybrid architecture

Share and Cite

MDPI and ACS Style

Huang, T.; Pan, C.; Liu, J.; Kang, Z. CNN–Transformer-Based Model for Maritime Blurred Target Recognition. Electronics 2025, 14, 3354. https://doi.org/10.3390/electronics14173354

AMA Style

Huang T, Pan C, Liu J, Kang Z. CNN–Transformer-Based Model for Maritime Blurred Target Recognition. Electronics. 2025; 14(17):3354. https://doi.org/10.3390/electronics14173354

Chicago/Turabian Style

Huang, Tianyu, Chao Pan, Jin Liu, and Zhiwei Kang. 2025. "CNN–Transformer-Based Model for Maritime Blurred Target Recognition" Electronics 14, no. 17: 3354. https://doi.org/10.3390/electronics14173354

APA Style

Huang, T., Pan, C., Liu, J., & Kang, Z. (2025). CNN–Transformer-Based Model for Maritime Blurred Target Recognition. Electronics, 14(17), 3354. https://doi.org/10.3390/electronics14173354

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop