Next Article in Journal
Effect of Blade Number on Tip Vortex Cavitation of Propeller
Previous Article in Journal
Sediment Resuspension in the Yellow River Subaqueous Delta During Gale Events
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

VIOS-Net: A Multi-Task Fusion System for Maritime Surveillance Through Visible and Infrared Imaging

1
Naval Architecture and Shipping College, Guangdong Ocean University, Zhanjiang 524005, China
2
Technical Research Center for Ship Intelligence and Safety Engineering of Guangdong Province, Zhanjiang 524005, China
3
Guangdong Provincial Key Laboratory of Intelligent Equipment for South China Sea Marine Ranching, Zhanjiang 524005, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work and should be considered co-first authors.
J. Mar. Sci. Eng. 2025, 13(5), 913; https://doi.org/10.3390/jmse13050913
Submission received: 23 March 2025 / Revised: 27 April 2025 / Accepted: 29 April 2025 / Published: 6 May 2025
(This article belongs to the Section Ocean Engineering)

Abstract

:
Automatic ship monitoring models leveraging image recognition have become integral to regulatory applications within maritime management, with multi-source image co-monitoring serving as the primary method for achieving comprehensive, round-the-clock surveillance. Despite their widespread use, the existing models predominantly train each data source independently or simultaneously train multiple sources without fully optimizing the integration of similar information. This approach, while capable of all-weather detection, results in the underutilization of data features from related sources and unnecessary repetition in model training, leading to excessive time consumption. To address these inefficiencies, this paper introduces a novel multi-task learning framework designed to enhance the utilization of data features from diverse information sources, thereby reducing training time, lowering costs, and improving recognition accuracy. The proposed model, VIOS-Net, integrates the advantages of both visible and infrared data sources to meet the challenges of all-weather, all-day ship monitoring under complex environmental conditions. VIOS-Net employs a Shared Bottom network architecture, utilizing both shared and specific feature extraction modules at the model’s lower and upper layers, respectively, to optimize the system’s recognition capabilities and maximize data utilization efficiency. The experimental results demonstrate that VIOS-Net achieves an accuracy of 96.20% across both visible and infrared spectral datasets, significantly outperforming the baseline ResNet-34 model, which attained accuracies of only 4.86% and 9.04% in visible and infrared data, respectively. Moreover, VIOS-Net reduces the number of parameters by 48.82% compared to the baseline, achieving optimal performance in multi-spectral ship monitoring. Extensive ablation studies further validate the effectiveness of the individual modules within the proposed framework.

1. Introduction

Ship monitoring constitutes a critical component of ship management [1,2,3,4,5], playing a pivotal role in ensuring maritime safety, environmental protection, resource management, operational efficiency, and the prevention of criminal activities [6,7,8,9,10,11,12]. Effective ship monitoring is essential for safeguarding maritime safety, as it enables the timely detection of maritime accidents and violations through the intelligent monitoring and identification of vessel targets, thereby preventing cargo mishaps and other maritime incidents. This function is particularly crucial in increasingly complex meteorological and marine environments, where the ability to capture abnormal behaviors and dangerous situations through early warning and preventive measures becomes even more pronounced. Moreover, ship monitoring contributes significantly to marine environmental protection by enabling the control and monitoring of pollutant emissions from vessels. This capability helps reduce marine pollution and preserves the health of the marine ecosystem. By facilitating the real-time monitoring and statistical analysis of pollutant discharges, such as exhaust gases and sewage, ship monitoring supports regulatory authorities in taking timely actions to mitigate environmental contamination. Ship monitoring also serves as an effective tool for the real-time management of marine resources, including fisheries, oil and gas, and mineral resources. By tracking the number and regional distribution of fishing vessels, it aids in the sustainable development and conservation of fishery resources. Additionally, monitoring vessel transport routes and cargo types enables the efficient management of marine mineral and oil and gas resources. In the context of port operations, ship monitoring provides a foundation for enhancing operational efficiency. It enables the accurate tracking and identification of vessel positions within ports, thereby equipping port management personnel with the necessary information for informed decision making and scheduling. This capability is instrumental in improving both port management and transportation efficiency. Furthermore, ship monitoring plays a crucial role in maritime law enforcement and crime prevention. Through vessel identification and route tracking, it assists maritime law enforcement officers in detecting and addressing violations and criminal activities, such as illegal immigration, drug trafficking, and smuggling, thereby enhancing the effectiveness of investigations and interventions. In maritime safety supervision, the real-time monitoring and positioning of vessels help prevent dangerous forced landings, thereby safeguarding lives and property.
With the continuous advancement of science and technology, ship monitoring based on sensor imaging (including satellite imagery, UAV-based aerial photography, and coastal CCTV systems) has emerged as a primary method for achieving all-weather, real-time detection of maritime conditions. While our deep learning model demonstrates strong adaptability across these heterogeneous image sources through transfer learning techniques, its performance remains contingent upon image resolution quality and stable data transmission conditions. The proposed framework is suitable for port surveillance systems with dedicated computing infrastructure, though deployment on resource-constrained edge devices may require additional optimization.
In the realm of ship image monitoring, multiple information sources are utilized to ensure comprehensive real-time surveillance, with visible light, infrared imagery, and synthetic aperture radar (SAR) being the most commonly employed technologies [13]. These sensors, operating on different imaging principles, generate image data with distinct characteristics.
Visible light sensors excel in capturing color, texture, and shape information, resulting in high-resolution images that offer a vivid, intuitive representation of the maritime environment. This capability enables the clear identification of vessels and their spatial relationships within the scene. In contrast, infrared imaging operates by detecting radiation differences between the target and its background, making it less susceptible to variations in lighting conditions. This unique imaging mechanism allows infrared sensors to penetrate certain materials, such as fog and smoke, making them particularly effective for night vision, target detection in extreme environments, and thermal signature identification [14]. On the other hand, SAR operates independently of external light sources by emitting microwave signals and receiving the signals reflected from the Earth’s surface. This distinctive working mechanism enables SAR to penetrate clouds, rain, and fog, providing continuous monitoring regardless of lighting and weather conditions. SAR’s ability to function in any environment, day or night, makes it an invaluable tool for uninterrupted maritime surveillance [15]. However, no sensor system is without limitations. Visible light sensors are highly dependent on ambient light, rendering them vulnerable to variations in lighting conditions, such as shadows, occlusions, and low light, which can compromise image quality. Infrared images, while effective in certain conditions, often lack color information and typically have lower resolution compared to visible light images, complicating the tasks of ship identification and scene interpretation. SAR images, due to their complex imaging mechanism, suffer from prolonged imaging times and reduced intuitive appeal. They do not provide direct color information on the Earth’s surface and require sophisticated processing and interpretation techniques [16].
By leveraging the combined use of two or more information sources, it is possible to achieve all-weather, round-the-clock ship monitoring, thereby enhancing maritime safety and operational efficiency. In contrast to SAR imagery, the temporal, spatial, feature, and processing consistency of data acquired from visible and infrared sensors significantly improves the training outcomes of deep learning-based ship monitoring systems. Our methodology specifically addresses pre-registered multi-source data scenarios, leveraging spatiotemporally synchronized imaging systems that maintain angular perspective consistency while tolerating minor parallax offsets inherent in practical maritime observation platforms. Visible and infrared images captured simultaneously ensure that the dynamic scene depicted is consistent across both modalities. This temporal consistency facilitates the effective processing of dynamic temporal information during training, allowing for the full utilization of the high-resolution and distinct features of visible imaging during daylight hours. At the same time, it mitigates the limitations of visible imaging in low-light conditions by transitioning to infrared imaging, which excels in night vision environments and in detecting thermal signatures for accurate ship classification. The residual architecture selection directly supports this configuration through its capacity for preserving cross-modal positional relationships—shallow convolutional layers capture alignment-sensitive low-level features while deeper layers maintain geometric correspondence through identity mapping connections. Moreover, when visible light and infrared images are captured from the same angle, the geometric relationships within the scene remain consistent, and the ship’s position is identical in both images. This consistency provides a crucial spatial reference for data fusion and model training, with ResNet-34’s skip connections proving particularly effective in preserving these geometric priors throughout the network depth. Our architectural analysis reveals that compared to deeper alternatives, ResNet-34 optimally balances positional sensitivity with contextual abstraction—its intermediate feature maps maintain sufficient spatial resolution (112 × 112 at residual block 3) to resolve typical parallax offsets ≤ 15 pixels observed in our dataset, while avoiding excessive downsampling that could degrade alignment-critical features. This consistency reduces errors in the fusion process and serves as a key factor in the simultaneous training of multiple data sources. It also plays a pivotal role in enhancing the accuracy and reliability of monitoring and identification. The geometric structure of ship images obtained from different sensors at the same moment and angle is consistent, preserving fundamental morphological features such as edges, contours, and basic shapes. This feature consistency is instrumental in enabling the model to better understand and integrate multiple sources of information for effective multi-task learning. Given the similarity in features, consistent or similar preprocessing and data augmentation strategies can be employed across multiple data sources to enhance data diversity while maintaining geometric consistency. This approach promotes feature sharing across different data sources and fosters cross-source feature learning, thereby optimizing the model’s performance.
In summary, the high correlation between ship information obtained from visible and infrared sensors, coupled with the complementary nature of the data they provide, makes it advantageous to combine these two sources. By integrating the information from both sensors, we can leverage their respective strengths to enhance the overall monitoring capabilities.
Deep learning-based multi-task automatic ship recognition technology not only facilitates the learning of common features from multi-source images obtained through different sensors but also transforms the traditional approach into a unified reinforcement framework. This integrated method integrates visible detection and infrared classification tasks, allowing for the fusion of features from individual information sources. As a result, the performance of both visible detection and infrared classification tasks is significantly enhanced when compared to single-source approaches. By combining multiple sources of information, this approach achieves higher monitoring outcomes, achieving state-of-the-art performance compared to the existing ship recognition frameworks.
Inspired by the success of multi-task learning, we aim to improve real-world visible and infrared ship recognition tasks through mutual reinforcement within a unified architecture. Unlike the existing models that treat these tasks independently, this paper introduces a unified multi-task ship recognition model, VIOS-Net, which facilitates interaction between multiple information sources by learning a set of bidirectionally complementary features from ship images relevant to both tasks. This mutual learning process enhances each task’s specific features, allowing the high-resolution imaging properties of visible light to be shared and integrated with the penetrating capabilities of infrared imaging. This is achieved using a multi-layer convolutional neural network with shared modules across multiple information sources, where a shared feature extractor is employed at the bottom layer. Two Unique Feature Extractors (UFEs) are positioned at the top to accommodate different task-specific representations and their corresponding parameters. Given that the variation in information content between visible and infrared sources might affect the effectiveness of joint reinforcement learning, we address this by fusing information from multiple sources and designing weighting factors for each source during training. By integrating multiple information sources, the simultaneous learning of visible and infrared ship recognition not only reduces the risk of overfitting—often present when training on a single source—but also addresses the challenge of differing information content between sources, thereby further enhancing recognition accuracy. The learned representations are more compact than those constructed from single-source surface features and demonstrate higher accuracy than the results obtained by merely fusing multiple information sources with linear weighted decisions. The experimental results indicate that the combined visible and infrared ship monitoring system significantly improves the performance of the multi-task ship recognition system VIOS-Net for each information source. This approach yields superior recognition results compared to models that rely solely on single-source information or that simply fuse multiple information sources without integrated decision making.
The experimental results demonstrate that our newly proposed multi-task deep ship recognition system, VIOS-Net, effectively establishes associations between multiple information sources, enabling joint reinforcement recognition. This significantly enhances the model’s learning efficacy and strengthens its ability to address the challenges posed by poor sensor imaging due to varying light, weather, and sea state conditions in real-world marine environments. Additionally, this paper offers four main contributions:
  • Integrated Multi-Task Fusion Learning Neural Network (VIOS-Net): This study introduces a multi-task fusion learning neural network (VIOS-Net) based on visible and infrared imaging for ship classification. To the best of our knowledge, this is the first model to integrate visible and infrared multi-information sources into a unified framework for complementary feature fusion and joint reinforcement within a deep ship recognition model. Our model achieves effective ship recognition using only an optical camera and an infrared imager, greatly enhancing recognition accuracy over single-sensor systems. This approach is particularly suited to the complex and dynamic maritime environment, ensuring around-the-clock maritime safety.
  • Superior Performance on Real-World Datasets: Using two real-world ship-monitoring datasets (visible and infrared), VIOS-Net achieves the best performance among the existing networks, with an accuracy of 96.199% on both datasets. Compared to the baseline ResNet-34 model, VIOS-Net shows significant improvements—25.688% in visible image recognition accuracy and 28.047% in infrared image recognition accuracy. Furthermore, compared to the simple linear weighted decision fusion of multiple information sources, VIOS-Net demonstrates a 6.00% improvement in accuracy on the visible dataset and a 10.80% improvement on the infrared dataset, highlighting the model’s generalization capabilities and its reduced risk of overfitting.
  • Impact of Weighting Coefficients on Performance: Through experimentation, we identified the influence of weighting coefficients on model accuracy. The accuracy of both visible and infrared tasks reached 96.199%, and adjustments to the weighting settings during training affected the highest accuracy by 4.150% for visible information sources and 4.650% for infrared information sources. The use of weight coefficients allows the model to more effectively transfer learned features from one task to another, addressing the limitations of neural networks in learning underlying features from a single task.
  • Effectiveness of Transfer Learning and Data Augmentation: Our experiments confirmed the effectiveness of transfer learning and data augmentation in improving model performance. Specifically, the use of transfer learning increased accuracy by 5.505% compared to not using it, while data augmentation led to a 2.097% increase in accuracy. These techniques enhance the model’s ability to generalize and improve overall recognition performance.
The remainder of this paper is structured as follows: Section 2 provides a comprehensive review of the related work. Section 3 offers an in-depth description of our fusion monitoring system designed for visible and infrared data sources. In Section 4, we present the experiments conducted and analyze the results obtained. Section 5 evaluates the effectiveness of the various components within our proposed architecture. Finally, Section 6 summarizes the findings and discusses potential directions for future research.

2. Related Works

This section provides a review of related works, which can be classified into two main types: (1) Traditional-Based Vessel Monitoring Method and (2) Image Information Source Vessel Monitoring Method.

2.1. Traditional-Based Vessel Monitoring Methods

Traditionally, radio communication devices aboard ships have been integral to maritime monitoring systems, playing a crucial role in facilitating effective shipping management by port authorities. Early research predominantly centered on enhancing the accuracy of maritime monitoring through the acquisition and optimization of data from radar systems, radio communication, and Automatic Identification System (AIS) equipment. These studies investigated various techniques to improve data collection and processing. As scientific and technological advancements have progressed, deep learning techniques have been increasingly incorporated into traditional ship monitoring systems. The integration of deep learning has significantly optimized the analysis of ship monitoring data, thereby reducing the need for human intervention and further improving the accuracy of shipping management.
Early maritime monitoring relied on shipboard radio devices (AIS, radar, and GPS) with three evolutionary milestones: the 1940s radar era established basic vessel tracking [17,18], 1990s digital communication enabled real-time reporting [19,20], and recent AI integration enhanced data processing [21,22]. Modern systems combine deep learning with legacy technologies: CNNs improve radar target detection in clutter [23], graph networks enhance spatiotemporal correlation [24], and transformer models detect AIS anomalies [25]. However, these device-dependent approaches face intrinsic limitations—over 25% of maritime accidents involve equipment failures according to EMSA 2024 reports [26].
In summary, while the aforementioned methods have proven effective in utilizing intelligent communication devices aboard ships for monitoring a limited number of vessels, several challenges remain. When these devices malfunction, maritime management agencies face significant obstacles in obtaining timely and accurate information about ship activities at sea. This limitation poses potential risks to maritime safety and regulatory oversight, as the inability to effectively monitor and manage ship behavior may introduce latent hazards that compromise both safety and compliance.

2.2. Image Information Source Vessel Monitoring Method

2.2.1. Single-Source-Based Vessel Monitoring Method

Deep learning techniques have significantly advanced the field of image recognition, and this progress is increasingly being integrated into traditional ship monitoring methodologies. The transition from monitoring conventional radio communication equipment to the analysis of ship images marks a pivotal development in maritime surveillance. This shift is largely driven by the superior capabilities of deep learning models in image recognition, where training on extensive datasets substantially enhances the precision and accuracy of ship identification. Unlike traditional methods, deep learning excels in extracting and discerning representative features from images, thereby improving classification and recognition.
The dominance of single-source approaches in maritime surveillance stems from their conceptual simplicity and direct sensor-specific optimization. The current methodologies can be categorized into three technical streams based on sensor modality and learning strategies: visible spectrum modeling, infrared thermal analysis, and SAR signal processing.
The visible technical stream, pioneered by Liu et al.’s dual-attention network [27], prioritizes RGB feature extraction through background-aware architectures, multi-scale pyramid networks (GLPM’s 7.2% accuracy gain [28]), and enhanced attention mechanisms (Xu et al.’s spatial-channel modules [29]). While achieving 89.3% mean accuracy in daylight conditions [30], these methods suffer severe performance degradation (<62% AP) under low-light scenarios due to photometric limitations.
The infrared technical stream, represented by Wu et al.’s SRCANet [31], employs thermal gradient encoding (Li et al.’s low-rank decomposition [32]), salience-guided detection (Chen’s morphological reconstruction [33]), and lightweight adaptation (YOLOv5-based nighttime detectors [34]). Though effective in darkness (84.5% recall rate [35]), fog/rain causes > 30% false positives from atmospheric thermal noise.
In the SAR technical stream, advanced through Zeng’s dual-polarized CNN [36], key innovations include speckle-robust convolutions (Li’s Faster R-CNN variant [37]) and transfer learning for small datasets (2.1× sample efficiency [38]). Despite high resolution (0.5 m precision [39]), SAR struggles with dense ship clusters (<70% detection rate in [40]).
The accumulated evidence underscores a critical insight: the physical limitations of individual sensors create irreversible information bottlenecks that no algorithmic refinement can overcome. This realization motivates our shift toward adaptive multi-source fusion, as detailed in Section 2.2.2.

2.2.2. Multi-Source-Based Vessel Monitoring Method

The evolution of maritime monitoring systems reveals an essential paradox: while single-sensor approaches dominate the current implementations (infrared/visible/SAR), their inherent limitations fundamentally contradict the dynamic complexity of marine environments. Traditional methods relying on single-source data exhibit three critical failures: spectral bias where modality-specific features (e.g., thermal signatures in infrared) overshadow complementary information; environmental fragility causing performance degradation under lighting/weather variations; and scale rigidity limiting adaptation to multi-resolution targets. These deficiencies originate from neural networks’ tendency to overfit sensor-specific patterns while ignoring cross-modal relationships.
As the demands on maritime management technology evolve, image classification algorithms grounded in computer vision are playing an increasingly critical role. To overcome challenges such as incomplete information, feature extraction biases, and limited generalization capabilities, researchers have developed methods that train models on data from multiple information sources simultaneously, resulting in more effective monitoring outcomes. Recent advancements address these challenges through three architectural paradigms: parallel modality processing, hybrid hierarchical fusion, and attention-based integration. Exemplified by Li et al. [41]’s dual-stream network for satellite/contour images, this approach achieves 12.6% accuracy gains but incurs 38% computational overhead due to separate processing branches. While transfer learning mitigates domain gaps, their fixed-weight fusion (0.5:0.5 ratio) fails to adapt to environmental changes. Similarly, Aziz et al. [42] introduced a recognition method based on a multi-source convolutional neural network (CNN) for visible and infrared spectra, although the method’s ship recognition accuracy remained relatively low. Chang et al. [40,43]’s YOLOv3 variants demonstrate how multi-scale feature extraction (through SPP modules) improves small ship detection (F1-score +9.3%). However, their exhaustive cross-sensor interactions lead to quadratic complexity growth (O(n2)), making real-time processing infeasible with >3 sensors. Additionally, Chen et al. [44]’s parametric edge convolution reduces resource consumption by 22%, yet their static attention mechanism cannot handle heterogeneous sensor sampling rates—a critical limitation in moving vessel scenarios.
The aforementioned studies provide a practical framework for continuous, all-weather ship monitoring to enhance ship monitoring accuracy, marking significant progress in maritime monitoring research. However, many of these approaches predominantly rely on single-source images—such as infrared, visible light, or SAR imagery—for ship identification, often neglecting the potential benefits of integrating information from multiple sources. Such integration could substantially improve the accuracy of ship identification at sea. Moreover, challenges persist in multi-source ship recognition methods, including suboptimal quality in serial feature fusion and high algorithmic complexity. However, four fundamental limitations persist: isomorphic feature treatment disregarding sensor-specific characteristics; fixed fusion weights failing to adapt to environmental dynamics; computation redundancy from exhaustive cross-sensor interaction; and lack of self-supervised alignment for modality gaps.
Additionally, the existing research has largely focused on deep learning tasks without sufficiently addressing the challenges inherent in multi-source ship recognition methods based on deep learning. This oversight significantly hinders the practical applicability of these models within maritime management systems. Addressing these limitations is crucial for developing more effective and feasible solutions for maritime monitoring and management.

3. VIOS-Net System

The overall workflow of the VIOS-Net network proposed in this paper is divided into three main stages: the preprocessing and augmentation unit (PAU), the shared feature extractor (SFE), the Unique Feature Extractor (UFE), and the Result Output Module, as illustrated in Figure 1 below. The first stage, termed the Data Preprocessing and Augmentation stage, involves the preprocessing and enhancement of infrared and visible light datasets from multiple information sources. In the PAU, input data are processed according to the specific characteristics of each information source. Subsequently, images of the same ship, captured by different sensors from multiple information sources at the same angle, are simultaneously fed into two channels—comprising a visible subnet and an infrared subnet—both of which are based on the ResNet-34 architecture. These two subnets together form the SFE, with shared weights between them during the training process, enabling the learning of effective classification features from the preprocessed and augmented images. The SFE Block performs matrix transformations on the two datasets, producing feature data that integrate shared information from multiple sources. These shared feature data are then input into the UFE, a fully connected layer corresponding to the visible and infrared subnets, where the SoftMax function is applied to extract effective categorical features, yielding predicted ship category labels at the output layer. The error between the actual ship category labels and the predicted labels is calculated, and a weight backpropagation algorithm is employed to iteratively update the weights and biases of the UFE in VIOS-Net, optimizing the implementation of the visible and infrared subnets. During the model performance evaluation phase, after training is completed, multi-source images are fed into the visible and infrared sub-networks separately to extract effective classification features. The optimal model is then used to test visible and infrared images, with the SoftMax function independently classifying each, resulting in final maritime vessel image recognition outcomes for the visible and infrared subnets. The proposed VIOS-Net system demonstrates the ability to perform joint visible and infrared monitoring with high accuracy, providing a practical and reliable solution for maritime monitoring.
The following sections present a comprehensive description of the VIOS-Net network model, covering the Problem Statement, preprocessing and augmentation unit, shared feature extractor, and Unique Feature Extractor.

3.1. Problem Statement

We defined the ship monitoring task as a hybrid task, which can be divided into a visible light monitoring subtask and an infrared monitoring subtask. The visible light monitoring subtask is defined as follows: given a ship’s visible light image, the model determines the type of ship depicted in the image. The visible light dataset is denoted as I i v , i { 1 , , δ } , where I i v is the i -th image in the dataset, and δ represents the total number of images in the visible light ship image dataset. The output of this subtask is denoted as f v : I i v y i v ^ , where y i v ^ denotes the predicted labels of the ship’s visible images. Similarly, the infrared monitoring subtask uses a dataset denoted as I i r , i { 1 , , λ } , where I i r is the i -th image in the dataset, and λ denotes the total number of images in the infrared ship image dataset. The output of this subtask is denoted as f r : I i r y i r ^ , where y i r ^ represents the predicted labels of the ship’s infrared images.

3.2. Preprocessing and Augmentation Unit

The preprocessing and augmentation unit (PAU) aims to enhance the quality of the images, enable the model to learn various features and changes in the data, and ultimately improve model performance. Additionally, the transformations applied during data augmentation generate diverse training samples and reduce the model’s dependency on the training data, making the model more robust, less susceptible to noisy data, and better at mitigating overfitting. This, in turn, improves the model’s generalization ability. The images generated by preprocessing and data augmentation are shown in Figure 2.
Figure 2. The images are generated based on Algorithms 1 and 2.
Figure 2. The images are generated based on Algorithms 1 and 2.
Jmse 13 00913 g002
Algorithm 1 Preprocess and Augmentation for Visible Light Images Algorithm
Input: Outboard Profile Visible Light Ship Inspection Dataset I i v , 1 , , δ ;
Require: Random rotation function, Random Rotation(); Flipping function, Flipping();
   Scaling function, Scaling(); Cropping function, Cropping().
1.   #Mean normalization
2.   I ¯ mean v = 1 δ i = 1 δ I i v
3.  for i in 1 δ do
4.     I i v = I i v I ¯ m e a n v
5.   end for
6.   # Data standardization
7.   σ = 1 δ i = 1 δ I i v 2 # σ is the standard deviation
8.  for i in 1 δ do
9.     I i v = I i v / σ
10.   end for
11.  # Augmentation
12.  SetRandomSeed(seed_value)
13.  for i in 1 δ do
14.     I i v 1 R a m d o m   R o t a t i o n I i v , θ [ 30 ° , 30 ° ] # Simulate vessel orientation variations under wave disturbances
15.     I i v 2 H o r i z o n t a l F l i p p i n g I i v # Leverage port-starboard symmetry in ship structures
16.     I i v 3 S c a l i n g I i v , l [ 0.8 , 1.2 ] # Emulate multi-scale observation from varying distances
17.     I i v 4 C r o p p i n g I i v , W 0.7 w , H 0.7 h # Simulate partial occlusion in camera-captured scenarios
18.  end for
19.   I i v = I i v 1 I i v 2 I i v 3 I i v 4 # Merge the original dataset of four data augmentation methods
Output: preprocess and augmented visible light dataset I i v
Algorithm 2 Preprocess and Augmentation for Infrared Images Algorithm
Input: Outboard Profile Infrared Ship Inspection Dataset I i r ,   1 , , λ ;
Require: Random rotation function, RandomRotation(); Flipping function, Flipping();
      Scaling function, Scaling(); Cropping function, Cropping();Grayscale inversion
      function(); Contrast adjusting function();Pseudo-color mapping function, Pseudo
      ColorMapping().
1.      # Mean normalization
2.      I ¯ mean r = 1 λ i = 1 λ I i r
3.     for i in 1 λ do
4.       I i r = I i r I ¯ m e a n r
5.      end for
6.      # Data standardization
7.      σ = 1 λ i = 1 λ I i r 2   #   σ   is   the   standard   deviation
8.     for i in 1 λ do
9.       I i r = I i r / σ
10.      end for
11.     # Augmentation
12.     SetRandomSeed(seed_value)
13.     for i in 1 λ do
14.       I i r 1 R a m d o m R o t a t i o n I i r , , θ [ 30 ° , 30 ° ] # Simulate vessel orientation variations under wave disturbances
15.       I i r 2 H o r i z o n t a l F l i p p i n g I i r # Leverage port-starboard symmetry in ship structures
16.       I i r 3 S c a l i n g I i r , , l [ 0.8 , 1.2 # Emulate multi-scale observation from varying distances
17.       I i r 4 C r o p p i n g I i r , W 0.7 w , H 0.7 h # Simulate partial occlusion in camera-captured scenarios
18.      end for
19.      I i r = I i r 1 I i r 2 I i r 3 I i r 4 # Merge the original dataset of four data augmentation methods
20.     for i in 1 λ do
21.       I i r = G r a y s c a l e I n v e r s i o n I i r
22.       I i r = C o n t r a s t A d j u s t i n g I i r
23.       I i r = P s e u d o C o l o r M a p p i n g I i r , I i r , I i r
24.      end for
Output: preprocess and augmented infrared dataset I i r

3.2.1. Preprocessing and Augmentation Unit for Visible Light Images

For the visible light image dataset, two preprocessing steps were performed: mean normalization and data standardization. Additionally, we implemented physics-constrained data augmentation through four maritime-compliant transformations: rotation, flipping, scaling, and cropping. The union operation (line 18) aggregates these mutually exclusive augmented samples into the training set, where each image undergoes only one transformation per iteration. This approach expands data diversity while maintaining maritime physical realism. The algorithm of the preprocessing and augmentation unit for the visible light dataset is described in Algorithm 1.
1.
Zero-Centering Normalization
Zero-Centering normalization adjusts the pixel value distribution of an image to be centered around zero, reducing data bias and improving model training. This technique helps the model converge more quickly and prevents gradient saturation when activation functions are used in neural network layers, thus enhancing training efficiency and performance.
2.
Data Standardization
Data standardization modifies each channel’s data so that the mean is zero and the standard deviation is one. This process ensures that the normalized data falls within the same range as the activation function, specifically between 0 and 1. Consequently, fewer non-zero gradients occur during the training of the visible light monitoring module, enabling faster learning by the network’s neurons.
3.
Rotating
Rotating an image can simulate the appearance of a ship at different angles, enhancing the robustness of the model to changes in orientation while increasing the diversity of the dataset. The rotation angles are constrained to ±30° to mimic realistic ship pitching motions in maritime environments. The formula for rotation is as follows:
w h = c o s θ s i n θ s i n θ c o s θ w 2 h 2 ( θ [ 30 ° , 30 ° ] )
where the given image size h is for height and w for width, the midpoint of the image is w 2 , h 2 , the angle of rotation is θ , and the new position of the midpoint is w , h .
4.
Horizontal flipping
Horizontal flipping an image increases the amount and diversity of training data, addressing the lack of samples in certain directions within the dataset. Specifically, while vertical flipping is deliberately excluded due to its contradiction with fundamental buoyancy principles in vascular dynamics, horizontal flipping maintains physical plausibility. The formula for horizontal flipping is as follows:
x = w x
y = y
where x , y are arbitrary coordinate points in the given image, and x , y are the flipped coordinates.
5.
Scaling
Random scaling of the image can simulate the change in the size of the ship at different distances and help the model to identify the ship effectively at different scales, the scaling formula for the image is as follows:
x = l · x   ( l [ 0.8 ,   1.2 ] )
y = l · y ( l [ 0.8 , 1.2 ] )
where x , y are the coordinates of a point in the image, x , y is the scaled coordinate, and l is the scaling factor.
6.
Cropping
Cropping involves selecting a random sub-image from an image, allowing the model to extract image data from different regions and enhancing its ability to learn diverse image information. The cropped image includes the following pixels from the original image:
x x 0 , x 0 + W ,   W [ 0.7 w , w ]
y y 0 , y 0 + H ,   H [ 0.7 h , h ]
where x 0 0 , w W , y 0 0 , h H , w , h is the size of the original image, W , H is the size of the cropped image, and x 0 , y 0 are the coordinates of the upper left corner.

3.2.2. Preprocessing and Augmentation Unit for Infrared Images

For the infrared (IR) image set, we applied two preprocessing steps: Zero-Centering normalization and data standardization. Additionally, six maritime-constrained data augmentation techniques were implemented: rotation, flipping, scaling, cropping, grayscale inversion, and contrast adjustment. We also achieved pseudo-color mapping by channel splicing the original IR image with the grayscale-inverted and contrast-adjusted images. The algorithm for the preprocessing and augmentation unit for the Infrared Dataset is detailed in Algorithm 2.
1.
Basic Preprocessing and Data Augmentation for Infrared Images
The preprocessing and data augmentation techniques applied to infrared images are the same as those used for visible light images, including rotation, horizontal flipping, scaling, and cropping.
2.
Pseudo-color Mapping
Pseudo-color mapping significantly enhances the informativeness and visual interpretability of infrared images, facilitating improved analysis and processing by both human observers and computer algorithms. This technique involves enhancing contrast and inverting the grayscale within the infrared data processing unit to enrich the visual content of the image. As a result, the features within the image become more prominent, simplifying observation, analysis, and interpretation, while also making the image more accessible for computer processing and deep learning applications. The pseudo-color mapping formula is expressed as follows:
I i r c o n c a t I i r , I i r , I i r
In this formula, I i r on the left side of the concat function denotes the original infrared image. I i r represents the infrared image after grayscale inversion, while I i r refers to the infrared image after histogram equalization. The result of the concat function is a new infrared image that has undergone the stitching process, effectively integrating the information from all three versions into a single, enhanced dataset.
  • Grayscale Inversion
Grayscale inversion facilitates the exchange of light and dark regions within the original image, thereby enhancing the level of detail available for pseudo-color mapping. This transformation also modifies the visual presentation of the image, making distinct features more easily identifiable. The formula for grayscale inversion is given by the following:
I i r x , y = 255 I i r x , y
where x , y represents any coordinate point within the image, while I i r x , y denotes the gray value at this coordinate in the original image. The resulting I i r x , y is the gray value at the corresponding point in the grayscale-inverted image I i r .
  • Histogram Equalization
Histogram equalization is employed to redistribute the gray levels within an image, effectively enhancing contrast and addressing issues such as the blurring of boundaries between the monitored ship and its background. This adjustment makes the ship more prominent within the image. The contrast transformation process is outlined as follows, beginning with the computation of the gray-level histogram:
n g = I g
In this equation, n g represents the frequency of occurrence for a given gray level g within the image. Here, g is the gray value at pixel I x , y in the ith image I i r , where g 0,255 . I g is a matrix that shares the same dimensions as the ith image I i r , and I g x , y is defined as follows:
I g x , y = 1 if   I x , y = i 0 else
Next, the probability density function (PDF) is calculated using the following formula:
p g = n g w × h
where w × h represents the total number of pixels in the image, with w and h corresponding to the image’s width and height, respectively. Following this, the cumulative distribution function (CDF) is determined as follows:
c g = j = 0 g p j
Finally, gray level mapping is performed, wherein the original image is mapped from its gray level I i r x , y to the new gray level I i r x , y using the CDF as follows:
I i r x , y = 255 · c g I i r x , y
where · represents the floor function, which rounds down to the nearest integer, yielding the histogram-equalized image I i r .

3.3. Shared Feature Extractor

The infrared (IR) and visible light images I i r and I i v , obtained from the preprocessing and augmentation units, are subsequently fed simultaneously into the shared feature extractor (SFE) for feature extraction. To accurately recognize a set of ship images relevant to both tasks, we employed the ResNet-34 model as the backbone, utilizing its residual network for robust feature extraction. The data are fed into the S F E through dual channels, as represented by the following formula:
s = S F E I i r , I i v
where s denotes the extracted feature, S F E stands for the shared feature extractor, and I i r and I i v represent the infrared and visible images, respectively.

3.3.1. Residual Networks Block Unit

These image representations are processed through multiple shared residual building blocks to achieve deeper feature extraction. The structure of these residual blocks is depicted in Figure 3, which includes a convolutional layer, a batch normalization layer, and an activation layer. Each residual building block follows two potential paths: the direct path Residual x and the shortcut path Residual x + x . The latter is implemented through feedforward neural networks with “shortcuts,” allowing the bypass of one or more layers, thereby facilitating efficient information transfer and reducing the risk of model overfitting.
After processing through the residual building blocks, the information is passed to the subsequent layer, the Average Pooling layer, for dimensionality reduction. This processed information is then forwarded to the Unique Feature Extractor (UFE) for specialized feature extraction. The detailed algorithm for the residual networks block unit in the shared feature extractor is provided in Algorithm 3.
Algorithm 3 Residual Networks for Shared Feature Extractor
Input: Preprocessed visible light image set I i v ,   { 1 , , δ } ;   I n f r a r e d   i m a g e   s e t   I i r , { 1 , , λ } ;
Require: basic block of residual networks, Residual(); convolution layer, Conv(); max
      pooling function, Maxpool(); avgpooling function, Avgpool(), Fully Connected
      Layer, FC()
1.      # Visible Image Feature Extracting
2.      s i v C o n v I i v
3.      s i v M a x p o o l s i v
4.     for i in [3,4,6,3] do #[3,4,6,3] is the number of the residual basic block
5.      for j in 1   i do
6.          s i v R e s i d u l e ( s i v )
7.       endfor
8.      endfor
9.      s i v A v g p o o l ( s i v )
10.      # Infrared Images Feature Extracting
11.      s i r C o n v I i r
12.      s i r M a x p o o l s i r
13.     for i in [3,4,6,3] do #[3,4,6,3] is the number of the residual basic block
14.      for j in 1   i do
15.          s i r R e s i d u l e s i r
16.       end for
17.      end for
18.      s i r A v g p o o l s i r
Output: Infrared images feature s i r ; visible light images feature s i v

3.3.2. Shared Feature Extracting Block Unit

Within the Shared Feature Extracting Block Unit, a set of related visible and infrared image information is concurrently fed into the SFE from both channels to obtain a more sophisticated representation. The core layer of this unit, the convolutional layer, which is integral to the convolutional neural network (CNN), scans the input data using convolutional kernels to extract high-level features. These extracted features are then processed through a max-pooling layer to identify the most representative features. The formula governing the convolutional layer operation is as follows:
s i j r = m n I i r i + m × w m , n
s i j v = m n I i v i + m × w m , n
where s denotes the extracted feature, I is the input image to the convolutional layer, w represents the weights of the convolutional kernel, i and j are the dimensions of the extracted information, and m and n are the dimensions of the convolutional kernel. After processing the IR and visible images through the SFE, their features are extracted and subsequently fused. The fusion process is represented by the following formula:
s s h a r e d = s i j r · s i j v
In this equation, s s h a r e d represents the shared feature derived from the fusion of visible and infrared image features, while s i j r and s i j v denote the features obtained from the SFE for the infrared and visible images, respectively.

3.4. Unique Feature Extractor

To systematically address platform-induced spatiotemporal misalignments between modalities, the Unique Feature Extractor (UFE) operates in tandem with the shared feature extractor (SFE) through a dual-branch architecture. While SFE focuses on multi-source data invariant patterns, UFE explicitly handles modality-specific distortion correction and feature recalibration through adaptive attention mechanisms—crucial for maintaining representation integrity under practical parallax conditions.

3.4.1. Unique Feature Extracting Block Unit

To systematically address platform-induced spatiotemporal misalignments between modalities, two sets of fully connected layers are employed to perform unique feature extraction for both visible and infrared images, tailored to the respective tasks. The UFE block processes multi-modal features through parallel transformation paths: employing F C r with temperature-aware kernel initialization to enhance thermal signature resolution and utilizing F C v with spectral sensitivity normalization for RGB feature adaptation. The extraction process is governed by the following formulas:
s i j r F C r s s h a r e d
s i j v F C v s s h a r e d
y i r ^ = S o f t m a x W r · s i j r + b r
y i v ^ = S o f t m a x W v · s i j v + b v
where s s h a r e d represents the data output from the shared feature extractor, while W r and b r denote the weight and bias for the infrared image extractor. Similarly, W v and b v represent the weight and bias of the visible light image extractor within the Unique Feature Extractor (UFE). Though sharing the input s s h a r e d , these fully connected layers contain disjoint parameters optimized for distinct sensor physics.

3.4.2. Unique Feature Extracting Weight Factor

In multi-task training involving multiple information sources, the visible light and infrared images originate from different sensors, which inevitably results in varying image information being captured by these sensors. As the deep learning model is trained on images from these diverse sources, it can lead to discrepancies in how the model attends to the information from each source, as well as differences in the training progress for the two tasks.
To address this, we introduced different weight coefficients during backpropagation for the tasks under varying information sources. This approach optimizes the model’s accuracy across both tasks by adjusting the focus of the learning process. In traditional multi-task training, each task is typically assigned the same weight during training. However, by configuring appropriate weighting coefficients and learning rates for each task, the model can be directed to focus more or less on certain tasks, thereby enhancing or diminishing its ability to capture task-specific features. This method helps the model develop a deeper and more precise task-specific representation. Moreover, due to the model’s shared representation learning nature, these adjustments not only impact the specific task for which the learning rate is configured but also positively influence other tasks, allowing the model to balance the learning across multiple tasks. The loss function for predicting results and calculating label loss is given by the following:
L r = β r · i = 1 N y i log y i ^
L v = ( 1 β r ) · i = 1 N y i log y i ^
L = m e a n ( L r + L v )
where L v and L r represents the losses between the predicted result and the true label for visible and infrared images, respectively. L represents the average loss of L v and L r , where β r serves as the weight factor for the infrared data source. The formula for updating the weights for unique feature extraction is as follows:
w w α L W
b b α L b
In these equations, w denotes the weight, b is the bias, and α is the learning rate. This approach enables the model to extract specific features from images derived from different information sources effectively.

4. Experiments and Results

4.1. Dataset

Due to the limited availability of infrared-visible related datasets in the maritime domain, we utilized two publicly available datasets that were balanced and augmented for both visible and infrared images. These datasets were used to train the VIOS-Net model, enabling the evaluation of its performance. The dataset categorizes ship targets into six types: medium-sized “other” ships, medium-sized passenger ships, merchant ships, sailboats, small boats, and tugboats. Given the imbalanced data distribution in both the infrared and visible light datasets, as shown in Table 1, with sailboats and small boats accounting for 24.6% and 39.1%, respectively, and other categories representing less than 20%, we addressed the issue of data imbalance to mitigate model training bias. Initially, we employed data augmentation techniques to achieve a more balanced dataset distribution across categories, as shown in Table 2. Following augmentation, the data were relatively evenly distributed and subsequently split into training and testing datasets at an 8:2 ratio, as depicted in Table 3. This resulted in 3052 training images and 763 testing images. An example of the dataset is illustrated in Figure 4.
The performance of the VIOS-Net system was quantitatively evaluated using several key metrics, including accuracy, precision, recall, and F1-score. These metrics are essential for assessing the effectiveness of the model in classification tasks. The evaluation metrics are defined as follows:
As a classification problem, six evaluation metrics are used in this paper to evaluate the performance of VIOS-Net, Accuracy, Recall, Precision, and F1-score (F1), these formulas are shown below:
Accuracy = (TP + TN)/(TP + FP + FP + FN)
Recall = TP/(TP + FN)
Precision = TP/(TP + FP)
F1 = 2 × (Precision × Recall)/(Precision + Recall)
In these formulas, TP (True Positive) refers to the number of samples correctly predicted as positive by VIOS-Net, FP (False Positive) is the number of samples incorrectly predicted as positive, TN (True Negative) represents the number of samples correctly predicted as negative, and FN (False Negative) denotes the number of samples incorrectly predicted as negative.

4.2. Experiment Setup

The VIOS-Net system was constructed using the ResNet-34 model, which was pre-trained on the ImageNet dataset through transfer learning and served as the backbone network for the monitoring module. The entire model was optimized using a proposed multivariate cross-entropy loss function, incorporating weighting coefficients with a ratio of 0.4:0.6 for infrared and visible tasks to enhance task-specific optimization. The optimizer chosen was Averaged Stochastic Gradient Descent (ASGD). Empirically, we set the batch size to 32, the learning rate to 0.0001, and the number of training iterations to 60.
The experiments were conducted on a system running a 64-bit Windows 11 operating system, powered by a 12th generation Intel (R) Core (TM) i7-12700 processor at 2.10 GHz, equipped with 32 GB of RAM and an NVIDIA GeForce RTX 3060 graphics card. The deep learning framework employed was PyTorch 1.13, with PyCharm 2024.1.1 (Community Edition) as the primary software tool, and Python 3.9 as the programming language.

4.3. Performance Comparison

This section presents a comparative analysis of our model against the existing baseline models, including widely used deep learning models for classification tasks such as AlexNet, VGG-13, VGG-16, GoogLeNet, ResNet-18, and ResNet-34, utilizing the VAIS dataset. All the comparative models were initialized with ImageNet pre-trained weights. The VIOS-Net framework not only outperformed traditional baselines but also surpassed the improved CNN and Dual CNN models across all the evaluated metrics. The comparison metrics—accuracy, precision, recall, and F1-score—are summarized in Table 4. The VIOS-Net framework demonstrated state-of-the-art performance on the VAIS dataset, underscoring its capability for high-accuracy monitoring and strong generalization across different datasets. The experimental outcome is presented in the last row of Table 4.
Among the evaluated models, ResNet-34 emerged as the best-performing backbone network for image recognition, which may be attributed to ResNet’s architecture that effectively mitigates the issues of gradient vanishing and exploding in deep neural network training. This allows for the successful training of very deep networks without performance degradation. As evidenced in Table 4, VIOS-Net achieved a substantial improvement in accuracy metrics, with an 8.010% and 7.732% increase for visible and infrared imaging, respectively, and an 8.059% and 7.879% improvement in F1-scores compared to ResNet-18. The multi-task learning approach employed by VIOS-Net facilitated complementary information sharing and feature extraction through the shared feature extractor (SFE), significantly enhancing recognition performance. Additionally, multi-task learning improved the model’s generalization ability and reduced the risk of overfitting by leveraging shared features across multiple tasks.
As depicted in Figure 5 and Figure 6, the ROC curves for the visible light and infrared monitoring task indicate that VIOS-Net achieved 96.20% accuracy in ship classification tasks. The macro average and micro average ROC curve values surpassed 0.99 for visible and infrared image recognition. Compared to the other networks, the superior monitoring performance of VIOS-Net in both the infrared and visible light detection tasks highlights its robust capability to provide stable performance across diverse ship recognition scenarios.

5. Quantitative Analysis

To validate the effectiveness of the various components in the VIOS-Net architecture, several ablation experiments were conducted using different variants of the model. These variants include VIOS-Augmentation—a variant of VIOS-Net that excludes data augmentation; VIOS-Transfer—a variant of VIOS-Net that excludes the transfer learning module; VIOS-Infrared—the infrared image single-task network within VIOS-Net; VIOS-Visible—the visible light single-task network within VIOS-Net; VIOS-SFE—a variant of VIOS-Net that excludes the Shared Feature Extracting module; VIOS-Weight—a variant of VIOS-Net that removes the effect of weight coefficients.

5.1. Preprocessing and Augmentation Unit Impact Study

To demonstrate the improvement brought about by the transfer learning and data augmentation strategies, an experiment was conducted comparing the performance of VIOS-Net with and without data augmentation. The results are presented in Table 5 and Figure 7. In the VIOS-Augmentation model, which lacks preprocessing and data augmentation, the original dataset I i r was directly input into VIOS-Net without any pseudo-color mapping or data augmentation. The results indicate that the absence of preprocessing and data augmentation led to a 2.097% decrease in accuracy on the visible light dataset and a 1.966% decrease in the infrared dataset compared to VIOS-Net. These findings underscore the effectiveness of the preprocessing and data augmentation strategies employed in VIOS-Net.

5.2. Transfer Learning Impact Study

To assess the impact of transfer learning on VIOS-Net, the transfer learning module was excluded, resulting in the VIOS-Transfer model. As demonstrated in Table 6 and Figure 8, the absence of transfer learning led to a decrease in prediction accuracy by 5.505% for visible light data and 5.374% for infrared data. Additionally, the AP values of the PR curves declined by 0.3 and 0.2, respectively. These results underscore the crucial role of transfer learning in enhancing information flow within the fusion network, thereby emphasizing its significance in improving the overall performance of VIOS-Net. As depicted in Figure 9, VIOS-Net exhibits distinct learning behaviors under three configurations: (1) baseline (without augmentation/transfer learning), (2) transfer learning only, and (3) combined augmentation-transfer learning.

5.3. Shared Feature Extraction Impact Study

As depicted in Table 7, VIOS-Net achieved a 25.688% improvement in accuracy over the VIOS-Visible variant and a 28.047% improvement over the VIOS-Infrared variant. In terms of the F1-score, VIOS-Net showed a 23.678% improvement over VIOS-Visible and a 27.075% improvement over VIOS-Infrared. These results indicate that both visible and infrared features are crucial for accurate target recognition, and that VIOS-Net effectively enhances fusion network performance by extracting shared feature representations from both datasets.
To further validate the efficiency of the shared feature extractor (SFE) in reducing model complexity, an additional experiment was conducted. In this experiment, the SFE was removed, resulting in the VIOS-SFE model. The findings, as presented in Table 7, show that removing the SFE from ResNet-34 increased the number of parameters by 48.818% compared to VIOS-Net. VIOS-Net’s SFE module allows for more efficient feature extraction with fewer parameters, reducing model complexity, training time, and overfitting risks. Additionally, fewer parameters translate to lower computational and storage costs during both the training and inference phases. Among the tested models, VIOS-Net strikes the best balance between parameter reduction and accuracy improvement.
As the scale of visual models continues to expand, their parameter sizes frequently reach billions or even trillions, imposing extremely high demands on computational and storage resources. Against this backdrop, the improvement in parameter efficiency is manifested in three key dimensions. In terms of computational cost, a lower parameter density (Params/Accuracy) can significantly reduce the amount of computation, effectively alleviating the computational pressure brought by large-scale models. Regarding storage efficiency, reducing parameter density not only avoids high-cost backpropagation in large-scale backbone networks, thereby substantially saving GPU memory to meet the storage constraints of edge computing devices (such as the shipborne NVIDIA Jetson Nano), but also, in practical applications involving large-scale models and numerous downstream tasks, changes the situation where full fine-tuning requires storing massive model weights for each task. By only storing a small number of parameters, it significantly reduces storage burden and costs. In terms of training efficiency, compared with the current mainstream lightweight methods, VIOS-Net achieves a parameter density of 0.232M/% while maintaining 96.199% classification accuracy, which demonstrates the superiority of this method in balancing accuracy and cost. These cost advantages are of great significance for practical deployment, enabling better promotion of the application and implementation of related technologies in real-world scenarios.

5.4. Weight Coefficient Impact Study

The impact of adjusting the weighting coefficients on the performance of VIOS-Net was examined using the VIOS-Weight variant. As shown in Table 8 and Figure 10, altering the weight settings affected model accuracy by up to 4.150% for the visible light dataset and 4.650% for the infrared dataset. Figure 10 illustrates the weighting pattern, which shows that by fine-tuning the weight coefficients, the model can better manage the learning progress of both tasks, leading to optimal performance in both visible and infrared image recognition tasks.

6. Conclusions

In this paper, we proposed a novel automatic ship monitoring model that leverages the fusion of multiple information sources to enhance maritime management effectiveness under various lighting conditions and adverse weather such as snow, rain, frost, dust, and haze. This model addresses the challenges posed by traditional visible light cameras in accurately identifying ship types. Specifically, the proposed method integrates a preprocessing and data augmentation module, a shared feature extraction module, and a specific feature extraction module.
The preprocessing and data augmentation module is designed to enhance the quality of the training data, thereby improving the model training process and its overall effectiveness. The shared feature extraction module extracts two sets of highly correlated features from multi-source data, enabling the complementary exchange of information between data sources through joint augmentation and weight sharing. This bidirectional feature extraction process significantly improves accuracy by 25.688% and 28.047% compared to separate training on visible and infrared datasets while reducing the number of parameters of the ResNet-34 backbone network by 48.818%. The specific feature extraction module further optimizes the fused features from both information sources, tailoring the dual-channel network for superior performance. Additionally, this study introduces a task weighting factor, uncovering the relationship between accuracy and the task weighting setup, thereby achieving targeted specific feature optimization. This approach effectively compensates for the limitations of visible light cameras in real-world environments, contributing valuable theoretical insights toward the realization of all-weather maritime ship management.
To further enhance the efficiency of maritime management, future research will explore two key directions: Firstly, we will investigate the incorporation of adaptive augmentation learning techniques during the fusion process of multiple information sources. This will allow the model to dynamically adjust data weights and feature extraction strategies in response to environmental changes, thereby improving the model’s robustness and generalization capabilities. Secondly, we will optimize the sensor arrangement and fusion strategies for image data within the same modality to enhance data complementarity and consistency, ultimately improving the overall performance of the model. Furthermore, the feature-sharing module will be extended to sensor data from different modalities, such as radar and sonar, to develop a more comprehensive multi-modal deep learning model capable of enhancing ship monitoring in more complex environments.
Thanks to the information-sharing and joint augmentation capabilities of our proposed VIOS-Net, it maintains excellent feature extraction performance even when processing data from multiple sources. This significantly mitigates the limitations of optical sensors under varying lighting conditions and adverse weather, offering valuable insights for the achievement of continuous, all-weather ship monitoring.

Author Contributions

Conceptualization, methodology, resources, software, validation, supervision, and writing—original draft preparation, J.Z.; methodology, conceptualization, software, validation, supervision, and writing—review and editing, J.L.; data curation and validation, J.S.; validation and software, L.W.; funding acquisition, methodology, and supervision, H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Young Innovative Talents Grants Program of Guangdong Province (Grant No. 2022KQNCX024), the Ocean Young Talent Innovation Program of Zhanjiang City (Grant No. 2022E05002), the National Natural Science Foundation of China (Grant No. 52171346), the Natural Science Foundation of Guangdong Province (Grant No. 2021A1515012618), the program for scientific research start-up funds of Guangdong Ocean University, the College Student Innovation Team of Guangdong Ocean University (Grant No. CXTD2024018), the Student Innovation Team of Port Industry TechAcademy (Grant No. GHCY2024008), and the Innovation and Entrepreneurship Training Program for College Students (Grant No. CXXL2024221).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, Y.; Wang, Z. biSAMNet: A Novel Approach in Maritime Data Completion Using Deep Learning and NLP Techniques. J. Mar. Sci. Eng. 2024, 12, 868. [Google Scholar] [CrossRef]
  2. Zhang, S.; Chen, J.; Wan, Z.; Yu, M.; Shu, Y.; Tan, Z.; Liu, J. Challenges and Countermeasures for International Ship Waste Management: IMO, China, United States, and EU. Ocean Coast. Manag. 2021, 213, 105836. [Google Scholar] [CrossRef]
  3. Li, J.; Sun, J.; Li, X.; Yang, Y.; Jiang, X.; Li, R. LFLD-CLbased NET: A Curriculum-Learning-Based Deep Learning Network with Leap-Forward-Learning-Decay for Ship Detection. J. Mar. Sci. Eng. 2023, 11, 1388. [Google Scholar] [CrossRef]
  4. Sun, J.; Li, J.; Li, R.; Wu, L.; Cao, L.; Sun, M. Addressing unfamiliar ship type recognition in real-scenario vessel monitoring: A multi-angle metric networks framework. Front. Mar. Sci. 2025, 11, 1516586. [Google Scholar] [CrossRef]
  5. Feng, Y.; Yin, H.; Zhang, H.; Wu, L.; Dong, H.; Li, J. Independent Tri-Spectral Integration for Intelligent Ship Monitoring in Ports: Bridging Optical, Infrared, and Satellite Insights. J. Mar. Sci. Eng. 2024, 12, 2203. [Google Scholar] [CrossRef]
  6. Li, X.; Li, D.; Liu, H.; Wan, J.; Chen, Z.; Liu, Q. A-BFPN: An Attention-Guided Balanced Feature Pyramid Network for SAR Ship Detection. Remote Sens. 2022, 14, 3829. [Google Scholar] [CrossRef]
  7. Li, W.; Ning, C.; Fang, Y.; Yuan, G.; Zhou, P.; Li, C. An Algorithm for Ship Detection in Complex Observation Scenarios Based on Mooring Buoys. J. Mar. Sci. Eng. 2024, 12, 1226. [Google Scholar] [CrossRef]
  8. Połap, D.; Włodarczyk-Sielicka, M.; Wawrzyniak, N. Automatic Ship Classification for a Riverside Monitoring System Using a Cascade of Artificial Intelligence Techniques Including Penalties and Rewards. ISA Trans. 2022, 121, 232–239. [Google Scholar] [CrossRef]
  9. Baek, W.-K.; Kim, E.; Jeon, H.-K.; Lee, K.-J.; Kim, S.-W.; Lee, Y.-K.; Ryu, J.-H. Monitoring Maritime Ship Characteristics Using Satellite Remote Sensing Data from Different Sensors. Ocean Sci. J. 2024, 59, 8. [Google Scholar] [CrossRef]
  10. Šaparnis, L.; Rapalis, P.; Daukšys, V. Ship Emission Measurements Using Multirotor Unmanned Aerial Vehicles: Review. J. Mar. Sci. Eng. 2024, 12, 1197. [Google Scholar] [CrossRef]
  11. Jiang, X.; Li, J.; Huang, Z.; Huang, J.; Li, R. Exploring the performance impact of soft constraint integration on reinforcement learning-based autonomous vessel navigation: Experimental insights. Int. J. Nav. Archit. Ocean. Eng. 2024, 16, 100609. [Google Scholar] [CrossRef]
  12. Li, J.; Jiang, X.; Zhang, H.; Wu, L.; Cao, L.; Li, R. Multi-joint adaptive control enhanced reinforcement learning for unmanned ship. Ocean Eng. 2025, 318, 120121. [Google Scholar] [CrossRef]
  13. Kim, H.; Kim, D.; Park, B.; Lee, S.-M. Artificial Intelligence Vision-Based Monitoring System for Ship Berthing. IEEE Access 2020, 8, 227014–227023. [Google Scholar] [CrossRef]
  14. Wang, Y.; Wang, B.; Huo, L.; Fan, Y. GT-YOLO: Nearshore Infrared Ship Detection Based on Infrared Images. J. Mar. Sci. Eng. 2024, 12, 213. [Google Scholar] [CrossRef]
  15. Wang, R.; Miao, K.; Deng, H.; Sun, J. Application of Computer Image Recognition Technology in Ship Monitoring Direction. In New Approaches for Multidimensional Signal Processing; Kountchev, R., Mironov, R., Nakamatsu, K., Eds.; Smart Innovation, Systems and Technologies; Springer: Singapore, 2022; Volume 270, pp. 233–241. ISBN 9789811685576. [Google Scholar]
  16. Zhang, Z.; Zhang, L.; Wu, J.; Guo, W. Optical and Synthetic Aperture Radar Image Fusion for Ship Detection and Recognition: Current State, Challenges, and Future Prospects. IEEE Geosci. Remote Sens. Mag. 2024, 12, 132–168. [Google Scholar] [CrossRef]
  17. Wang, J.; Li, S. Maritime Radar Target Detection in Sea Clutter Based on CNN With Dual-Perspective Attention. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  18. Wang, J.; Liu, J.; Zhao, W.; Xiao, N. Performance Analysis of Radar Detection for Fluctuating Targets Based on Coherent Demodulation. Digit. Signal Process 2022, 122, 103371. [Google Scholar] [CrossRef]
  19. Chen, X.; Su, N.; Huang, Y.; Guan, J. False-Alarm-Controllable Radar Detection for Marine Target Based on Multi Features Fusion via CNNs. IEEE Sens. J. 2021, 21, 9099–9111. [Google Scholar] [CrossRef]
  20. Su, N.; Chen, X.; Guan, J.; Huang, Y. Maritime Target Detection Based on Radar Graph Data and Graph Convolutional Network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  21. Jing, H.; Cheng, Y.; Wu, H.; Wang, H. Radar Target Detection With Multi-Task Learning in Heterogeneous Environment. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  22. Li, W.; Wang, K.; You, L.; Huang, Z. A New Deep Learning Framework for HF Signal Detection in Wideband Spectrogram. IEEE Signal Process. Lett. 2022, 29, 1342–1346. [Google Scholar] [CrossRef]
  23. Lin, H.; Dai, X. Real-Time Multisignal Detection and Identification in Known and Unknown HF Channels: A Deep Learning Method. Wirel. Commun. Mob. Comput. 2022, 2022, 1–29. [Google Scholar] [CrossRef]
  24. Li, C.; Liu, Y.; Wang, Y.; Ye, X. Short-Term Prediction Method of HF Frequency Based on Deep Learning Network. In Proceedings of the 2019 International Conference on Artificial Intelligence and Advanced Manufacturing (AIAM), IEEE, Dublin, Ireland, 6–9 October 2019; pp. 360–363. [Google Scholar]
  25. Li, H.; Jiao, H.; Yang, Z. AIS Data-Driven Ship Trajectory Prediction Modelling and Analysis Based on Machine Learning and Deep Learning Methods. Transp. Res. Part E Logist. Transp. Rev. 2023, 175, 103152. [Google Scholar] [CrossRef]
  26. European Maritime Safety Agency. Annual Overview of Marine Casualties and Incidents 2024. 2024. Available online: https://www.emsa.europa.eu/publications/reports/item/5352-annual-overview-of-marine-casualties-and-incidents-2024.html (accessed on 27 April 2025).
  27. Liu, B.; Xiao, Q.; Zhang, Y.; Ni, W.; Yang, Z.; Li, L. Intelligent Recognition Method of Low-Altitude Squint Optical Ship Target Fused with Simulation Samples. Remote Sens. 2021, 13, 2697. [Google Scholar] [CrossRef]
  28. Meng, H.; Tian, Y.; Ling, Y.; Li, T. Fine-Grained Ship Recognition for Complex Background Based on Global to Local and Progressive Learning. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  29. Xu, M.-Z.; Yao, Z.-X.; Kong, X.-P.; Xu, Y.-C. Ships Classification Using Deep Neural Network Based on Attention Mechanism. In Proceedings of the 2021 OES China Ocean Acoustics (COA), IEEE, Harbin, China, 14 July 2021; pp. 1052–1055. [Google Scholar]
  30. Zhang, T.; Zhang, X.; Shi, J.; Wei, S.; Wang, J.; Li, J.; Su, H.; Zhou, Y. Balance Scene Learning Mechanism for Offshore and Inshore Ship Detection in SAR Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  31. Wu, P.; Huang, H.; Qian, H.; Su, S.; Sun, B.; Zuo, Z. SRCANet: Stacked Residual Coordinate Attention Network for Infrared Ship Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  32. Li, Y.; Liu, G.; Xiong, J.; Jia, M.; Zhang, W. Infrared Ship Detection Based on Latent Low-Rank Decomposition and Salient Feature Fusion. In Proceedings of the 5th International Conference on Computer Information Science and Application Technology (CISAT 2022); Zhao, F., Ed.; SPIE: Chongqing, China, 2022; p. 233. [Google Scholar]
  33. Chen, X.; Qiu, C.; Zhang, Z. A Multiscale Method for Infrared Ship Detection Based on Morphological Reconstruction and Two-Branch Compensation Strategy. Sensors 2023, 23, 7309. [Google Scholar] [CrossRef]
  34. Li, L.; Yu, J.; Chen, F. TISD: A Three Bands Thermal Infrared Dataset for All Day Ship Detection in Spaceborne Imagery. Remote Sens. 2022, 14, 5297. [Google Scholar] [CrossRef]
  35. Wu, P.; Su, S.; Tong, X.; Guo, R.; Sun, B.; Zuo, Z.; Zhang, J. SARFB: Strengthened Asymmetric Receptive Field Block for Accurate Infrared Ship Detection. IEEE Sens. J. 2023, 23, 5028–5044. [Google Scholar] [CrossRef]
  36. Zeng, L.; Zhu, Q.; Lu, D.; Zhang, T.; Wang, H.; Yin, J.; Yang, J. Dual-Polarized SAR Ship Grained Classification Based on CNN With Hybrid Channel Feature Loss. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  37. Li, J.; Qu, C.; Shao, J. Ship Detection in SAR Images Based on an Improved Faster R-CNN. In Proceedings of the 2017 SAR in Big Data Era: Models, Methods and Applications (BIGSARDATA), IEEE, Beijing, China, 21–25 May 2017; pp. 1–6. [Google Scholar]
  38. Zhenzhen, L.; Baojun, Z.; Linbo, T.; Zhen, L.; Fan, F. Ship Classification Based on Convolutional Neural Networks. J. Eng. 2019, 2019, 7343–7346. [Google Scholar] [CrossRef]
  39. Ren, Y.; Wang, X.; Yang, J. Maritime Ship Recognition Based on Convolutional Neural Network and Linear Weighted Decision Fusion for Multimodal Images. Math. Biosci. Eng. 2023, 20, 18545–18565. [Google Scholar] [CrossRef] [PubMed]
  40. Chang, L.; Chen, Y.-T.; Hung, M.-H.; Wang, J.-H.; Chang, Y.-L. YOLOV3 Based Ship Detection in Visible and Infrared Images. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, IEEE, Brussels, Belgium, 11 July 2021; pp. 3549–3552. [Google Scholar]
  41. Li, J.; Yang, Y.; Li, X.; Sun, J.; Li, R. Knowledge-Transfer-Based Bidirectional Vessel Monitoring System for Remote and Nearshore Images. J. Mar. Sci. Eng. 2023, 11, 1068. [Google Scholar] [CrossRef]
  42. Pan, B.; Jiang, Z.; Wu, J.; Zhang, H.; Luo, P. Ship Recognition Based on Active Learning and Composite Kernel SVM. In Advances in Image and Graphics Technologies; Tan, T., Ruan, Q., Wang, S., Ma, H., Di, K., Eds.; Communications in Computer and Information Science; Springer: Berlin, Germany, 2015; Volume 525, pp. 198–207. ISBN 978-3-662-47790-8. [Google Scholar]
  43. Chang, L.; Chen, Y.-T.; Wang, J.-H.; Chang, Y.-L. Modified Yolov3 for Ship Detection with Visible and Infrared Images. Electronics 2022, 11, 739. [Google Scholar] [CrossRef]
  44. Wang, Y.; Wang, C.; Zhang, H. Combining a Single Shot Multibox Detector with Transfer Learning for Ship Detection Using Sentinel-1 SAR Images. Remote Sens. Lett. 2018, 9, 780–788. [Google Scholar] [CrossRef]
  45. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  46. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  47. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, Boston, MA, USA, 8–10 June 2015; pp. 1–9. [Google Scholar]
  48. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  49. Howard, A.; Sandler, M.; Chen, B.; Wang, W.; Chen, L.-C.; Tan, M.; Chu, G.; Vasudevan, V.; Zhu, Y.; Pang, R.; et al. Searching for MobileNetV3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), IEEE, Seoul, Republic of Korea, 7–10 October 2019; pp. 1314–1324. [Google Scholar]
  50. Ma, N.; Zhang, X.; Zheng, H.-T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. In Computer Vision—ECCV 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2018; Volume 11218, pp. 122–138. ISBN 978-3-030-01263-2. [Google Scholar]
  51. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-Level Accuracy with 50× Fewer Parameters and <0.5 MB Model Size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  52. Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. arXiv 2022, arXiv:2201.03545. [Google Scholar]
Figure 1. Process architecture diagram of VIOS-Net.
Figure 1. Process architecture diagram of VIOS-Net.
Jmse 13 00913 g001
Figure 3. The diagram of the residual block architecture.
Figure 3. The diagram of the residual block architecture.
Jmse 13 00913 g003
Figure 4. Example images from the VAIS dataset.
Figure 4. Example images from the VAIS dataset.
Jmse 13 00913 g004
Figure 5. ROC curves across different methods on the visible light dataset: (a) AlexNet; (b) VGG-11; (c) VGG-13; (d) GoogLeNet; (e) ResNet-18; (f) ResNet-34; (g) MobileNet-V2; (h) MobileNet-V3Small; (i) ShuffleNet-V2_x0_5; (j) SqueezeNet; (k) ConvNeXt; (l) VIOS-Net(ours).
Figure 5. ROC curves across different methods on the visible light dataset: (a) AlexNet; (b) VGG-11; (c) VGG-13; (d) GoogLeNet; (e) ResNet-18; (f) ResNet-34; (g) MobileNet-V2; (h) MobileNet-V3Small; (i) ShuffleNet-V2_x0_5; (j) SqueezeNet; (k) ConvNeXt; (l) VIOS-Net(ours).
Jmse 13 00913 g005
Figure 6. ROC curves across different methods on the infrared dataset: (a) AlexNet; (b) VGG-11; (c) VGG-13; (d) GoogLeNet; (e) ResNet-18; (f) ResNet-34; (g) MobileNet-V2; (h) MobileNet-V3Small; (i) ShuffleNet-V2_x0_5; (j) SqueezeNet; (k) ConvNeXt; (l) VIOS-Net(ours).
Figure 6. ROC curves across different methods on the infrared dataset: (a) AlexNet; (b) VGG-11; (c) VGG-13; (d) GoogLeNet; (e) ResNet-18; (f) ResNet-34; (g) MobileNet-V2; (h) MobileNet-V3Small; (i) ShuffleNet-V2_x0_5; (j) SqueezeNet; (k) ConvNeXt; (l) VIOS-Net(ours).
Jmse 13 00913 g006
Figure 7. Data augmentation impact result: (a) training result with data augmentation technique (left: visible light dataset; right: infrared dataset); (b) training result without data augmentation technique (left: visible light dataset; right: infrared dataset).
Figure 7. Data augmentation impact result: (a) training result with data augmentation technique (left: visible light dataset; right: infrared dataset); (b) training result without data augmentation technique (left: visible light dataset; right: infrared dataset).
Jmse 13 00913 g007
Figure 8. Impact of transfer learning: (a) training results with transfer learning technique (left: visible light dataset; right: infrared dataset); (b) training results without transfer learning technique (left: on the infrared dataset; right: on the visible light dataset).
Figure 8. Impact of transfer learning: (a) training results with transfer learning technique (left: visible light dataset; right: infrared dataset); (b) training results without transfer learning technique (left: on the infrared dataset; right: on the visible light dataset).
Jmse 13 00913 g008
Figure 9. Training and validation curves of VIOS-Net on infrared and visible data sources (without data augmentation, without transfer learning, and with both techniques).
Figure 9. Training and validation curves of VIOS-Net on infrared and visible data sources (without data augmentation, without transfer learning, and with both techniques).
Jmse 13 00913 g009
Figure 10. Pattern and impact of weight coefficient (a) on the visible light dataset and (b) on the infrared dataset.
Figure 10. Pattern and impact of weight coefficient (a) on the visible light dataset and (b) on the infrared dataset.
Jmse 13 00913 g010
Table 1. Distribution of datasets prior to data balancing.
Table 1. Distribution of datasets prior to data balancing.
DatasetCategoryTrainTestAll
Visible
Dataset
medium ‘other’14837185
medium passenger11228140
merchant14436180
sailing33082412
small524131655
tug8020100
Infrared
Dataset
medium ‘other’14837185
medium passenger11228140
merchant14436180
sailing33082412
small524131655
tug8020100
Table 2. Comparation of dataset distribution pre- and post-data balancing.
Table 2. Comparation of dataset distribution pre- and post-data balancing.
CategoryOriginal DataBalanced DataCategoryOriginal DataBalanced Data
medium ‘other’185650medium ‘other’185650
medium passenger140560medium passenger140560
merchant180650merchant180650
sailing412650sailing412650
small655655small655655
tug100650tug100650
Table 3. Distribution of datasets after data balancing.
Table 3. Distribution of datasets after data balancing.
DatasetVisibleInfrared
CategoryTrain Set1Test Set1Train Set2Test Set2
medium ‘other’520130520130
medium passenger448112448112
merchant520130520130
sailing520130520130
small524131524131
tug520130520130
Table 4. Comparison of quantitative evaluation across different methods.
Table 4. Comparison of quantitative evaluation across different methods.
MethodVisible Images (%)Infrared Images (%)
AccuracyPrecisionRecallF1-ScoreAccuracyPrecisionRecallF1-Score
Alex Net [45]88.97689.50088.65088.82086.23986.38086.21085.950
VGG-11 [46]83.20283.79083.30083.04084.14284.60084.22083.780
VGG-13 [46]84.51489.98084.34084.82084.53584.46084.43084.400
GoogLeNet [47]88.45189.63088.51088.55085.71479.99078.79078.200
ResNet-18 [48]88.18988.50088.05088.14088.46788.52088.45088.320
ResNet-34 [48]91.33991.62091.41091.28087.15686.92087.22086.960
MobileNetV2 [49]86.08986.79085.92085.97084.92884.70085.02084.700
MobileNetV3 Small [49]90.81491.19090.78090.82085.45285.38085.39085.100
ShuffleNetV2_x0_5 [50]89.23889.78089.28089.25087.13988.23087.19087.210
SqueezeNet [51]86.35286.80086.38085.91076.01675.20075.84074.860
ConvNeXt [52]89.50191.05089.21089.47084.92885.63084.98084.880
Improved CNN [39]90.20091.60089.10090.00085.40082.90085.80083.500
Dual CNN [39]93.60095.50092.00093.50093.60095.50092.00093.500
VIOS-Net96.19996.21496.24196.22796.19996.23896.26396.250
Table 5. Comparison of quantitative evaluation on data augmentation.
Table 5. Comparison of quantitative evaluation on data augmentation.
Visible Images (%)Infrared Images (%)
AccuracyPrecisionRecallF1-ScoreAccuracyPrecisionRecallF1-Score
With Data Augmentation Technique96.19996.21496.24196.22796.19996.23896.26396.250
Without Data Augmentation Technique94.10294.48494.13294.30894.23394.49894.28194.389
Table 6. Comparison of quantitative evaluation on transfer learning.
Table 6. Comparison of quantitative evaluation on transfer learning.
Visible Images (%)Infrared Images (%)
AccuracyPrecisionRecallF1-ScoreAccuracyPrecisionRecallF1-Score
With Transfer Learning Technique96.19996.21496.24196.22796.19996.23896.26396.250
Without Transfer Learning Technique90.69491.04290.32490.68290.82591.15090.47590.811
Table 7. Comparison of parameters across different methods.
Table 7. Comparison of parameters across different methods.
MethodAlex NetVGG-11VGG-13GoogLeNetResNet-18ResNet-34MobileNetV2MobileNetV3 SmallShuffleNetV2_x0_5SqueezeNetConvNeXtVIOS-Net (OURS)
Visible61.10M128.14M128.32M6.99M11.68M21.79M3.96M662.032k641.46k6.49M2.81M11.11M
Infrared61.10M128.14M128.32M6.99M11.68M21.79M3.96M662.032k641.46k6.49M2.81M11.11M
Total Parameters122.20M256.38M256.65M13.99M23.39M43.59M7.92M1.32M1.28M12.98M5.62M22.31M
Acc87.60783.67284.52587.08388.32889.24785.50988.13388.18981.18487.21596.199
Parameter/Acc (M/%)1.393.063.040.160.260.490.090.010.010.160.060.23
Table 8. Comparison of quantitative evaluation of variations in weight coefficients.
Table 8. Comparison of quantitative evaluation of variations in weight coefficients.
Weight Coefficients Pair0.1,0.9 0.2,0.8 0.3,0.7 0.4,0.6 0.5,0.5 0.6,0.4 0.7,0.3 0.8,0.2 0.9,0.1
Visible accuracy (%)92.04995.56996.05996.19996.01995.73995.73995.58992.049
Infrared accuracy (%)91.54995.26995.87996.19995.89995.66995.54995.48995.349
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhan, J.; Li, J.; Wu, L.; Sun, J.; Yin, H. VIOS-Net: A Multi-Task Fusion System for Maritime Surveillance Through Visible and Infrared Imaging. J. Mar. Sci. Eng. 2025, 13, 913. https://doi.org/10.3390/jmse13050913

AMA Style

Zhan J, Li J, Wu L, Sun J, Yin H. VIOS-Net: A Multi-Task Fusion System for Maritime Surveillance Through Visible and Infrared Imaging. Journal of Marine Science and Engineering. 2025; 13(5):913. https://doi.org/10.3390/jmse13050913

Chicago/Turabian Style

Zhan, Junquan, Jiawen Li, Langtao Wu, Jiahua Sun, and Hui Yin. 2025. "VIOS-Net: A Multi-Task Fusion System for Maritime Surveillance Through Visible and Infrared Imaging" Journal of Marine Science and Engineering 13, no. 5: 913. https://doi.org/10.3390/jmse13050913

APA Style

Zhan, J., Li, J., Wu, L., Sun, J., & Yin, H. (2025). VIOS-Net: A Multi-Task Fusion System for Maritime Surveillance Through Visible and Infrared Imaging. Journal of Marine Science and Engineering, 13(5), 913. https://doi.org/10.3390/jmse13050913

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop