You are currently viewing a new version of our website. To view the old version click .
Journal of Marine Science and Engineering
  • Article
  • Open Access

21 October 2025

Automatic Identification of Static Harbor Objects Based on Camera Images from a Highly Autonomous Dredger Ship

,
and
1
Department of Computing Science, Carl von Ossietzky University Oldenburg, Ammerländer Heerstraße 114-118, 26129 Oldenburg, Germany
2
German Aerospace Center (DLR), Institute of Systems Engineering for Future Mobility, Escherweg 2, 26121 Oldenburg, Germany
*
Author to whom correspondence should be addressed.
This article belongs to the Section Ocean Engineering

Abstract

Possible collisions with port infrastructure are a big challenge in the automation of commercial shipping. The first step to avoiding these collisions is identifying static port infrastructure. To minimize the risk of collisions of automated vessels with port infrastructure, this study aims to develop a model for automatically detecting static harbor objects (quay walls and piles) in port areas using a YOLOv5-based deep learning architecture. The existing architecture is adapted by generating a port-specific image dataset using image obfuscation techniques that simulate real-world operational scenarios, additionally improving robustness. To determine optimal hyperparameters, such as image resolution, batch size, or selection of optimization algorithm, multiple experiments were conducted and evaluated. As the proposed system is used in a time critical environment, the evaluation is performed on the basis of model performance as well as inference time.

1. Introduction

As research and current accident data show, the majority of accidents are attributed to human errors [1,2,3,4,5]. From 2014 to 2020, up to 53% of ship accidents were due to human errors in the European waters [1]. The overall collision risk is especially high in coastal waters and congested areas, like ports [6].
Especially in ports, navigating safely in confined waters is challenging due to other traffic, port infrastructure, and the environmental conditions, like gusty winds or low visibility. A prevalent approach to automated obstacle detection involves utilizing image recognition algorithms. In recent years, AI-based systems leveraging neural networks have been increasingly employed for image recognition [7]. These algorithms utilize image data from onboard camera systems to proactively identify potential obstacles. Such information can then be provided to the onboard crew as decision support, aiding in making informed navigational choices. These commercially available products are already being used on ships today, primarily targeting large shipping companies for deep-sea freight transportation.
However, these systems are less suitable for obstacle detection in port areas, where static harbor objects, such as quay walls, piles, and buoys, pose a significant threat to vessels operating within harbor basins or close to the coastline.
This results from a very limited selection of suitable candidates [8,9]. Currently, only Fa few datasets are available that are suitable for training AI models based on neural networks—examples are the “Singapore Maritime Dataset” (SMD) and the “SeaShips7000” dataset [8]. Most other datasets are either too small, not optimized for object recognition use cases, or utilize satellite images [8]. However, both the SMD and the SeaShips7000 datasets focus on recognizing various types of ships [8]. The SMD comprises a total of ten classes, including six ship classes [9]. Among static harbor objects, only a “buoy” class is represented [10]. The SeaShips7000 dataset exclusively covers ship types and is specialized for object recognition on rivers [8,9,11].
This work presents a framework to create port-specific datasets needed to account for the differences in harbor infrastructure, which differ from harbor to harbor. As a proof of concept, this paper is going to consider the port of Emden as the operational area of an autonomous vessel. Therefore, the training dataset is going to contain objects specific to this area. However, the proposed method can be adapted to any other port by tailoring the training dataset to the specific conditions on site.

3. Method

As shown in Section 2, there is currently a research gap that exists due to the lack of training data for static infrastructure commonly found in ports. Even though some of these structures are locally occurring phenomena, autonomous vessels that operate in these areas need to be able to correctly identify these to guarantee safe operation. The process is divided into four steps. The first step (see Figure 1) of the project was to build a domain-specific object detection dataset for the detection of static port objects based on camera images from the port of Emden. This was necessary as currently there is no freely available object detection dataset that includes static port objects, such as quay walls and piles. Prior to annotating these images, frames had to be extracted and processed from the provided raw video material. Subsequently, the static port objects—in this case, quay walls and piles—were labeled using bounding boxes. Once this process was completed, the annotations underwent quality control to correct any errors. In the next step, the dataset was prepared by applying a data augmentation.
Figure 1. Methodical steps of creating a port-specific object detection algorithm.
Afterward, the next step was to select a suitable AI model. It was decided to implement the object detection model “YOLOv5s6” to answer the research question. The model was then parameterized to achieve optimal adaptation to the specific use case—the automatic detection of static port objects using camera images from the port of Emden. Once the parameterization was completed, the performance of the final model iteration was evaluated using appropriate performance metrics. This approach allowed for an assessment of how well the implemented model was suited for automated detection of static port objects. Additionally, it facilitated the quantification and evaluation of the model’s strengths and weaknesses. As part of the performance evaluation, a final step involves investigating the extent to which the model is capable of reliably detecting images of static harbor objects from different port areas.
As part of the performance evaluation, a final step was to investigate to what extent the model is capable of reliably detecting quay walls and piles on images recorded in other port areas.

3.1. Processing of the Raw Video Material

The preparation of raw video material involves obtaining the images ready for labeling. Figure 2 illustrates the four steps that needed to be performed during the data preparation process.
Figure 2. Processing steps of the raw material from video footage to filtered, labeled images. Red bars represent discarded video sections and images, while green bars indicate discarded ones.
In the first step, it was essential to ensure that the three camera streams were extracted from the provided videos. An ML algorithm learns where target objects appear more frequently in the images. Training on the unprocessed raw material would lead to the model not being adequately prepared for use in the port of Emden. During training, the ML algorithm would adapt to the specific locations of objects in the three camera streams of the raw material. However, in reality, there are no images that combine all three camera views in one frame. The large black areas between the camera streams are not present in actual use cases. In the productive environment, the model works on frames that fill the entire image. Training on the unprocessed raw data would lead the model to assume that the target objects are much smaller than they actually are in reality. All of this would negatively impact the model’s accuracy in the productive environment.
In the second step, unsuitable video passages needed to be removed from the previously extracted video streams. This primarily involved removing segments where the camera boat was not moving. This measure ensured that not too many nearly identical images were integrated into the dataset. Integrating such similar images into the dataset would cause the model to overly adapt to recognizing the same quay walls and piles repeatedly. Consequently, the model would lose its ability to generalize effectively.
Next, the videos needed to be divided into individual frames. It is also essential to ensure that too many nearly identical images are not created. To prevent this, only one image per second was extracted from the videos; all other frames were discarded. This approach is recommended to avoid overfitting the model [10].
In the last step, the images that will be part of the dataset were selected. The selection was performed randomly from the pool of extracted frames. This measure ensured that all segments of the videos were represented in the dataset to increase diversity and improve the model’s robustness.

3.2. Conceptualization of the Dataset

In this step, the quay walls and piles contained in the images from the port of Emden were annotated using bounding boxes. After completing the labeling process, a final quality control of the dataset was conducted. During this control, the quality of the labels was reviewed, and any inaccurate or erroneous bounding boxes were corrected. This included annotations with incorrect class labels or improperly placed bounding boxes. Additionally, any objects that were overlooked during the initial annotation process were manually reannotated.
It was crucial that the dataset represents the two target classes in a diverse manner to prepare the trained model for real-world deployment. To achieve this, the objects should be depicted from various perspectives and angles, at different distances, under varying weather conditions (sunny, cloudy, rainy, or foggy), and lighting conditions, as well as during different times of the day. Overlaps between objects, intersections with other objects in the surroundings, and the image borders should also be included to enhance the robustness of the AI model. However, the existing images from the port of Emden only partially fulfill these diversity requirements. A beneficial aspect is that the recordings were captured by three different cameras, providing three different perspectives and angles of the quay walls and piles. The cameras also have a slight variation in their tilt, resulting in varying inclination angles of the depicted quay walls and piles. Overlaps between piles and quay walls, as well as intersections with other objects in the surroundings, were also represented. On the downside, the recordings were exclusively taken in broad daylight and under sunny or cloudy skies. This limitation could have potentially impacted the model’s performance in scenarios with different lighting conditions or weather situations.
To improve the diversity of the dataset and prepare the model for real-world deployment, data augmentation was used. Data augmentation allows artificial expansion of the size of the dataset by adding modified copies of the original images [25]. By applying data augmentation, the diversity of the dataset increases, which reduces the risk of overfitting the model [25]. Another advantage is that this technique enables the simulation of various weather and lighting conditions, as well as changes in perspective.
Figure 3 illustrates the transformation steps that have been applied as part of the data augmentation as demonstrated in [26]. First, rotation ([−5°, +5°]) and shearing ([−5°, +5°]) were performed to simulate fluctuations caused by strong winds and/or rough seas. These fluctuations lead to changes in the camera’s perspective and angle on the port objects (slight tilting). The brightness augmentation ([−60%, 60%]) was used to imitate different lighting conditions. Artificial darkening of the images allowed the model to recognize objects in dim light conditions during real-world deployment. The implementation of image noise (10%) helped simulate rain, hail, snow, or fog and prepare the model for technical disturbances [27]. The artificial blurring (2 Pixel) of images was used to simulate water droplets on the camera lens. Moreover, it can simulate scenarios where one or more cameras lose focus.
Figure 3. Image augmentations to represent real-world environmental conditions.
The resulting labeled dataset contains 6811 images in total that contain 20,110 annotations. The dataset was divided into a train-validation-test split set in a 60:20:20 manner as recommended by [28]. This results in 4087 images used in the training set, 1357 images in the validation set, and 1362 images in the test set. After augmentation the dataset contains 10,240 images containing 30,315 annotations. The size of the training set is increased to 7514 images, while the size of the validation and test set stay the same.

3.3. Hyperparameter Optimization

In order to detect the quay walls and piles on the previously constructed dataset, an in-house pre-trained YOLOv5s6 instance was used. This approach aimed to reduce the risk of model overfitting and accelerate the training process [29].
The YOLO model has a variety of adjustable hyperparameters. It is possible to manually pass a portion of these parameters during the execution of the training script. Examples of these parameters include the resolution of input images, the number of trained epochs, the batch size, and the optimization algorithm used. The batch size indicates how many training examples are processed by the model before the internal model parameters are updated by the optimizer [28,30]. These four model parameters significantly impact the eventual model performance and the time and computational power required for training.
In addition to these “primary” hyperparameters, there are numerous other “secondary” hyperparameters that can be optimized for the YOLOv5 model. Unlike primary hyperparameters, these secondary parameters cannot be manually passed as arguments to the training script; instead, they must be provided in a separate hyperparameter configuration file. Examples of secondary hyperparameters include the learning rate, momentum, or the Intersection over Union (IoU) threshold used for training. Due to the large number of secondary hyperparameters, it is not practical to manually try out all possible hyperparameter combinations. Instead, a genetic algorithm was used to increase the likelihood of finding an optimal hyperparameter combination.
Figure 4 illustrates the process of hyperparameter optimization. A total of four experiments were conducted. The division into four separate experiments was aimed at clarifying the cause-and-effect relationship between individual hyperparameter settings and the model performance.
Figure 4. Hyperparameter optimization process.
The first three experiments concentrated on tuning the three primary hyperparameters: image resolution, batch size, and the chosen optimization algorithm. In the fourth step, the optimization of secondary hyperparameters took place. To establish a basis for comparison, the model instances parameterized within the scope of the four experiments were trained for 100 epochs each.
In the first experiment, the goal was to determine an appropriate resolution for the input images. For this purpose, the three resolutions [640 × 640], [1280 × 1280], and [1280 × 720] were compared. [640 × 640] corresponds to the standard image resolution of the YOLOv5 model. YOLOv5 offers an additional model family alongside the standard model variants, supporting a maximum image resolution of [1280 × 1280]. YOLOv5 also accommodates the processing of rectangular images with a 16:9 aspect ratio. The [1280 × 720] resolution corresponds to the native resolution of the images in the previously created dataset. By comparing the resolutions [640 × 640] and [1280 × 1280], the aim is to evaluate whether an increase in image resolution indeed leads to improved performance, particularly for smaller objects. The investigation of the [1280 × 720] resolution aims to assess the extent to which using the native image resolution impacts the model’s performance and the training process.
In the second experiment, the influences of different batch sizes on required training resources and model performance were investigated. Generally, it is recommended to set the batch size as high as possible [31]. In this experiment, the training began with a batch size of 16. Subsequently, the value was doubled as is common practice in the field (32, 64, 128, etc.). This process continued for as long as it was supported by the hardware.
In the third experiment, the selection of an appropriate optimization algorithm took place. All optimizers provided as default choices by the YOLOv5 model were examined. YOLOv5 offers the option to choose between three optimization algorithms: “SGD with Nesterov momentum”, “Adam”, and “AdamW”.
In the final experiment, the secondary hyperparameters were optimized using the genetic algorithm integrated into YOLOv5. Subsequently, the extent to which this approach can contribute to improving model performance will be investigated.
For training the final model, the parameters determined to be optimal through hyperparameter optimization were used. In all previous experiments, the models were trained for 100 epochs each. However, in this context, the epoch limit was removed to determine the optimal number of epochs for training the final model. To identify the optimal number of epochs, a technique known as “early stopping” was implemented. Early stopping involves terminating training prematurely when there is no further improvement in performance over a certain number of epochs [30]. To prevent overfitting, it was important to ensure that the validation loss remained consistently low and did not start increasing again in the final epochs [28].
In the context of the conducted experiments, the impacts of different hyperparameter configurations on the predictive accuracy and speed of the implemented model were examined. To evaluate predictive accuracy, two performance metrics, mAP_0.5 and mAP_0.5:0.95, were used. Model speed was evaluated through the metrics of inference time and frames per second (FPS). Besides these primary performance metrics, additional metrics such as precision–recall curves and loss curves were also used, if necessary, to interpret the results. In addition to the model performance, the influences of different hyperparameters on required training resources were investigated. For performance metrics, both training duration and memory requirements (system RAM and GPU memory) were collected and analyzed.
Within the context of performance evaluation, an investigation was also conducted to determine if the observed performance metrics of the final model could be transferred to other use cases. Since no suitable dataset existed to depict quay walls and piles from other harbor areas, a collection of 30 internet images was assembled for this purpose. These images were then processed by the trained model to make predictions. This approach helped assess the model’s capability to recognize foreign harbor objects and determine the extent to which the model can be adapted to different scenarios.
All training and evaluation were conducted on a single NVIDIA A100-SXM4-40GB with 40 GB of VRAM, an AMD EPYC 7742 CPU, 83.5 GB RAM, and storage capacity of 166.8 GB.

4. Evaluation

Due to space constraints, only the results of the last two experiments will be described in this section. Based on these research findings in experiment one, the decision has been made to use an image resolution of [1280 × 720] in subsequent training runs. This parameter setting allows for a favorable cost–benefit ratio. On the one hand, the model could achieve a significantly higher accuracy—especially in detecting the smaller piles objects—compared to the model using an image resolution of [640 × 640]. On the other hand, the training resources required for this model can be reduced by up to 50 percent compared to the [1280 × 720] image resolution model. Another advantage is that the camera images generated in operational use do not require any additional preprocessing steps in this case, as would be necessary for the other two models. This helps to improve the responsiveness of the model.
Based on the research findings in experiment two, the decision was made to use a batch size of 64 for training the models in subsequent experiments. While this parameter setting does have the drawback of requiring more system and GPU RAM, it can lead to an improved accuracy in detecting quay walls and piles within the IoU interval of 0.5 to 0.95. Additionally, utilizing the maximum supported batch size allows for a reduction in the training duration.

4.1. Optimization of the Used Optimization Algorithm

In the first step, the ability of the three optimization algorithms to minimize the box loss of the training and the validation set was evaluated. It was evident that in both cases, the most effective reduction in loss could be achieved when the model employed the SGD optimizer. After 100 epochs, the AdamW and SGD optimizers achieved box loss values of 0.0259 and 0.0232, respectively, on the validation set. The Adam optimizer yielded a significantly higher loss of 0.0383 (+48% and +65%) in comparison. In the training set, the box loss of the AdamW optimizer was approximately 35% higher than that of the SGD optimizer; however, in the validation set, this difference was only 12%.
The respective ability to minimize the loss functions was also reflected in the predictive accuracy of the three models. Figure 5 illustrates the development of the mAP_0.5 metric across the 100 epochs. It was evident that the mAP_0.5 for the SGD optimizer experienced an exponential increase in the first 40 epochs. On the other hand, a linear progression was noticeable for the AdamW optimizer. The final mAP_0.5 for the AdamW optimizer was approximately two percentage points lower than the SGD model (97.6% vs. 95.8%). Using the Adam optimizer only achieved a mAP_0.5 value of 67.8%. A similar ranking of the three models was also evident in the mAP_0.5:0.95 metric.
Figure 5. Mean average precision (mAP) (IoU = 0.5).
A look at Table 1 reveals that the Adam optimizer had difficulties in accurately predicting both quay walls and piles. Particularly in the detection of quay walls, a significant performance difference was noticeable: the model only achieved a mean average precision of 20.4 percent.
Table 1. Experiment 3: Mean average precision.
The use of the Adam optimizer not only leads to the model’s unreliable localization of the harbor objects but also decreases the model’s classification accuracy. The classification loss in the training set, unlike the other two models, is least efficiently minimized with the Adam optimizer. While the model using the Adam optimizer also reaches the highest final loss value in this regard, the gaps between the three models are notably smaller compared to the box loss.
In summary, it can be concluded that the use of the Adam optimizer in the current application leads to inaccurate and flawed predictions, especially at high IoU thresholds. Additionally, the model is incapable of minimizing the box and classification losses as rapidly and effectively as achieved by using the AdamW or SGD optimizers. An examination of the performance metrics reveals that the Adam optimizer particularly struggles to achieve a high recall (see Figure 6). After 100 epochs, the model attains a final recall value of only 59.8%, whereas the SGD optimizer achieves a recall of 94.1%.
Figure 6. Recall curve: Adam optimizer.
As depicted in Figure 6, the recall stagnates within the confidence range of 50 to 100 percent. This observation arises from the fact that the model fails to reliably localize many quay wall objects within that range. Due to generating a high number of false negatives, the recall could not be increased further in this confidence range (Recall = TP/TP + FN). In essence, the model overlooks numerous quay wall objects for which ground truth bounding boxes exist.
In comparison to the other optimization algorithms, the SGD optimizer proved to be the most effective in improving the efficiency of the gradient descent algorithm in the current application. It ensured a swift and sustainable minimization of the loss functions. Particularly in the reduction in the box loss, the SGD optimizer outperformed the other two algorithms. Consequently, the SGD model achieved the highest mean average precision values. Since the choice of optimization algorithm had no additional impact on the required training resources and the predictive accuracy of the models, the decision was made to use the SGD optimizer with Nesterov momentum for the subsequent experiments.

4.2. Evolutionary Optimization

Up to this point, only the default configuration provided by YOLOv5 for the instantiation of the secondary hyperparameters has been utilized. In the context of this experiment, the integrated genetic algorithm within the YOLOv5 repository was employed to automatically determine an optimal hyperparameter combination.
The current YOLOv5 model involves a total of 29 adjustable secondary hyperparameters. The model employs an initial (l0) and final learning rate (lf) to control the strength of weight adjustments over time. Additional secondary hyperparameters encompass the duration of the warm-up phase, the strength of the used momentum and warm-up momentum, the warm-up learning rate, the IoU threshold applied for training, the weight decay, and the weighting factors for the loss functions.
The evolutionary algorithm was configured to train the model for ten epochs each. After every ten epochs, the model’s predictive accuracy was evaluated. The genetic algorithm ran for a total of 100 epochs. Hence, the hyperparameter optimization process took up 1000 epochs (100 generations × 10 epochs per generation). The creation of a new offspring proceeded by selecting a specific solution from the top n solution candidates across all previous generations. In the initial generation, there was only one solution candidate—the default hyperparameter configuration. Over time, more hyperparameter combinations were generated. The algorithm dictated that during parent selection, at most the top five solutions could be considered. Each of these maximum five solution candidates was assigned a weight. Solutions with higher fitness values were given more weight than those with lower fitness values. These weights were subsequently utilized to select a concrete offspring for the current generation from the best n solutions. The weighting of candidates ensured that solutions with high fitness were preferred choices [32].
The hyperparameters of the determined offspring were subsequently mutated component-wise. The probability of a hyperparameter being altered was set at 80 percent. The Gaussian mutation operator was employed for the mutation process. In this mutation operator, an initial random value was drawn from a normal distribution (variance = 0, mean = 1) [33]. This value was then multiplied by the so-called mutation rate, which controls the strength of the mutation [33]. In this scenario, the mutation rate was 0.2. The resulting value represented the actual mutation factor, which was applied to the corresponding hyperparameter [32]. Subsequently, the performance of the generated solution was evaluated through a fitness function. The fitness function calculated the weighted sum of the two performance metrics, mAP_0.5 (weight = 10%) and mAP_0.5:0.95 (weight = 90%) [32].
The most effective hyperparameter configuration was discovered after 78 generations. This candidate achieved the highest mAP_0.5 value of 93.3 percent in the first ten epochs. The hyperparameter optimization has resulted in the model being better equipped to minimize the box loss. This implies that the model achieves improved results more rapidly when adapting to the training dataset and previously unseen input images. Even the box loss achieved after 100 epochs is lower. For the training set, this value was reduced to 0.0127 after optimization, whereas it was almost double at 0.0228 before. In the validation set, the box loss after 100 epochs was reduced by almost 45 percent through hyperparameter optimization. Similar trends are evident in the classification loss in the validation set. The final classification loss was reduced from 0.0001837 to 0.00005904 (−68 percent) through the adjusted hyperparameters. However, the adaptability to training images could hardly be improved.
It is notable that precision, recall, and mAP metrics can be improved in less time through the adjustment of hyperparameters. The original model with the default hyperparameter configuration required more epochs initially to achieve similar high-performance values. Additionally, it is evident that the performance values of the two models gradually converge, and, after 100 epochs, there was no significant difference in mean average precision measurable anymore. The enhancement in efficiency in minimizing the loss functions was attributed to the adaptation of hyperparameters within the framework of evolutionary optimization (see Table 2).
Table 2. Evolution of selected hyperparameters.
On the one hand, the warm-up phase was reduced from 3 to 1.87 epochs. By decreasing the initial learning rate from 0.01 to 0.00746, the momentum from 0.937 to 0.853, and the warm-up momentum, the likelihood of the model skipping valleys in the loss function at the beginning of the learning process decreased. It seems that without the adjusted hyperparameters, the model might have become stuck in a less optimal minimum after a certain point, as opposed to the optimized model. This is why the final loss scores of the two models—particularly the box loss—differ after 100 epochs.
In summary, it can be concluded that the optimization of the secondary hyperparameters aided in improving the YOLO model’s adaptation to training images and previously unseen data. While the hyperparameter optimization did not lead to an enhancement in the final model performance, it did bring about an efficiency improvement in the training process. The model is now capable of sustaining a stronger minimization of the loss functions and achieving quicker increases in precision, recall, and mean average precision in the initial epochs. The convergence of final mean average precision values can be attributed to the differing weightings of the loss functions (box loss: 0.05 (before) vs. 0.0279 (after); class loss: 0.5 (before) vs. 0.261 (after)).

4.3. Final Model

The final model utilizes the optimal settings for hyperparameters determined within the scope of the four experiments (image resolution: [1280 × 720], batch size: 64, optimization algorithm: SGD with Nesterov momentum, and the optimized hyperparameters). The implementation of early stopping led to the premature termination of training after 306 epochs. Table 3 presents the mean average precision (mAP) values and inference time of the final model.
Table 3. Performance metrics of the final model.
It is evident that the model’s prediction accuracy in detecting the quay walls is higher in both the mAP_0.5 and mAP_0.5:0.95 metrics compared to detecting the piles. While the performance difference in class-specific mAP_0.5 is merely one percentage point, it amounts to 6.6 percentage points in the mAP_0.5:0.95 metric. This observation has been consistent throughout the preceding experiments. The size of the objects can be attributed as a reason for this performance difference. Piles are generally smaller than quay wall objects. Due to the functioning of the YOLO architecture, instances of imprecise localization of such objects are more frequent [34,35]. This issue becomes more pronounced, especially when higher Intersection over Union (IoU) thresholds are applied.
The final model achieves a prediction accuracy (mAP_0.5) of 97.7 percent while having a low inference time of 7.7 milliseconds per image, enabling the final model to process approximately 130 frames per second (FPS). This allows the model to be used in a real-time context, like early warning and navigation systems.
As outlined in Section 3, an investigation was conducted to determine whether the implemented model in this study could be utilized for the automated detection of quay walls and piles in different port areas. For this purpose, a total of 30 images was collected from the internet, depicting quay walls and piles in other port areas. Subsequently, the final YOLO model was employed to make predictions on these previously unseen images. The evaluation of the results revealed that the model’s performance can only be transferred to other use cases to a very limited extent. On 14 out of the 30 images, the quay walls and piles were not detected at all due to their differing appearance (e.g., color or shape).

5. Summary and Conclusions

Within the scope of this study, it has been demonstrated that the automated detection of static harbor objects within port areas has not been sufficiently investigated thus far. Current research and commercially available solutions predominantly focus on ship detection. To address this research gap, the central research question of this study aimed to be answered: How can static harbor objects be automatically identified in camera images?
During the hyperparameter optimization process, noteworthy observations were made. Specifically, the reliability in detecting smaller objects like piles could be enhanced by increasing the input image resolution. Furthermore, employing a native image resolution of [1280 × 720] offers the advantage of directly processing images in operational use without requiring an additional preprocessing step, leading to improved model prediction speed. In the context of the second experiment, it was evident that a higher batch size correlated with improved model accuracy and reduced training duration. However, this also demands more memory space. The third experiment revealed that the “stochastic gradient descent with momentum” algorithm is best suited for detecting static harbor objects in the given application. In the final step, the remaining secondary hyperparameters of the model were fine-tuned using a genetic algorithm. This hyperparameter optimization facilitated a more sustainable minimization of loss functions, particularly the box loss, resulting in a model that better adapts to the underlying images.
It is important to note that the optimal hyperparameter configuration derived from these experiments may not be universally transferable to other use cases. The suitability of this hyperparameter combination for the current application does not guarantee optimal outcomes for all other scenarios (No Free Lunch theorem) [28]. This especially holds true for the employed optimization algorithm and the determined secondary hyperparameters.
The evaluation of the final model’s performance revealed that all specified functional mandatory requirements were successfully met. Both performance-related requirements regarding the achieved prediction accuracy (mAP_0.5 ≥ 95%) and speed (≥30 FPS) of the model were surpassed. However, it was also demonstrated that the performance of the implemented model cannot be readily transferred to other port areas. The reason for this is that the appearance of harbor objects varies significantly from one port area to another, differing from the quay walls and piles in the port of Emden. As a result, the model fails to recognize these objects in the unseen images. The model’s predictions made in these cases are largely inaccurate and erroneous.
Although the model implemented in this study is not suitable for detecting static harbor objects in other port areas, the conceptual framework developed in this study can still serve as a guide for undertaking appropriate measures when building a custom dataset and implementing an individual YOLOv5s6 model. The outlined steps for data preprocessing, dataset labeling, and data augmentation can be readily applied to the detection of various static harbor objects in different port areas, if necessary, with minimal adjustments. The recommended practices for setting image resolution and batch size also hold universal applicability. In such cases, only the underlying training images need to be substituted. Moreover, the dataset created as part of this work provides a foundational resource to propel further research in the realm of static harbor object detection.
In conclusion, it can be affirmed that the formulated research question has been successfully addressed. Through the evaluation of model performance, it has been demonstrated that the deployed YOLOv5 model is capable of making reliable and swift predictions, effectively adapting to the specific challenges within the maritime domain.

Author Contributions

M.S. carried out conceptualization, review, and supervision of the study. T.S. study design, implemented software, evaluation, and writing of original draft. C.S. review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The datasets presented in this article are not readily available because of data protection regulation. Requests to access the datasets should be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. European Maritime Safety Organization. Annual Overview of Marine Casualties and Incidents 2022. Available online: https://www.emsa.europa.eu/newsroom/latest-news/item/4867-annual-overview-of-marine-casualties-and-incidents-2021.html (accessed on 23 September 2024).
  2. Qiao, W.; Liu, Y.; Ma, X.; Liu, Y. Human Factors Analysis for Maritime Accidents Based on a Dynamic Fuzzy Bayesian Network. Risk Anal. 2020, 40, 957–980. [Google Scholar] [CrossRef] [PubMed]
  3. Shi, X.; Zhuang, H.; Xu, D. Structured survey of human factor-related maritime accident research. Ocean Eng. 2021, 237, 109561. [Google Scholar] [CrossRef]
  4. Zaib, A.; Yin, J.; Khan, R.U. Determining Role of Human Factors in Maritime Transportation Accidents by Fuzzy Fault Tree Analysis (FFTA). J. Mar. Sci. Eng. 2022, 10, 381. [Google Scholar] [CrossRef]
  5. Allianz Commercial. Safety and Shipping Review 2022. AGCS. Available online: https://commercial.allianz.com/news-and-insights/news/safety-shipping-review-2022.html (accessed on 23 September 2024).
  6. Wróbel, K.; Montewka, J.; Kujala, P. Towards the assessment of potential impact of unmanned vessels on maritime transportation safety. Reliab. Eng. Syst. Saf. 2017, 165, 155–169. [Google Scholar] [CrossRef]
  7. Wu, P.; He, X.; Dai, W.; Zhou, J.; Shang, Y.; Fan, Y.; Hu, T. A Review on Research and Application of AI-Based Image Analysis in the Field of Computer Vision. IEEE Access 2025, 13, 76684–76702. [Google Scholar] [CrossRef]
  8. Spraul, R.; Sommer, L.; Schumann, A. A comprehensive analysis of modern object detection methods for maritime vessel detection. In Proceedings of the Artificial Intelligence and Machine Learning in Defense Applications II, Online, 21–25 September 2020; SPIE: Bellingham, WA, USA, 2020; pp. 13–24. [Google Scholar]
  9. Iancu, B.; Soloviev, V.; Zelioli, L.; Lilius, J. ABOships—An Inshore and Offshore Maritime Vessel Detection Dataset with Precise Annotations. Remote Sens. 2021, 13, 988. [Google Scholar] [CrossRef]
  10. Moosbauer, S.; Konig, D.; Jakel, J.; Teutsch, M. A Benchmark for Deep Learning Based Object Detection in Maritime Environments. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; IEEE: New York, NY, USA, 2019; pp. 916–925, ISBN 978-1-7281-2506-0. [Google Scholar]
  11. Shao, Z.; Wu, W.; Wang, Z.; Du, W.; Li, C. SeaShips: A Large-Scale Precisely Annotated Dataset for Ship Detection. IEEE Trans. Multimedia 2018, 20, 2593–2604. [Google Scholar] [CrossRef]
  12. Khatab, E.; Onsy, A.; Varley, M.; Abouelfarag, A. Vulnerable objects detection for autonomous driving: A review. Integration 2021, 78, 36–48. [Google Scholar] [CrossRef]
  13. Chen, Z.; Yang, J.; Chen, L.; Li, F.; Feng, Z.; Jia, L.; Li, P. RailVoxelDet: A Lightweight 3-D Object Detection Method for Railway Transportation Driven by Onboard LiDAR Data. IEEE Internet Things J. 2025, 12, 37175–37189. [Google Scholar] [CrossRef]
  14. Kanchana, B.; Peiris, R.; Perera, D.; Jayasinghe, D.; Kasthurirathna, D. Computer Vision for Autonomous Driving. In Proceedings of the 2021 3rd International Conference on Advancements in Computing (ICAC), Colombo, Sri Lanka, 9–11 December 2021; IEEE: New York, NY, USA, 2021; pp. 175–180, ISBN 978-1-6654-0862-2. [Google Scholar]
  15. Singh, K.B.; Arat, M.A. Deep Learning in the Automotive Industry: Recent Advances and Application Examples. arXiv 2019, arXiv:1906.08834v2. [Google Scholar] [CrossRef]
  16. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  17. Xue, C.; Zhong, B.; Liang, Q.; Zheng, Y.; Li, N.; Xue, Y.; Song, S. Similarity-Guided Layer-Adaptive Vision Transformer for UAV Tracking. arXiv 2025, arXiv:2503.06625. [Google Scholar]
  18. Statheros, T.; Howells, G.; Maier, K.M. Autonomous Ship Collision Avoidance Navigation Concepts, Technologies and Techniques. J. Navig. 2008, 61, 129–142. [Google Scholar] [CrossRef]
  19. Nanda, A.; Cho, S.W.; Lee, H.; Park, J.H. KOLOMVERSE: KRISO Open large-scale image dataset for object detection in the maritime universe. arXiv 2022, arXiv:2206.09885. [Google Scholar] [CrossRef]
  20. Fefilatyev, S.; Goldgof, D.; Lembke, C. Tracking Ships from Fast Moving Camera through Image Registration. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; IEEE: New York, NY, USA, 2010; pp. 3500–3503, ISBN 978-1-4244-7542-1. [Google Scholar]
  21. Zhang, Y.; Ge, H.; Lin, Q.; Zhang, M.; Sun, Q. Research of Maritime Object Detection Method in Foggy Environment Based on Improved Model SRC-YOLO. Sensors 2022, 22, 7786. [Google Scholar] [CrossRef] [PubMed]
  22. Kaido, N.; Yamamoto, S.; Hashimoto, T. Examination of automatic detection and tracking of ships on camera image in marine environment. In Proceedings of the 2016 Techno-Ocean (Techno-Ocean), Kobe, Japan, 6–8 October 2016; IEEE: New York, NY, USA, 2016; pp. 58–63, ISBN 978-1-5090-2445-2. [Google Scholar]
  23. Bovcon, B.; Mandeljc, R.; Perš, J.; Kristan, M. Stereo obstacle detection for unmanned surface vehicles by IMU-assisted semantic segmentation. Robot. Auton. Syst. 2018, 104, 1–13. [Google Scholar] [CrossRef]
  24. Fernandes Ramos, R.; Strauhs, M.; Neto, S.V.; Paulino de Lira, A.R.; Cepeda, M.A.F.S.; Guaycuru de Carvalho, L.F.; Marques de Oliveira Moita, J.V.; Caprace, J.-D. A Review of Deep Learning Application for Computational Vision within the Maritime Industry. In Proceedings of the Anais do 12º Seminário Internacional de Transporte e Desenvolvimento Hidroviário Interior, Online, 19–21 October 2021; Galoa Science: Campinas, SP, Brazil, 2021. [Google Scholar]
  25. TensorFlow. Data Augmentation. 19 July 2024. Available online: https://www.tensorflow.org/tutorials/images/data_augmentation (accessed on 23 September 2024).
  26. Wang, S. Development of approach to an automated acquisition of static street view images using transformer architecture for analysis of Building characteristics. Sci. Rep. 2025, 15, 29062. [Google Scholar] [CrossRef] [PubMed]
  27. Create Augmented Images. Roboflow. Available online: https://docs.roboflow.com/datasets/image-augmentation (accessed on 23 September 2024).
  28. Glassner, A.S. Deep Learning: A Visual Approach, 1st ed.; No Starch Press Inc: San Francisco, CA, USA, 2021; ISBN 1718500726. [Google Scholar]
  29. Lee, S.J.; Roh, M.I.; Oh, M.J. Image-based ship detection using deep learning. Ocean. Syst. Eng. 2020, 10, 415–434. [Google Scholar]
  30. Chollet, F. Deep Learning with Python; Manning: Shelter Island, NY, USA, 2018; ISBN 9781617294433. [Google Scholar]
  31. Ultralytics. Tips for Best Training Results. 12 November 2023. Available online: https://docs.ultralytics.com/yolov5/tutorials/tips_for_best_training_results/ (accessed on 23 September 2024).
  32. Ultralytics. Hyperparameter Evolution. 12 November 2023. Available online: https://docs.ultralytics.com/yolov5/tutorials/hyperparameter_evolution/ (accessed on 23 September 2024).
  33. Kramer, O. Genetic Algorithm Essentials; Springer International Publishing: Cham, Switzerland, 2017; ISBN 978-3-319-52155-8. [Google Scholar]
  34. Zou, Z.; Chen, K.; Shi, Z.; Guo, Y.; Ye, J. Object Detection in 20 Years: A Survey. Proc. IEEE 2023, 111, 257–276. [Google Scholar] [CrossRef]
  35. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. arXiv 2015, arXiv:1506.02640. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.