Next Article in Journal
A Novel COLREGs-Based Automatic Berthing Scheme for Autonomous Surface Vessels
Previous Article in Journal
Multi-AUV Hunting Strategy Based on Regularized Competitor Model in Deep Reinforcement Learning
Previous Article in Special Issue
Neural Unilateral Nussbaum Gain Sliding Mode Control for Uncertain Ship Course Keeping with an Unknown Control Direction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Marine Voyage Optimization and Weather Routing with Deep Reinforcement Learning

by
Charilaos Latinopoulos
1,*,
Efstathios Zavvos
1,*,
Dimitrios Kaklis
2,
Veerle Leemen
1 and
Aristides Halatsis
1
1
VLTN BV, De Keyserlei 58-60 bus 19, 2018 Antwerp, Belgium
2
Danaos Shipping Co., Ltd., 14 Akti Kondyli, 18545 Piraeus, Greece
*
Authors to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2025, 13(5), 902; https://doi.org/10.3390/jmse13050902
Submission received: 9 April 2025 / Revised: 28 April 2025 / Accepted: 29 April 2025 / Published: 30 April 2025
(This article belongs to the Special Issue Autonomous Marine Vehicle Operations—3rd Edition)

Abstract

Marine voyage optimization determines the optimal route and speed to ensure timely arrival. The problem becomes particularly complex when incorporating a dynamic environment, such as future expected weather conditions along the route and unexpected disruptions. This study explores two model-free Deep Reinforcement Learning (DRL) algorithms: (i) a Double Deep Q Network (DDQN) and (ii) a Deep Deterministic Policy Gradient (DDPG). These algorithms are computationally costly, so we split optimization into an offline phase (costly pre-training for a route) and an online phase where the algorithms are fine-tuned as updated weather data become available. Fine tuning is quick enough for en-route adjustments and for updating the offline planning for different dates where the weather might be very different. The models are compared to classical and heuristic methods: the DDPG achieved a 4% lower fuel consumption than the DDQN and was only outperformed by Tabu Search by 1%. Both DRL models demonstrate high adaptability to dynamic weather updates, achieving up to 12% improvement in fuel consumption compared to the distance-based baseline model. Additionally, they are non-graph-based and self-learning, making them more straightforward to extend and integrate into future digital twin-driven autonomous solutions, compared to traditional approaches.
Keywords: voyage optimization; weather routing; deep reinforcement learning; maritime energy efficiency voyage optimization; weather routing; deep reinforcement learning; maritime energy efficiency

Share and Cite

MDPI and ACS Style

Latinopoulos, C.; Zavvos, E.; Kaklis, D.; Leemen, V.; Halatsis, A. Marine Voyage Optimization and Weather Routing with Deep Reinforcement Learning. J. Mar. Sci. Eng. 2025, 13, 902. https://doi.org/10.3390/jmse13050902

AMA Style

Latinopoulos C, Zavvos E, Kaklis D, Leemen V, Halatsis A. Marine Voyage Optimization and Weather Routing with Deep Reinforcement Learning. Journal of Marine Science and Engineering. 2025; 13(5):902. https://doi.org/10.3390/jmse13050902

Chicago/Turabian Style

Latinopoulos, Charilaos, Efstathios Zavvos, Dimitrios Kaklis, Veerle Leemen, and Aristides Halatsis. 2025. "Marine Voyage Optimization and Weather Routing with Deep Reinforcement Learning" Journal of Marine Science and Engineering 13, no. 5: 902. https://doi.org/10.3390/jmse13050902

APA Style

Latinopoulos, C., Zavvos, E., Kaklis, D., Leemen, V., & Halatsis, A. (2025). Marine Voyage Optimization and Weather Routing with Deep Reinforcement Learning. Journal of Marine Science and Engineering, 13(5), 902. https://doi.org/10.3390/jmse13050902

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop