Recent Trends in the Optimization of Logistics Systems Through Discrete-Event Simulation and Deep Learning
Abstract
1. Introduction
2. Advantages of the Combination of Discrete-Event Simulation and Deep Learning in Logistics
2.1. The Significance of Discrete-Event Simulation in Logistics
2.2. The Advantages Resulting from the Combination of Discrete-Event Simulation and Deep Learning in Logistics
3. Description of the Applied Methodology
3.1. Inclusion and Exclusion Criteria
- Only original research papers were considered.
- The included studies had to provide a solution to a logistics or material flow problem through the combined application of Discrete-Event Simulation and Deep Learning. In the context of Deep Learning, this meant the inclusion of only those works that either at least partially utilized Artificial Neural Networks (ANNs) with a minimum of two hidden layers (of course, architectures like Long Short-Term Memory (LSTM) are, by definition, considered Deep Neural Networks) or provided a framework that facilitates such applications. However, in relation to logistics and material flow, a wider scope of problems is considered, with the only specific criterion being that the work had to deal with the optimization of some type of logistical problem (e.g., warehouse management, production scheduling, transport optimization, healthcare logistics, etc.).
- The included studies had to be published between 2015 and 2025 (in some cases, due to publication practice, this criterion also allowed the inclusion of conference papers from 2014).
3.2. Search Phrases and General Results
4. Analysis of the Identified Literature
4.1. Systematization of the Included Papers
4.2. Detailed Overview of the Included Papers
4.2.1. Deep Reinforcement Learning: Deep Q-Learning
4.2.2. Deep Reinforcement Learning
4.2.3. Combined Methods
4.2.4. Generative AI
4.2.5. Deep Neural Network
4.2.6. Recurrent Neural Network
4.2.7. Convolutional Neural Network
4.2.8. Transformer
5. Discussion
5.1. Evolution of the Identified Research Trends
5.2. Identification of the Research Gaps
6. Conclusions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
A2C | Advanced Actor–Critic |
A3C | Asynchronous Actor–Critic |
AGV | Automated Guided Vehicle |
ARS | Augmented Random Search |
AI | Artificial Intelligence |
AMR | Autonomous Mobile Robots |
ANN | Artificial Neural Network |
API | Application Programming Interface |
ASMG | Automated Simulation Model Generation |
BDQ | Branching Duelling Q-Network |
CNN | Convolutional Neural Network |
CONWIP | Constant Work In Process |
DDDQN | Duelling Double Deep Q-network |
DDMRP | Demand-Driven Materials Requirements Planning |
DDQL | Double Deep Q-learning |
DDQN | Double Deep Q-network |
DEPS | Discrete Event Probabilistic Simulation |
DES | Discrete-Event Simulation |
DL | Deep Learning |
DNN | Deep Neural Network |
DQL | Deep Q-learning |
DQN | Deep Q-network |
DRL | Deep Reinforcement Learning |
DT | Digital Twin |
EDD | earliest due date |
FIFO | First-in, first-out |
GAN | Generative Adversarial Network |
GGMM | Generative Graphical Metamodel |
GRASP | Greedy Randomized Adaptive Search Procedures |
GNN | Graph Neural Network |
HP | highest priority first |
IoT | Internet of Things |
JSON | JavaScript Object Notation |
LDF | lowest-distance-first |
LLM | Large Language Model |
LMDM | Large Manufacturing Decision Model |
LSTM | Long Short-Term Memory |
MADRL | Multi-agent Deep Reinforcement Learning |
MAPE | Mean Absolute Percentual Error |
MDN | Mixture Density Network |
MDP | Markov Decision Process or Markovian Decision-Making Process |
ML | Machine Learning |
MLPNN | Multilayer Perceptron Neural Network |
MTRL | Multi-task Reinforcement Learning |
NARX | Nonlinear Autoregressive with External Input |
PPO | Proximal Policy Optimization |
QR-DQN | Quantile Regression DQN |
RL | Reinforcement Learning |
RNN | Recurrent Neural Network |
RNN-ED | Recurrent Encoder–Decoder |
SAC | Soft Actor Critic |
SAN | Stochastic Activity Network |
SL | Supervised Learning |
SPT | shortest processing time |
TH | throughput |
TRPO | Trust Region Policy Optimization |
TTR | Time to Recover |
UAS | Unmanned Aircraft Systems |
UTMC | Urban Traffic Management and Control |
WIP | Work In Process |
References
- Agalianos, K.; Ponis, S.T.; Aretoulaki, E.; Plakas, G.; Efthymiou, O. Discrete event simulation and digital twins: Review and challenges for logistics. Procedia Manuf. 2020, 51, 1636–1641. [Google Scholar] [CrossRef]
- Kogler, C.; Maxera, P. A literature review of supply chain analyses integrating discrete simulation modelling and machine learning. J. Simul. 2025, 1–25. [Google Scholar] [CrossRef]
- Makarov, V.L.; Bakhtizin, A.R.; Beklaryan, G.L.; Akopov, A.S. Digital plant: Methods of discrete-event modeling and optimization of production characteristics. Bus. Inform. 2021, 15, 7–20. [Google Scholar] [CrossRef]
- Lang, S.; Reggelin, T.; Müller, M.; Nahhas, A. Open-source discrete-event simulation software for applications in production and logistics: An alternative to commercial tools? Procedia Comput. Sci. 2021, 180, 978–987. [Google Scholar] [CrossRef]
- Müller, M.; Reggelin, T.; Kutsenko, I.; Zadek, H.; Reyes-Rubiano, L.S. Towards Deadlock Handling with Machine Learning in a Simulation-Based Learning Environment. In Proceedings of the 2022 Winter Simulation Conference (WSC), Singapore, 11–14 December 2022; pp. 1485–1496. [Google Scholar]
- Rabe, M.; Dross, F.; Wuttke, A.; Duarte, A.; Juan, A.; Lourenço, H.R. Combining a discrete-event simulation model of a logistics network with deep reinforcement learning. In Proceedings of the MIC and MAEB 2017 Conferences, Barcelona, Spain, 4–7 July 2017; pp. 765–774. [Google Scholar]
- Feldkamp, N.; Bergmann, S.; Strassburger, S. Simulation-based deep reinforcement learning for modular production systems. In Proceedings of the 2020 Winter Simulation Conference (WSC), Orlando, FL, USA, 14–18 December 2020; pp. 1596–1607. [Google Scholar]
- de Carvalho, J.P.; Dimitrakopoulos, R. Integrating production planning with truck-dispatching decisions through reinforcement learning while managing uncertainty. Minerals 2021, 11, 587. [Google Scholar] [CrossRef]
- Altenmüller, T.; Stüker, T.; Waschneck, B.; Kuhnle, A.; Lanza, G. Reinforcement learning for an intelligent and autonomous production control of complex job-shops under time constraints. Prod. Eng. 2020, 14, 319–328. [Google Scholar] [CrossRef]
- Eriksson, K.; Ramasamy, S.; Zhang, X.; Wang, Z.; Danielsson, F. Conceptual framework of scheduling applying discrete event simulation as an environment for deep reinforcement learning. Procedia CIRP 2022, 107, 955–960. [Google Scholar] [CrossRef]
- Tang, J.; Salonitis, K. A deep reinforcement learning based scheduling policy for reconfigurable manufacturing systems. Procedia CIRP 2021, 103, 1–7. [Google Scholar] [CrossRef]
- Dittrich, M.A.; Fohlmeister, S. A deep q-learning-based optimization of the inventory control in a linear process chain. Prod. Eng. 2021, 15, 35–43. [Google Scholar] [CrossRef]
- Silva, T.; Azevedo, A. Production flow control through the use of reinforcement learning. Procedia Manuf. 2019, 38, 194–202. [Google Scholar] [CrossRef]
- Hofmann, C.; Krahe, C.; Stricker, N.; Lanza, G. Autonomous production control for matrix production based on deep Q-learning. Procedia CIRP 2020, 88, 25–30. [Google Scholar] [CrossRef]
- Duhem, L.; Benali, M.; Martin, G. Parametrization of a demand-driven operating model using reinforcement learning. Comput. Ind. 2023, 147, 103874. [Google Scholar] [CrossRef]
- Lee, W.; Oh, S.C.; Park, J.; Lee, C.; Fan, H.T.; Arinez, J.; An, S.; Do Noh, S. Matrix Assembly System Scheduling Optimization in Automotive Manufacturing: A Deep Q-Network Approach. In Proceedings of the 2024 Winter Simulation Conference (WSC), Orlando, FL, USA, 15–18 December 2024; pp. 3434–3445. [Google Scholar]
- Noriega, R.; Pourrahimian, Y.; Askari-Nasab, H. Deep Reinforcement Learning based real-time open-pit mining truck dispatching system. Comput. Oper. Res. 2025, 173, 106815. [Google Scholar] [CrossRef]
- Alexopoulos, K.; Mavrothalassitis, P.; Bakopoulos, E.; Nikolakis, N.; Mourtzis, D. Deep Reinforcement Learning for Selection of Dispatch Rules for Scheduling of Production Systems. Appl. Sci. 2024, 15, 232. [Google Scholar] [CrossRef]
- Waschneck, B.; Reichstaller, A.; Belzner, L.; Altenmüller, T.; Bauernhansl, T.; Knapp, A.; Kyek, A. Optimization of global production scheduling with deep reinforcement learning. Procedia CIRP 2018, 72, 1264–1269. [Google Scholar] [CrossRef]
- Lang, S.; Behrendt, F.; Lanzerath, N.; Reggelin, T.; Müller, M. Integration of deep reinforcement learning and discrete-event simulation for real-time scheduling of a flexible job shop production. In Proceedings of the 2020 Winter Simulation Conference (WSC), Orlando, FL, USA, 14–18 December 2020; pp. 3057–3068. [Google Scholar]
- Valet, A.; Altenmüller, T.; Waschneck, B.; May, M.C.; Kuhnle, A.; Lanza, G. Opportunistic maintenance scheduling with deep reinforcement learning. J. Manuf. Syst. 2022, 64, 518–534. [Google Scholar] [CrossRef]
- Nikkerdar, M.; ElMaraghy, W.; ElMaraghy, H. Smart adaptable assembly line rebalancing based on reinforcement learning. Int. J. Adv. Manuf. Technol. 2025, 138, 4155–4173. [Google Scholar] [CrossRef]
- Shi, D.; Fan, W.; Xiao, Y.; Lin, T.; Xing, C. Intelligent scheduling of discrete automated production line via deep reinforcement learning. Int. J. Prod. Res. 2020, 58, 3362–3380. [Google Scholar] [CrossRef]
- Marchesano, M.G.; Guizzi, G.; Santillo, L.C.; Vespoli, S. Dynamic scheduling in a flow shop using deep reinforcement learning. In Proceedings of the IFIP International Conference on Advances in Production Management Systems, Nantes, France, 31 August 2021; pp. 152–160. [Google Scholar]
- Vespoli, S.; Mattera, G.; Marchesano, M.G.; Nele, L.; Guizzi, G. Adaptive manufacturing control with Deep Reinforcement Learning for dynamic WIP management in industry 4.0. Comput. Ind. Eng. 2025, 202, 110966. [Google Scholar] [CrossRef]
- Nahhas, A.; Kharitonov, A.; Turowski, K. Deep reinforcement learning for solving allocation problems in supply chain: An image-based observation space. Procedia Comput. Sci. 2024, 232, 2570–2579. [Google Scholar] [CrossRef]
- Lim, J.B.; Jeong, J. Factory Simulation of Optimization Techniques Based on Deep Reinforcement Learning for Storage Devices. Appl. Sci. 2023, 13, 9690. [Google Scholar] [CrossRef]
- Jiang, C.; Li, X.; Lin, J.R.; Liu, M.; Ma, Z. Adaptive control of resource flow to optimize construction work and cash flow via online deep reinforcement learning. Autom. Constr. 2023, 150, 104817. [Google Scholar] [CrossRef]
- Yang, S.L.; Wang, J.Y.; Xin, L.M.; Xu, Z.G. Verification of intelligent scheduling based on deep reinforcement learning for distributed workshops via discrete event simulation. Adv. Prod. Eng. Manag. 2022, 17, 401–412. [Google Scholar] [CrossRef]
- Bitsch, G.; Senjic, P. Intelligent scheduling based on discrete-time simulation using machine learning. Procedia CIRP 2024, 126, 48–51. [Google Scholar] [CrossRef]
- Krenczyk, D. Deep reinforcement learning and discrete simulation-based digital twin for cyber–physical production systems. Appl. Sci. 2024, 14, 5208. [Google Scholar] [CrossRef]
- Bretas, A.M.C.; Mendes, A.; Chalup, S.; Jackson, M.; Clement, R.; Sanhueza, C. Addressing deadlock in large-scale, complex rail networks via multi-agent deep reinforcement learning. Expert Syst. 2025, 42, e13315. [Google Scholar] [CrossRef]
- Liu, Y.; Fan, J.; Zhao, L.; Shen, W.; Zhang, C. Integration of deep reinforcement learning and multi-agent system for dynamic scheduling of re-entrant hybrid flow shop considering worker fatigue and skill levels. Robot. Comput.-Integr. Manuf. 2023, 84, 102605. [Google Scholar] [CrossRef]
- Tremblay, F.A.; Durand, A.; Morin, M.; Marier, P.; Gaudreault, J. Deep reinforcement learning for continuous wood drying production line control. Comput. Ind. 2024, 154, 104036. [Google Scholar] [CrossRef]
- Kumar, A.; Kumar, R. Improving efficiency of steel converter facility: A digital twin and reinforcement learning approach. Int. J. Prod. Res. 2025, 1–16. [Google Scholar] [CrossRef]
- Göppert, A.; Mohring, L.; Schmitt, R.H. Predicting performance indicators with ANNs for AI-based online scheduling in dynamically interconnected assembly systems. Prod. Eng. 2021, 15, 619–633. [Google Scholar] [CrossRef]
- Klar, M.; Schworm, P.; Wu, X.; Simon, P.; Glatt, M.; Ravani, B.; Aurich, J.C. Transferable multi-objective factory layout planning using simulation-based deep reinforcement learning. J. Manuf. Syst. 2024, 74, 487–511. [Google Scholar] [CrossRef]
- Klar, M.; Simon, P.M.; Ravani, B.; Aurich, J.C. Simulation-based brownfield factory planning using deep reinforcement learning. Digit. Eng. 2024, 3, 100026. [Google Scholar] [CrossRef]
- Zuercher, P.D.; Bohné, T. Discrete Event Probabilistic Simulation (DEPS) integrated into a Reinforcement Learning Framework for Optimal Production. Procedia CIRP 2024, 130, 662–670. [Google Scholar] [CrossRef]
- Panzer, M.; Gronau, N. Designing an adaptive and deep learning based control framework for modular production systems. J. Intell. Manuf. 2024, 35, 4113–4136. [Google Scholar] [CrossRef]
- Seah, S.X.; Srigrarom, S. Multiple UAS Traffic Planning Based on Deep Q-Network with Hindsight Experience Replay and Economic Considerations. Aerospace 2023, 10, 980. [Google Scholar] [CrossRef]
- Zhou, W.; Fotouhi, H.; Miller-Hooks, E. Decision support through deep reinforcement learning for maximizing a courier’s monetary gain in a meal delivery environment. Decis. Support Syst. 2025, 190, 114388. [Google Scholar] [CrossRef]
- Leon, J.F.; Li, Y.; Martin, X.A.; Calvet, L.; Panadero, J.; Juan, A.A. A hybrid simulation and reinforcement learning algorithm for enhancing efficiency in warehouse operations. Algorithms 2023, 16, 408. [Google Scholar] [CrossRef]
- Göppert, A.; Rachner, J.; Kaven, L.; Schmitt, R.H. Modular software architecture for interfacing online scheduling agents with assembly planning and control systems. In Proceedings of the Conference on Production Systems and Logistics: CPSL, Vancouver, BC, Canada, 17–20 May 2022; Publish-Ing.: Hannover, Germany, 2022; pp. 206–216. [Google Scholar]
- Lang, S.; Kuetgens, M.; Reichardt, P.; Reggelin, T. Modeling production scheduling problems as reinforcement learning environments based on discrete-event simulation and openai gym. IFAC-PapersOnLine 2021, 54, 793–798. [Google Scholar] [CrossRef]
- Rumin, R.; Duda, J.; Blaut, J.; Pękała, D.; Merolla, T. Utilizing Artificial Intelligence for Energy-Efficient Route Planning in Hyperloop Low-Pressure Capsule Transit: A Study in Alignment with Sustainable Development Goals; Scientific Papers of Silesian University of Technology: Organization and Management Series No. 204; Silesian University of Technology Publishing House: Gliwice, Poland, 2023. [Google Scholar] [CrossRef]
- Cen, W.; Haas, P.J. Enhanced simulation metamodeling via graph and generative neural networks. In Proceedings of the 2022 Winter Simulation Conference (WSC), Singapore, 11–14 December 2022; pp. 2748–2759. [Google Scholar]
- Zhang, Y.; Zhang, J.; Wang, X.; Zeng, T. Multi-task deep reinforcement learning for dynamic scheduling of large-scale fleets in earthmoving operations. Autom. Constr. 2025, 174, 106123. [Google Scholar] [CrossRef]
- Ogunsoto, O.V.; Olivares-Aguila, J.; ElMaraghy, W. A conceptual digital twin framework for supply chain recovery and resilience. Supply Chain. Anal. 2025, 9, 100091. [Google Scholar] [CrossRef]
- Kulmer, F.; Wolf, M.; Ramsauer, C. Planning Assistant for Medium-term Capacity Management using Deep Reinforcement Learning. IFAC-PapersOnLine. 2024, 58, 31–36. [Google Scholar] [CrossRef]
- Fischer, A.; Beiderwellen Bedrikow, A.; Tommelein, I.D.; Nübel, K.; Fottner, J. From activity recognition to simulation: The impact of granularity on production models in heavy civil engineering. Algorithms 2023, 16, 212. [Google Scholar] [CrossRef]
- Reider, R.; Lang, J.P.D.I.S. GenAI for Simulation Modeling: A Synergistic Approach. In Proceedings of the 18th International Doctoral Students Workshop on Logistics, Supply Chain and Production Management, Magdeburg, Germany, 17 June 2025. [Google Scholar]
- Fei, Y.; Song, M.; Jiao, R.J. AI-assisted Intelligent Discrete-Event Simulation Platform for Manufacturing Systems. In Proceedings of the 15th International Conference on Industrial Engineering and Operations Management, Singapore, 18–20 February 2025; pp. 2748–2759. [Google Scholar]
- Li, X.; Nassehi, A.; Hu, S.J.; Joung, B.G.; Gao, R.X. A large manufacturing decision model for human-centric decision-making. CIRP Ann. 2025, 74, 621–625. [Google Scholar] [CrossRef]
- Chen, Y.C.; Loh, C.H.; Wang, F.C.; Chen, Z.J.; Fu, S.H.; Wang, C.Y. Application of Generative Adversarial Network to Optimize Vehicle Allocation at Dispatch Stations of Paratransit Services. Electronics 2022, 11, 423. [Google Scholar] [CrossRef]
- Montevechi, J.A.B.; Gabriel, G.T.; Campos, A.T.; dos Santos, C.H.; Leal, F.; Machado, M.E. Using generative adversarial networks to validate discrete event simulation models. In Proceedings of the 2022 Winter Simulation Conference (WSC), Singapore, 11–14 December 2022; pp. 2772–2783. [Google Scholar]
- Feldkamp, N.; Bergmann, S.; Conrad, F.; Strassburger, S. A method using generative adversarial networks for robustness optimization. ACM Trans. Model. Comput. Simul. (TOMACS) 2022, 32, 1–22. [Google Scholar] [CrossRef]
- Jackson, I.; Jesus Saenz, M.; Ivanov, D. From natural language to simulations: Applying AI to automate simulation modelling of logistics systems. Int. J. Prod. Res. 2024, 62, 1434–1457. [Google Scholar] [CrossRef]
- Kumar, G.S.; Gite, P.; Prasad, M.; Madiajagan, M.; Selvakanmani, S.; Saravanan, V. Design of Autonomous Production Using Deep Neural Network for Complex Job-Shops. Int. J. Grid Distrib. Comput. 2021, 14, 813–824. [Google Scholar]
- Molnár, Z.; Tamás, P.; Illés, B. The always changing data problem of using AI in manufacturing-Using synthetic data from the digital twin to feed AI models. Adv. Logist. Syst.-Theory Pract. 2024, 18, 19–28. [Google Scholar] [CrossRef]
- Zhang, M.; Tao, F.; Nee, A.Y.C. Digital twin enhanced dynamic job-shop scheduling. J. Manuf. Syst. 2021, 58, 146–156. [Google Scholar] [CrossRef]
- Reed, S.; Löfstrand, M.; Andrews, J. Modelling stochastic behaviour in simulation digital twins through neural nets. J. Simul. 2022, 16, 512–525. [Google Scholar] [CrossRef]
- Vespoli, S.; Grassi, A.; Guizzi, G.; Popolo, V. Generalised performance estimation in novel hybrid MPC architectures: Modeling the CONWIP flow-shop system. Appl. Sci. 2023, 13, 4808. [Google Scholar] [CrossRef]
- Yildirim, S.; Murat, A.E.; Yildirim, M.; Arslanturk, S. Process Knowledge Driven Change Point Detection for Automated Calibration of Discrete Event Simulation Models Using Machine Learning. arXiv 2020, arXiv:2005.05385. [Google Scholar] [CrossRef]
- Elkhaili; Mohame, D.; Terrada, L.; Ouajji, H.; Daaif, A. Towards a green supply chain based on smart urban traffic using deep learning approach. Stat. Optim. Inf. Comput. 2022, 10, 25–44. [Google Scholar] [CrossRef]
- Woerrlein, B.; Strassburger, S. On the Usage of Deep Learning for Modelling Energy Consumption in Simulation Models. Simul. Notes Eur. 2020, 30, 165–174. [Google Scholar] [CrossRef]
- Ashraf, M.; Eltawil, A.; Ali, I. Time-to-recovery prediction in a disrupted three-echelon supply chain using LSTM. IFAC-PapersOnLine. 2022, 55, 1319–1324. [Google Scholar] [CrossRef]
- Sommer, M.; Stjepandić, J.; Stobrawa, S.; von Soden, M. Automated generation of digital twin for a built environment using scan and object detection as input for production planning. J. Ind. Inf. Integr. 2023, 33, 100462. [Google Scholar] [CrossRef]
- Innuphat, S.; Toahchoodee, M. The implementation of discrete-event simulation and demand forecasting using Temporal Fusion Transformers to validate spare parts inventory policy for petrochemicals industry. ECTI Trans. Comput. Inf. Technol. (ECTI-CIT) 2022, 16, 247–258. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 6000–6010. [Google Scholar]
Reference | Application Area | Applied Method | General Methodology | Overall Number of Included Papers |
---|---|---|---|---|
[5] | AGV routing | Deep Q-Learning | Deep Reinforcement Learning: Deep Q-Learning | 21 |
[6] | Logistics network | Deep Q-Learning (Convolutional Neural Network) | ||
[7] | AGV routing | Double Deep Q-Learning | ||
[8] | Fleet management | Double Deep Q-Learning | ||
[9] | Production control | Deep Q-Learning | ||
[10] | Production scheduling | Deep Q-Learning | ||
[11] | Manufacturing system reconfiguration | Duelling Double DQN | ||
[12] | Inventory control | Deep Q-Learning | ||
[13] | Production flow control | Deep Q-Learning | ||
[14] | Production control | Deep Q-Learning | ||
[15] | Material requirements planning | Branching Duelling Q-Network | ||
[16] | Production scheduling | Deep Q-Learning | ||
[17] | Fleet management | Double Deep Q-Learning | ||
[18] | Production scheduling | Deep Q-Learning | ||
[19] | Production scheduling | Deep Q-Learning | ||
[20] | Production scheduling | Deep Q-Learning (LSTM) | ||
[21] | Maintenance scheduling | Deep Q-Learning | ||
[22] | Assembly line rebalancing | Deep Q-Learning | ||
[23] | Production scheduling | Deep Q-Learning | ||
[24] | Production scheduling | Deep Q-Learning | ||
[25] | Production control | Deep Q-Learning | ||
[26] | Supply chain optimization | Deep Reinforcement Learning (A3C, PPO) | Deep Reinforcement Learning | 10 |
[27] | Storage optimization | Deep Reinforcement Learning (PPO) | ||
[28] | Resource flow control | Deep Reinforcement Learning (PPO) | ||
[29] | Production scheduling | Deep Reinforcement Learning (A2C) | ||
[30] | Production scheduling | Deep Reinforcement Learning (PPO) | ||
[31] | Production logistics | Deep Reinforcement Learning (PPO, A2C) | ||
[32] | Traffic management | Multi-Agent Deep Reinforcement Learning | ||
[33] | Production scheduling | Multi-Agent Deep Reinforcement Learning | ||
[34] | Production control | Deep Reinforcement Learning (PPO) | ||
[35] | Material flow optimization | Deep Reinforcement Learning (PPO) | ||
[36] | Production scheduling | Combined (AlphaZero and greedy agent) | Combined | 16 |
[37] | Factory planning | Combined (GNN, CNN, and Rainbow DQN) | ||
[38] | Brownfield factory planning | Combined (GNN, CNN, and Rainbow DQN) | ||
[39] | Production optimization | Combined (DQN, DDQN, DDDQN, and SAC) | ||
[40] | Production control | Combined (DQN and heuristics) | ||
[41] | Traffic planning | Combined (DQN, and CNN) | ||
[42] | Profit optimization | Combined (DQN, DDQN, Duelling DQN, D3QN, and A2C) | ||
[43] | Warehouse management | Combined (A2C, PPO, and DQN) | ||
[44] | Production scheduling | Combined (Software Architecture for Online Agents) | ||
[45] | Production scheduling | Combined (Integration of OpenAI Gym with DES) | ||
[46] | Route planning | Combined (Integration of OpenAI Gym with DES) | ||
[47] | Metamodeling | Combined (Graph Neural Network and Generative Neural Network) | ||
[48] | Fleet management | Combined (Multi-Task Reinforcement Learning, LSTM, and A3C) | ||
[49] | supply chain management | Combined (LSTM and Multilayer Perceptron Neural Network) | ||
[50] | Production planning | Combined (TRPO, PPO, Recurrent PPO, DQN, QR-DQN, ARS, and A2C) | ||
[51] | Construction process optimization | Combined (CNN, LSTM, and Softmax) | ||
[52] | Automated simulation model generation | Large Language Model | Generative AI | 7 |
[53] | Automated simulation model generation | Large Language Model | ||
[54] | Manufacturing system reconfiguration | Large Manufacturing Decision Model | ||
[55] | Fleet management | Generative Adversarial Network | ||
[56] | DES model validation | Generative Adversarial Network | ||
[57] | Robustness optimization | Generative Adversarial Network | ||
[58] | Automated simulation model generation | Large Language Model | ||
[59] | Production control | Deep Neural Network | Deep Neural Network | 5 |
[60] | Production logistics | Deep Neural Network | ||
[61] | Production scheduling | Deep Neural Network | ||
[62] | Vehicle operation | Deep Neural Network (Mixture Density Network) | ||
[63] | Production control | Deep Neural Network | ||
[64] | Automated calibration of DES models | NARX (Nonlinear Autoregressive Exogenous) Neural Network | Recurrent Neural Network | 4 |
[65] | Urban traffic | Long Short-Term Memory | ||
[66] | Energy consumption modelling | Recurrent Encoder–Decoder Model | ||
[67] | Supply chain management | Long Short-Term Memory | ||
[68] | Automated simulation model generation | Convolutional Neural Network | Convolutional Neural Network | 1 |
[69] | Inventory management | Temporal Fusion Transformer | Transformer | 1 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Skapinyecz, R. Recent Trends in the Optimization of Logistics Systems Through Discrete-Event Simulation and Deep Learning. Algorithms 2025, 18, 573. https://doi.org/10.3390/a18090573
Skapinyecz R. Recent Trends in the Optimization of Logistics Systems Through Discrete-Event Simulation and Deep Learning. Algorithms. 2025; 18(9):573. https://doi.org/10.3390/a18090573
Chicago/Turabian StyleSkapinyecz, Róbert. 2025. "Recent Trends in the Optimization of Logistics Systems Through Discrete-Event Simulation and Deep Learning" Algorithms 18, no. 9: 573. https://doi.org/10.3390/a18090573
APA StyleSkapinyecz, R. (2025). Recent Trends in the Optimization of Logistics Systems Through Discrete-Event Simulation and Deep Learning. Algorithms, 18(9), 573. https://doi.org/10.3390/a18090573