Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (357)

Search Parameters:
Keywords = RLS technique

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
24 pages, 3185 KB  
Article
A Hybrid Optimization Approach for Multi-Generation Intelligent Breeding Decisions
by Mingxiang Yang, Ziyu Li, Jiahao Li, Bingling Huang, Xiaohui Niu, Xin Lu and Xiaoxia Li
Information 2026, 17(1), 106; https://doi.org/10.3390/info17010106 - 20 Jan 2026
Viewed by 124
Abstract
Multi-generation intelligent breeding (MGIB) decision-making is a technique used by plant breeders to select mating individuals to produce new generations and allocate resources for each generation. However, existing research remains scarce on dynamic optimization of resources under limited budget and time constraints. Inspired [...] Read more.
Multi-generation intelligent breeding (MGIB) decision-making is a technique used by plant breeders to select mating individuals to produce new generations and allocate resources for each generation. However, existing research remains scarce on dynamic optimization of resources under limited budget and time constraints. Inspired by advances in reinforcement learning (RL), a framework that integrates evolutionary algorithms with deep RL was proposed to fill this gap. The framework combines two modules: the Improved Look-Ahead Selection (ILAS) module and Deep Q-Networks (DQNs) module. The former employs a simulated annealing-enhanced estimation of the distribution algorithm to make mating decisions. Based on the selected mating individual, the latter module learns multi-generation resource allocation policies using DQN. To evaluate our framework, numerical experiments were conducted on two realistic breeding datasets, i.e., Corn2019 and CUBIC. The ILAS outperformed LAS on corn2019, increasing the maximum and mean population Genomic Estimated Breeding Value (GEBV) by 9.1% and 7.7%. ILAS-DQN consistently outperformed the baseline methods, achieving significant and practical improvements in both top-performing and elite-average GEBVs across two independent datasets. The results demonstrated that our method outperforms traditional baselines, in both generalization and effectiveness for complex agricultural problems with delayed rewards. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Graphical abstract

26 pages, 2235 KB  
Article
Climate-Resilient Reinforcement Learning Control of Hybrid Ventilation in Mediterranean Offices Under Future Climate Scenarios
by Hussein Krayem, Jaafar Younes and Nesreen Ghaddar
Sustainability 2026, 18(2), 1037; https://doi.org/10.3390/su18021037 - 20 Jan 2026
Viewed by 118
Abstract
This study develops an explainable reinforcement learning (RL) control framework for hybrid ventilation in Mediterranean office buildings to enhance thermal comfort, energy efficiency, and long-term climate resilience. A working environment was created Using EnergyPlus to represent an office test cell equipped with natural [...] Read more.
This study develops an explainable reinforcement learning (RL) control framework for hybrid ventilation in Mediterranean office buildings to enhance thermal comfort, energy efficiency, and long-term climate resilience. A working environment was created Using EnergyPlus to represent an office test cell equipped with natural ventilation and air conditioning. The RL controller, based on Proximal Policy Optimization (PPO), was trained exclusively on present-day Typical Meteorological Year (TMY) data from Beirut and subsequently evaluated, without retraining, under future 2050 and 2080 climate projections (SSP1-2.6 and SSP5-8.5) generated using the Belcher morphing technique, in order to quantify robustness under projected climate stressors. Results showed that the RL control achieved consistent, though moderate, annual HVAC energy reductions (6–9%), and a reduction in indoor overheating degree (IOD) by about 35.66% compared to rule-based control, while maintaining comfort and increasing natural ventilation hours. The Climate Change Overheating Resistivity (CCOR) improved by 24.32%, demonstrating the controller’s resilience under warming conditions. Explainability was achieved through Kernel SHAP, which revealed physically coherent feature influences consistent with thermal comfort logic. The findings confirmed that physics-informed RL can autonomously learn and sustain effective ventilation control, remaining transparent, reliable, and robust under future climates. This framework establishes a foundation for adaptive and interpretable RL-based hybrid ventilation control, enabling long-lived office buildings in Mediterranean climates to reduce cooling energy demand and mitigate overheating risks under future climate change. Full article
(This article belongs to the Section Energy Sustainability)
Show Figures

Figure 1

16 pages, 4339 KB  
Article
Reinforcement Learning Technique for Self-Healing FBG Sensor Systems in Optical Wireless Communication Networks
by Rénauld A. Dellimore, Jyun-Wei Li, Hung-Wei Huang, Amare Mulatie Dehnaw, Cheng-Kai Yao, Pei-Chung Liu and Peng-Chun Peng
Appl. Sci. 2026, 16(2), 1012; https://doi.org/10.3390/app16021012 - 19 Jan 2026
Viewed by 159
Abstract
This paper proposes a large-scale, self-healing multipoint fiber Bragg grating (FBG) sensor network that employs reinforcement learning (RL) techniques to enhance the resilience and efficiency of optical wireless communication networks. The system features a mesh-structured, self-healing ring-mesh architecture employing 2 × 2 optical [...] Read more.
This paper proposes a large-scale, self-healing multipoint fiber Bragg grating (FBG) sensor network that employs reinforcement learning (RL) techniques to enhance the resilience and efficiency of optical wireless communication networks. The system features a mesh-structured, self-healing ring-mesh architecture employing 2 × 2 optical switches, enabling robust multipoint sensing and fault tolerance in the event of one or more link failures. To further extend network coverage and support distributed deployment scenarios, free-space optical (FSO) links are integrated as wireless optical backhaul between central offices and remote monitoring sites, including structural health, renewable energy, and transportation systems. These FSO links offer high-speed, line-of-sight connections that complement physical fiber infrastructure, particularly in locations where cable deployment is impractical. Additionally, RL-based artificial intelligence (AI) techniques are employed to enable intelligent path selection, optimize routing, and enhance network reliability. Experimental results confirm that the RL-based approach effectively identifies optimal sensing paths among multiple routing options, both wired and wireless, resulting in reduced energy consumption, extended sensor network lifespan, and improved transmission delay. The proposed hybrid FSO–fiber self-healing sensor system demonstrates high survivability, scalability, and low routing path loss, making it a strong candidate for future services and mission-critical applications. Full article
(This article belongs to the Section Electrical, Electronics and Communications Engineering)
Show Figures

Figure 1

39 pages, 4912 KB  
Systematic Review
Grid-Scale Battery Energy Storage and AI-Driven Intelligent Optimization for Techno-Economic and Environmental Benefits: A Systematic Review
by Nipon Ketjoy, Yirga Belay Muna, Malinee Kaewpanha, Wisut Chamsa-ard, Tawat Suriwong and Chakkrit Termritthikun
Batteries 2026, 12(1), 31; https://doi.org/10.3390/batteries12010031 - 17 Jan 2026
Viewed by 298
Abstract
Grid-Scale Battery Energy Storage Systems (GS-BESS) play a crucial role in modern power grids, addressing challenges related to integrating renewable energy sources (RESs), load balancing, peak shaving, voltage support, load shifting, frequency regulation, emergency response, and enhancing system stability. However, harnessing their full [...] Read more.
Grid-Scale Battery Energy Storage Systems (GS-BESS) play a crucial role in modern power grids, addressing challenges related to integrating renewable energy sources (RESs), load balancing, peak shaving, voltage support, load shifting, frequency regulation, emergency response, and enhancing system stability. However, harnessing their full potential and lifetime requires intelligent operational strategies that balance technological performance, economic viability, and environmental sustainability. This systematic review examines how artificial intelligence (AI)-based intelligent optimization enhances GS-BESS performance, focusing on its techno-economic, environmental impacts, and policy and regulatory implications. Following Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines, we review the evolution of GS-BESS, analyze its advancements, and assess state-of-the-art applications and emerging AI techniques for GS-BESS optimization. AI techniques, including machine learning (ML), predictive modeling, optimization algorithms, deep learning (DL), and reinforcement learning (RL), are examined for their ability to improve operational efficiency and control precision in GS-BESSs. Furthermore, the review discusses the benefits of advanced dispatch strategies, including economic efficiency, emissions reduction, and improved grid resilience. Despite significant progress, challenges persist in data availability, model generalization, high computational requirements, scalability, and regulatory gaps. We conclude by identifying emerging opportunities to guide the next generation of intelligent energy storage systems. This work serves as a foundational resource for researchers, engineers, and policymakers seeking to advance the deployment of AI-enhanced GS-BESS for sustainable, resilient power systems. By analyzing the latest developments in AI applications and BESS technologies, this review provides a comprehensive perspective on their synergistic potential to drive sustainability, cost-effectiveness, and energy systems reliability. Full article
(This article belongs to the Special Issue AI-Powered Battery Management and Grid Integration for Smart Cities)
Show Figures

Figure 1

50 pages, 3712 KB  
Article
Explainable AI and Multi-Agent Systems for Energy Management in IoT-Edge Environments: A State of the Art Review
by Carlos Álvarez-López, Alfonso González-Briones and Tiancheng Li
Electronics 2026, 15(2), 385; https://doi.org/10.3390/electronics15020385 - 15 Jan 2026
Viewed by 205
Abstract
This paper reviews Artificial Intelligence techniques for distributed energy management, focusing on integrating machine learning, reinforcement learning, and multi-agent systems within IoT-Edge-Cloud architectures. As energy infrastructures become increasingly decentralized and heterogeneous, AI must operate under strict latency, privacy, and resource constraints while remaining [...] Read more.
This paper reviews Artificial Intelligence techniques for distributed energy management, focusing on integrating machine learning, reinforcement learning, and multi-agent systems within IoT-Edge-Cloud architectures. As energy infrastructures become increasingly decentralized and heterogeneous, AI must operate under strict latency, privacy, and resource constraints while remaining transparent and auditable. The study examines predictive models ranging from statistical time series approaches to machine learning regressors and deep neural architectures, assessing their suitability for embedded deployment and federated learning. Optimization methods—including heuristic strategies, metaheuristics, model predictive control, and reinforcement learning—are analyzed in terms of computational feasibility and real-time responsiveness. Explainability is treated as a fundamental requirement, supported by model-agnostic techniques that enable trust, regulatory compliance, and interpretable coordination in multi-agent environments. The review synthesizes advances in MARL for decentralized control, communication protocols enabling interoperability, and hardware-aware design for low-power edge devices. Benchmarking guidelines and key performance indicators are introduced to evaluate accuracy, latency, robustness, and transparency across distributed deployments. Key challenges remain in stabilizing explanations for RL policies, balancing model complexity with latency budgets, and ensuring scalable, privacy-preserving learning under non-stationary conditions. The paper concludes by outlining a conceptual framework for explainable, distributed energy intelligence and identifying research opportunities to build resilient, transparent smart energy ecosystems. Full article
Show Figures

Figure 1

60 pages, 3790 KB  
Review
Autonomous Mobile Robot Path Planning Techniques—A Review: Metaheuristic and Cognitive Techniques
by Mubarak Badamasi Aremu, Gamil Ahmed, Sami Elferik and Abdul-Wahid A. Saif
Robotics 2026, 15(1), 23; https://doi.org/10.3390/robotics15010023 - 14 Jan 2026
Viewed by 257
Abstract
Autonomous mobile robots (AMRs) require robust, efficient path planning to operate safely in complex, often dynamic environments (e.g., logistics, transportation, and healthcare). This systematic review focuses on advanced metaheuristic and learning- and reasoning-based (cognitive) techniques for AMR path planning. Drawing on approximately 230 [...] Read more.
Autonomous mobile robots (AMRs) require robust, efficient path planning to operate safely in complex, often dynamic environments (e.g., logistics, transportation, and healthcare). This systematic review focuses on advanced metaheuristic and learning- and reasoning-based (cognitive) techniques for AMR path planning. Drawing on approximately 230 articles published between 2018 and 2025, we organize the literature into two prominent families, metaheuristic optimization and AI-based navigation, and introduce and apply a unified taxonomy (planning scope, output type, and constraint awareness) to guide the comparative analysis and practitioner-oriented synthesis. We synthesize representative approaches, including swarm- and evolutionary-based planners (e.g., PSO, GA, ACO, GWO), fuzzy and neuro-fuzzy systems, neural methods, and RL/DRL-based navigation, highlighting their operating principles, recent enhancements, strengths, and limitations, and typical deployment roles within hierarchical navigation stacks. Comparative tables and a compact trade-off synthesis summarize capabilities across static/dynamic settings, real-world validation, and hybridization trends. Persistent gaps remain in parameter tuning, safety, and interpretability of learning-enabled navigation; sim-to-real transfer; scalability under real-time compute limits; and limited physical experimentation. Finally, we outline research opportunities and open research questions, covering benchmarking and reproducibility, resource-aware planning, multi-robot coordination, 3D navigation, and emerging foundation models (LLMs/VLMs) for high-level semantic navigation. Collectively, this review provides a consolidated reference and practical guidance for future AMR path-planning research. Full article
(This article belongs to the Section Sensors and Control in Robotics)
Show Figures

Figure 1

20 pages, 570 KB  
Review
Reinforcement Learning Techniques for the Flavor Problem in Particle Physics
by Alessio Giarnetti and Davide Meloni
Symmetry 2026, 18(1), 131; https://doi.org/10.3390/sym18010131 - 9 Jan 2026
Viewed by 164
Abstract
This short review discusses recent applications of Reinforcement Learning (RL) techniques to the flavor problem in particle physics. Traditional approaches to fermion masses and mixing often rely on extensions of the Standard Model based on horizontal symmetries, but the vast landscape of possible [...] Read more.
This short review discusses recent applications of Reinforcement Learning (RL) techniques to the flavor problem in particle physics. Traditional approaches to fermion masses and mixing often rely on extensions of the Standard Model based on horizontal symmetries, but the vast landscape of possible models makes systematic exploration infeasible. Recent works have shown that RL can efficiently navigate this landscape by constructing models that reproduce observed quark and lepton observables. These approaches demonstrate that RL not only rediscovers models already proposed in the literature but also uncovers new, phenomenologically acceptable solutions. Full article
(This article belongs to the Special Issue Neutrinos and Symmetry: Theoretical Developments and New Directions)
Show Figures

Figure 1

25 pages, 2211 KB  
Article
When Demand Uncertainty Occurs in Emergency Supplies Allocation: A Robust DRL Approach
by Weimeng Wang, Junchao Fan, Weiqiao Zhu, Yujing Cai, Yang Yang, Xuanming Zhang, Yingying Yao and Xiaolin Chang
Appl. Sci. 2026, 16(2), 581; https://doi.org/10.3390/app16020581 - 6 Jan 2026
Viewed by 194
Abstract
Emergency supplies allocation is a critical task in post-disaster response, as ineffective or delayed decisions can directly lead to increased human suffering and loss of life. In practice, emergency managers must make rapid allocation decisions over multiple periods under incomplete information and highly [...] Read more.
Emergency supplies allocation is a critical task in post-disaster response, as ineffective or delayed decisions can directly lead to increased human suffering and loss of life. In practice, emergency managers must make rapid allocation decisions over multiple periods under incomplete information and highly unpredictable demand, making robust and adaptive decision support essential. However, existing allocation approaches face several challenges: (1) Those traditional approaches rely heavily on predefined uncertainty sets or probabilistic models, and are inherently static, making them unsuitable for multi-period, dynamically allocation problems; and (2) while reinforcement learning (RL) technique is inherently suitable for dynamic decision-making, most existing RL-base approaches assume fixed demand, making them unable to cope with the non-stationary demand patterns seen in real disasters. To address these challenges, we first establish a multi-period and multi-objective emergency supplies allocation problem with demand uncertainty and then formulate it as a two-player zero-sum Markov game (TZMG). Demand uncertainty is modeled through an adversary rather than predefined uncertainty sets. We then propose RESA, a novel RL framework that uses adversarial training to learn robust allocation policies. In addition, RESA introduces a combinatorial action representation and reward clipping methods to handle high-dimensional allocations and nonlinear objectives. Building on RESA, we develop RESA_PPO by employing proximal policy optimization as its policy optimizer. Experiment results with realistic post-disaster data show that RESA_PPO achieves near-optimal performance, with an average gap of only 3.7% in terms of the objective value of the formulated problem, from the theoretical optimum derived by exact solvers. Moreover, RESA_PPO outperforms all baseline methods, including heuristic and standard RL methods, by at least 5.25% on average. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

58 pages, 4657 KB  
Review
Machine Learning for Energy Management in Buildings: A Systematic Review on Real-World Applications
by Panagiotis Michailidis, Federico Minelli, Iakovos Michailidis, Mehmet Kurucan, Hasan Huseyin Coban and Elias Kosmatopoulos
Energies 2026, 19(1), 219; https://doi.org/10.3390/en19010219 - 31 Dec 2025
Viewed by 473
Abstract
Machine learning (ML) is becoming a key enabler in building energy management systems (BEMS), yet most existing reviews focus on simulations and fail to reflect the realities of real-world deployment. In response to this limitation, the present work aims to present a systematic [...] Read more.
Machine learning (ML) is becoming a key enabler in building energy management systems (BEMS), yet most existing reviews focus on simulations and fail to reflect the realities of real-world deployment. In response to this limitation, the present work aims to present a systematic review dedicated entirely to experimental, field-tested applications of ML in BEMS, covering systems such as Heating, Ventilation & Air-conditioning (HVAC), Renewable Energy Systems (RES), Energy Storage Systems (ESS), Ground Heat Pumps (GHP), Domestic Hot Water (DHW), Electric Vehicle Charging (EVCS), and Lighting Systems (LS). A total of 73 real-world deployments are analyzed, featuring techniques like Model Predictive Control (MPC), Artificial Neural Networks (ANNs), Reinforcement Learning (RL), Fuzzy Logic Control (FLC), metaheuristics, and hybrid approaches. In order to cover both methodological and practical aspects, and properly identify trends and potential challenges in the field, current review uses a unified framework: On the methodological side, it examines key-attributes such as algorithm design, agent architectures, data requirements, baselines, and performance metrics. From a practical standpoint, the study focuses on building typologies, deployment architectures, zones scalability, climate, location, and experimental duration. In this context, the current effort offers a holistic overview of the scientific landscape, outlining key trends and challenges in real-world machine learning applications for BEMS research. By focusing exclusively on real-world implementations, this study offers an evidence-based understanding of the strengths, limitations, and future potential of ML in building energy control—providing actionable insights for researchers, practitioners, and policymakers working toward smarter, grid-responsive buildings. Findings reveal a maturing field with clear trends: MPC remains the most deployment-ready, ANNs provide efficient forecasting capabilities, RL is gaining traction through safer offline–online learning strategies, FLC offers simplicity and interpretability, and hybrid methods show strong performance in multi-energy setups. Full article
Show Figures

Figure 1

38 pages, 9342 KB  
Review
Monitoring and Control of the Direct Energy Deposition (DED) Additive Manufacturing Process Using Deep Learning Techniques: A Review
by Yonghui Liu, Haonan Ren, Qi Zhang, Peng Yuan, Hui Ma, Yanfeng Li, Yin Zhang and Jiawei Ning
Materials 2026, 19(1), 89; https://doi.org/10.3390/ma19010089 - 25 Dec 2025
Viewed by 566
Abstract
Directed Energy Deposition (DED), as a core branch of additive manufacturing, encompasses two typical processes: laser directed energy deposition (LDED) and wire and arc additive manufacturing (WAAM), which are widely used in manufacturing aerospace engine blades and core components of high-end equipment. In [...] Read more.
Directed Energy Deposition (DED), as a core branch of additive manufacturing, encompasses two typical processes: laser directed energy deposition (LDED) and wire and arc additive manufacturing (WAAM), which are widely used in manufacturing aerospace engine blades and core components of high-end equipment. In recent years, with the increasing adoption of deep learning (DL) technologies, the research focus in DED has gradually shifted from traditional “process parameter optimization” to “AI-driven process optimization” and “online real-time monitoring”. Given the complex and distinct influence mechanisms of key parameters (such as laser power/arc current, scanning/travel speed) on melt pool behavior and forming quality in the two processes, the introduction of artificial intelligence to address both common and specific issues has become particularly necessary. This review systematically summarizes the application of DL techniques in both types of DED processes. It begins by outlining DL frameworks, such as artificial neural networks (ANNs), recurrent neural networks (RNNs), convolutional neural networks (CNNs), and reinforcement learning (RL), and their compatibility with DED data. Subsequently, it compares the application scenarios, monitoring accuracy, and applicability of AI in DED process monitoring across multiple dimensions, including process parameters, optical, thermal fields, acoustic signals, and multi-sensor fusion. The review further explores the potential and value of DL in closed-loop parameter adjustment and reinforcement learning control. Finally, it addresses current bottlenecks such as data quality and model interpretability, and outlines future research directions, aiming to provide theoretical and engineering references for the intelligent upgrade and quality improvement of both DED processes. Full article
Show Figures

Graphical abstract

26 pages, 900 KB  
Article
Quality Management for AI-Generated Self-Adaptive Resource Controllers
by Claus Pahl, Hamid R. Barzegar and Nabil El Ioini
Machines 2026, 14(1), 25; https://doi.org/10.3390/machines14010025 - 24 Dec 2025
Viewed by 353
Abstract
Many complex systems requires the use of controllers to allow an automated, self-adaptive management of components and resources. Controllers are software components that observe a system, analyse its quality, and recommend and enact decisions to maintain or improve quality. While controllers have been [...] Read more.
Many complex systems requires the use of controllers to allow an automated, self-adaptive management of components and resources. Controllers are software components that observe a system, analyse its quality, and recommend and enact decisions to maintain or improve quality. While controllers have been for many years, recently Artificial Intelligence (AI) techniques such as Machine Learning (ML) and specifically reinforcement learning (RL) are used to construct these controllers, causing uncertainties about the quality of them due to their construction. We investigate quality metrics for RL-constructed software-based controllers that allow for their continuous quality control, which is particularly motivated by increasing automation and also the usage of artificial intelligence and control theoretic solutions for controller construction and operation. We introduce self-adaptation and control principles and define a quality-oriented controller reference architecture for controllers for self-adaptive systems. This forms the basis for the central contribution, a quality analysis metrics framework for controllers themselves. Full article
Show Figures

Figure 1

17 pages, 834 KB  
Article
Predefined-Time Tracking Control of Servo Hydraulic Cylinder Based on Reinforcement Learning
by Tao Han, Xiaohua Nie, Ninan Que, Jie Lu, Jianyong Yao and Xiaochuan Yu
Actuators 2026, 15(1), 9; https://doi.org/10.3390/act15010009 - 24 Dec 2025
Viewed by 233
Abstract
Electro-hydraulic servo systems are characterized by significant nonlinearities. Reinforcement learning (RL), known for its model-free nature and adaptive learning capabilities, presents a promising approach for handling uncertainties inherent in such systems. This paper proposes a predefined-time tracking control scheme based on RL, which [...] Read more.
Electro-hydraulic servo systems are characterized by significant nonlinearities. Reinforcement learning (RL), known for its model-free nature and adaptive learning capabilities, presents a promising approach for handling uncertainties inherent in such systems. This paper proposes a predefined-time tracking control scheme based on RL, which achieves fast and accurate tracking performance. The proposed design employs an actor–critic neural network strategy to actively compensate for system uncertainties. Within a conventional backstepping framework, a command-filtering technique is integrated to construct a predefined-time control structure. This not only circumvents the issue of differential explosion but also guarantees system convergence within a predefined time, which can be specified independently by the designer. Simulation results and comparisons validate the enhanced control performance of the proposed controller. Full article
Show Figures

Figure 1

32 pages, 8971 KB  
Systematic Review
Systematic Review of Reinforcement Learning in Process Industries: A Contextual and Taxonomic Approach
by Marco Antonio Paz Ramos and Axel Busboom
Appl. Sci. 2025, 15(24), 12904; https://doi.org/10.3390/app152412904 - 7 Dec 2025
Viewed by 1407
Abstract
The process industry (PI) plays a vital role in the global economy and faces mounting pressure to enhance sustainability, operational agility, and resource efficiency amid tightening regulatory and market demands. Although artificial intelligence (AI) has been explored in this domain for decades, its [...] Read more.
The process industry (PI) plays a vital role in the global economy and faces mounting pressure to enhance sustainability, operational agility, and resource efficiency amid tightening regulatory and market demands. Although artificial intelligence (AI) has been explored in this domain for decades, its adoption in industrial practice remains limited. Recently, machine learning (ML) has gained momentum, particularly when integrated with core PI systems such as process control, instrumentation, quality management, and enterprise platforms. Among ML techniques, reinforcement learning (RL) has emerged as a promising approach to tackle complex operational challenges. In contrast to conventional data-driven methods that focus on prediction or classification, RL directly addresses sequential decision making under uncertainty, a defining characteristic of dynamic process operations. Given RL’s growing relevance, this study conducts a systematic literature review to evaluate its current applications in the PI, assess methodological developments, and identify barriers to broader industrial adoption. The review follows the PRISMA methodology, a structured framework for identifying, screening, and selecting relevant publications. This approach ensures alignment with a clearly defined research question and minimizes bias, focusing on studies that demonstrate meaningful industrial applications of RL. The findings reveal that RL is transitioning from a theoretical construct to a practical tool, particularly in the chemical sector and for tasks such as process control and scheduling. Methodological maturity is improving, with algorithm selection increasingly tailored to problem-specific requirements and a trend toward hybrid models that integrate RL with established control strategies. However, most implementations remain confined to simulated environments, underscoring the need for real-world deployment, safety assurances, and improved interpretability. Overall, RL exhibits the potential to serve as a foundational component of next-generation smart manufacturing systems. Full article
Show Figures

Figure 1

23 pages, 1181 KB  
Article
Robust Regularized Recursive Least-Squares Algorithm Based on Third-Order Tensor Decomposition
by Radu-Andrei Otopeleanu, Constantin Paleologu, Jacob Benesty, Cristian-Lucian Stanciu, Laura-Maria Dogariu and Ruxandra-Liana Costea
Algorithms 2025, 18(12), 768; https://doi.org/10.3390/a18120768 - 5 Dec 2025
Viewed by 303
Abstract
The decomposition-based adaptive filtering algorithms have recently gained increasing interest due to their capability to reduce the parameter space. In this context, the third-order tensor (TOT) decomposition technique reformulates the conventional approach that involves a single (usually long) adaptive filter by using a [...] Read more.
The decomposition-based adaptive filtering algorithms have recently gained increasing interest due to their capability to reduce the parameter space. In this context, the third-order tensor (TOT) decomposition technique reformulates the conventional approach that involves a single (usually long) adaptive filter by using a combination of three shorter filters via the Kronecker product. This leads to a twofold gain in terms of both performance and complexity. Thus, it can be applied efficiently when operating with more complex algorithms, like the recursive least-squares (RLS) approach. In this paper, we develop an RLS-TOT algorithm with improved robustness features due to a novel regularization method that considers the contribution of the external noise and the so-called model uncertainties (which are related to the system). Simulation results obtained in the framework of echo cancelation support the performance of the proposed algorithm, which outperforms the existing RLS-TOT counterparts, as well as the conventional RLS algorithm that uses the specific regularization technique. Full article
(This article belongs to the Special Issue Recent Advances in Numerical Algorithms and Their Applications)
Show Figures

Figure 1

21 pages, 1500 KB  
Article
Intelligent Multi-Objective Path Planning for Unmanned Surface Vehicles via Deep and Fuzzy Reinforcement Learning
by Ioannis A. Bartsiokas, Charis Ntakolia, George Avdikos and Dimitris Lyridis
J. Mar. Sci. Eng. 2025, 13(12), 2285; https://doi.org/10.3390/jmse13122285 - 30 Nov 2025
Viewed by 575
Abstract
Unmanned Surface Vehicles (USVs) are increasingly employed in maritime operations requiring high levels of autonomy, safety, and energy efficiency. However, traditional path planning techniques struggle to simultaneously address multiple conflicting objectives such as fuel consumption, trajectory smoothness, and obstacle avoidance in dynamic maritime [...] Read more.
Unmanned Surface Vehicles (USVs) are increasingly employed in maritime operations requiring high levels of autonomy, safety, and energy efficiency. However, traditional path planning techniques struggle to simultaneously address multiple conflicting objectives such as fuel consumption, trajectory smoothness, and obstacle avoidance in dynamic maritime environments. To overcome these limitations, this paper introduces a Deep Q-Learning (DQN) framework and a novel Fuzzy Deep Q-Learning (F-DQN) algorithm that integrates Mamdani-type fuzzy reasoning into the reinforcement-learning (RL) reward model. The key contribution of the proposed approach lies in combining fuzzy inference with deep reinforcement learning (DRL) to achieve adaptive, interpretable, and multi-objective USV navigation—overcoming the fixed-weight reward limitations of existing DRL methods. The study develops a multi-objective reward formulation that jointly considers path deviation, curvature smoothness, and fuel consumption, and evaluates both algorithms in a simulation environment with varying obstacle densities. The results demonstrate that the proposed F-DQN model significantly improves trajectory optimality, convergence stability, and energy efficiency, achieving over 35% reduction in path length and approximately 70–80% lower fuel consumption compared with the baseline DQN, while maintaining comparable success rates. Overall, the findings highlight the effectiveness of fuzzy-augmented reinforcement learning in enabling efficient and interpretable autonomous maritime navigation. Full article
(This article belongs to the Special Issue Advanced Control Strategies for Autonomous Maritime Systems)
Show Figures

Figure 1

Back to TopTop