This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Open AccessArticle
A Reinforcement Learning Guided Oppositional Mountain Gazelle Optimizer for Time–Cost–Risk Trade-Off Optimization Problems
1
Civil Engineering Department, Karadeniz Technical University, 61080 Trabzon, Türkiye
2
Centre for Wireless Technology, CoE for Intelligent Network, Faculty of Artificial Intelligence & Engineering, Multimedia University, Persiaran Multimedia, Cyberjaya 63100, Selangor, Malaysia
3
Department of Computer Science and Engineering, Graphic Era Deemed to be University, Dehradun 248002, Uttarakhand, India
4
Faculty of Engineering, Technology and Built Environment, UCSI University, Cheras, Kuala Lumpur 56000, Malaysia
*
Author to whom correspondence should be addressed.
Buildings 2026, 16(1), 144; https://doi.org/10.3390/buildings16010144 (registering DOI)
Submission received: 22 November 2025
/
Revised: 24 December 2025
/
Accepted: 25 December 2025
/
Published: 28 December 2025
Abstract
Existing metaheuristic approaches often struggle to maintain an effective exploration–exploitation balance and are prone to premature convergence when addressing highly conflicting time–cost–safety–risk trade-off problems (TCSRTPs) under complex construction project constraints, which can adversely affect project productivity, safety, and the provision of decent jobs in the construction sector. To overcome these limitations, this study introduces a hybrid metaheuristic called the Q-Learning Inspired Mountain Gazelle Optimizer (QL-MGO) for solving multi-objective TCSRTPs in construction project management, supporting the delivery of resilient infrastructure and resilient building projects. QL-MGO enhances the original MGO by integrating Q-learning with an opposition-based learning strategy to improve the balance between exploration and exploitation while reducing computational effort and enhancing resource efficiency in construction scheduling. Each gazelle functions as an adaptive agent that learns effective search behaviors through a state–action–reward structure, thereby strengthening convergence stability and preserving solution diversity. A dynamic switching mechanism represents the core innovation of the proposed approach, enabling Q-learning to determine when opposition-based learning should be applied based on the performance history of the search process. The performance of QL-MGO is evaluated using 18- and 37-activity construction scheduling problems and compared with NDSII-MGO, NDSII-Jaya, NDSII-TLBO, the multi-objective genetic algorithm (MOGA), and NDSII-Rao-2. The results demonstrate that QL-MGO consistently generates superior Pareto fronts. For the 18-activity project, QL-MGO achieves the highest hypervolume (HV) value of 0.945 with a spread of 0.821, outperforming NDSII-Rao-2, MOGA, and NDSII-MGO. Similar results are observed for the 37-activity project, where QL-MGO attains the highest HV of 0.899 with a spread of 0.674, exceeding the performance of NDSII-Jaya, NDSII-TLBO, and NDSII-MGO. Overall, the integration of Q-learning significantly enhances the search capability of MGO, resulting in faster convergence, improved solution diversity, and more reliable multi-objective trade-off solutions. QL-MGO therefore serves as an effective and computationally efficient decision-support tool for construction scheduling that promotes safer, more reliable, and resource-efficient project delivery.
Share and Cite
MDPI and ACS Style
Eirgash, M.A.; Tiang, J.-J.; Ateş, B.; Sharma, A.; Lim, W.H.
A Reinforcement Learning Guided Oppositional Mountain Gazelle Optimizer for Time–Cost–Risk Trade-Off Optimization Problems. Buildings 2026, 16, 144.
https://doi.org/10.3390/buildings16010144
AMA Style
Eirgash MA, Tiang J-J, Ateş B, Sharma A, Lim WH.
A Reinforcement Learning Guided Oppositional Mountain Gazelle Optimizer for Time–Cost–Risk Trade-Off Optimization Problems. Buildings. 2026; 16(1):144.
https://doi.org/10.3390/buildings16010144
Chicago/Turabian Style
Eirgash, Mohammad Azim, Jun-Jiat Tiang, Bayram Ateş, Abhishek Sharma, and Wei Hong Lim.
2026. "A Reinforcement Learning Guided Oppositional Mountain Gazelle Optimizer for Time–Cost–Risk Trade-Off Optimization Problems" Buildings 16, no. 1: 144.
https://doi.org/10.3390/buildings16010144
APA Style
Eirgash, M. A., Tiang, J.-J., Ateş, B., Sharma, A., & Lim, W. H.
(2026). A Reinforcement Learning Guided Oppositional Mountain Gazelle Optimizer for Time–Cost–Risk Trade-Off Optimization Problems. Buildings, 16(1), 144.
https://doi.org/10.3390/buildings16010144
Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details
here.
Article Metrics
Article metric data becomes available approximately 24 hours after publication online.