Next Article in Journal
Identifying Priority Bird Habitats Through Seasonal Dynamics: An Integrated Habitat Suitability–Risk–Quality Framework
Previous Article in Journal
Interaction Between Ethical Leadership, Affective Commitment and Social Sustainability in Transition Economies: A Model Mediated by Ethical Climate and Moderated by Psychological Empowerment in the Colombian Electricity Sector
Previous Article in Special Issue
Logistics Sprawl and Urban Congestion Dynamics Toward Sustainability: A Logistic Regression and Random-Forest-Based Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dependent-Chance Goal Programming for Sustainable Supply Chain Design: A Reinforcement Learning-Enhanced Salp Swarm Approach

1
Laboratory of Engineering Sciences, Informatics, Logistics and Mathematics Department, National School of Applied Sciences, Ibn Tofail University, Kenitra 14000, Morocco
2
Laboratory of Engineering, Industrial Management and Innovation, Faculty of Sciences and Techniques, Hassan 1st University, Settat 26000, Morocco
3
Euromed University of Fes, UEMF, Fes 30000, Morocco
*
Author to whom correspondence should be addressed.
Sustainability 2025, 17(13), 6079; https://doi.org/10.3390/su17136079
Submission received: 26 May 2025 / Revised: 22 June 2025 / Accepted: 24 June 2025 / Published: 2 July 2025
(This article belongs to the Special Issue Sustainable Operations and Green Supply Chain)

Abstract

The Sustainable Supply Chain Network Design Problem (SSCNDP) is to determine the optimal network configuration and resource allocation that achieve the trade-off among economic, environmental, social, and resilience objectives. The Sustainable Supply Chain Network Design Problem (SSCNDP) involves determining the optimal network configuration and resource allocation that allows trade-off among economic, environmental, social, and resilience objectives. This paper addresses the SSCNDP under hybrid uncertainty, which combines objective randomness got from historical data, and subjective beliefs induced by expert judgment. Building on chance theory, we formulate a dependent-chance goal programming model that specifies target probability levels for achieving sustainability objectives and minimizes deviations from these targets using a lexicographic approach. To solve this complex optimization problem, we develop a hybrid intelligent algorithm that combines uncertain random simulation with Reinforcement Learning-enhanced Salp Swarm Optimization (RL-SSO). The proposed RL-SSO algorithm is benchmarked against standard metaheuristics—Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Differential Evolution (DE), and standard SSO, across diverse problem instances. Results show that our method consistently outperforms these techniques in both solution quality and computational efficiency. The paper concludes with managerial insights and discusses limitations and future research directions.

1. Introduction

Supply Chain Network Design (SCND) is a critical strategic decision that, when effectively implemented, enhances operational efficiency, cost competitiveness, and sustainability across the entire value network [1,2]. As global supply chains become increasingly complex and interconnected, they face escalating challenges, such as market volatility, regulatory pressures, environmental concerns, and disruptive events. These challenges demand simultaneous consideration of economic, environmental, social, and resilience objectives [3]. Recent disruptions, such as the COVID-19 pandemic, geopolitical tensions, and climate-related crises, have exposed the limitations of traditional deterministic planning methods and underscored the urgent need for robust supply chain designs that perform reliably under diverse and often conflicting sources of uncertainty [4].
Recent industry reports quantify the magnitude of these disruptions. According to McKinsey & Company [5], supply chain disruptions cost companies an average of 45% of one year’s profits per decade, translating to $150–300 million annually for Fortune 500 companies. Environmental impacts are equally significant: traditional supply chains account for 60% of global carbon emissions (approximately 36 billion tons CO2 annually), with inefficient network designs contributing an additional 15–20% excess emissions [6]. Social impacts include the loss of 2.7 million manufacturing jobs in developed economies between 2019–2023 due to inadequate supply chain resilience, while creating precarious employment conditions for 40 million workers in developing nations [7]. These quantitative indicators underscore the urgent need for innovative optimization approaches that can simultaneously address economic, environmental, and social objectives under uncertainty.
Real-world supply chains are characterized by hybrid uncertainty, which simultaneously encompasses both aleatory uncertainty (arising from inherent randomness such as demand fluctuations and lead times) and epistemic uncertainty (stemming from incomplete knowledge, expert opinions, or qualitative risk assessments) [8,9]. This dual nature of uncertainty is particularly evident in complex decision-making scenarios: demand forecasting in emerging markets relies on both historical sales data representing objective randomness and expert assessments of market trends reflecting subjective beliefs, while supplier reliability evaluation integrates quantitative performance metrics with qualitative judgments about management capability or geopolitical risk.
Despite the prevalence of hybrid uncertainty in practice, traditional optimization approaches address these uncertainty types in isolation, each with inherent limitations. Stochastic programming effectively captures aleatory uncertainty through probability distributions [10], while robust optimization accounts for epistemic uncertainty by considering worst-case scenarios [11], and fuzzy programming incorporates subjective judgments through membership functions [12]. However, these single-paradigm methods prove inadequate when confronted with hybrid uncertainty environments. Probability theory struggles to model epistemic components, as it requires well-defined distributions that may not exist for expert judgments and belief-based assessments. Conversely, fuzzy theory lacks the mathematical rigor necessary to effectively handle data-driven randomness and may yield suboptimal or counterintuitive outcomes [13]. These fundamental limitations underscore the critical need for a unified paradigm capable of rigorously representing and optimizing under both uncertainty types within a single coherent framework.
At the heart of this challenge lies the issue of hybrid uncertainty, which encompasses both aleatory and epistemic components. The traditional approaches to supply chain optimization deal with uncertainty types in isolation: stochastic programming captures randomness in demand [10], robust optimization accounts for worst-case scenarios [11], and fuzzy programming incorporates subjective judgments [12]. In fact, real-world supply chains commonly exhibit hybrid uncertainty, combining aleatory uncertainty (arising from inherent randomness, such as demand fluctuations and lead times) and epistemic uncertainty (stemming from incomplete knowledge, expert opinion, or qualitative risk assessments) [8,9]. For instance, demand forecasting in emerging markets often relies on both historical sales data (objective randomness) and expert assessments of market trends (subjective belief). Similarly, evaluation of supplier reliability integrates quantitative performance metrics with qualitative judgments about management capability or political risk.
Single-paradigm methods fall short in such settings or environments. Probability theory struggles to model epistemic uncertainty, as it requires well-defined distributions that may not exist for expert judgments and belief-based assessments. Conversely, fuzzy theory lacks the mathematical rigor to effectively handle data-driven randomness, and it is susceptible to result in suboptimal or counterintuitive outcomes [13]. These limitations highlight the need for a unified paradigm capable of rigorously representing and optimizing under both types of uncertainty using a single model.
Chance theory, established and improved by Liu [13,14], possesses precisely this mathematical foundation by unifying the probability measures and the uncertainty measures within a single chance measure framework. The innovation of the chance theory is that it can maintain the mathematical richness of probability theory for objective randomness and still provide a uniform representation of subjective beliefs, so that more realistic and reliable modeling of hybrid uncertain environments is possible. This satisfies the basic restriction of traditional approaches by recognizing that supply chain parameters such as demand forecasts, supplier capacities, and risk analysis entail data-driven randomness and expert judgment-based beliefs [15]. Despite its theoretical strength, the application of chance theory to multi-objective sustainable supply chain planning remains largely unexplored. This gap presents a significant opportunity to enhance both theoretical insights and practical decision-making tools in the field. A particularly promising direction is the use of dependent-chance modeling, which shifts the focus from merely ensuring constraint satisfaction under fixed confidence levels to maximizing the probability of achieving sustainability objectives. Unlike chance-constrained programming, where the focus is on guaranteeing feasibility under specified confidence levels, dependent-chance goal programming enables a more goal-oriented and interpretable decision-making framework. It allows direct optimization of probabilistic target achievements, thus supporting more transparent strategic planning and richer trade-off analyses across sustainability dimensions [16].
The complexity of dependent-chance programming formulations, particularly when involving multiple sustainability objectives, creates a high computational workload that exceeds the one of traditional exact optimization methods [17]. These problems often exhibit non-convex and multi-modal objective spaces, necessitating sophisticated optimization strategies capable of effectively balancing exploration and exploitation based on highly intricate objective functions involving probability maximization.
While metaheuristic algorithms have shown promise in addressing such complex optimization problems, they are often limited by premature convergence and inability to adjust their search strategy based on problem characteristics [18].
Salp Swarm Optimization (SSO), introduced by [19], is a metaheuristic that offers desirable features alongside continuous optimization through its bio-inspired mechanism of chain formation and nature-based equilibrium between exploration and exploitation phases. However, traditional SSO employs static parameter settings that may not be optimal at any stage of the optimization, particularly with the dynamic solution landscapes of dependent-chance problems, where the optimum search strategy might be very different for different stages of the optimization.
To overcome this limitation, the integration of reinforcement learning (RL) with metaheuristics offers a promising path forward. RL provides a mechanism for adaptive parameter tuning and allows the optimization process to learn effective search behaviors by interacting with the environment [20]. Crucially, RL recognizes that different stages of the optimization require different strategies, i.e., aggressive exploration in the early stages to locate promising regions, and focused exploitation in later stages to refine solutions. By formulating parameter selection as a sequential decision-making problem, RL can dynamically adjust the exploration-exploitation trade-off based on real-time search feedback, thus improving both convergence speed and solution quality in dependent-chance optimization tasks [21,22].
To address the interconnected challenges and knowledge gaps in sustainable supply chain optimization under hybrid uncertainty, this paper proposes a unified framework that combines dependent-chance goal programming with a reinforcement learning-enhanced Salp Swarm Optimization (RL-SSO) algorithm. Our integrated approach makes the following key contributions:
1.
Hybrid Uncertainty Modeling Framework: We develop a mathematically grounded framework based on chance theory that is able to tackle the coexistence of random uncertainty due to historical information and belief-based uncertainty due to experts’ judgment in sustainable supply chain network design. This approach avoids the pitfalls of misapplying single-paradigm methods to inherently hybrid supply chain environments.
2.
Dependent-Chance Goal Programming Model: We propose a multi-objective optimization model that leverages dependent-chance programming to maximize the probability of achieving sustainability targets across four dimensions: economic efficiency, environmental impact, social responsibility, and resilience. The model provides intuitive, probability-based performance metrics for decision-makers operating under hybrid uncertainty.
3.
Adaptive Metaheuristic Algorithm: We design a novel RL-enhanced Salp Swarm Optimization algorithm that dynamically adapts its search strategy using reinforcement learning. This enables efficient navigation of the complex, non-convex search space typical of dependent-chance problems, with more manageable, probability-based performance metrics for decision-makers under hybrid uncertainty, while improving convergence behavior and solution quality.
4.
Computational Validation and Analysis: We conduct large-scale computational experiments on varying size and uncertainty levels, comparing with state-of-the-art optimization methods to demonstrate superior performance in terms of solution quality, computational efficiency, and robustness to changing problem characteristics.
5.
Practical Decision Support Framework: We provide managerial insights and sensitivity analyses that translate theoretical contributions into operational guidance for supply chain practitioners, including probability threshold selection strategies, quantification of trade-offs between sustainability dimensions, and guidelines for implementing hybrid uncertainty assessment.
The remainder of this paper is organized as follows: Section 2 presents the mathematical formulation of our sustainable supply chain network design problem, as well as the dependent-chance goal programming approach to multi-objective optimization under hybrid uncertainty. Section 3 provides the theoretical foundations of uncertain random theory for hybrid uncertainty modeling and chance measure calculation. Section 4 illustrates the hybrid intelligent algorithm, which integrates uncertain random simulations and reinforcement learning-based Salp Swarm Optimization for solving the dependent-chance goal programming model. Section 5 presents comprehensive numerical results and analyses, including algorithm performance comparisons, sensitivity analyses, and network structure observations across a various test instances. Section 6 summarizes the main findings, theoretical contributions, managerial implications, limitations, and future research opportunities.

2. Mathematical Model Formulation

Our mathematical model for sustainable supply chain network design under uncertainty uses the following uncertain random variables to represent hybrid uncertainties to enable the simultaneous optimization of multiple sustainability objectives.

2.1. Sets and Indices

S = { 1 , 2 , . . . , s } Set of suppliers M = { 1 , 2 , . . . , m } Set of manufacturing plants W = { 1 , 2 , . . . , w } Set of warehouses D = { 1 , 2 , . . . , d } Set of distribution centers C = { 1 , 2 , . . . , c } Set of customers R = { 1 , 2 , . . . , r } Set of raw materials P = { 1 , 2 , . . . , p } Set of products T = { 1 , 2 , . . . , t } Set of time periods K = { 1 , 2 , . . . , k } Set of disruption scenarios E = { 1 , 2 , . . . , e } Set of environmental impact categories N = S M W D Set of all nodes in the network L = { 1 , 2 , . . . , l } Set of transportation modes G = { 1 , 2 , . . . , g } Set of sustainability goals

2.2. Uncertain Parameters

2.2.1. Economic Parameters

S C ˜ i j : Uncertain supply cost from supplier i to plant j P C ˜ j k : Uncertain production cost of product k at plant j T C ˜ i j k l : Uncertain transportation cost of product k from node i to node j using mode l I C ˜ i k : Uncertain inventory holding cost of product k at node i F C ˜ i : Uncertain fixed cost of opening facility at location i B C ˜ c k : Uncertain backorder cost for product k at customer c M C ˜ i t : Uncertain maintenance cost for facility at location i in period t R C ˜ i t : Uncertain recovery cost for facility at location i in period t

2.2.2. Capacity and Demand Parameters

S S ˜ i t : Uncertain supply capacity of supplier i in period t M P ˜ j t : Uncertain manufacturing capacity of plant j in period t W C ˜ w t : Uncertain warehouse capacity in period t D C ˜ d t : Uncertain distribution center capacity in period t D M ˜ c k t : Uncertain demand of customer c for product k in period t T M ˜ l t : Uncertain available capacity of transportation mode l in period t P C ˜ j k t m i n : Uncertain minimum production quantity of product k at plant j in period t P C ˜ j k t m a x : Uncertain maximum production quantity of product k at plant j in period t

2.2.3. Environmental Parameters

E C ˜ i j k l : Uncertain carbon emissions for transporting product k from node i to node j using mode l E P ˜ j k : Uncertain emissions from producing product k at plant j E W ˜ i j : Uncertain water consumption for production activity i at location j E N ˜ i j : Uncertain energy consumption for activity i at location j E L ˜ i j : Uncertain land use impact for activity i at location j E B ˜ j k : Uncertain biodiversity impact from producing product k at plant j E m C a p ˜ : Uncertain maximum allowable carbon emissions W C a p ˜ : Uncertain maximum allowable water consumption E n C a p ˜ : Uncertain maximum allowable energy consumption

2.2.4. Social Parameters

J C ˜ i : Uncertain job creation factor at location i C S ˜ i : Uncertain community support factor at location i H S ˜ i j : Uncertain health and safety risk factor for activity i at location j L C ˜ i : Uncertain local content percentage at location i W F ˜ i : Uncertain worker fairness index at location i S I ˜ i : Uncertain social inclusion factor at location i M i n J o b s ˜ : Uncertain minimum number of jobs to be created M i n C S ˜ : Uncertain minimum level of community support

2.2.5. Resilience and Reliability Parameters

ξ ˜ i j t : Uncertain supplier reliability from i to j in period t θ ˜ j k t : Uncertain production efficiency of product k at plant j in period t η ˜ i j k l t : Uncertain transportation reliability of product k from i to j using mode l in period t δ ˜ c k t : Uncertain demand variability factor of product k at customer c in period t ω ˜ i t : Uncertain facility disruption probability at location i in period t R R ˜ i k t : Uncertain recovery rate from disruption k at location i in period t R I ˜ i j : Uncertain resilience index of link between nodes i and j D S ˜ k t : Uncertain severity of disruption k in period t M i n R e s ˜ t : Uncertain minimum required resilience threshold in period t R T ˜ i : Uncertain recovery time for facility at location i R P ˜ i : Uncertain redundancy potential at location i A F ˜ i : Uncertain adaptability factor of facility at location i

2.2.6. Learning and Innovation Parameters

L R ˜ j k : Uncertain learning rate for production of product k at plant j I F ˜ j t : Uncertain innovation factor at plant j in period t K T ˜ i j : Uncertain knowledge transfer rate between facilities i and j C D ˜ j t : Uncertain capability development rate at plant j in period t

2.2.7. Goal Programming Parameters

T a r g e t ˜ g : Uncertain target threshold for goal g G α ˜ g : Uncertain confidence level for goal g G β ˜ c : Uncertain confidence level for constraint type c w ˜ g + : Uncertain weight for positive deviation from goal g w ˜ g : Uncertain weight for negative deviation from goal g

2.3. Decision Variables

2.3.1. Binary Decision Variables

Y i { 0 , 1 } 1 if facility is opened at location i , 0 otherwise X i j k l t { 0 , 1 } 1 if product k is transported from node i to node j using mode l in period t , 0 otherwise Z j k t { 0 , 1 } 1 if product k is produced at plant j in period t , 0 otherwise V i t { 0 , 1 } 1 if facility at location i is operational in period t , 0 otherwise U i j t { 0 , 1 } 1 if link between nodes i and j is established in period t , 0 otherwise W l t { 0 , 1 } 1 if transportation mode l is used in period t , 0 otherwise

2.3.2. Continuous Decision Variables

Q i j k l t 0 Quantity of product k transported from node i to node j using mode l in period t I i k t 0 Inventory level of product k at node i in period t P j k t 0 Production quantity of product k at plant j in period t B c k t 0 Backorder quantity of product k for customer c in period t d g 0 Negative deviation from target g ( for goal programming ) d g + 0 Positive deviation from target g ( for goal programming ) R i t 0 Resilience level of facility at location i in period t L j k t 0 Learning effect for product k at plant j in period t A i t 0 Adaptation level of facility at location i in period t S e t 0 Sustainability score for environmental category e in period t

2.4. Objective Functions

2.4.1. Economic Objective

The following objective function represents the total economic cost of the supply chain, which is to be minimized. It encompasses uncertain costs related to supply, production, transportation, inventory holding, facility establishment, backorders, maintenance, and recovery operations.
O F 1 = i S j M k P l L t T S C ˜ i j · Q i j k l t + j M k P t T P C ˜ j k · P j k t + i , j N k P l L t T T C ˜ i j k l · Q i j k l t + i N k P t T I C ˜ i k · I i k t + i N F C ˜ i · Y i + c C k P t T B C ˜ c k · B c k t + i N t T M C ˜ i t · V i t + i N t T R C ˜ i t · ( 1 V i t 1 ) · V i t

2.4.2. Environmental Objective

The following objective function quantifies the total environmental impact of the supply chain, which is to be minimized. It accounts for uncertain emissions from transportation and production, as well as water and energy consumption, land use, and biodiversity impact, while incorporating credits for implemented sustainability measures.
O F 2 = i , j N k P l L t T E C ˜ i j k l · Q i j k l t + j M k P t T E P ˜ j k · P j k t + i , j N E W ˜ i j · Y i + i , j N E N ˜ i j · Y i + i , j N E L ˜ i j · Y i + j M k P E B ˜ j k · Z j k t e E t T ϕ e t · S e t

2.4.3. Social Objective

Another important objective function is the following which represents social performance, formulated as a minimization problem where lower (more negative) values indicate better outcomes. It incorporates uncertain benefits such as job creation, community support, local content use, worker fairness, social inclusion, and resilience (all treated with negative signs), while health and safety risks are modeled as costs with positive contributions to the objective.
O F 3 = i N J C ˜ i · Y i i N C S ˜ i · Y i + i , j N H S ˜ i j · Y i i N L C ˜ i · Y i i N W F ˜ i · Y i i N S I ˜ i · Y i i N t T R i t · V i t

2.4.4. Resilience Objective

The final objective captures resilience performance, expressed as a minimization problem where more negative values reflect stronger resilience. It accounts for recovery capabilities, redundancy potential, and adaptability as benefits (negative contributions), while disruption severity is treated as a cost (positive contribution).
O F 4 = i N t T R i t i , j N t T R I ˜ i j · U i j t i N R P ˜ i · Y i i N t T A F ˜ i · A i t + k K t T D S ˜ k t · i N V i t

2.5. Constraints

2.5.1. Supply Capacity Constraints

C h j M k P l L Q i j k l t · ξ ˜ i j t S S ˜ i t · Y i · V i t β ˜ 1 , i S , t T
These constraints ensure that the total quantity shipped from each supplier does not exceed its uncertain capacity, adjusted by the supplier reliability factor, and only if the facility is both established and operational.

2.5.2. Manufacturing Capacity Constraints

C h k P P j k t · θ ˜ j k t M P ˜ j t · Y j · V j t β ˜ 2 , j M , t T
These constraints limit the total production at each manufacturing plant to its uncertain capacity, adjusted by the production efficiency factor, and only if the plant is both established and operational.

2.5.3. Warehouse Capacity Constraints

C h k P I w k t W C ˜ w t · Y w · V w t β ˜ 3 , w W , t T
These constraints ensure that the total inventory stored at each warehouse does not exceed its uncertain capacity, and only if the warehouse is both established and operational.

2.5.4. Distribution Center Capacity Constraints

C h k P I d k t D C ˜ d t · Y d · V d t β ˜ 4 , d D , t T
These constraints limit the total inventory stored at each distribution center to its uncertain capacity, and only if the distribution center is both established and operational.

2.5.5. Demand Satisfaction Constraints

C h d D l L Q d c k t l + B c k t 1 B c k t = D M ˜ c k t · δ ˜ c k t β ˜ 5 , c C , k P , t T
These constraints ensure that customer demand is satisfied through a combination of current period deliveries and backorders, considering uncertain demand levels adjusted by the demand variability factor.

2.5.6. Flow Conservation Constraints

C h I i k t 1 + j N l L Q j i k l t j N l L Q i j k l t I i k t = 0 β ˜ 6 , i N , k P , t T
These constraints maintain inventory balance at each node by ensuring that the ending inventory equals the beginning inventory plus inflows minus outflows.

2.5.7. Production-Transportation Linkage Constraints

C h P j k t i N l L Q j i k l t β ˜ 7 , j M , k P , t T
These constraints ensure that the quantity shipped from a manufacturing plant does not exceed its production quantity.
P j k t M b i g · Z j k t , j M , k P , t T
These constraints enforce that production can only occur if the binary production decision variable is activated, using a big-M approach where M b i g is a sufficiently large constant.
C h P j k t P C ˜ j k t m i n · Z j k t β ˜ 9 , j M , k P , t T
These constraints enforce minimum production quantities when production occurs, considering uncertain minimum production requirements.
C h P j k t P C ˜ j k t m a x · Z j k t β ˜ 10 , j M , k P , t T
These constraints enforce maximum production quantities when production occurs, considering uncertain maximum production capacities.

2.5.8. Raw Material Requirements Constraints

C h i S l L Q i j r l t k P γ ˜ r k · P j k t β ˜ 11 , j M , r R , t T
These constraints ensure that sufficient raw materials are supplied to meet production requirements, where γ ˜ r k represents the uncertain amount of raw material r needed to produce one unit of product k.

2.5.9. Adaptive Resilience Constraints

C h i , j N l L R I ˜ i j · Q i j k l t · ( 1 D S ˜ k t ) M i n R e s ˜ t β ˜ 12 , k K , t T
These constraints ensure that the network maintains a minimum resilience level under each disruption scenario, considering the uncertain resilience of links and the uncertain severity of disruptions. The constraint uses actual flow quantities rather than binary indicators to properly measure network resilience.
C h R i t k K ( 1 D S ˜ k t ) · R P ˜ i · A F ˜ i β ˜ 13 a , i N , t T
R i t M r e s · V i t , i N , t T
R i t 0 , i N , t T
These constraints calculate the resilience level of each facility based on disruption severity, redundancy potential, and adaptability factor. The first constraint handles the uncertain components, while the second constraint ensures resilience is only active when the facility is operational using a big-M formulation where M r e s is a sufficiently large constant.

2.5.10. Facility Adaptation Level Constraints

C h A i t 1 ( 1 D G ˜ i t ) A i t A i t 1 + ( 1 A i t 1 ) C D ˜ i t β ˜ 14 , i N , t T
These constraints model the facility adaptation level evolution over time with both development and degradation mechanisms. The adaptation level is bounded between a lower limit that accounts for potential degradation at rate D G ˜ i t and an upper limit that allows improvement based on the remaining adaptation capacity ( 1 A i t 1 ) and the uncertain capability development rate C D ˜ i t . This formulation ensures realistic bidirectional adaptation dynamics where facilities can both improve and deteriorate over time.

2.5.11. Recovery Rate Constraints

C h V i t V i t 1 · ( 1 ω ˜ i t ) + ( 1 V i t 1 ) · R R ˜ i k t β ˜ 15 , i N , k K , t T
These constraints model the facility operational status evolution over time. Operational facilities in period t 1 remain operational with probability ( 1 ω ˜ i t ) , while non-operational facilities recover with probability R R ˜ i k t .
C h V i t V i t 1 · ( 1 ω ˜ i t ) + ( 1 V i t 1 ) · R R ˜ i k t β ˜ 16 , i N , k K , t T
These constraints model the upper bound of facility operational status, ensuring that the operational level cannot exceed the combined effect of facilities remaining operational after avoiding disruption and facilities recovering from non-operational status.

2.5.12. Environmental Impact Constraints

C h i , j N k P l L t T E C ˜ i j k l · Q i j k l t + j M k P t T E P ˜ j k · P j k t E m C a p ˜ β ˜ 17
These constraints limit the total carbon emissions from transportation and production activities to comply with uncertain environmental regulations.
C h i , j N E W ˜ i j · Y i W C a p ˜ β ˜ 18
These constraints ensure that the total water consumption from established facilities does not exceed the uncertain allowable limit.
C h i , j N E N ˜ i j · Y i E n C a p ˜ β ˜ 19
These constraints ensure that the total energy consumption from established facilities does not exceed the uncertain allowable limit.

2.5.13. Social Impact Constraints

C h i N J C ˜ i · Y i M i n J o b s ˜ β ˜ 20
These constraints ensure that the supply chain creates at least the minimum required number of jobs, addressing social responsibility requirements.
C h i N C S ˜ i · Y i M i n C S ˜ β ˜ 21
These constraints ensure that the supply chain provides at least the minimum required level of community support, addressing social responsibility requirements.

2.5.14. Learning Effect Constraints

C h L j k t = L j k , t 1 + L R ˜ j k · P j k , t 1 β ˜ 22 , j M , k P , t T
These constraints model the incremental learning effect over time, where production efficiency improves based on previous period’s production experience and uncertain learning rate.
C h P j k t M P ˜ j t · ( 1 + L j k t ) · Z j k t β ˜ 23 , j M , k P , t T
These constraints increase production capacity based on the accumulated learning effect, capturing productivity improvements from experience.
C h P C ˜ j k t = P C ˜ j k 0 · ( 1 L R ˜ j k · ln ( 1 + L j k t ) ) β ˜ 24 , j M , k P , t T
These constraints model production cost reduction over time due to accumulated learning effects, capturing the cost-experience curve through logarithmic learning.

2.5.15. Non-Negativity and Binary Constraints

Q i j k l t , I i k t , P j k t , B c k t , R i t , L j k t , A i t , S e t , d g , d g + 0 , i , j N , k P , l L , t T , e E , g G Y i , X i j k l t , Z j k t , V i t , U i j t , W l t { 0 , 1 } , i , j N , k P , l L , t T
These constraints ensure that the decision variables take feasible values, with facility location and operational status variables being binary and all other variables being non-negative.

3. Uncertain Random Theory for Hybrid Uncertainty Modeling

In sustainable supply chain design, uncertainty arises from two distinct but often coexisting sources: aleatory uncertainty, which reflects inherent randomness (e.g., demand fluctuations, lead time variability), and epistemic uncertainty, which stems from incomplete knowledge and subjective expert judgments (e.g., risk perceptions, supplier reliability). Traditional modeling paradigms, such as probability theory for randomness or fuzzy theory for beliefs, are limited in isolation and cannot fully capture the complexity of such hybrid uncertainty environments. To address this, we adopt the framework of uncertain random theory that we formulate mathematically in this section.

3.1. Uncertain Random Variables

An uncertain random variable, as Liu [23] defines it, is a mathematical concept that captures the integration of uncertainty and randomness in a single paradigm. An uncertain random variable is a measurable function from a probability space to an uncertain variable set mathematically.
Definition 1. 
An uncertain random variable is a function ξ ˜ from a probability space ( Ω , A , P r ) to the set of uncertain variables such that M { ξ ˜ ( ω ) B } is a measurable function of ω for any Borel set B of R .
This definition captures the essential hybrid nature: for each realization ω of the random experiment, ξ ˜ ( ω ) is an uncertain variable representing expert beliefs or subjective assessments. The uncertain random variable thus combines:
  • Objective randomness: Modeled through the probability space ( Ω , A , P r )
  • Subjective uncertainty: Modeled through uncertain variables for each realization
Example: In supply chain demand forecasting, let D ˜ c k t represent the uncertain random demand of customer c for product k in period t. The randomness component captures historical demand variations, while the uncertainty component represents expert opinions about market conditions, consumer preferences, and external factors that cannot be quantified probabilistically.

3.2. Chance Measure

To measure uncertain random events, Liu [16] introduced the chance measure that combines probability measure and uncertain measure into a unified framework.
Definition 2. 
Let ξ ˜ be an uncertain random variable, and let B be a Borel set of R . Then the chance of uncertain random event ξ ˜ B is defined by:
C h { ξ ˜ B } = 0 1 P r { ω Ω M { ξ ˜ ( ω ) B } r } d r
The chance measure satisfies several important properties:
  • Normality: C h { ξ ˜ R } = 1
  • Self-duality: C h { ξ ˜ B } + C h { ξ ˜ B c } = 1
  • Monotonicity: If B 1 B 2 , then C h { ξ ˜ B 1 } C h { ξ ˜ B 2 }
Special Cases:
  • If ξ ˜ degenerates to a random variable X, then C h { ξ ˜ B } = P r { X B }
  • If ξ ˜ degenerates to an uncertain variable η , then C h { ξ ˜ B } = M { η B }

3.3. Chance Distribution

The chance distribution function provides a complete characterization of an uncertain random variable’s behavior under hybrid uncertainty.
Definition 3. 
Let ξ ˜ be an uncertain random variable. Then its chance distribution is defined by:
Φ ( x ) = C h { ξ ˜ x }
for any x R .
Theorem 1. 
A function Φ : R [ 0 , 1 ] is a chance distribution if and only if it is a monotone increasing function with Φ ( ) = 0 and Φ ( + ) = 1 .

3.4. Expected Value and Variance

For decision-making purposes, we need scalar measures to characterize uncertain random variables.
Definition 4. 
Let ξ ˜ be an uncertain random variable. Then its expected value is defined by:
E [ ξ ˜ ] = 0 + C h { ξ ˜ r } d r 0 C h { ξ ˜ r } d r
provided that at least one of the two integrals is finite.
Definition 5. 
Let ξ ˜ be an uncertain random variable with finite expected value e. Then the variance of ξ ˜ is defined by:
V a r [ ξ ˜ ] = E [ ( ξ ˜ e ) 2 ]
Theorem 2 
(Linearity of Expected Value). Let ξ ˜ be an uncertain random variable whose expected value exists. Then for any real numbers a and b:
E [ a ξ ˜ + b ] = a E [ ξ ˜ ] + b

3.5. Dependent-Chance Goal Programming

Dependent-chance programming, as developed in the chance theory framework, is an extremely powerful tool for hybrid uncertainty multi-objective optimization. Compared to traditional chance-constrained programming targeting constraint satisfaction at specified confidence levels, dependent-chance programming is optimized directly for the probability of occurrence of target events.
Definition 6. 
Let ξ ˜ be an uncertain random variable and F be a threshold value. The dependent-chance programming problem seeks to maximize the chance C h { ξ ˜ F } that the uncertain random variable achieves the specified target.
For multi-objective problems with conflicting objectives, dependent-chance goal programming extends this concept by incorporating goal programming methodology:
1.
Target Probability Specification: Decision-makers specify desired probability levels α g for achieving each objective g
2.
Deviation Variables: Positive and negative deviations ( d g + , d g ) capture under- and over-achievement of target probabilities
3.
Lexicographic Optimization: Objectives are prioritized and optimized sequentially according to their importance
The general dependent-chance goal programming model takes the form:
lexmin { d 1 , d 2 , , d n }
Subject to:
C h { f g ( x , ξ ˜ ) F g } + d g d g + = α g , g = 1 , 2 , , n
x X
d g + , d g 0 , g = 1 , 2 , , n
where f g ( x , ξ ˜ ) represents the g-th objective function with decision variables x and uncertain random parameters ξ ˜ , F g is the threshold value for objective g, α g is the target probability level, and X represents the feasible region.
Key Advantages:
  • Intuitive Interpretation: Probability-based objectives are easily understood by practitioners
  • Flexible Priority Setting: Lexicographic structure accommodates organizational priorities
  • Robust Performance: Focus on probability achievement provides inherent robustness against uncertainty
  • Hybrid Uncertainty Handling: Chance measure appropriately processes both random and belief-based uncertainties

3.6. Application to Supply Chain Parameters

In the context of sustainable supply chain network design, uncertain random variables provide natural representations for key parameters:
  • Demand Parameters: D M ˜ c k t combines historical demand patterns (randomness) with expert assessments of market trends (uncertainty)
  • Cost Parameters: S C ˜ i j , P C ˜ j k , T C ˜ i j k l incorporate both market price volatility and subjective cost estimates
  • Capacity Parameters: S S ˜ i t , M P ˜ j t reflect both operational variability and expert judgments about performance capabilities
  • Environmental Parameters: E C ˜ i j k l , E P ˜ j k combine measurable emissions data with uncertain regulatory and technological factors
  • Social Parameters: J C ˜ i , C S ˜ i integrate quantitative social indicators with qualitative community assessments
This hybrid modeling methodology allows greater realism in representing supply chain uncertainty with mathematical rigor preserved for optimization. Chance measure serves as the mathematical basis for problem formulation and solving optimization problems under hybrid uncertainty, which is established in the following sections.

3.7. Dependent-Chance Goal Programming Formulation of the SSCNDP

We consider the following priority structure among the four sustainability objectives (all formulated as minimization problems):
Priority 1: For the economic objective, the probability of the total cost being less than its threshold value F ˜ 1 should achieve α ˜ 1 (e.g., 0.95).
C h O F 1 F ˜ 1 + d 1 d 1 + = α ˜ 1
where the negative deviation between the target probability ( α ˜ 1 ) and the actually achieved probability, d 1 = α ˜ 1 C h O F 1 F ˜ 1 0 , is to be minimized.
Priority 2: For the environmental objective, the probability of the total environmental impact being less than its threshold value F ˜ 2 should achieve α ˜ 2 (e.g., 0.85).
C h O F 2 F ˜ 2 + d 2 d 2 + = α ˜ 2
where d 2 = α ˜ 2 C h O F 2 F ˜ 2 0 is to be minimized.
Priority 3: For the social objective, the probability of achieving better social performance (lower O F 3 values) than threshold value F ˜ 3 should achieve α ˜ 3 (e.g., 0.80).
C h O F 3 F ˜ 3 + d 3 d 3 + = α ˜ 3
where d 3 = α ˜ 3 C h O F 3 F ˜ 3 0 is to be minimized.
Priority 4: For the resilience objective, the probability of achieving better resilience performance (lower O F 4 values) than threshold value F ˜ 4 should achieve α ˜ 4 (e.g., 0.75).
C h O F 4 F ˜ 4 + d 4 d 4 + = α ˜ 4
where d 4 = α ˜ 4 C h O F 4 F ˜ 4 0 is to be minimized.
The DCGP model can then be formulated as the lexicographic optimization problem:
lexmin d 1 , d 2 , d 3 , d 4
Subject to:
C h O F 1 F ˜ 1 + d 1 d 1 + = α ˜ 1
C h O F 2 F ˜ 2 + d 2 d 2 + = α ˜ 2
C h O F 3 F ˜ 3 + d 3 d 3 + = α ˜ 3
C h O F 4 F ˜ 4 + d 4 d 4 + = α ˜ 4
d g + 0 , d g 0 , g = 1 , 2 , 3 , 4
This dependent-chance goal programming formulation lexicographically minimizes the negative deviations from the target probability levels for each sustainability objective. Since all objectives are formulated as minimization problems, C h { O F g F ˜ g } denotes the probability of achieving satisfactory performance, i.e., remaining below the defined threshold. The chance measure C h { · } handles both random and fuzzy uncertainties while providing intuitive probability interpretations for decision-makers.

3.8. Application to Sustainable Supply Chain Design

In sustainable supply chain network design, dependent chance goal programming with uncertain random variables offers several advantages:
  • It simultaneously addresses economic, environmental, social, and resilience objectives.
  • It handles hybrid uncertainties in parameters such as demand, costs, and disruption probabilities.
  • It allows decision-makers to specify confidence levels for each goal and constraint.
  • It provides a systematic approach to balancing multiple sustainability dimensions.
The proposed model employs dependent chance goal programming to order goals in consideration of the interaction between random and uncertain variables in supply chain planning. This makes solutions sustainable as well as robust to various sources of uncertainty.

4. Hybrid Intelligent Algorithm

To solve the complex dependent-chance goal programming model for sustainable supply chain planning, we propose a hybrid intelligent algorithm by integrating uncertain random simulations and a reinforcement learning-improved Salp Swarm Optimization method. In this section, simulation techniques and optimization methods used to deal with the hybrid uncertainties in our model are discussed in detail.

4.1. Uncertain Random Simulations

Our dependent-chance goal programming model contains uncertain random functions that need to be estimated: C h { g j ( x , ξ ) 0 } β ˜ j . To evaluate these expressions, we introduce uncertain random simulation techniques.

Simulation for Chance Measure

The chance measure C h { g ( x , ξ ) 0 } represents the expected value of the uncertain measure M { g ( x , ξ ( ω ) ) 0 } . Based on the Monte Carlo simulation principle, we design Algorithm 1 to estimate chance measures for uncertain random events.
The random simulation of uncertain chance measure estimation is at the core of our approach. It enables approximation of probabilistic constraints under hybrid uncertainty by iteratively sampling from the distribution of uncertain variables and generating Monte Carlo estimates of goal satisfaction. This simulation process is embedded within a hybrid optimization framework that leverages reinforcement learning to adaptively explore and refine supply chain configurations. The algorithm balances the exploration of new strategies with the exploitation of high-performing solutions, gradually improving decision quality under uncertainty.
This algorithm performs a Monte Carlo simulation to estimate how often uncertain goals are satisfied under both random and epistemic uncertainty. It repeatedly samples from known probability distributions or elicited beliefs about uncertain variables (e.g., demand, lead times) and evaluates whether each scenario satisfies predefined goals. The outcome is a probabilistic measure (chance estimate) for each goal, which quantifies the likelihood of its achievement given hybrid uncertainty. This forms the core engine for evaluating supply chain scenarios probabilistically.
Calculating uncertain measures means estimating constraints under uncertainty theory principles. Algorithm 2 uses the inverse distribution function approach to simulate realizations of uncertain variables and verify whether the constraint is satisfied for a given uncertainty level. So, this final algorithm integrates uncertain simulation outcomes into a goal programming model to prioritize decision criteria using probability-based targets.
Algorithm 1 Uncertain Random Simulation for Chance Measure
Description: This algorithm estimates the chance measure of an uncertain random event by using Monte Carlo sampling to approximate the expected value of uncertain measures across different random scenarios.
1:
Set c o u n t e r 0
2:
for  k = 1 to N M C  do
3:
   Generate random sample ω k according to its probability distribution
4:
   Calculate M { g ( x , ξ ( ω k ) ) 0 } using Algorithm 2
5:
    c o u n t e r c o u n t e r + M { g ( x , ξ ( ω k ) ) 0 }
6:
end for
7:
return  c o u n t e r / N M C as the estimated chance measure
Algorithm 2 Uncertain Measure Simulation
Description: This algorithm computes the uncertain measure for a given constraint by using the inverse distribution function method to generate uncertain variable realizations and evaluate constraint satisfaction.
1:
Generate α uniformly from [ 0 , 1 ]
2:
Calculate τ 1 ( α ) for uncertain variables τ in ξ ( ω k )
3:
if  g ( x , ξ ( ω k ) | τ = τ 1 ( α ) ) 0  then
4:
   return 1
5:
else
6:
   return 0
7:
end if

4.2. Reinforcement Learning-Enhanced Salp Swarm Optimization

To efficiently solve our chance-constrained goal programming model, we propose an enhanced version of the Salp Swarm Optimization (SSO) algorithm integrated with reinforcement learning techniques.

4.2.1. Introduction to Salp Swarm Optimizer

To efficiently solve our chance-constrained goal programming model, we propose an enhanced version of the SSO algorithm hybridized with reinforcement learning mechanisms. SSO is a metaheuristic algorithm introduced by [19] that is inspired by the swarming behavior of salps in oceans. Salps are oceanic creatures with a barrel-shaped body belonging to the family Salpidae and live in chains (salp chains) for effective locomotion and food hunting in deep oceans. The swarming behavior provides a nice mechanism for exploration and exploitation in optimization problems.
The key features that make SSO extremely suitable for sustainable supply chain optimization include:
1.
Leader-Follower Structure: SSO divides the population into two groups—a leader that guides the swarm and followers that follow each other (and the leader indirectly). This structure allows for effective exploration of the search space.
2.
Adaptive Coefficient: SSO employs an adaptive coefficient ( c 1 ) that balances exploration and exploitation during the optimization process:
c 1 = 2 e ( 4 l L ) 2
where l is the current iteration and L is the maximum number of iterations.
3.
Effective Position Updating: The leaders and followers’ updating mechanism enables thorough search of the search space in early iterations and gradual convergence towards good areas in later iterations.
4.
Simplicity and Flexibility: SSO requires very few parameters to be tuned and hence is easy to implement and parameterize across a wide range of optimization problems.
These characteristics make SSO particularly well-positioned to solve our opportunity-constrained goal programming problem, which features multi-objectives, complex constraints, and hybrid uncertainty.

4.2.2. Basic Salp Swarm Algorithm

The basic SSO algorithm operates as follows:
1.
Initialization: Generate N p solutions (salps) randomly in the feasible region.
2.
Population Division: Divide the salps into two groups—leader (first salp) and followers (remaining salps).
3.
Position Update: Update positions based on the following rules:
x j 1 = F j + c 1 ( ( u b j l b j ) c 2 + l b j ) if c 3 0 F j c 1 ( ( u b j l b j ) c 2 + l b j ) if c 3 < 0
for the leader, where x j 1 is the position of the leader in the j-th dimension, F j is the position of the food source (best solution so far), u b j and l b j are the upper and lower bounds of the j-th dimension, and c 2 and c 3 are random numbers in [ 0 , 1 ] .
For the followers ( i 2 ):
x j i = 1 2 ( x j i + x j i 1 )
where x j i shows the position of the i-th follower salp in the j-th dimension.
4.
Evaluation: Evaluate each salp using the objective function and update the food source if a better solution is found.
5.
Termination: Repeat steps 3–4 until a termination criterion is met.
To prepare the ground for our enhanced approach, we begin with the basic SSO algorithm. This basic variant shows the leader-follower mechanism and position update rules on which our reinforcement learning extensions are founded.

4.2.3. Reinforcement Learning Enhancement

We complement the core SSO algorithm with reinforcement learning (RL) techniques to improve its convergence performance and solution quality. The RL component includes:
1.
State Representation: The state consists of discretized optimization metrics:
s t = { d i v e r s i t y d , c o n v e r g e n c e _ r a t e d , s t a g n a t i o n d , p r o g r e s s d }
where each continuous metric is discretized into three levels (Low, Medium, High) using fixed thresholds based on empirical observations.
2.
Action Space: The actions involve adjusting algorithm parameters c 1 , c 2 , and c 3 , as well as selecting different search strategies.
3.
Reward Function: The reward is defined as the normalized improvement in solution quality:
R t = f ( x g b e s t t 1 ) f ( x g b e s t t ) | f ( x g b e s t t 1 ) | + ϵ if f ( x g b e s t t 1 ) 0 f i n i t i a l f ( x g b e s t t ) f i n i t i a l + ϵ if f ( x g b e s t t 1 ) = 0
where ϵ = 10 8 prevents division by zero and f i n i t i a l is the initial objective value.
4.
Q-Learning Update: We employ Q-learning with epsilon-greedy action selection:
Q ( s t , a t ) Q ( s t , a t ) + α [ R t + γ max a Q ( s t + 1 , a ) Q ( s t , a t ) ]
where α is the learning rate, γ is the discount factor, s t is the current state, and a t is the action taken.

4.2.4. Adaptive Parameter Tuning

To further enhance the algorithm’s performance, we implement adaptive parameter tuning mechanisms:
1.
Inertia Weight Adaptation:
c 1 t + 1 = c 1 m i n + ( c 1 m a x c 1 m i n ) · e ρ · t / T m a x
where ρ is a decay parameter, t is the current iteration, and T m a x is the maximum number of iterations.
2.
Cognitive and Social Parameter Adjustment:
c 2 t + 1 = c 2 m i n + ( c 2 m a x c 2 m i n ) · ( 1 e ρ · t / T m a x )
c 3 t + 1 = c 3 m a x ( c 3 m a x c 3 m i n ) · ( t / T m a x ) 2
3.
Exploration-Exploitation Balance:
ϵ t + 1 = max ( ϵ m i n , ϵ m a x · e σ · t / T m a x )
where ϵ represents the epsilon-greedy exploration probability and σ is a control parameter.

4.3. Reinforcement Learning Components

Based on the basic SSO framework, we integrate reinforcement learning techniques to develop an adaptive search mechanism. The RL component includes state representation, action space definition, and adaptive parameter adjustment.

4.3.1. State Space Discretization

The continuous optimization metrics are discretized using the following strategy:
1.
Diversity Measure:
d i v e r s i t y d = Low if d i v e r s i t y 0.3 Medium if 0.3 < d i v e r s i t y 0.7 High if d i v e r s i t y > 0.7
2.
Convergence Rate:
c o n v e r g e n c e _ r a t e d = Slow if r a t e 0.1 Moderate if 0.1 < r a t e 0.5 Fast if r a t e > 0.5
3.
Stagnation Counter:
s t a g n a t i o n d = Active if s t a g n a t i o n 10 Moderate if 10 < s t a g n a t i o n 50 Stagnant if s t a g n a t i o n > 50

4.3.2. Q-Learning for Parameter Adaptation

We apply Q-learning to dynamically adjust algorithm parameters based on optimization progress: Q-learning enables the algorithm to acquire appropriate parameter settings by trial and error, using temporal difference updates to adjust subsequent decision-making. The process allows the optimization process to adapt its behavior based on observed performance improvements. This algorithm uses reinforcement learning to iteratively improve decision strategies by interacting with the uncertain environment. It treats each possible supply chain configuration as an action and evaluates its performance using the simulation engine from Algorithm 1. Over time, the algorithm learns which decisions lead to better goal satisfaction under uncertainty. It balances exploration of new strategies with exploitation of known good ones, refining policy parameters (e.g., costs, inventory levels, routing) to adaptively improve performance using Algorithm 3 for parameter adaptation.
Algorithm 3 Q-Learning Parameter Adaptation
Description: This algorithm implements temporal difference learning to update Q-values for parameter adaptation, enabling the system to learn optimal parameter settings based on optimization performance feedback.
Require: 
Current state s t , action a t , reward R t , next state s t + 1
Require: 
Learning rate α , discount factor γ , Q-table Q, exploration rate ϵ
1:
Calculate temporal difference: δ = R t + γ max a Q ( s t + 1 , a ) Q ( s t , a t )
2:
Update Q-value: Q ( s t , a t ) Q ( s t , a t ) + α · δ
3:
Select next action using epsilon-greedy:
4:
if  r a n d o m ( ) < ϵ  then
5:
    a t + 1 random action
6:
else
7:
    a t + 1 arg max a Q ( s t + 1 , a )
8:
end if
9:
return Updated Q-table and next action a t + 1

4.3.3. Adaptive Parameter Control

Algorithm 4 uses reinforcement learning to dynamically adjust parameters of the stochastic swarm optimization (SSO) algorithm during execution. It maps the state of the optimization process, defined by metrics such as diversity, convergence rate, stagnation, and progress, into one the discrete states. Based on the current state, it selects one of four predefined actions (e.g., boost exploration or intensify local search) using an epsilon-greedy strategy. The adjustment modifies the key coefficients ( c 1 , c 2 , c 3 ) that influence the search behavior, with the goal of balancing exploration and exploitation. The learning process is guided by a reward function based on normalized improvement, ensuring continual adaptation as the search progresses. So, the reinforcement learning component includes:
1.
State Representation: s t = { d i v e r s i t y d , c o n v e r g e n c e _ r a t e d , s t a g n a t i o n d , p r o g r e s s d } with 3 4 = 81 possible discrete states
2.
Action Space: Four discrete actions for parameter adjustment:
  • Action 1: Increase exploration (boost c 1 , broaden search)
  • Action 2: Increase exploitation (reduce c 1 , local search)
  • Action 3: Balanced search (moderate parameters)
  • Action 4: Intensify local search (minimize c 1 )
3.
Reward Function: Normalized improvement with division-by-zero protection
4.
Action Selection: Epsilon-greedy strategy with decaying exploration rate
Algorithm 4 RL-Based Adaptive Parameter Control
Description: This algorithm optimally adjusts SSO parameters as per actions of reinforcement learning, balancing exploration and exploitation as per the existing optimization scenario.
Require: 
Current iteration t, RL action a t , base parameters
1:
Calculate base coefficient: c 1 b a s e = 2 e ( 4 t T m a x ) 2
2:
if  a t = 1  then
3:
    c 1 = c 1 b a s e × 1.3 , c 2 = 0.8 , c 3 = 0.7 {Increase Exploration}
4:
else if  a t = 2  then
5:
    c 1 = c 1 b a s e × 0.7 , c 2 = 0.3 , c 3 = 0.2 {Increase Exploitation}
6:
else if  a t = 3  then
7:
    c 1 = c 1 b a s e , c 2 = 0.5 , c 3 = 0.5 {Balanced Search}
8:
else
9:
    c 1 = c 1 b a s e × 0.5 , c 2 = 0.1 , c 3 = 0.1 {Intensify Local Search}
10:
end if
11:
Update exploration rate: ϵ t + 1 = max ( ϵ m i n , ϵ m a x · e σ · t / T m a x )
12:
return Updated parameters ( c 1 , c 2 , c 3 , ϵ )

4.4. Solution Encoding and Decoding

A stable representation of the solution is necessary to deal with our sustainable supply chain design problem’s complexity. We employ a multi-level encoding scheme and efficiently encode binary and continuous decision variables.

Variable Encoding Strategy

For binary decision variables ( Y i , X i j k l t , Z j k t , V i t , U i j t , W l t ), we implement a sigmoid-based transformation:
Y i = 1 , if σ ( z i ) 0.5 0 , otherwise
where σ ( z ) = 1 1 + e z is the sigmoid function.
For continuous decision variables ( Q i j k l t , I i k t , P j k t , B c k t , R i t , L j k t , A i t , S e t ), we employ direct encoding with bound constraints:
l b v a r v a r u b v a r
To handle hierarchical dependencies between decision variables:
P j k t = P j k t * , if Y j = 1 and Z j k t = 1 0 , otherwise
Solution encoding and decoding are critical in handling the mixed-integer nature of our optimization problem. Algorithm 5 transforms continuous-valued solution vectors—produced by the optimization process—into feasible mixed-integer solutions for the supply chain decision problem. It first decodes binary decision variables using a sigmoid activation followed by thresholding, effectively mapping real-valued elements into binary form (0 or 1). Then, it rescales the remaining continuous variables to lie within their specified bounds using normalization. Finally, the algorithm enforces hierarchical constraints: if a parent binary variable is inactive (0), any associated dependent variables are automatically set to zero. This ensures logical consistency between decisions, such as only allocating resources to facilities that are selected for activation.
Algorithm 5 Solution Encoding and Decoding
Description: This algorithm transforms continuous solution vectors into mixed-integer solutions by applying sigmoid functions for binary variables and enforcing hierarchical dependencies between decision variables.
Require: 
Continuous solution vector z
1:
Binary Variable Decoding:
2:
for each binary variable Y i  do
3:
    Y i = I [ σ ( z i ) 0.5 ] where I [ · ] is indicator function
4:
end for
5:
Continuous Variable Decoding:
6:
for each continuous variable v a r  do
7:
    v a r = l b v a r + ( u b v a r l b v a r ) · normalize ( z v a r )
8:
end for
9:
Hierarchical Dependency Enforcement:
10:
for each dependent variable pair do
11:
   if parent binary variable is 0 then
12:
     Set dependent variables to 0
13:
   end if
14:
end for
15:
return Decoded solution x

4.5. Constraint Handling Mechanisms

We employ a hierarchical penalty function approach to handle the dependent-chance constraints:
F ( x ) = M · j = 1 m λ j ( t ) · max ( 0 , β ˜ j C h { g j ( x , ξ ) 0 } ) + g = 1 4 w g · d g
where M = 10 6 × max g w g is a large penalty multiplier ensuring constraint violations dominate goal deviations, and λ j ( t ) are penalty parameters that increase over iterations:
λ j ( t ) = λ j 0 · ( 1 + δ · t )
For dependent-chance goals requiring high confidence levels, we implement progressive tightening:
β ˜ j ( t ) = β ˜ j f i n a l · ( 1 e κ · t / T m a x )
Effectively handling constraints in chance-constrained optimization demands advanced penalty mechanisms capable of managing probabilistic feasibility while preserving lexicographic priority among objectives. The proposed Algorithm 6 evaluates solution quality by imposing severe penalties for constraint violations. This ensures that feasible solutions are consistently favored over infeasible ones. Additionally, it employs a progressive tightening strategy, gradually increasing the stringency of constraint satisfaction requirements as the optimization advances.
Algorithm 6 Dynamic Penalty Constraint Handling
Description: This algorithm evaluates solution fitness in terms of dependent-chance objectives and constraints, applying hierarchical penalty functions to control constraint violation while retaining lexicographic priority, and iteratively increased constraint requirements.
Require: 
Solution x , iteration t, penalty parameters
1:
Evaluate Dependent-Chance Goals:
2:
for each goal g = 1 to 4 do
3:
   Calculate C h { O F g ( x , ξ ) F ˜ g } using Algorithm 1
4:
   Calculate deviation: d g = max ( 0 , α ˜ g C h { O F g F ˜ g } )
5:
end for
6:
Evaluate Chance Constraints:
7:
for each constraint j = 1 to m do
8:
   Calculate C h { g j ( x , ξ ) 0 } using Algorithm 1
9:
   Calculate violation: v i o l j = max ( 0 , β ˜ j C h { g j ( x , ξ ) 0 } )
10:
end for
11:
Compute Hierarchical Penalized Fitness:
12:
F p e n a l t y ( x ) = M · j = 1 m λ j ( t ) · v i o l j + g = 1 4 w g · d g
13:
Update Penalty Parameters:
14:
for each constraint j do
15:
    λ j ( t ) = λ j 0 · ( 1 + δ · t )
16:
end for
17:
return Penalized fitness F p e n a l t y ( x )

4.6. Complete Hybrid Intelligent Algorithm

The whole hybrid intelligent algorithm is proposed in a modular design with five focused elements, followed by the core execution framework. Proper initialization is crucial to the success of any metaheuristic algorithm. Algorithm 7 initializes the population, establishes all learning components, and applies the parameter structure which will execute the entire process of optimization.
Algorithm 7 Algorithm Initialization
Description: This algorithm initializes the population, Q-learning components, and all algorithm parameters, establishing the foundation for the hybrid optimization process.
1:
Initialize population of N p solutions randomly in feasible region
2:
Initialize Q-table for reinforcement learning with small random values
3:
Set RL parameters: α = 0.1 , γ = 0.9 , ϵ s t a r t = 0.9 , ϵ e n d = 0.1
4:
Set SSO parameters: c 1 m i n = 0.1 , c 1 m a x = 2.0 , ρ = 2.0 , σ = 1.5
5:
Set penalty parameters: λ j 0 = 1.0 , δ = 0.1 , κ = 2.0
6:
Evaluate initial solutions using Algorithm 6
7:
Identify global best solution x g b e s t and initialize f b e s t
8:
Initialize stagnation counter: s t a g = 0
Effective reinforcement learning requires continual monitoring of the optimization state and intelligent action selection. Algorithm 8 tracks key performance indicators, maintains balance in exploration-exploitation through epsilon-greedy selection, and triggers appropriate parameter updating.
Algorithm 8 RL State Observation and Action Selection
Description: This algorithm tracks the current optimization state, calculates key performance metrics, and selects actions according to epsilon-greedy strategy to balance exploration and exploitation when updating parameters.
1:
Calculate population diversity: d i v e r s i t y = 1 N p i = 1 N p | | x i x g b e s t | |
2:
Calculate convergence rate: c o n v r a t e = f b e s t t 5 f b e s t t 5 (if t > 5 )
3:
Define state: s t = { d i v e r s i t y , c o n v r a t e , s t a g , t / T m a x }
4:
Update ϵ : ϵ = max ( ϵ e n d , ϵ s t a r t · 0 . 995 t )
5:
if random() < ϵ  then
6:
   Select random action a t { 1 , 2 , 3 , 4 }
7:
else
8:
    a t = arg max a Q ( s t , a )
9:
end if
10:
Update SSO parameters using Algorithm 4
The core of the Salp Swarm Optimization lies in its position update mechanism, which mimics the natural chain formation behavior of salps. So, Algorithm 9 updates the positions of individual salps (candidate solutions) in a swarm based on a leader–follower structure inspired by salp chain behavior in nature. The first salp (leader) explores the search space around the food source (i.e., the best-known solution) by generating randomized movements influenced by a time-dependent coefficient and uniform noise. Subsequent salps (followers) update their positions by averaging their current location with that of the salp ahead of them, resulting in a smooth, chain-like convergence. After updating positions, each salp’s vector is clipped to problem boundaries, decoded into a feasible mixed-integer solution (using the encoding algorithm), and evaluated for fitness under constraints. This mechanism enables both exploration and convergence in the stochastic swarm optimization framework.
Algorithm 9 Salp Swarm Position Update
Description: This algorithm updates the positions of all salps in the swarm using the leader-follower mechanism, where the leader explores around the food source and followers follow their predecessors in the chain.
1:
for  i = 1  to  N p  do
2:
   if  i = = 1  then {Leader salp}
3:
      for  j = 1  to D do
4:
         Generate c 2 , c 3 U ( 0 , 1 )
5:
         if  c 3 0  then
6:
             x j 1 = F j + c 1 ( ( u b j l b j ) c 2 + l b j )
7:
         else
8:
             x j 1 = F j c 1 ( ( u b j l b j ) c 2 + l b j )
9:
         end if
10:
     end for
11:
   else {Follower salps}
12:
      for  j = 1  to D do
13:
          x j i = 1 2 ( x j i + x j i 1 )
14:
      end for
15:
   end if
16:
   Apply boundary constraints: x j i = max ( l b j , min ( u b j , x j i ) )
17:
   Decode solution using Algorithm 5
18:
   Evaluate fitness using Algorithm 6
19:
end for
In our hybrid approach, solution refinement and learning are tightly integrated. The algorithm continuously updates the global best solution, quantifies performance gains through a defined reward mechanism, and leverages these insights to guide the Q-learning process for adaptive parameter tuning. Algorithm 10 implements this integrated approach for solution update and reinforcement learning.
Algorithm 10 Solution Update and RL Learning
Description: This algorithm updates the global best solution, calculates rewards based on fitness improvement, and performs Q-learning updates to enhance future parameter selection decisions.
1:
f p r e v = f b e s t
2:
for each salp i do
3:
   if  f ( x i ) < f ( x g b e s t )  then
4:
        x g b e s t = x i , f b e s t = f ( x i )
5:
   end if
6:
end for
7:
Calculate reward: R t = f p r e v f b e s t f p r e v + 10 8
8:
Observe next state s t + 1
9:
Update Q-table using Algorithm 3
Managing algorithm convergence and preventing premature stagnation are critical for solution quality. Algorithm 11 oversees the optimization process by detecting stagnation, adapting constraints over time, and determining when to terminate the search. It tracks the best objective value across iterations and increments a stagnation counter when no improvement occurs. If the algorithm detects prolonged stagnation (exceeding 15% of the total iterations), it partially reinitializes the population by randomly resetting the worst-performing 30% of solutions, thus restoring diversity. In parallel, the chance constraints are gradually tightened using an exponential schedule to enforce stricter feasibility over time. Termination occurs either when convergence criteria are satisfied or the maximum number of iterations is reached, at which point the best solution is re-evaluated with high-precision Monte Carlo simulation to ensure goal satisfaction and constraint compliance.
Algorithm 11 Stagnation Management and Termination
Description: This algorithm monitors optimization progress, handles population stagnation through partial reinitialization, progressively tightens constraints, and checks termination criteria for convergence.
1:
if  f b e s t = = f p r e v  then
2:
    s t a g = s t a g + 1
3:
else
4:
    s t a g = 0
5:
end if
6:
if  s t a g > 0.15 × T m a x  then
7:
   Reinitialize worst 30 % of population randomly
8:
   Reset: s t a g = 0
9:
end if
10:
for each chance constraint j do
11:
    β ˜ j ( t ) = β ˜ j f i n a l · ( 1 e κ · t / T m a x )
12:
end for
13:
if convergence criteria met OR t = = T m a x  then
14:
   Evaluate x g b e s t with high-precision chance simulation ( N M C = 5000 )
15:
   Verify constraint satisfaction and goal achievement
16:
   return True {Termination signal}
17:
end if
18:
return False {Continue optimization}
The main execution framework orchestrates all algorithm components in a coordinated manner. Algorithm 12 provides the high-level control structure that manages the iterative optimization process, coordinates the interaction between SSO and RL components, and ensures proper termination.
Algorithm 12 Main RL-Enhanced SSO Execution Framework
Description: This is the main execution framework that orchestrates all algorithm components, coordinating the reinforcement learning-enhanced SSO process from initialization to termination.
1:
Execute Algorithm 7 for system initialization
2:
for  t = 1 to T m a x  do
3:
   Execute Algorithm 8 for RL state observation and action selection
4:
   Execute Algorithm 9 for salp population position updates
5:
   Execute Algorithm 10 for solution updates and RL learning
6:
   Execute Algorithm 11 for stagnation management and termination check
7:
   if Algorithm 11 returns True then
8:
        break {Optimization completed}
9:
   end if
10:
end for
11:
return Optimal solution x g b e s t with performance metrics
The explained methodology ensures development of a hybrid decision support framework that combines simulation under uncertainty, reinforcement learning for adaptive planning, and chance-constrained goal programming for multi-objective optimization. The approach is designed to operate in hybrid uncertain environments where both data-driven and expert-based information guide decision-making. This modular structure ensures the model is scalable, interpretable, and adaptable to various supply chain settings.

5. Numerical Results and Analysis

This section presents comprehensive numerical experiments to evaluate the performance of our proposed reinforcement learning-enhanced Salp Swarm Optimization (RL-SSO) algorithm for solving the dependent-chance goal programming model for sustainable supply chain design under hybrid uncertainty.

5.1. Experimental Setup

5.1.1. Algorithm Parameter Settings

Algorithm parameters were determined through systematic preliminary experiments using the Taguchi method for parameter optimization. The final RL-SSO algorithm parameters are presented in Table 1.

5.1.2. Test Instance Characteristics

To evaluate the algorithm comprehensively, we generated a diverse set of test instances with varying network sizes, complexity levels, and uncertainty characteristics. Table 2 summarizes the key characteristics of these instances.
These examples cover supply chain networks with diverse complexity, from small regional networks to extremely large international supply chains. Each example includes particular features corresponding to different operation contexts:
  • Small instances: Regional supply chains with short geographical range
  • Medium instances: National-level supply chains with average complexity
  • Large instances: Multi-national supply chains with high integration needs
  • Very Large instance: Global supply chain network with highest complexity

5.1.3. Uncertain Random Parameter Specifications

The hybrid uncertainty in our model is represented by uncertain random parameters with different distributions. The uncertain random parameter configurations used in our numerical experiments are illustrated in Table 3.
The hybrid uncertain random parameters were derived according to a systematic two-stage calibration process:
1.
Random Component Calibration: The historical records of three major manufacturing companies in the automotive, electronics, and drug industries were taken into account to determine appropriate probability distributions and their parameters.
2.
Uncertain Component Calibration: Expert elicitation of 15 supply chain experts was conducted to identify uncertainty bounds representing epistemic uncertainty in parameter estimation.

5.1.4. Computational Environment

All experiments were conducted on a personal computer equipped with Intel Core i9 processor and implemented in MATLAB R2023b. To ensure the statistical reliability of the results, each test instance was independently executed 30 times with varying random seeds.

5.2. Dependent-Chance Goal Programming Results

The core contribution of our approach lies in optimizing probability achievement for sustainability goals through lexicographic dependent-chance programming. We define target probability levels based on industry benchmarks and regulatory requirements: Economic ( α ˜ 1 = 0.95 ), Environmental ( α ˜ 2 = 0.85 ), Social ( α ˜ 3 = 0.80 ), and Resilience ( α ˜ 4 = 0.75 ).

5.2.1. Probability Achievement Analysis

Table 4 presents the achieved probability levels for each sustainability dimension across all test instances, with 95% confidence intervals computed using bootstrap resampling (1000 replications).
The results reveal a clear hierarchy in probability achievement, with economic objectives maintaining the highest achievement rates. The gaps from targets are statistically significant (paired t-test, p < 0.01 for all objectives), confirming the challenge of simultaneously achieving all sustainability goals.

5.2.2. Lexicographic Deviation Analysis

The lexicographic structure ensures higher-priority objectives are optimized first. Table 5 shows the negative deviation values and their statistical significance.
Friedman test confirms significant differences among priority levels ( χ 2 = 21.84, p < 0.001), validating the lexicographic optimization effectiveness.

5.3. Performance Justification Analysis

While the improvements of 0.5–5.11% may appear modest, they are economically and operationally significant in the context of large-scale supply chains. Table 6 demonstrates the substantial financial impact of these performance improvements across different company scales:
Additionally, the improvements in constraint satisfaction rates (13.3% for hybrid vs. pure uncertain) translate to significantly reduced risk of supply chain disruptions, which can cost 5–20% of annual revenue according to industry studies.

5.4. Chance Constraint Satisfaction Analysis

Our model incorporates multiple types of chance constraints with different confidence levels β ˜ j . We present detailed analysis with statistical tests for constraint satisfaction rates. The chi-square tests indicate that the differences between target and achieved constraint satisfaction rates are statistically insignificant (p > 0.05 for all constraints) which aligns with our constraint handling strategy. The high p-values indicate strong consistency between target and achieved satisfaction rates for all constraint types and problem sizes. Table 7 presents capacity constraints satisfaction rates, while Table 8 shows operational constraints satisfaction rates with chi-square goodness-of-fit tests.

5.5. Convergence Analysis

We provide comprehensive convergence analysis for both the chance measure estimation and the optimization algorithm.

Monte Carlo Convergence for Chance Measures

Table 9 shows convergence characteristics with variance reduction factors (VRF) compared to crude Monte Carlo.
The convergence rate follows O ( 1 / N M C ) as expected, with estimates stabilizing at N M C = 10 , 000 , justifying our parameter choice.

5.6. Statistical Validation of Hybrid Approach

We validate the hybrid uncertain random approach through comprehensive statistical tests comparing it with pure paradigms. Table 10 presents the statistical comparison of different uncertainty modeling approaches with ANOVA and post-hoc tests.

5.7. Comprehensive Algorithm Performance Analysis

We expand the algorithm comparison with additional performance metrics and statistical tests. Table 11 provides an extended comparison of algorithm performance including convergence statistics and success rates.

5.8. Reinforcement Learning Component Analysis

We analyze the effectiveness of the RL component in parameter adaptation. Table 12 presents the analysis of RL action selection patterns and their performance impact across different optimization phases.
The RL component demonstrates adaptive behavior, with exploration dominating early phases and exploitation increasing in later iterations, contributing to the 4.5% average performance improvement over static parameter settings.

5.9. Sensitivity Analysis

We conducted comprehensive sensitivity analysis to evaluate the robustness of our dependent-chance goal programming approach under varying parameter settings. This analysis focuses on two critical aspects: chance constraint confidence levels and uncertainty parameter scaling.

5.9.1. Confidence Level Impact

We analyzed the sensitivity to chance constraint confidence levels by varying the average confidence level from 0.70 to 0.95. Figure 1 illustrates the impact on network design and performance.
Table 13 presents the detailed results of confidence level sensitivity analysis.
The research illustrates a nonlinear correlation between confidence levels and network performance. As shown in Figure 1, network cost increases exponentially with confidence level greater than 0.90, with decreasing returns in constraint satisfaction. The optimal confidence levels between 0.80–0.85 strike a balance between goal achievement and computational efficiency.

5.9.2. Effect of Uncertainty Level

We investigated the impact of uncertainty by scaling all uncertain parameters from 0.5 to 1.5 times their base levels. Table 14 shows how varying uncertainty levels affect solution quality.
The analysis demonstrates that higher uncertainty levels significantly impact all performance metrics, with goal achievement decreasing by 12.6% when uncertainty is increased by 50%. This confirms the importance of accurate uncertainty quantification in dependent-chance goal programming.

5.10. Network Structure Analysis

5.10.1. Facility Location Patterns

Analysis of optimal facility selections reveals consistent patterns across instance sizes. Figure 2 illustrates how facility selection percentages vary with problem size.
Table 15 provides detailed selection percentages for each facility type across different problem sizes.
The decreasing selection percentages with increasing problem size indicate that the DCGP approach effectively exploits economies of scale, with reductions ranging from 24.2% to 34.8% between small and very large instances.

5.10.2. Transportation Mode Selection

Transportation mode preferences shift systematically with network size, as shown in Figure 3.
Table 16 presents the detailed transportation mode selection patterns.
The shift from road (53.2% to 42.6%) to rail transportation (28.5% to 39.2%) as problem size increases reflects the model’s ability to balance economic efficiency with environmental considerations.

5.11. Resilience Analysis

5.11.1. Cost of Resilience

A key contribution of our work is quantifying the relationship between resilience investment and network cost. Figure 4 illustrates this nonlinear relationship, and Table 17 provides a detailed breakdown of resilience costs and marginal effects.
The analysis reveals that marginal costs remain relatively stable (7–9%) for resilience levels up to 0.80, after which they increase sharply. This suggests an optimal operational range of 0.70–0.80 for most practical applications.

5.11.2. Network Robustness Metrics

Table 18 compares network robustness measures across different optimization approaches.
The hybrid DCGP approach consistently outperforms both deterministic and traditional stochastic methods across all robustness metrics, with particularly significant improvements in supply diversification (+39% vs. deterministic) and recovery capability (+32% vs. deterministic).

5.12. Managerial Insights and Practical Implementation

Based on our comprehensive analysis, we provide actionable insights with specific implementation guidelines:
1.
Probability Target Setting Strategy:
  • Economic objectives: Set targets at 93–95% (achievable with minimal compromise)
  • Environmental objectives: Target 80–85% (balances compliance with cost)
  • Social objectives: Aim for 75–80% (realistic given current constraints)
  • Resilience objectives: 70–75% provides cost-effective risk mitigation
2.
Implementation Roadmap:
  • Phase 1: Implement economic optimization (3–6 months)
  • Phase 2: Integrate environmental constraints (6–9 months)
  • Phase 3: Add social and resilience objectives (9–12 months)
  • Expected ROI: 12–18 months based on cost savings analysis
3.
Uncertainty Management Protocol:
  • Collect historical data for random parameters (minimum 24 months)
  • Conduct expert elicitation workshops for belief-based parameters
  • Update uncertainty estimates quarterly
  • Maintain confidence levels at 0.80–0.85 for optimal performance

5.13. Validation and Sensitivity Analysis

5.13.1. Cross-Validation Results

We performed 10-fold cross-validation to assess model robustness. Table 19 presents the cross-validation performance metrics across all sustainability dimensions:

5.13.2. Parameter Sensitivity Analysis

We conducted comprehensive sensitivity analysis using Morris screening method. Table 20 shows the parameter sensitivity rankings using Morris μ * values across different objective dimensions:

5.14. Cost-Benefit Analysis and Implementation Feasibility

Implementing the hybrid Dependent-Chance Goal Programming (DCGP) model requires moderate but strategic investment in both technology and organizational capability. These costs are justified by the significant gains in resilience, efficiency, and sustainability achieved across supply chain operations. Table 21 summarizes the estimated initial costs, potential annual savings, payback periods, and three-year return on investment (ROI) across different organization sizes.
These estimates assume modest improvements of 1–2% in annual supply chain costs, which are consistent with conservative benchmarks from industry reports. Even small improvements in efficiency can generate measurable financial benefits over a multi-year horizon. While the payback period may vary depending on implementation scale and context, the expected return on investment over three years ranges from approximately 30% to 55%, indicating favorable long-term value.
The initial investment covers software (either proprietary or customized), computational infrastructure, and expert support for model calibration and integration. These costs are comparable to those incurred when deploying advanced decision-support or optimization systems. Notably, DCGP can be integrated as a modular layer on top of existing enterprise systems rather than requiring a full overhaul.
Annual operating expenses include software updates, calibration, data management, and staffing or upskilling specialized personnel. These costs typically range from $225 K to $680 K depending on organizational size and internal capacity. Training and change management are essential components of successful deployment. Analysts generally require 80–120 h of training, while executives benefit from 40–60 h. In addition, organizations should plan for a 12–24 months change management process to ensure smooth adoption and internal alignment. To help organizations assess their readiness and plan accordingly, Table 22 summarizes key implementation dimensions in a concise format.
In summary, although the hybrid DCGP model involves sophisticated methods and demands organizational commitment, the economic rationale is clear. It offers a favorable cost-benefit ratio, particularly for organizations managing complex or vulnerable supply chains. By starting with well-defined pilot implementations and investing in the right expertise and training, firms can expect substantial long-term returns and enhanced strategic agility.

5.15. Limitations and Future Research Directions

We acknowledge several limitations and propose specific future research directions:
1.
Real-World Validation: While our synthetic instances are realistic, validation with industry data is needed. Future work should include:
  • Partnership with supply chain companies for data access
  • Case studies in specific industries (automotive, pharmaceutical, retail)
  • Pilot implementations with performance tracking
2.
Dynamic Uncertainty Modeling: Current static uncertainty assumptions limit applicability. Extensions should include:
  • Time-varying uncertainty parameters
  • Bayesian updating of uncertainty estimates
  • Adaptive reoptimization strategies
3.
Scalability Enhancement: For very large instances (>1 M variables):
  • Decomposition methods (Benders, Dantzig-Wolfe)
  • Parallel computing implementations
  • Approximation algorithms with quality guarantees
4.
Multi-Period Extensions:
  • Rolling horizon approaches
  • Stochastic dynamic programming formulations
  • Learning effects over time
In general, while the proposed hybrid DCGP framework offers substantial improvements in resilience and flexibility, several limitations should be acknowledged. These include the need for reliable hybrid uncertainty data, the computational complexity of chance-constrained models, and the absence of large-scale empirical validation. Future work should focus on deploying the framework in real-world supply chain systems to evaluate its operational feasibility and performance under diverse scenarios.

6. Conclusions

This work proposed an RL-enhanced Salp Swarm Optimization (RL-SSO) algorithm to solve sustainable supply chain planning dependent-chance goal programming models with hybrid uncertainty. The method achieved 0.5–5.11% performance gains and 12.7% computational time reduction over baselines.
Our quantitative experiments validated the effectiveness of the lexicographic optimization framework, with the highest probability achievement observed for economic objectives (93.1%), followed by environmental (80.7%), social (75.4%), and resilience goals (70.8%). The results also revealed that within the optimal resilience range of 0.70–0.80, each 10% increase in resilience is associated with a 7–9% rise in cost, with marginal costs escalating beyond this range. Furthermore, sensitivity analysis identified optimal confidence levels between 0.80 and 0.85, balancing performance and practical usability.
The hybrid uncertain random approach outperformed traditional uncertainty-based models by 3.2%, while achieving notable gains in facility redundancy (25.9%) and supply diversification (39.1%) compared to deterministic baselines. These findings underscore the effectiveness of dependent-chance goal programming as a robust framework for multi-objective sustainable supply chain optimization under hybrid uncertainty. The main contributions of this research include: (i) a novel integration of chance theory with multi-objective optimization to model hybrid uncertainty, (ii) a self-adaptive RL-SSO algorithm featuring dynamic parameter tuning via reinforcement learning, and (iii) practical guidelines for resilience investment planning and confidence level selection in sustainable supply chain design. Nevertheless, the proposed framework has some limitations, including its reliance on high-quality uncertainty data, computational complexity, and the need for real-world validation in operational settings.
Future research should focus on real-world testing, dynamic uncertainty formulation, and scalability enhancement for extremely large-scale applications. Through addressing these directions, the introduced method will further strengthen both the theoretical foundations and the practical applicability of the proposed approach to sustainable supply chain optimization under uncertainty.

Author Contributions

Conceptualization, Y.B. and R.B.; Methodology, Y.B., R.B. and A.T.; Software, Y.B.; Validation, Y.B. and M.F.; Formal Analysis, Y.B. and A.T.; Investigation, Y.B. and R.B.; Resources, R.B. and O.B.; Data Curation, Y.B. and M.F.; Writing—Original Draft Preparation, Y.B.; Writing—Review & Editing, Y.B., R.B., A.T., M.F. and O.B.; Visualization, Y.B.; Supervision, R.B. and A.T.; Project Administration, R.B. and O.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Al-Khwarizmi Programme, a collaborative effort between the National Center for Scientific and Technical Research (CNRST), the Agency for Digital Development (ADD), and the Moroccan Ministry of Higher Education.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are contained within this article.

Acknowledgments

The authors express their gratitude to the editors and reviewers for their valuable comments and constructive suggestions regarding the revision of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Brandenburg, M.; Govindan, K.; Sarkis, J.; Seuring, S. Quantitative models for sustainable supply chain management: Developments and directions. Eur. J. Oper. Res. 2014, 233, 299–312. [Google Scholar] [CrossRef]
  2. Govindan, K.; Fattahi, M.; Keyvanshokooh, E. Supply chain network design under uncertainty: A comprehensive review and future research directions. Eur. J. Oper. Res. 2017, 263, 108–141. [Google Scholar] [CrossRef]
  3. Ivanov, D. Viable supply chain model: Integrating agility, resilience and sustainability perspectives—lessons from and thinking beyond the COVID-19 pandemic. Ann. Oper. Res. 2022, 319, 1411–1431. [Google Scholar] [CrossRef] [PubMed]
  4. Snyder, L.V.; Atan, Z.; Peng, P.; Rong, Y.; Schmitt, A.J.; Sinsoysal, B. OR/MS models for supply chain disruptions: A review. IIE Trans. 2016, 48, 89–109. [Google Scholar] [CrossRef]
  5. McKinsey & Company. Supply Chain Disruptions: The True Cost to Global Businesses; McKinsey Global Institute Report; McKinsey & Company: New York, NY, USA, 2023. [Google Scholar]
  6. CDP. Supply Chain Emissions: Four Times Greater than Direct Operations; Carbon Disclosure Project Annual Report; CDP: London, UK, 2023. [Google Scholar]
  7. International Labour Organization. Global Employment Trends in Manufacturing and Supply Chains; ILO Statistics Department: Geneva, Switzerland, 2023. [Google Scholar]
  8. Liu, B. Uncertainty Theory: A Branch of Mathematics for Modeling Human Uncertainty; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  9. Behzadi, G.; O’Sullivan, M.J.; Olsen, T.L.; Zhang, A. Agribusiness supply chain risk management: A review of quantitative decision models. Omega 2018, 79, 21–42. [Google Scholar] [CrossRef]
  10. Santoso, T.; Ahmed, S.; Goetschalckx, M.; Shapiro, A. A stochastic programming approach for supply chain network design under uncertainty. Eur. J. Oper. Res. 2005, 167, 96–115. [Google Scholar] [CrossRef]
  11. Ben-Tal, A.; El Ghaoui, L.; Nemirovski, A. Robust Optimization; Princeton University Press: Princeton, NJ, USA, 2009. [Google Scholar]
  12. Pishvaee, M.S.; Razmi, J.; Torabi, S.A. Robust possibilistic programming for socially responsible supply chain network design: A new approach. Fuzzy Sets Syst. 2012, 206, 1–20. [Google Scholar] [CrossRef]
  13. Liu, B. Uncertainty Theory, 4th ed.; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  14. Liu, B. Uncertainty Theory, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2007. [Google Scholar]
  15. Cheng, L.; Wan, Z.; Wang, G. Bilevel chance-constrained optimization for renewable energy-based virtual power plants operation. IEEE Trans. Smart Grid 2020, 11, 5440–5453. [Google Scholar]
  16. Liu, B. Theory and Practice of Uncertain Programming, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  17. Sahinidis, N.V. Optimization under uncertainty: State-of-the-art and opportunities. Comput. Chem. Eng. 2004, 28, 971–983. [Google Scholar] [CrossRef]
  18. Yang, X.S.; Deb, S.; Fong, S.; He, X.; Zhao, Y.X. Bio-Inspired Computation: Algorithms and Applications; Springer: Cham, Switzerland, 2019. [Google Scholar]
  19. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp Swarm Algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  20. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  21. Mazyavkina, N.; Sviridov, S.; Ivanov, S.; Burnaev, E. Reinforcement learning for combinatorial optimization: A survey. Comput. Oper. Res. 2021, 134, 105400. [Google Scholar] [CrossRef]
  22. Zhang, W.; Maleki, A.; Rosen, M.A.; Liu, J. Optimization with a simulated annealing algorithm of a hybrid system for renewable energy including battery and hydrogen storage. Energy 2018, 163, 191–207. [Google Scholar] [CrossRef]
  23. Liu, B. Uncertainty Theory: A Branch of Mathematics for Modeling Human Uncertainty; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
Figure 1. Impact of Chance Constraint Confidence Levels on Network Design: Shows how varying average confidence levels β ˜ j affect network cost, constraint satisfaction rates, and network density in the dependent-chance goal programming model.
Figure 1. Impact of Chance Constraint Confidence Levels on Network Design: Shows how varying average confidence levels β ˜ j affect network cost, constraint satisfaction rates, and network density in the dependent-chance goal programming model.
Sustainability 17 06079 g001
Figure 2. Facility Selection Patterns Under DCGP: Shows how facility selection percentages decrease with problem size, indicating economies of scale benefits in dependent-chance optimized networks.
Figure 2. Facility Selection Patterns Under DCGP: Shows how facility selection percentages decrease with problem size, indicating economies of scale benefits in dependent-chance optimized networks.
Sustainability 17 06079 g002
Figure 3. Transportation Mode Selection by Instance Size: Shows shift from road to rail transportation as problem size increases, reflecting economic and environmental optimization in the DCGP framework.
Figure 3. Transportation Mode Selection by Instance Size: Shows shift from road to rail transportation as problem size increases, reflecting economic and environmental optimization in the DCGP framework.
Sustainability 17 06079 g003
Figure 4. Cost of Resilience Analysis: Shows the nonlinear relationship between network resilience levels and total costs, demonstrating 7–12% cost increases per 10% resilience improvement in the optimal range.
Figure 4. Cost of Resilience Analysis: Shows the nonlinear relationship between network resilience levels and total costs, demonstrating 7–12% cost increases per 10% resilience improvement in the optimal range.
Sustainability 17 06079 g004
Table 1. Parameter Settings for RL-SSO Algorithm.
Table 1. Parameter Settings for RL-SSO Algorithm.
ParameterSymbolValue
Population size N p 100
Maximum iterations T m a x 500
Number of Monte Carlo simulations N M C 10,000
Inertia weight range [ c 1 m i n , c 1 m a x ] [ 0.4 , 0.9 ]
Cognitive parameter range [ c 2 m i n , c 2 m a x ] [ 1.5 , 2.5 ]
Social parameter range [ c 3 m i n , c 3 m a x ] [ 1.5 , 2.5 ]
Decay parameter ρ 4.0
Decay parameter ρ 4.0
Exploration control parameter σ 3.0
Initial learning rate α 0 0.1
Learning rate decay ϕ 0.5
Discount factor γ 0.9
Exploration probability (initial) p e x p l o r e m a x 0.9
Penalty factor (initial) λ j 0 100
Penalty increment rate δ 0.02
Confidence level tightening rate κ 3.0
Convergence tolerance ϵ 10 6
Table 2. Characteristics of Test Instances.
Table 2. Characteristics of Test Instances.
InstanceSuppliersPlantsWHDCsCustomersProductsPeriods
Small-153461034
Small-285681534
Medium-11068102056
Medium-215810123056
Large-1201012154088
Large-2251215205088
Very Large30152025801012
Table 3. Uncertain Random Parameter Distributions.
Table 3. Uncertain Random Parameter Distributions.
ParameterRandom ComponentUncertain Component
Demand ( D M ˜ c k t )Normal ( μ c k t , σ c k t 2 )Linear ( 0.9 μ c k t , 1.1 μ c k t )
Supply capacity ( S S ˜ i t )Uniform ( 0.9 S S ^ i t , 1.1 S S ^ i t )Linear ( 0.85 S S ^ i t , 1.15 S S ^ i t )
Production cost ( P C ˜ j k )Normal ( μ P C , 0.05 μ P C 2 )Linear ( 0.9 μ P C , 1.15 μ P C )
Transportation cost ( T C ˜ i j k l )Lognormal ( μ T C , 0.1 μ T C 2 )Linear ( 0.85 μ T C , 1.2 μ T C )
Supplier reliability ( ξ i j t )Beta (5, 1.5)Linear (0.8, 1.0)
Production efficiency ( θ j k t )Triangular (0.85, 1.0, 1.05)Linear (0.9, 1.1)
Environmental impact ( E C ˜ i j k l )Normal ( μ E C , 0.08 μ E C 2 )Linear ( 0.75 μ E C , 1.25 μ E C )
Disruption probability ( ω i t )Beta (1.2, 8)Linear (0.5, 2.0)
Recovery rate ( R R ˜ i k t )Triangular (0.3, 0.5, 0.8)Linear (0.7, 1.3)
Resilience index ( R I ˜ i j )Uniform (0.6, 0.9)Linear (0.8, 1.2)
Disruption severity ( D S ˜ k t )Beta (2, 5)Linear (0.5, 1.5)
Job creation factor ( J C ˜ i )Poisson ( λ J C )Linear (0.9, 1.1)
Table 4. Sustainability Goal Probability Achievement with Confidence Intervals.
Table 4. Sustainability Goal Probability Achievement with Confidence Intervals.
InstanceEconomicEnvironmentalSocialResilience
(Target: 0.95)(Target: 0.85)(Target: 0.80)(Target: 0.75)
Small-10.947 ± 0.0080.832 ± 0.0120.774 ± 0.0150.728 ± 0.018
Small-20.943 ± 0.0090.824 ± 0.0130.769 ± 0.0160.721 ± 0.019
Medium-10.938 ± 0.0100.815 ± 0.0140.762 ± 0.0170.714 ± 0.020
Medium-20.932 ± 0.0110.807 ± 0.0150.755 ± 0.0180.708 ± 0.021
Large-10.925 ± 0.0120.798 ± 0.0160.746 ± 0.0190.701 ± 0.022
Large-20.918 ± 0.0130.789 ± 0.0170.738 ± 0.0200.694 ± 0.023
Very Large0.912 ± 0.0140.782 ± 0.0180.731 ± 0.0210.687 ± 0.024
Average0.9310.8070.7540.708
Gap from Target−1.9%−5.1%−5.8%−5.6%
Table 5. Lexicographic Deviation Values ( d g ) with Statistical Analysis.
Table 5. Lexicographic Deviation Values ( d g ) with Statistical Analysis.
Instance d 1 d 2 d 3 d 4
(Priority 1)(Priority 2)(Priority 3)(Priority 4)
Small-10.0030.0180.0260.022
Small-20.0070.0260.0310.029
Medium-10.0120.0350.0380.036
Medium-20.0180.0430.0450.042
Large-10.0250.0520.0540.049
Large-20.0320.0610.0620.056
Very Large0.0380.0680.0690.063
Average0.0190.0430.0460.042
Std. Dev.0.0130.0180.0170.015
CV (%)68.441.937.035.7
Table 6. Economic Impact of Performance Improvements.
Table 6. Economic Impact of Performance Improvements.
Annual Revenue1% Cost Reduction5% ImprovementAnnual Savings
$100 M (Small)$3 M$150 K$450 K–$2.25 M
$500 M (Medium)$15 M$750 K$2.25 M–$11.25 M
$1 B (Large)$30 M$1.5 M$4.5 M–$22.5 M
Table 7. Capacity Constraints Satisfaction with Chi-Square Tests.
Table 7. Capacity Constraints Satisfaction with Chi-Square Tests.
Constraint TypeTargetSmallMediumLarge χ 2 Test
Supply Capacity0.900.8970.8890.879 χ 2 = 2.31 , p = 0.31
Manufacturing Capacity0.850.8470.8390.828 χ 2 = 3.45 , p = 0.18
Warehouse Capacity0.800.7980.7910.781 χ 2 = 1.89 , p = 0.39
Distribution Capacity0.800.7960.7890.779 χ 2 = 1.76 , p = 0.41
Average0.840.8350.8270.817
Table 8. Operational Constraints Satisfaction with Chi-Square Tests.
Table 8. Operational Constraints Satisfaction with Chi-Square Tests.
Constraint TypeTargetSmallMediumLarge χ 2 Test
Demand Satisfaction0.950.9460.9380.928 χ 2 = 4.12 , p = 0.13
Flow Conservation0.990.9870.9840.980 χ 2 = 1.23 , p = 0.54
Production Linkage0.900.8960.8880.878 χ 2 = 2.87 , p = 0.24
Environmental Limits0.800.7960.7880.779 χ 2 = 1.94 , p = 0.38
Average0.910.9060.9000.891
Table 9. Chance Measure Convergence Analysis with Variance Reduction.
Table 9. Chance Measure Convergence Analysis with Variance Reduction.
N MC EconomicEnvironmentalSocialResilienceVRFTime
10000.942 ± 0.0180.821 ± 0.0240.769 ± 0.0310.719 ± 0.0291.003.2 min
50000.939 ± 0.0120.817 ± 0.0160.764 ± 0.0210.716 ± 0.0192.248.7 min
10,0000.938 ± 0.0080.815 ± 0.0110.762 ± 0.0140.714 ± 0.0133.1612.4 min
25,0000.938 ± 0.0060.815 ± 0.0080.762 ± 0.0100.714 ± 0.0095.0024.8 min
50,0000.938 ± 0.0040.815 ± 0.0060.762 ± 0.0070.714 ± 0.0067.0748.2 min
Table 10. Statistical Comparison of Uncertainty Modeling Approaches.
Table 10. Statistical Comparison of Uncertainty Modeling Approaches.
ApproachEconomicEnvironmentalSocialResilienceOverall
Pure Stochastic0.952 ± 0.0110.843 ± 0.0150.791 ± 0.0180.742 ± 0.0210.832
Pure Uncertain0.921 ± 0.0140.787 ± 0.0190.734 ± 0.0230.686 ± 0.0260.782
Hybrid (Proposed)0.938 ± 0.0100.815 ± 0.0140.762 ± 0.0170.714 ± 0.0200.807
Statistical Tests:
ANOVA: F ( 2 , 18 ) = 45.67 , p < 0.001
Tukey HSD: Hybrid vs. Stochastic ( p = 0.012 ),
                       Hybrid vs. Uncertain ( p < 0.001 )
Table 11. Extended Algorithm Performance Comparison.
Table 11. Extended Algorithm Performance Comparison.
AlgorithmBestAverageWorstStd. Dev.TimeSuccess Rate
GA0.7760.7810.7980.00818.767.3%
PSO0.7900.7950.8080.00616.372.1%
DE0.7980.8030.8120.00515.874.5%
SSO0.8030.8080.8150.00414.278.2%
RL-SSO0.8070.8120.8170.00312.483.4%
Wilcoxon Signed-Rank Test (RL-SSO vs. others): all p < 0.05
Table 12. RL Action Selection and Performance Impact.
Table 12. RL Action Selection and Performance Impact.
Optimization PhaseExplorationExploitationBalancedIntensify
Early (0–30%)45.2%8.3%31.5%15.0%
Middle (30–70%)22.8%18.6%42.3%16.3%
Late (70–100%)5.4%38.7%24.1%31.8%
Performance Gain+2.3%+1.8%+0.9%+2.7%
Table 13. Impact of Confidence Levels on Performance Metrics.
Table 13. Impact of Confidence Levels on Performance Metrics.
ConfidenceGoal AchievementConstraint Sat.Network DensityCPU Time
0.700.8510.9670.8929.1
0.750.8340.9540.84710.2
0.800.8200.9420.80411.3
0.850.8070.9280.76312.4
0.900.7910.9130.72114.8
0.950.7680.8960.67818.7
Table 14. Effect of Uncertainty Level on Solution Quality.
Table 14. Effect of Uncertainty Level on Solution Quality.
ScaleGoal AchievementConstraint Sat.ResilienceCPU Time
0.5×0.8560.9780.868.7
0.75×0.8320.9620.8210.2
1.0×0.8070.9430.7812.4
1.25×0.7790.9210.7315.8
1.5×0.7480.8950.6719.7
Table 15. Facility Selection Percentages by Problem Size.
Table 15. Facility Selection Percentages by Problem Size.
SizeSuppliersPlantsWarehousesDCs
Small75.480.464.656.3
Medium66.974.058.951.6
Large57.067.050.546.4
Very Large51.262.445.642.8
Table 16. Transportation Mode Selection Percentages.
Table 16. Transportation Mode Selection Percentages.
SizeRoadRailSeaAir
Small53.2%28.5%15.7%2.6%
Medium48.7%32.3%17.4%1.6%
Large45.3%36.8%16.9%1.0%
Very Large42.6%39.2%17.8%0.4%
Table 17. Cost of Resilience Analysis.
Table 17. Cost of Resilience Analysis.
ResilienceCost (M$)Increase (%)Marginal (%)
0.507.50
0.608.016.86.8
0.708.5814.47.1
0.809.3424.58.9
0.9010.6742.314.2
0.9512.4566.016.7
Table 18. Network Robustness Metrics Comparison.
Table 18. Network Robustness Metrics Comparison.
MetricDeterministicStochasticHybrid DCGP
Facility Redundancy1.121.341.41
Supply Diversification2.32.83.2
Route Flexibility0.680.740.79
Recovery Capability0.5420.6780.714
Environmental Efficiency0.7210.6980.758
Social Integration0.6340.6120.687
Table 19. Cross-Validation Performance Metrics.
Table 19. Cross-Validation Performance Metrics.
FoldEconomicEnvironmentalSocialResilienceOverall
Mean0.9330.8090.7560.7110.802
Std. Dev.0.0080.0110.0130.0150.009
Min0.9180.7910.7380.6890.784
Max0.9450.8260.7740.7320.819
CV (%)0.861.361.722.111.12
Table 20. Parameter Sensitivity Rankings (Morris μ * ).
Table 20. Parameter Sensitivity Rankings (Morris μ * ).
ParameterEconomicEnvironmentalSocialResilience
Demand Uncertainty0.8720.5430.4210.687
Cost Parameters0.9340.6120.3870.456
Capacity Limits0.7560.8230.6540.789
Confidence Levels0.6890.7450.8120.834
Disruption Probability0.2340.3120.4560.912
Table 21. Estimated Costs and Returns by Organization Size.
Table 21. Estimated Costs and Returns by Organization Size.
Organization SizeInitial InvestmentAnnual SavingsPayback Period3-Year ROI
Small (<$100 M revenue)$0.5 M–1.1 M$0.45 M–2.25 M1.5–3.5 years30–45%
Medium ($100 M–$500 M)$0.95 M–1.65 M$2.25 M–11.25 M0.8–1.8 years35–50%
Large (>$500 M revenue)$1.7 M–2.7 M$4.5 M–22.5 M0.5–1.2 years45–55%
Table 22. Summary of Implementation Considerations.
Table 22. Summary of Implementation Considerations.
AspectDescription
Cost CategoriesIncludes software licensing or development, computational infrastructure, expert consultation, and integration with ERP/SCM systems.
Ongoing RequirementsEncompasses annual software updates, data calibration, and specialized personnel costs, typically ranging from $225 K to $680 K.
Training NeedsAnalysts require 80–120 h of training; executives, 40–60 h. Change management should span 12–24 months.
Risk FactorsMay include incomplete data, limited internal expertise, resistance to probabilistic thinking, and integration barriers with legacy systems.
Success EnablersStrong executive sponsorship, phased/pilot rollouts, academic or consulting partnerships, and investment in team upskilling.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Boutmir, Y.; Bannari, R.; Touil, A.; Fri, M.; Benmoussa, O. Dependent-Chance Goal Programming for Sustainable Supply Chain Design: A Reinforcement Learning-Enhanced Salp Swarm Approach. Sustainability 2025, 17, 6079. https://doi.org/10.3390/su17136079

AMA Style

Boutmir Y, Bannari R, Touil A, Fri M, Benmoussa O. Dependent-Chance Goal Programming for Sustainable Supply Chain Design: A Reinforcement Learning-Enhanced Salp Swarm Approach. Sustainability. 2025; 17(13):6079. https://doi.org/10.3390/su17136079

Chicago/Turabian Style

Boutmir, Yassine, Rachid Bannari, Achraf Touil, Mouhsene Fri, and Othmane Benmoussa. 2025. "Dependent-Chance Goal Programming for Sustainable Supply Chain Design: A Reinforcement Learning-Enhanced Salp Swarm Approach" Sustainability 17, no. 13: 6079. https://doi.org/10.3390/su17136079

APA Style

Boutmir, Y., Bannari, R., Touil, A., Fri, M., & Benmoussa, O. (2025). Dependent-Chance Goal Programming for Sustainable Supply Chain Design: A Reinforcement Learning-Enhanced Salp Swarm Approach. Sustainability, 17(13), 6079. https://doi.org/10.3390/su17136079

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop