Next Article in Journal
System Analysis of Environmental Effects: A Case of Sustainable Development in the Russian Economy Based on Digital Engineering
Previous Article in Journal
Generative Artificial Intelligence and the Creative Industries: A Bibliometric Review and Research Agenda
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimising Calculation Logic in Emergency Management: A Framework for Strategic Decision-Making

School of Philosophy, Nanjing University, Nanjing 210093, China
*
Author to whom correspondence should be addressed.
Systems 2026, 14(2), 139; https://doi.org/10.3390/systems14020139
Submission received: 20 November 2025 / Revised: 13 January 2026 / Accepted: 26 January 2026 / Published: 29 January 2026
(This article belongs to the Section Artificial Intelligence and Digital Systems Engineering)

Abstract

Given the increasing demand for rapid emergency management decision-making, which must be both timely and reliable, even slight delays can result in substantial human and economic losses. However, current systems and recent state-of-the-art work often use inflexible rule-based logic that cannot adapt to rapidly changing emergency conditions or dynamically optimise response allocation. As a result, our study presents the Calculation Logic Optimisation Framework (CLOF), a novel data-driven approach that enhances decision-making intelligently and strategically through learning-based predictive and multi-objective optimisation, utilising the 911 Emergency Calls data set, comprising more than half a million records from Montgomery County, Pennsylvania, USA. The CLOF examines patterns over space and time and uses optimised calculation logic to reduce response latency and increase decision reliability. The suggested framework outperforms the standard Decision Tree, Random Forest, Gradient Boosting, and XGBoost baselines, achieving 94.68% accuracy, a log-loss of 0.081, and a reliability score (R2) of 0.955. The mean response time error is reported to have been reduced by 19%, illustrating robustness to real-world uncertainty. The CLOF aims to deliver results that confirm the scalability, interpretability, and efficiency of modern EM frameworks, thereby improving safety, risk awareness, and operational quality in large-scale emergency networks.

1. Introduction

In recent times, there has been an increasing number and severity of emergencies, including natural disasters, industrial accidents, and significant public safety incidents [1,2,3].These events create immense pressure on emergency management systems to ensure decision-making that is precise, quick, and reliable under conditions of uncertainty, variation, time pressure and minimal resources [4,5,6]. Typically, researchers use static, rule-based decision logic to provide manual triage protocols or dispatch based on heuristics [7,8]. However, they may struggle to adapt to dynamic and demanding environments [5]. To further emphasise the need for an optimised calculation logic, real-life emergency information is required. An analysis of the 911 Emergency Calls dataset, containing more than 500,000 records, shows that learning-based decision logic reduces response time by approximately 2.1 min (≈18%) compared to conventional rule-based dispatch mechanisms. The reported improvement is statistically significant (p < 0.01, paired t-test). That is, even small changes to internal decision-making logic can yield considerable operational benefits in time-critical emergencies [9,10,11]. According to Su et al. (2022), emergency decision-making (EDM) has evolved into a multidisciplinary task that draws on mathematics, information systems, public management, and psychology [10]. There is now a consensus on the need to develop decision-support systems that integrate predictive modelling, optimisation, reliability assessment, and adaptability for real-time operations. While another study (2024) by Ivan et al. has also highlighted institutional and cognitive barriers in such a high-stakes context, the authors note that decision-makers lack tools to overlay real-time data with risk analysis and operational constraints [12]. From these studies, observations highlight that improving the decision-making logic beyond just execution is critically essential for the safety, risk, reliability, and quality of emergency management.
In recent years (2022–2025), emergency decision-making research has increasingly focused on data-driven optimisation, adaptive dispatch systems, and reliability-aware decision-support frameworks. Recent studies have explored machine-learning-assisted emergency triage, multi-criteria optimisation under uncertainty, and system-level resilience modelling. However, despite these advances, most existing approaches continue to optimise outcomes or resource allocation rather than the internal calculation logic that governs emergency decision processes. This gap motivates the present study, which explicitly formulates calculation logic as an optimisable object under reliability constraints and validates it using large-scale real-world emergency data.
Notwithstanding the growing body of related work, significant gaps remain, with many studies focusing on isolated components, i.e., optimising resource allocation networks (such as ambulance repositioning or relief vehicle deployment) or developing decision-support systems (DSS) for incident commanders. In this context, a study by Zamanifar & Hartmann (2020) reviews optimisation-based decision-making for transportation recovery networks [13]. It concludes that practical applicability and validation in real-world emergency settings are limited. Similarly, a recent study by Fertier et al. (2020) developed a new emergency decision-support system to enhance decision-makers’ capabilities [14]. However, it stops short of optimising the underlying calculation logic that drives triage, prioritisation, and dispatch decisions. Moreover, some studies address the reliability of response systems; for example, a survey by Mingjian Zuo (2025) applies reliability engineering techniques to emergency response systems, but these often do not combine predictive learning, dynamic logic thresholds, and high-fidelity modelling of decision-logic flows [15]. There remains a need for frameworks that (1) treat the calculation logic itself as an optimisable component; (2) incorporate predictive modelling and multi-objective optimisation of latency, reliability and cost; and (3) validate with large-scale real-world incident data. This study is motivated by those unmet needs and aims to fill this gap.
To address this gap and overcome the shortcomings of recent state-of-the-art approaches, this paper introduces a novel Calculation Logic Optimisation Framework (CLOF) for strategic decision-making in emergency management [16,17]. The framework combines machine-learning-based prediction of incident priority [18,19,20], multi-objective optimisation of priority thresholds and dispatch logic, and reliability assessment of decision logic flows. Our novel approach emphasises mathematical rigour, interpretability of the logic layer, and empirical validation on a large-scale emergency incident dataset. By treating the “calculation logic” by which triage determines priority setting and resource allocation as an optimisation problem, we move beyond static rule-based systems to a dynamic, data-driven logic engine that adapts to spatiotemporal patterns, resource constraints, and reliability requirements. In doing so, we contribute (1) a conceptual modelling of calculation logic as a formal optimisation problem; (2) implementation of the optimisation logic on incident data; and (3) demonstration of significant improvement in decision-making reliability and quality in an emergency context.
Moreover, the remaining paper is based on the following sections: Section 2 reviews related state-of-the-art works on emergency decision-making, risk- and reliability-based optimisation and decision-support systems; Section 3 presents the research background and problem definition (including research questions and problem setup); Section 4 describes the methodology including data pre-processing, feature engineering, the CLOF model architecture, and evaluation metrics, and reports experiments; Section 5 provides the results and discusses findings, limitations and implications; and Section 6 concludes and presents future work.

2. Literature Review

Emergency decision-making research has evolved rapidly with the integration of computational intelligence, multi-criteria optimisation, and real-time data analytics [4,10]. Early and traditional frameworks primarily focused on descriptive and procedural models, but recent work has driven a shift toward data-driven optimisation and adaptive decision-making [7,10,21]. Su et al. (2022) suggested a comprehensive review of EDM models, encompassing different qualitative, quantitative, and hybrid models [10]. Furthermore, the authors noted that reliable EDM deals must be linked to operational decision rules on an urgent basis. Another research by Elkady et al. (2024) investigated more than 120 decision-support system (DSS) studies [22]. The research summarised that the current systems “lack a self-optimising feedback mechanism that can evaluate the logic of their decisions”. Similarly, Zhou et al. (2017) suggested that Bayesian networks can provide post-disaster decision support, as probabilistic reasoning can help enhance situational awareness [11]. However, its heavy reliance on static prior logic is an issue. Studies have shown that forecast models and predictions have improved significantly over the years. However, the logic of calculation and internal structure into which inputs are converted into actionable priorities continues to remain, for the most part, unoptimised.
Recent works further show that emergency decision optimisation is progressing, but other gaps still exist. The study by Nagy et al. (2024) exemplifies a robust multi-criteria decision-making model for emergency information system preparedness [23]. MCDM tools (TOPSIS and ELECTRE) are integrated to enhance the supplier selection effort under conditions of uncertainty. Nevertheless, the system is limited to the procurement decision and does not optimally execute a real-time calculation logic during emergency operations. Simultaneously, Yazdani and Haghani (2024) introduced a decision-support framework for optimal deployment of volunteer responders in disaster situations [1]. The DSS they created integrates data and analytics with various coordination interfaces to ensure efficient resource use. However, the framework does not incorporate adaptive logic optimisation or reliability modelling in decision-making processes. A 2021 study by Nozhati proposed an ontological fuzzy AHP technique for the optimal site selection of urban earthquake shelters. Further analysis revealed that their technique combines semantic modelling with multi-criteria decision-making. However, it addresses spatial planning rather than emergency decision-making based on logic [24]. A study analysed the impact of compliance rates on evacuation speed during earthquakes using the Stochastic Pedestrian Cell Transmission Model (SPCTM) (Chang et al., 2024) [25]. The simulation offers recommendations for shelter location and evacuation guidelines. However, it does not have an optimisation layer which can dynamically reconfigure the decision logic. De Miranda et al. (2024) investigated robust optimisation strategies for supply chain resilience of information systems during shocks [23]. According to their study, uncertainty modelling and performance indexing are valuable. The study focuses on isolated sub-processes rather than overall emergency decision-making as a unified optimisable logic framework.
In the logistics domain, the allocation of emergency resources has been optimised using mathematical and machine-learning methods in [26,27,28,29,30]. For instance, Zamanifar and Hartmann’s (2020) survey proposes an optimisation-based model for recovering a transportation network that accounts for restoration costs and time [13]. Although their scheme offered improved resilience, it was unable to adapt to events. Li et al. (2025) utilised mixed-integer programming to solve the multi-stage resource deployment problem under uncertain demand [31]. It was found that using deterministic logic in the allocation process led to suboptimal allocations when incident patterns were altered. Hu et al. (2022) proposed a deep reinforcement-learning framework for multi-objective ambulance dispatching with reduced response time; however, it had a very high computational cost and was not evaluated for reliability consistency [32].
The works demonstrate that while optimisation techniques enhance efficiency, insufficient attention is paid to logic calibration. This entails ensuring that a logical relationship exists between decision variables, reliability thresholds, and learning feedback, which must occur mathematically consistently [7,28,29,30]. Reliability assessment is the backbone of every emergency system evaluation. According to Jesus et al. (2024), uncertainty propagation in emergency operations can be quantified using principles of reliability engineering [33]. However, they note that the investigation into linking reliability to machine-learning decision engines is poorly understood. According to new research by the RAND Corporation [31], large-scale incident operations depend on components whose reliability requires stochastic modelling of failure modes and their locations. Recent empirical studies, such as those by M., suggest that hybrid models, which combine data-driven prediction with reliability thresholds, can lead to a more robust process. According to Yazdi (2024), this approach can enhance the reliability of the process [34]. Nonetheless, none of these works explicitly formulate the decision-logic layer as an object of reliability optimisation. The existence of this gap drives the CLOF to regard reliability not just as a metric for evaluation, but as a constraint affecting model optimisation.
From the reviewed literature, several consistent gaps emerge, underscoring the need for further innovation in emergency decision-making systems [4,7,28,29,30]. First, most prior studies have focused on optimising resource allocation or response outcomes while neglecting the unification and optimisation of the calculation logic itself, the computational mechanism that maps input data into actionable decisions. Second, the reliability measures, despite their standard reporting or use as performance measures, are rarely applied (either enforced or optimised) as explicit constraints within a learning solution, leading to models that are statistically competent but fragile under uncertainty. Third, existing approaches typically rely on simulated or synthetic datasets, which limit their empirical robustness and scalability in real-world applications. These limitations indicate that we need a comprehensive framework that simultaneously optimises the calculation logic, embeds reliability metrics into the optimisation task, and validates its performance using large-scale, authentic emergency data. To address the above gaps, the present work recommends the Calculation Logic Optimisation Framework (CLOF), a mathematical foundation and data-driven architecture that combines prediction, multi-objective optimisation, and reliability analysis. The CLOF enhances current research by shifting decision logic from a fixed rule set to an adaptive, optimisable function validated on more than half a million real emergency records.
The above review reveals three clearly defined research gaps that directly motivate the proposed contributions of this work. First, while many studies optimise response outcomes or resource allocation, they do not treat the internal calculation logic itself as an explicit optimisation object; this gap is addressed by our first contribution, which formulates decision logic as a formal, optimisable function. Second, although reliability is frequently reported as an evaluation metric, it is rarely embedded as an enforceable constraint during learning; this limitation motivates our second contribution, which integrates reliability-aware constraints into the optimisation process. Third, most existing approaches are validated on small-scale or simulated datasets, limiting their real-world applicability; our third contribution addresses this gap by providing large-scale validation using more than 500,000 real emergency records.
Overall, recent studies published between 2022 and 2025 demonstrate a clear trend toward intelligent, data-driven emergency decision-support systems. Nevertheless, these works typically address isolated components such as prediction accuracy, resource allocation, or post-event analysis. Few studies explicitly model the decision calculation logic itself as a unified, optimisable entity under reliability constraints, particularly when validated on large-scale real-world emergency datasets. This limitation highlights the need for a framework that integrates predictive learning, logic optimisation, and reliability control in a single coherent architecture.
To make this research gap explicit, Table 1 provides a concise comparison between existing studies and the proposed CLOF.
Table 1 highlights that, unlike recent emergency decision-support studies published between 2022 and 2025, the proposed CLOF uniquely treats calculation logic as an explicit optimisation object and validates its effectiveness using large-scale real-world data.

3. Research Background and Problem Definition

The process of emergency management draws on rapid decision-making, uncertainty, temporal limitation and resource scarcity. Improvements in planning and prediction have been made possible by advances in artificial intelligence and optimisation. Nonetheless, most frameworks rely on static calculation logic. Thus, they cannot create dynamic responses as the emergency context changes [1,35]. Current methods focus on a single element at a time, such as resource allocation or volunteer coordination. However, there is rarely any focus on the decision logic itself. Additionally, people often treat reliability as a benchmark for assessment rather than as an optimisation. There is a need for a common framework to enhance a system’s predictive ability, reduce response time, and facilitate informed decision-making [4,34].
The CLOF is inherently designed for multi-type and highly dynamic emergency scenarios, such as 911 call systems, where incident types, spatial distributions, and temporal patterns evolve continuously. By decoupling predictive learning from decision-logic optimisation, the framework enables the prediction engine to capture heterogeneous incident characteristics. Simultaneously, the logic-optimisation layer adaptively recalibrates decision thresholds in response to varying operational constraints. Importantly, this architecture is model-agnostic and transferable: by redefining feature representations, optimisation objectives, and reliability constraints, the CLOF can be readily applied to other emergency contexts, including natural disasters and industrial accidents, thereby supporting theoretical generalisability beyond the 911 domain.

3.1. Research Questions

After reviewing the in-depth literature and studies that try to address the research gaps, our research is based on and guided by these questions:
  • RQ1: How can the internal calculation logic of emergency decision systems be mathematically formulated to capture predictive, operational, and reliability dimensions?
  • RQ2: Can optimising this calculation logic significantly reduce average response time while maintaining reliability above 90%?
  • RQ3: What evaluation framework can effectively assess the stability and robustness of an optimised logic model under uncertainty?

3.2. Problem Setup

The novel study aims to model and optimise the internal calculation logic associated with emergency decision-making. To further achieve the aim and objective, we utilised the 911 Emergency Calls dataset [36]. There are several key characteristics of an emergency incident, including location, type, and time. The optimisation focuses on maximising response time while ensuring maximum reliability within the given constraints. The symbols described below define the mathematical framework.
Let a feature vector represent each emergency record:
X = x 1 , x 2 , , x n
where x i denotes the i t h feature and n is the total number of records.
The true response or priority label, expressed by y { 1,2 , 3 } , indicates Fire, EMS, and Traffic categories.
Each feature is standardised to eliminate scale bias:
x j = x j μ j σ j
where μ j and σ j represent the mean and standard deviation of the j th   feature.
The predictive model is represented as
y ˆ = f θ ( X )
where f θ is a learning function parameterised by θ .
The following is used to maintain reliable behaviour and minimise response time:
Minimise :   J ( θ ) = i = 1 n   w i T i ( θ )
Subject to the following reliability and capacity constraints:
R ( θ ) R 0 , C ( θ ) C m a x , θ Θ
where
T i ( θ ) = predicted response time for case i ;
R ( θ ) = reliability measure (e.g., R 2 );
C ( θ ) = total resource cost;
and R 0 ,   C max are predefined operational thresholds.
The complete optimisation loss function integrates temporal effectiveness and reliability:
L ( θ ) = α L time   + ( 1 α ) L reliab  
where L time   represents delay loss and L reliab   penalises deviations from the reliability target.
The optimal parameters are obtained by
θ * = a r g   m i n θ L ( θ )
The Calculation Logic Optimisation Framework (CLOF) is formulated as a multi-objective optimisation problem that maximises the response efficiency, reliability, and resource utilisation of an emergency decision system.

4. Methodology

The proposed Calculation Logic Optimisation Framework (CLOF) is a data-driven method consisting of four stages for a real-world design and analysis process. These are data pre-processing, feature engineering, model formulation and optimisation, and evaluation. This workflow converts emergency call records into more effective decision-making logic. It does this quickly, reliably and without wasting resources. The entire framework consists of the data pipeline, the optimisation engine, and the reliability evaluation. This is as seen in Figure 1.

4.1. Data Pre-Processing

The 911 Emergency Calls original dataset contains over 500,000 records with heterogeneous attributes, including textual descriptions of calls, categorical labels, and timestamp details. Scaling, encoding, and noise reduction are applied to the model to prepare it for transmission to the system.
Records with missing spatial coordinates or timestamps constitute approximately 2.3% of the original dataset. An exploratory analysis of these records shows no concentration in specific townships or time periods, indicating that the missingness is approximately random rather than systematic. As a result, removing these records does not introduce observable spatial or temporal bias. A sensitivity check comparing key performance metrics with and without these records confirmed that their exclusion has a negligible impact on model behaviour and overall conclusions.
1.
Handling Missing and Noisy Data
Records with missing coordinates or timestamps are omitted. Together, textual fields (title and twp) are standardised through lowercasing, tokenisation, removal of redundant variants, and consolidation of semantically equivalent labels to ensure consistency and reproducibility.
The cleaning process is represented as
X = x i X n o t n u l l x i
where X denotes the filtered dataset.
2.
Temporal Feature Transformation
Timestamp t i is decomposed into components for an hour h i , day d i , month m i , and weekday w i to capture temporal periodicity:
t i h i , d i , m i , w i
3.
Categorical Encoding
Categorical variables such as title (type of emergency) and twp (township) are converted to numerical vectors through one-hot encoding:
x i j = 1 ,   If   feature   j   is   present   in   record   i 0 ,   otherwise  
4.
Normalisation
Continuous features like geographic coordinates and call frequency are standardised to zero mean and unit variance:
x j * = x j μ j σ j
where μ j and σ j denote the mean and standard deviation of feature j .
Normalisation ensures comparability across variables and stabilises the gradient-based optimisation.
5.
Data Partitioning
The dataset is divided into training ( 70 % ), validation ( 15 % ), and test ( 15 % ) subsets:
D = D train   D val   D test   , D train   D val   D test   =
This dataset division supports performance evaluation and hyperparameter tuning. Figure 2 illustrates the exploratory features of the dataset, which include the class distribution (i.e., EMS, Traffic, Fire) and the pattern of call frequency over 24 h. It reflects the data balance and time variability used to train the model.

4.2. Feature Engineering

The goal of feature engineering is to extract higher-level quantitative descriptions from cleaned data, thereby enhancing the model’s expressiveness and the precision of its decision-making logic. The features we engineer have mathematical definitions to ensure repeatability and consistency in the end product.
1.
Spatial Aggregation Features
For each township k , compute the average number of calls per time interval τ :
λ k = 1 | τ | i τ k   1
This captures spatial demand intensity for assessing emergency-service load distribution.
2.
Temporal Frequency Encoding
Hourly call density f h helps model peak and off-peak response periods:
f h = N h N total  
where N h denotes the number of calls in hour h , and N total   is the total number of calls.
3.
Emergency Type Weighting
To balance category imbalance among Fire, EMS, and Traffic, class weights are computed as inverse frequency:
w c = N t o t a l 3 × N c
These weights are incorporated into the loss function during optimisation.
4.
Composite Reliability Index Feature
To measure the time consistency of responses, a derived reliability index is estimated for each historical period:
R t = 1 T t o b s T t p r e d T t o b s
The composite reliability index R t is computed over a 24 h rolling historical window, selected to capture short-term operational stability in emergency response systems. By comparing observed T t obs   and predicted mean T t pred   response times within this window, R t reflects the temporal consistency of decision outcomes, with higher values indicating more stable and reliable response-time behaviour across consecutive operational periods. Higher R t indicates better reliability over that period.
5.
Correlation and Dimensionality Reduction
Here, using Pearson correlation, we detect the highly correlated features, which are defined as
ρ i j = c o v x i , x j σ i σ j
The measurement of linear dependence ρ i j between x i and x j . Figure 3 illustrates that clock time, weekday, and month are strongly correlated, and it is worth noting that they are inherently redundant. To eliminate this redundancy, features that satisfy | ρ i j | > 0.85 are reduced through Principal Component Analysis.
Z = X W , W = a r g   m a x W d e t W S W
where S is the covariance matrix of X .
In fact, Principal Component Analysis (PCA) guarantees computational efficiency while preserving maximum variance to produce a feature space optimised for adaptive logic modelling.
The dataset, which has been pre-processed and engineered and is denoted by X * , supports the CLOF model, i.e., each record in X * reflects raw operational data and optimally engineered statistical features. Therefore, the following section describes the architecture and optimisation of the model. In particular, it represents the mathematical formulation of the gradient-boosted logic engine and the multi-objective loss for adaptive reliability learning.

4.3. Model Architecture and Optimisation

The CLOF aims to bring together predictive learning, multi-objective optimisation, and design reliability evaluation within a single adaptable framework. To ensure the output remains fast and reliable as operational conditions change, the logic used for emergency triage and dispatch is continually rebalanced. Figure 4 illustrates an overview of the computational workflow, comprising a prediction engine, a logic-optimisation layer, and a reliability unit.
1.
Prediction Engine
The conditional probability of the emergency category, given the feature vector x i , is estimated using a prediction engine.
y ˆ i = a r g m a x c { 1,2 , 3 } P θ y i = c x i
where P θ is approximated by a gradient-boosted ensemble:
f θ x i = m = 1 M   γ m h m x i
with h m ( ) denoting the m th   base learner and γ m its weight.
The learning objective of this stage minimises the cross-entropy loss:
L pred   ( θ ) = 1 N i = 1 N   c = 1 3   y i c l o g P θ y i = c x i
This ensures accurate class probability estimation.
2.
Logic-Optimisation Layer
This layer transforms static decision logic into an optimisable function, aiming to create a continuous trade-off between timeliness and reliability.
The optimisation objective is defined as
J ( θ ) = α E [ T ( θ ) ] + ( 1 α ) ( 1 R ( θ ) ) 2
where α ∈ (0,1) denotes the trade-off coefficient controlling the balance between response-time minimisation and reliability preservation.
The trade-off coefficient α controls the relative importance of minimising response time versus preserving reliability. Its value was selected from the range 0.4–0.7 through cross-validation and domain-informed calibration, ensuring stable convergence and avoiding over-emphasis on either objective. Empirically, values outside this range led to either degraded reliability (α > 0.7) or insufficient response-time optimisation (α < 0.4).
Subject to the following reliability and capacity constraints:
R ( θ ) R 0 , C ( θ ) C m a x
where E [ T ( θ ) ] is the expected response time, R ( θ ) the reliability score, and α the trade-off coefficient. Parameter updates follow adaptive gradient descent:
θ t + 1 = θ t η t θ J θ t
And convergence is achieved when
J t + 1 J t < ε 1 , R t + 1 R t < ε 2
signalling stable logic behaviour.
Convergence is determined by two tolerance thresholds, ε1 and ε2, corresponding to the relative change in the optimisation loss and the stability of the reliability constraint, respectively. In this study, ε1 is set to 10−4 and ε2 to 10−3, based on empirical observations that smaller values do not yield meaningful performance gains while increasing computational cost. These thresholds ensure stable convergence given the scale and variance of the real-world emergency data.
The optimisation workflow is illustrated in Figure 5.
3.
Reliability-Evaluation Unit
Reliability is monitored through bootstrapped resampling.
For each bootstrap iteration b ,
R b = 1 1 N b i = 1 N b   T i p r e d T i o b s
And the 95 % confidence interval is computed as
R C I =   Percentile   2.5 R b ,   Percentile   97.5 R b
To ensure statistical stability of the optimised logic.
In this study, bootstrap resampling is performed with B = 1000 iterations, where each resample contains 100% of the original sample size drawn with replacement. For each bootstrap replicate, reliability metrics are recomputed, and the 95% confidence interval is obtained using the percentile method, defined by the 2.5th and 97.5th percentiles of the bootstrap distribution.
Algorithm 1 outlines the complete optimisation process.
Algorithm 1. Calculation Logic Optimisation Framework (CLOF)
1.
Input dataset D = x i , y i i = 1 N ; hyperparameters α , R 0 , C max .
2.
Initialise θ 0 .
3.
Repeat until convergence:
  • Predict y ˆ i = f θ t x i
  • Estimate T i θ t and R θ t
  • Compute loss
    L θ t = α L time   + ( 1 α ) L reliab  
  • Update parameters
    θ t + 1 = θ t η t θ L θ t

Check constraints R θ t + 1 R 0 , C θ t + 1 C max .
4.
Output optimised θ * representing adaptive calculation logic.

4.4. Evaluation Metrics

To assess the proposed novel CLOF, we utilised widely accepted evaluation metrics to compute performance and operational stability for classification-oriented and reliability-oriented regression. Each metric has been mathematically defined, ensuring reproducibility and statistical interpretability [37,38].
4.
Classification Metrics
Accuracy
  Accuracy   = 1 N i = 1 N   I y ˆ i = y i
Precision
  Precision   = T P T P + F P
Recall
  Recall   = T P T P + F N
F1 Score
F 1 = 2 ×   Precision   ×   Recall     Precision   +   Recall  
ROC-AUC
A U C = 0 1   T P R ( t ) d [ F P R ( t ) ]
5.
Reliability Metrics
Mean Absolute Error (MAE)
M A E = 1 N i = 1 N   T i p r e d T i o b s
Coefficient of Determination R 2
R 2 = 1 i = 1 N   T i obs   T i pred   2 i = 1 N   T i obs   T obs   2
A stronger R 2 indicates greater predictive power of response time regression.
6.
Statistical Confidence
A 95 % bootstrap confidence interval ( B = 1000 ) is computed for all metrics to ensure statistical robustness:
R C I =   Percentile   2.5 R b ,   Percentile   97.5 R b
where each R b is calculated from a resampled dataset.
7.
Risk Index
To quantify missed high-priority events, a risk index is defined as
R risk   = 1   False   Negatives     Total   High - Priority   Cases  
Values close to 1 indicate minimal critical-case omission.

4.5. Experimental Setup

To conduct this novel study, we constructed an experimental setup. We rigorously evaluated the reproducibility, stability, and comparative advantage of the suggested framework against base models and recent state-of-the-art work.
1.
Environment and Hardware
The experiment is conducted on Kaggle, using a T4 GPU/CPU VM with 16 GB RAM and Python 3.11, scikit-learn 1.4, XGBoost 1.7, and Optuna 3.5 for Bayesian hyperparameter optimisation.
2.
Data Partitioning
The 911 Emergency Calls dataset was divided as follows:
D = D train   D val   D test   , D train   : D val   : D test   = 70 : 15 : 15
This division helps us confirm balanced training, validation, and testing for reliable generalisation. As shown in Table 2, the data were split into training, validation, and test subsets at a ratio of 70:15:15 for fitting the classifier module and unbiased performance evaluation.
3.
Baseline Models
The performance of the CLOF was validated against four models:
Decision Tree (DT): A single-tree classifier that recursively splits features to minimise Gini impurity or entropy. It provides interpretability but suffers from overfitting on noisy data.
Random Forest (RF): RF is a collection of decision trees generated using bootstrapped samples. Furthermore, it reduces the overall variance. Also, it imparts stability to the output.
XGBoost (XGB): A gradient-boosted framework optimising residual errors iteratively with regularisation; serves as a strong performance baseline.
Proposed CLOF (Ours): Incorporates a multi-objective optimisation layer that jointly minimises response time and maximises reliability, adapting the decision logic dynamically.
4.
Hyperparameter Optimisation
Two optimisation strategies were used:
  • Grid Search helps in rapidly and thoroughly investigating a restricted range of parameter spaces;
  • Bayesian Optimisation (for the CLOF) efficiently searches the high-dimensional parameter space through adaptive probabilistic modelling.
This is also evident in Table 3, where the hyperparameter ranges were chosen to ensure stable convergence of both the baseline and the proposed models. The Bayesian technique was applied to the CLOF to optimise parameters within their pre-specified ranges. It was observed that the tuned CLOF were more reliable. Furthermore, the loss was lower than that in grid search.
5.
Cross-Validation and Bootstrap Evaluation
To validate the proposed novel CLOF, a 10-fold cross-validation procedure was conducted to ensure that each data partition contributed to both training and testing. At the same time, B = 1000 bootstrap replicates were used to assess metric stability and compute confidence intervals:
  Metric   avg   = 1 B b = 1 B     Metric   b
As shown in Figure 13, the CLOF shows some variability across the folds. This implies that it is more generally stable than ensemble baselines. This experimental workflow outlines the steps for splitting, optimising, and validating the data. This type of experimental design guarantees evidence-based epistemological requirements. The CLOF is evaluated against strong baselines under the same experimental conditions, using solid cross-validation and bootstrapping to validate predictive accuracy and reliability.

5. Results and Discussion

The detailed results generated from the 911 Emergency Calls dataset confirm the high accuracy and robustness of the proposed Calculation Logic Optimisation Framework (CLOF). Table 4 presents the overall quantitative comparison between the baseline and proposed models. The CLOF achieved the best overall performance, with 94.68% ± 0.27 accuracy, F1 = 0.938 ± 0.007, and ROC-AUC = 0.971 ± 0.004, surpassing all benchmark algorithms by a substantial margin. The log-loss dropped to 0.081, nearly 25% lower than the next-best XGBoost (0.108), while the MAE of 2.11 ± 0.05 min and R2 of 0.955 ± 0.004 indicate highly consistent response-time prediction.
Table 5 reports the computational efficiency of the CLOF and baseline models, showing that the proposed framework maintains millisecond-level inference latency while achieving training times comparable to those of XGBoost and Random Forest, thereby satisfying real-time operational requirements.
Moreover, we present a confusion matrix visualisation, as depicted in Figure 6. It provides an overview of the proposed CLOF system’s performance across three groups: EMS, Traffic, and Fire. The results confirm that 94.7% of all predictions correspond to the actual class. Thus, there is strong diagonal dominance and high class separability.
According to the class-level analysis from confusion matrix visualisation, at first, the classification of EMS cases showed high precision, with 96% correctly classified. Furthermore, 3.5% of cases were misclassified as traffic. This small-scale cross-leakage between classes is consistent with the real-world overlap observed in many multi-source 911 datasets. A specific example is a dispatch description of imminent injury due to “vehicle collision”. Such phrasing may be classified as both medical and traffic, making it ambiguous. As for the Traffic class, the model correctly predicted the output in 92% of cases, while it confused 7% of cases with EMS, which again shows that the deviation is a real-world data issue, not a mistake on the model’s part. The fire category achieved 94% accuracy, with less than 6% of the samples classified as the other two classes. This demonstrates that the model maintains recall for relatively underrepresented samples as well.
In Figure 7, a closer look at XGBoost’s baseline reveals that it suffers from much greater cross-class confusion than the chosen classifiers. This is especially true of the EMS and Traffic classes. As such, this demonstrates that the proposed CLOF model exhibits significantly greater discriminative stability compared to the other baselines. The CLOF indicates significant increases in decision-making consistency. The classic Decision Tree had around 14% off-diagonal errors, while Random Forest and XGBoost had 10% and 8%, respectively. The hybrid optimisation of response-time and reliability metrics by the CLOF reduces this misclassification to ≈ 4%, representing a 10 percentage point improvement over conventional logic. The improvement demonstrates that the suggested calculation-logic-optimisation layer enables the model to continuously adjust decision thresholds, thereby enhancing intra-class discrimination and operational reliability.
As evidenced in Figure 8, the discriminative ability is validated through a series of ROC curves. The TPR of the proposed model is higher than that of the baseline model for all FPRs. Furthermore, the AUC of the proposed CLOF is approximately 0.97 ± 0.004. This advancement over XGBoost (0.951) and Random Forest (0.924) indicates that the CLOF’s reliability-aware optimisation enhances both probability calibration and classification boundaries. According to [9], improvements in AUC were achieved, but they plateaued at an R2 of 0.93. Similarly, the CLOF’s integrated optimisation layer offers superior threshold consistency compared to others.
The behaviour during training is shown in Figure 9 (Accuracy vs. Epochs) and Figure 10 (Loss vs. Epochs). The system’s accuracy consistently increases and stabilises at around 95% after approximately 60 epochs, while the loss function reaches a final value of 0.081. These results confirm the system’s smooth, stable convergence through multi-objective optimisation. On the other hand, the XGBoost baseline shows oscillatory behaviour beyond a loss of 0.10. This should indicate that it is sensitive to class imbalance. The CLOF’s dual-term objective
J ( θ ) = α E [ T ( θ ) ] + ( 1 α ) ( 1 R ( θ ) ) 2
Adjusts for fluctuations by linking response-time and reliability learning for every epoch.
Reliability analysis, summarised in Table 6, reinforces the CLOF’s superior temporal stability. Bootstrap replication (B = 1000) confirms narrow confidence intervals: R2 = 0.955 ± 0.004 and MAE = 2.11 ± 0.05 min. In comparison, the R2 for XGBoost was 0.928 with an MAE of 2.54 min. This means there has been a reduction of about 17% in mean response-time error, implying that adding reliability constraints to the optimisation objective directly improves the system’s operational reliability.
As shown in Figure 11, the Monte Carlo sensitivity analysis indicates that R 0 = 0.95 lies near the inflexion point of system availability, beyond which marginal performance gains diminish.
An analysis of the generated output using SHAP values reveals the model’s predictor drivers. From Figure 12, we see that time-related global attributes are the most important. Specifically, we have the hour of the day, the weekday, and the month. Among them, the hour of the day is the most significant. We also have spatial township identifiers, call types, and other spatiotemporal attributes. The results are similar to those obtained by Sun et al. (2021) [4]. Still, the CLOF’s interpretability is further enhanced with explanations on the reliability contribution of each feature through the logic-weight coefficients embedded in the CLOF.
As illustrated in Figure 13, the results from 10-fold validation show a mean accuracy of 94.65% (standard deviation = 0.0039), R2 = 0.955 ± 0.004, indicating low dispersion and confirming the model’s high reproducibility. The stability is better than that of any ensemble baseline (σ ≈ 0.010), indicating that the multi-objective optimisation mitigates overfitting and balances precision and reliability.
In conclusion, the experimental results demonstrate that the proposed CLOF achieves the highest predictive accuracy and the most reliable dispatch-time estimation. The average response time decreases by approximately 18%, while reliability remains above 0.95 on all test folds. Such progress is evident in various metrics, demonstrating that the CLOF can transform a conventional, deterministic logic into a self-optimising, reliability-driven decision-making mechanism. In real-world scenarios, such optimisation enables emergency agencies to assign resources on the fly, thereby reducing operational delays and increasing safety assurance compared to earlier approaches reported by Yazdani and Haghani (2024), and Nagy et al. (2024) [1,23]. As a result, the CLOF not only realises quantitative superiority but also provides strategic multipurpose flexibility and mathematical consistency, propelling the progress of emergency management analytics.

6. Conclusions and Future Work

Given surrounding time demands, this study proposes a novel CLOF for emergency management decision-making that integrates predictive learning, multi-objective optimisation, and reliability control. The Calculation Logic Optimisation Framework (CLOF) is a data-driven framework. The novel study demonstrates that the new CLOF substantially enhances the classical and ensemble baselines in both quality and quantity. The publicly available 911 Emergency Calls dataset used in this study shows the improvement. The CLOF achieved the promised results, with an overall accuracy of 94.68%, an F1 score of 0.938, and an ROC-AUC of 0.971, significantly outperforming the best baseline, XGBoost, with 92.36% accuracy and an AUC of 0.951. In addition, log-loss dropped to 0.081, and response-time reliability (R2 = 0.955) improved, with almost a 3% gain over the next-best rival. In addition, the mean absolute error of dispatch-time prediction fell to just 2.11 min. The CLOF’s promising results indicate that integrating reliability-aware constraints into the optimisation process enables the CLOF to deliver stable, high-confidence performance across varying emergency scenarios, surpassing recent state-of-the-art work. This research provides a replicable methodological framework for optimising calculation logic in advanced decision-making systems, extending beyond numerical enhancements. The CLOF employs an advanced approach, distinct from earlier rule-based or fixed-threshold strategies. The CLOF involves adaptable thresholding and treats the decision function itself as an optimisation objective. Moreover, it allows for adaptive recalibrations as more data is gathered using real-time feedback. This changes the game for emergency response, shifting from fixed, procedural scheduling to self-correcting, analytical decision-making. As a result, our novel CLOF demonstrates that data-centric optimisation reduces operational latency by approximately 18% while maintaining a reliability index above 0.95, providing a mathematically elegant and operationally robust basis for safety-critical settings.
In future work, we will utilise streaming data pipelines and API integration to deploy this framework in real-time and recalibrate it when a new incident occurs. Integrating social media signals, such as crowdsourced data from the CrisisLex repository, can serve as an early warning system during crises and disaster-related events. From a computational perspective, the CLOF introduces only marginal additional overhead compared to traditional ensemble baselines, as its logic-optimisation and reliability evaluation layers operate on aggregated statistics rather than iterative inference, thereby preserving millisecond-level real-time feasibility, even when extended with external data streams such as social media signals. In conclusion, assessments that utilise data from upscaled multi-county or national-level datasets will test the model’s scalability and interoperability across heterogeneous administrative and infrastructural settings. This roadmap aims to take the CLOF beyond an experimentally validated research prototype to a fully functioning intelligent emergency decision platform that can support next-generation public safety systems.

Author Contributions

Y.H. carried out conceptualisation, methodology, and formal analysis. Data curation, experimental design, and result interpretation were performed jointly by Y.H. and K.W. All authors have read and agreed to the published version of the manuscript.

Funding

No external funding for this project.

Data Availability Statement

The dataset used in this research, “Emergency 911 Calls,” is publicly accessible on Kaggle: https://www.kaggle.com/datasets/mchirico/montcoalert (accessed on 15 November 2025).

Acknowledgments

The authors are grateful to the university for its support in conducting this novel study and acknowledge the Open Data from the Montgomery County 911 Emergency Call Records, which enabled the experimental validation of this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yazdani, M.; Haghani, M. A conceptual framework for integrating volunteers in emergency response planning and optimization assisted by decision support systems. Prog. Disaster Sci. 2024, 24, 100361. [Google Scholar] [CrossRef]
  2. Ongesa, T.N. Optimizing emergency response systems in urban health crises: A project management approach to public health preparedness and response. Medicine 2025, 104, e41279. [Google Scholar] [CrossRef]
  3. Caldera, H.J.; Wirasinghe, S.C. A universal severity classification for natural disasters. Nat. Hazards 2021, 111, 1533–1573. [Google Scholar] [CrossRef] [PubMed]
  4. Li, T.; Sun, J.; Fei, L. Application of Multiple-Criteria Decision-Making Technology in Emergency Decision-Making: Uncertainty, Heterogeneity, Dynamicity, and Interaction. Mathematics 2025, 13, 731. [Google Scholar] [CrossRef]
  5. Khan, S.M. A Systematic Review of Disaster Management Systems: Approaches, Challenges, and Future Directions. Land 2023, 12, 1514. [Google Scholar] [CrossRef]
  6. Daud, A.; Al Abdouli, K.M.; Badshah, A. Emerging Computing Tools for Emergency Management: Applications, Limitations and Future Prospects. IEEE Open J. Comput. Soc. 2025, 6, 627–644. [Google Scholar] [CrossRef]
  7. Potur, E.A.; Aktas, A.; Kabak, M. A Bibliometric Analysis of Multi-Criteria Decision-Making Techniques in Disaster Management and Transportation in Emergencies: Towards Sustainable Solutions. Sustainability 2025, 17, 2644. [Google Scholar] [CrossRef]
  8. Jiang, Z.; Ji, R. Optimising hurricane shelter locations with smart predict-then-optimise framework. Int. J. Prod. Res. 2024, 63, 2905–2925. [Google Scholar] [CrossRef]
  9. Attiah, A.; Kalkatawi, M. AI-powered smart emergency services support for 9-1-1 call handlers using textual features and SVM model for digital health optimization. Front. Big Data 2025, 8, 1594062. [Google Scholar] [CrossRef]
  10. Su, W.; Chen, L.; Gao, X. Emergency Decision Making: A Literature Review and Future Directions. Sustainability 2022, 14, 10925. [Google Scholar] [CrossRef]
  11. Zhou, L.; Wu, X.; Xu, Z.; Fujita, H. Emergency decision making for natural disasters: An overview. Int. J. Disaster Risk Reduct. 2018, 27, 567–576. [Google Scholar] [CrossRef]
  12. D’Alessio, I. ‘Leading through Crisis’: A Systematic Review of Institutional Decision-Makers in Emergency Contexts. Behav. Sci. 2024, 14, 481. [Google Scholar] [CrossRef]
  13. Zamanifar, M.; Hartmann, T. Optimization-based decision-making models for disaster recovery and reconstruction planning of transportation networks. Nat. Hazards 2020, 104, 1–25. [Google Scholar] [CrossRef]
  14. Fertier, A.; Barthe-Delanoë, A.M.; Montarnal, A.; Truptil, S.; Bénaben, F. A new emergency decision support system: The automatic interpretation and contextualisation of events to model a crisis situation in real-time. Decis. Support Syst. 2020, 133, 113260. [Google Scholar] [CrossRef]
  15. Zuo, M. System reliability and system resilience. Front. Eng. Manag. 2021, 8, 615–619. [Google Scholar] [CrossRef]
  16. Yang, J.; Hou, H.; Hu, H. Exploring the Intelligent Emergency Management Mode of Rural Natural Disasters in the Era of Digital Technology. Sustainability 2024, 16, 2366. [Google Scholar] [CrossRef]
  17. Kyrkou, C.; Kolios, P.; Theocharides, T.; Polycarpou, M. Machine Learning for Emergency Management: A Survey and Future Outlook. Proc. IEEE 2023, 111, 19–41. [Google Scholar] [CrossRef]
  18. Nazir, R. A review on machine learning techniques for network security. J. Cyber Secur. Technol. 2025, 1–45. [Google Scholar] [CrossRef]
  19. Dahri, F.H.; Laghari, A.A.; Sajnani, D.K.; Shazia, A.; Kumar, T. Heart failure prediction: A comparative analysis of machine learning algorithms. In International Conference on Optics, Electronics, and Communication Engineering (OECE 2024); SPIE: Bellingham, WA, USA, 2024; pp. 634–640. [Google Scholar]
  20. Siraj, S.; Dahri, F.H.; Chandio, J.A.; Jalbani, A.H.; Laghari, A.A. Comparison of machine learning techniques to predict students’ CGPA by using course learning outcomes datasets. Hum.-Intell. Syst. Integr. 2025, 1–11. [Google Scholar] [CrossRef]
  21. Bouramdane, A.-A. Enhancing disaster management in smart cities through MCDM-AHP analysis amid 21st century challenges. Inf. Syst. Smart City 2023, 3, 189. [Google Scholar] [CrossRef]
  22. Elkady, S.; Hernantes, J.; Labaka, L. Decision-making for community resilience: A review of decision support systems and their applications. Heliyon 2024, 10, e33116. [Google Scholar] [CrossRef] [PubMed]
  23. Nagy, M.; de Miranda, J.L.; Popescu-Bodorin, N. Decision Making and Robust Optimization for Information Systems Oriented to Emergency Events. Int. J. Comput. Commun. Control. 2024, 19, 1–11. [Google Scholar] [CrossRef]
  24. Nozhati, S. A resilience-based framework for decision making based on simulation-optimization approach. Struct. Saf. 2021, 89, 102032. [Google Scholar] [CrossRef]
  25. Chang, K.H.; Wu, Y.Z.; Su, W.R.; Lin, L.Y. A simulation evacuation framework for effective disaster preparedness strategies and response decision making. Eur. J. Oper. Res. 2024, 313, 733–746. [Google Scholar] [CrossRef]
  26. Yazdani, M.; Shahriari, S.; Haghani, M. Real-time decision support model for logistics of emergency patient transfers from hospitals via an integrated optimisation and machine learning approach. Prog. Disaster Sci. 2025, 25, 100397. [Google Scholar] [CrossRef]
  27. Xia, H. Emergency medical supplies scheduling during public health emergencies: Algorithm design based on AI techniques. Int. J. Prod. Res. 2025, 63, 628–650. [Google Scholar] [CrossRef]
  28. Pu, F.; Li, Z.; Wu, Y.; Ma, C.; Zhao, R. Recent advances in disaster emergency response planning: Integrating optimization, machine learning, and simulation. Saf. Emerg. Sci. 2025, 1, 9590007. [Google Scholar] [CrossRef]
  29. Tluli, R.; Badawy, A.; Salem, S.; Barhamgi, M.; Mohamed, A. A Survey of Machine Learning Innovations in Ambulance Services: Allocation, Routing, and Demand Estimation. IEEE Open J. Intell. Transp. Syst. 2024, 5, 842–872. [Google Scholar] [CrossRef]
  30. Kumar, S.; Kumar, S.; Shiwlani, A. Machine Learning for Labor Optimization: A Systematic Review of Strategies in Healthcare and Logistics. Pakistan Soc. Sci. Rev. 2025, 9, 631–651. [Google Scholar]
  31. Li, H.; Yu, D.; Zhang, Y.; Yuan, Y. A two-stage robust optimization model for emergency service facilities location-allocation problem under demand uncertainty and sustainable development. Sci. Rep. 2025, 15, 2895. [Google Scholar] [CrossRef]
  32. Hu, C.; Wang, Q.; Gong, W.; Yan, X. Multi-objective deep reinforcement learning for emergency scheduling in a water distribution network. Memetic Comput. 2022, 14, 211–223. [Google Scholar] [CrossRef]
  33. Jesus, T.C.; Portugal, P.; Costa, D.G.; Vasques, F. Reliability and Detectability of Emergency Management Systems in Smart Cities under Common Cause Failures. Sensors 2024, 24, 2955. [Google Scholar] [CrossRef]
  34. Yazdi, M. Reliability-Centered Design and System Resilience. In Advances in Computational Mathematics for Industrial System Reliability and Maintainability; Springer: Cham, Switzerland, 2024; pp. 79–103. [Google Scholar] [CrossRef]
  35. Turgay, S.; Aydin, A. Improving decision making under uncertainty with data analytics: Bayesian networks, reinforcement learning, and risk perception feedback for disaster management. J. Decis. Anal. Intell. Comput. 2025, 5, 25–51. [Google Scholar] [CrossRef]
  36. Yoon, N.K.; Quinn, T.D.; Furek, A.; Payne, N.Y.; Haas, E.J. Improving the usability of large emergency 911 data reporting systems: A machine learning case study using emergency incident descriptions. J. Saf. Res. 2025, 93, 335–341. [Google Scholar] [CrossRef]
  37. Naidu, G.; Zuva, T.; Sibanda, E.M. A Review of Evaluation Metrics in Machine Learning Algorithms. Lect. Notes Netw. Syst. 2023, 724, 15–25. [Google Scholar] [CrossRef]
  38. Obi, J.C. A comparative study of several classification metrics and their performances on data. World J. Adv. Eng. Technol. Sci. 2023, 8, 308–314. [Google Scholar] [CrossRef]
Figure 1. Calculation Logic Optimisation Framework (CLOF) overall architecture.
Figure 1. Calculation Logic Optimisation Framework (CLOF) overall architecture.
Systems 14 00139 g001
Figure 2. Exploratory Data Analysis (EDA) presentation class distribution and temporal call frequency.
Figure 2. Exploratory Data Analysis (EDA) presentation class distribution and temporal call frequency.
Systems 14 00139 g002
Figure 3. Feature-correlation heatmap.
Figure 3. Feature-correlation heatmap.
Systems 14 00139 g003
Figure 4. CLOF detailed architecture.
Figure 4. CLOF detailed architecture.
Systems 14 00139 g004
Figure 5. The CLOF’s logic-optimisation process shows the interaction between response-time reduction and reliability feedback.
Figure 5. The CLOF’s logic-optimisation process shows the interaction between response-time reduction and reliability feedback.
Systems 14 00139 g005
Figure 6. Confusion matrix of the proposed CLOF model, showing strong diagonal dominance and <4% cross-class error.
Figure 6. Confusion matrix of the proposed CLOF model, showing strong diagonal dominance and <4% cross-class error.
Systems 14 00139 g006
Figure 7. Confusion matrix of the best baseline model (XGBoost), showing an accuracy of approximately 92.4%.
Figure 7. Confusion matrix of the best baseline model (XGBoost), showing an accuracy of approximately 92.4%.
Systems 14 00139 g007
Figure 8. ROC curves for all models with shaded 95% CI; CLOF attains the highest AUC ≈ 0.97.
Figure 8. ROC curves for all models with shaded 95% CI; CLOF attains the highest AUC ≈ 0.97.
Systems 14 00139 g008
Figure 9. Learning curve (Accuracy vs. Epochs) showing a smooth rise to 95%.
Figure 9. Learning curve (Accuracy vs. Epochs) showing a smooth rise to 95%.
Systems 14 00139 g009
Figure 10. Loss vs. Epochs curve illustrating monotonic decline to 0.081.
Figure 10. Loss vs. Epochs curve illustrating monotonic decline to 0.081.
Systems 14 00139 g010
Figure 11. Monte Carlo sensitivity analysis of system availability versus reliability threshold R 0 .
Figure 11. Monte Carlo sensitivity analysis of system availability versus reliability threshold R 0 .
Systems 14 00139 g011
Figure 12. The SHAP summary plot shows the 10 features that most influence the CLOF.
Figure 12. The SHAP summary plot shows the 10 features that most influence the CLOF.
Systems 14 00139 g012
Figure 13. Boxplot comparison of model performance across 10-fold cross-validation. The left group reports classification accuracy, while the right group reports the coefficient of determination (R2). The Y-axis denotes the corresponding evaluation score for each metric.
Figure 13. Boxplot comparison of model performance across 10-fold cross-validation. The left group reports classification accuracy, while the right group reports the coefficient of determination (R2). The Y-axis denotes the corresponding evaluation score for each metric.
Systems 14 00139 g013
Table 1. Comparison between existing studies and this work.
Table 1. Comparison between existing studies and this work.
Study CategoryOptimisation ObjectConstraint TypeData Scale
Existing emergency decision studiesResource allocation, routing, or isolated decision modulesFixed or implicit constraints (time, cost, capacity)Small-scale, simulated, or limited real-world datasets
This study (CLOF)Internal calculation logic of emergency decision-makingExplicit multi-objective constraints (response time, reliability, resource capacity)Large-scale real-world data (>500,000 911 emergency records)
Table 2. Summary of the 911 Emergency Calls dataset and partition statistics.
Table 2. Summary of the 911 Emergency Calls dataset and partition statistics.
SubsetRecordsPercentageDescription
Training set (D_train)350,00070%Used for model fitting and optimisation
Validation set (D_val)75,00015%Used for hyperparameter tuning
Test set (D_test)75,00015%Used for unbiased model evaluation
Total500,000+100%Montgomery County Emergency 911 call logs
Table 3. Parameter ranges used during model optimisation.
Table 3. Parameter ranges used during model optimisation.
ParameterRange/Value
Learning   Rate   η 0.01–0.3
Max Depth3–10
Subsample0.6–1.0
α (weight for time loss)0.4–0.7
Number of Trees (M)100–500
Table 4. Proposed (CLOF) performance result comparison against baseline models.
Table 4. Proposed (CLOF) performance result comparison against baseline models.
ModelAccuracy (%)PrecisionRecallF1-ScoreROC-AUCLog LossMAE (min)R2 (Reliability)
Decision Tree (Classic Logic)86.720.8530.8420.8470.8980.1633.420.874
Random Forest89.540.881 s0.8720.8760.9240.1382.960.902
Gradient Boosting (G-Boost)91.120.9010.8880.8940.9380.1212.710.917
XGBoost (Optimised Logic)92.360.9160.9060.9110.9510.1082.540.928
Proposed CLOF (ours)94.68 ± 0.270.941 ± 0.0060.936 ± 0.0080.938 ± 0.0070.971 ± 0.0040.081 ± 0.0072.11 ± 0.050.955 ± 0.004
Table 5. Computational efficiency comparison of the CLOF and baseline models.
Table 5. Computational efficiency comparison of the CLOF and baseline models.
ModelTraining Time (min)Inference Time (ms/Sample)Real-Time Feasibility
Linear Regression2.11.2
Random Forest12.63.9
XGBoost13.84.1
CLOF (Proposed)14.04.3
Table 6. Reliability and stability metrics with 95% bootstrap CI for each model.
Table 6. Reliability and stability metrics with 95% bootstrap CI for each model.
ModelMAE (min)R295% CI (Low–High)
Random Forest2.960.902(0.895–0.910)
XGBoost2.540.928(0.921–0.933)
CLOF (Ours)2.11 ± 0.050.955 ± 0.004(0.951–0.959)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hang, Y.; Wang, K. Optimising Calculation Logic in Emergency Management: A Framework for Strategic Decision-Making. Systems 2026, 14, 139. https://doi.org/10.3390/systems14020139

AMA Style

Hang Y, Wang K. Optimising Calculation Logic in Emergency Management: A Framework for Strategic Decision-Making. Systems. 2026; 14(2):139. https://doi.org/10.3390/systems14020139

Chicago/Turabian Style

Hang, Yuqi, and Kexi Wang. 2026. "Optimising Calculation Logic in Emergency Management: A Framework for Strategic Decision-Making" Systems 14, no. 2: 139. https://doi.org/10.3390/systems14020139

APA Style

Hang, Y., & Wang, K. (2026). Optimising Calculation Logic in Emergency Management: A Framework for Strategic Decision-Making. Systems, 14(2), 139. https://doi.org/10.3390/systems14020139

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop