5. Results and Discussion
In this section, a synthesis of existing research is presented on the quantification of the multidimensional impact of cyber security attacks. The synthesis details existing research findings investigating the quantification of the multidimensional impact of cyber security attacks. These findings show the key areas of consensus and disagreement within the existing literature and reflect the addressable gaps. The findings also provide a comprehensive overview of the current state-of-the-art in quantifying the multidimensional impact of cyber security attacks.
RQ1: What are the various cyber security impact terminologies used in academic research papers?
To address this research question, this paper completed a thematic mapping of cyber impact terminologies. The thematic mapping provided a lens to identify the most frequent, infrequent, and absent terminologies, excluding duplicate terms with identical meanings. The themes used within the thematic mapping are physical/digital, economic, psychological, reputational, and societal (
Figure 1). We adopted these themes from Bouveret [
32], Greenfield et al. [
34], Lis et al. [
35], and Agrafiotis et al.’s [
30] work due to its robust coverage of all the cyber security impact terminologies. These themes and their associated cyber impact terminologies are as follows:
This aspect describe the consequences of cyber attacks on material or information technology. We analysed the usage of physical and digital impact terminology across relevant academic research work. Subsequently, we compared the retrieved terms against a comprehensive set of terms gathered from the reviewed research. This mapping revealed the most frequent and infrequent terminology related to physical and digital impacts. The comprehensive set of physical and digital impact terminology used are compromised, infected, corrupted, unavailable, theft, exposed or leaked, bodily injury, loss of life and damage/destruction of equipment. This comprehensive set of terms represents a unique list of core physical and digital cyber security impact terminologies, excluding secondary impacts and synonyms. The analysis of the physical and digital cyber security impact terminologies is presented in
Table 5 including citations, and the comprehensive set of physical/digital impact terminologies used across the selected articles.
An analysis of related academic papers reveals that “theft” and “compromised” are the most frequent physical/digital cyber security terms, appearing in 19 and 16 out of 23 research papers, respectively. Following in frequency were “damaged/destruction” (6) and “infected” (5). The least frequent terminologies, each appearing less than 2 across the 23 papers, were “corrupted” [
30], “unavailable” [
30], “exposed” [
30,
71], “bodily injury” [
30], and “loss of life” [
30,
35,
77]. A plausible reason for this frequency distribution could be because “theft” and “compromised” have established legal definitions as referenced in the Theft Act 1968 [
82] and the Computer Misuse Act [
83] making them safe words. Other reasons could be researcher bias or lack of a defined domain for physical/digital impacts within the taxonomy of cyber security impacts. As this research has evidenced that no defined domain exists for physical/digital impacts within the taxonomy of cyber security impacts, it should be the most likely cause for this contextual gap. Moreover, terms like “corrupted” [
30], “unavailable” [
30], “exposed” [
30,
71], “bodily injury” [
30], and “loss of life” [
30,
35,
77], are valid and unique terminologies for describing the physical/digital impacts of cyber security attacks.
This aspect describes the consequences of cyber attacks on financial assets, business operations, and market stability. We analysed the usage of economic impact terminology across relevant academic research work within this research area. Subsequently, this research compared the retrieved terms against a comprehensive set of terms gathered from the reviewed research. This mapping revealed the most frequent and infrequent terminology related to economic impacts. The comprehensive set of economic impact terminology used includes disrupted operations, loss of revenue, reduced customer base, theft or loss of finances/Capital, regulatory fines, extortion payments and a fall in stock price. This comprehensive set of terms represents a unique list of core economics, excluding secondary impacts and synonyms. The analysis of the economic cyber security impact terminologies is presented in
Table 6.
An analysis of related academic papers reveals that “loss of revenue” and “reduced customers” are the most frequent economic cyber security terms, appearing in 22 and 20 out of 25 research papers, respectively. Following in frequency were “extortion payments” [
16,
30,
35,
55,
59] (5), “disrupted operations” [
30,
39,
56,
61] (4) and “regulatory fines” [
30,
34,
58] (3). The least frequent terminologies, each appearing once across 25 papers, were “fall in stock price” [
30] and “theft or loss of finance capital” [
30]. A plausible reason for this frequency distribution could be because “loss of revenue” and “reduced customers” are easier to quantify, widely reported, and often emphasised by board and organisation leaders. Meanwhile, “fall in stock price” or “theft of capital” is delayed/less directly attributable and rarely broken down in breach disclosure reports. Another reason could be a lack of a defined domain for economic impacts within the taxonomy of cyber security impacts. However, for this, both reasons are justifiable later due to the tendency for researchers to naturally gravitate toward the most readily observable and quantifiable metrics in the absence of economic impacts within the taxonomy of cyber security impacts.
This aspect describes the mental, emotional, and behavioural consequences individuals or organisations experience due to cybe rattacks. This research analysed the usage of psychological impact terminology across relevant academic research work within this research area. Subsequently, this research compared the retrieved terms against a comprehensive set of terms gathered from the reviewed research. This mapping revealed the most frequent and infrequent terminology related to psychological impacts. The comprehensive set of psychological impact terminology used includes confusion, frustration, anxiety/worry, feeling upset, embarrassment, shame, and guilt. This comprehensive set of terms represents a unique list of core psychological, excluding secondary impacts and synonyms. The analysis of the psychological cyber security impact terminologies is presented in
Table 7.
An analysis of related academic papers reveals that “worry or anxiety” and “frustration” are the most frequent psychological cyber security terms, appearing in 20 and 18 out of 27 research papers, respectively. Following in frequency were “feeling upset” [
30,
35,
48,
55,
71,
78] (6), “guilty” [
16,
30,
55,
58,
59] (5), “confusion” [
30,
39,
56,
61] (4), and “shame” [
16,
30,
43,
78] (4). The least frequent terminology is “embarrassment”, appearing only twice across 27 papers. A plausible reason for this frequency distribution could be that “worry or anxiety” and “frustration” are the initial reactions to cyber security incidents. Other reasons may be due to the focus on organisational or outcomes, and terms like “guilty”, “shame”, and “embarrassment” do come with a potential stigma or vulnerability. The lack of a defined domain for psychological impacts within cyber security taxonomy definitely contributes to this uneven focus, as it hinders comprehensive investigation of the psychological spectrum.
This aspect describes the consequences on the public image, public perception and trustworthiness of individuals or organisations due to cyber attacks or cyber-related incidents. This research analysed the usage of reputational impact terminology across relevant academic research work within this research area. Subsequently, this research compared the retrieved terms against a comprehensive set of terms gathered from the reviewed research. This mapping revealed the most frequent and infrequent terminology related to psychological impacts. The comprehensive set of psychological impact terminology used includes negative public image, diminished corporate reputation, negative customer sentiment, eroded supplier trust, reduced employer appeal, increased employee turnover, and revocation of credentials. This comprehensive set of terms represents a unique list of core reputational cyber security impacts, excluding secondary impacts and synonyms. The analysis of the reputational cyber security impact terminologies is presented in
Table 8.
An analysis of related academic papers reveals that “eroded supplier trust” and “negative customer sentiment” are the most frequent reputational cyber security terms, appearing in 21 and 17 out of 27 research papers, respectively. Following in frequency were “increased employee turnover” [
16,
30,
34,
35,
55,
59] (6), “reduced employer appeal” [
30,
48,
71,
77,
78] (5), and “negative public image” [
30,
39,
56,
61] (4). The least frequent terminologies, each appearing less than thrice across 27 papers, were “revocation of credentials” [
30,
34,
55], and “diminished corporate reputation” [
30]. A plausible reason for this frequency distribution could be that “eroded supplier trust” and “negative customer sentiment” are organisation performance metrics primarily included in external financial reports. Meanwhile, “increased employee turnover” and “reduced employer appeal” are included in internal performance reports. The lack of a defined domain for reputational impacts within the cyber security taxonomy contributes to this uneven focus on reputational cyber security impacts.
This aspect describes the consequences of cyber attacks on communities, societies, and even at national levels. This research analysed the usage of societal impact terminology across relevant academic research work within this research area. Subsequently, this research compared the retrieved terms against a comprehensive set of terms gathered from the reviewed research. This mapping revealed the most frequent and infrequent terminology related to psychological impacts. The comprehensive set of societal impact terminology used includes the erosion of public trust, disruption in daily life activities, negative impact on the nation and weakened organisational cohesion. This comprehensive set of terms represents a unique list of core societal cyber security impacts, excluding secondary impacts and synonyms. The analysis of the reputational cyber security impact terminologies is presented in
Table 9.
An analysis of related academic papers reveals that “negative impact on nation” and “weakened organisational cohesion” are the most frequent societal cyber security terms, appearing 18 both out of 27 research papers. The least frequent terminologies, each appearing less than 4 across the 27 papers, were “erosion of public trust” [
30,
39,
56,
61], and “disruption in daily life activities” [
30,
40,
58]. The frequent appearance of “negative impact on nation” and “weakened organisational cohesion” may be due to their applicability of public used digital services and technology. Conversely, the infrequent terms “erosion of public trust” and “disruption in daily life activities” might be less prioritised in due to a current focus on more immediate operational and national security ramifications. Also, the lack of a clearly defined domain for societal impacts within existing cyber security taxonomies may make societal consequences implicitly discussed within other categories rather than explicitly identified and counted as distinct terms.
In summary, the various cyber security impact terminologies used in academic research papers are presented in
Figure 5. The thematic mapping categorised under physical/digital, economic, psychological, reputational, and societal themes revealed a varied landscape in terminology usage. Notably, terms with established legal definitions (“theft,” “compromised”) tend to be more frequent in the physical/digital domain. At the same time, readily quantifiable metrics (“loss of revenue,” “reduced customers”) dominate the economic discourse. Psychological and reputational impacts see “worry or anxiety”, “frustration”, “eroded supplier trust”, and “negative customer sentiment” as the most frequent terms, respectively, potentially reflecting initial reactions and externally reported metrics. Societal impacts highlight “negative impact on nation” and “weakened organisational cohesion” as prevalent. This research also shows that there are terminology gaps across all the themes of the multidimensional impact of cyber attacks implying gaps in consistent and comprehensive articulation of cyber security impacts within academic literature. The identified lack of clearly defined domains for physical/digital, economic, psychological, reputational, and societal impacts within existing cyber security taxonomies is a likely contributing factor to this uneven distribution and potential contextual gaps. An accepted framework for categorising and describing the multidimensional impacts of cyber attacks will support academic researchers with more defined and quantifiable terms for academic research.
RQ2: What are the state-of-the-art quantification techniques for the multidimensional impact of cyber security attacks?
To answer this research question, this paper retrieved and analysed research publications on quantifying the multidimensional impact of cyber security attacks. The results are presented in a chronological order, as shown in
Table 10, including the citation, summary of contribution and limitations. The landscape of cyber security impact quantification techniques presents a diverse interplay of methodologies aimed at capturing the multifaceted impacts of cyber attacks. Across the spectrum, these methods can be categorised into pre-attack (PRE) and post-attack (POST) frameworks, each with distinct approaches, strengths, and limitations.
Pre-attack quantification models, such as those in Assen et al. [
40], Bentley et al. [
56], and Couce-Vieira et al. [
59], emphasise proactive assessment, focusing on threat modelling, scenario planning, and business impact analysis. These techniques aim to preemptively estimate potential losses by analysing dependencies [
56], business objectives [
40], or specific impact categories [
59]. The commonality lies in their reliance on structured frameworks, such as CIAA classifications and multivariate models, to map threats to quantifiable impacts. The key contribution or pre-attack quantification is in providing actionable insights to guide resource allocation and prioritise mitigative efforts by integrating threat modelling within business objectives. The emerging need in pre-attack quantification is more automation, scalability, and integration into operational workflows to enhance practical applicability [
40,
56]. Post-attack quantification models dominate the research landscape, with contributions such as VaR estimation [
39,
42,
48,
60], event-study methodologies [
58,
65,
69], and scenario-based case studies [
72,
78]. These approaches aim to quantify actual damages, including financial losses, reputational harm, and operational impacts. They commonly utilise statistical methods (e.g., copula functions in Bentley et al. [
56], Monte Carlo simulations in Thomas et al. [
78]) or industry benchmarks in Tahmasebi et al. [
77]. The key contribution is that it leverages empirical data to provide accurate cost estimates, which is critical for informing risk mitigation and insurance strategies. The emerging need is that post-attack quantification is hindered due to its data limitations, real-time updates, and holistic impact evaluation, as seen in Dongre et al. [
60], and Thomas et al. [
78].
The convergence across these research publications is visible in their methodological approach to financial metrics, event-driven insights, and scenario-based approaches. For financial metrics, research papers including Aldasoro et al. [
39] (VaR-based calculation), Bentley et al. [
56] (copula functions for dependency modelling), and Orlando et al. [
48] (Cy-VaR with frequency and severity distributions), emphasise financial quantification. They aim to distill complex impacts into monetary values, often leveraging statistical or actuarial models. For event-driven insights, post-attack models (e.g., Cavusoglu et al. [
58], Kamiya et al. [
69], and Portela et al. [
72]) use event-study methodologies to analyse stock market impacts and immediate financial losses from cyber security breaches. These highlight the real-time repercussions on affected firms and industries. For scenario-based approaches, both pre-cyber attack impact and post-cyber attack impact quantification techniques (e.g., Assen et al. [
40]’s threat modelling and Thomas et al. [
78]’s Monte Carlo simulation) leverage scenario-driven methods to estimate potential costs and resource implications, enabling a forward-looking perspective or retrospective analysis. The divergence across these research publications is visible in their differentiating techniques across temporary focus, data granularity/sources, methodological depth and impact scope. For temporal focus, pre-attack techniques (e.g., Assen et al. [
40]’s QuantTM and Couce-Vieira et al. [
59]’s category-specific quantification) prioritise predicting and mitigating risks before incidents occur. These models often incorporate organisational strategy alignment, threat modelling, and resource prioritisation. At the same time, the post-attack techniques (e.g., Franco et al. [
42]’s RCVaR and Greenfield et al. [
34]’s harm evaluation framework) aim to assess realised impacts, often analysing incident data to refine future risk management strategies. For granularity and sources, Anderson et al. [
55] and Tahmasebi et al. [
77] rely on industry reports and surveys, offering broad but sometimes shallow insights due to limited datasets, while others, such as Portela et al. [
72]’s health sector case study, employ domain-specific and often hypothetical scenarios, which can yield tailored but less generalisable findings. For methodological depth, models such as Bentley et al. [
56]’s copula-based multivariate approach explore advanced statistical techniques to capture dependencies and variances, and more straightforward ordinal-based approaches like Facchinetti et al. [
63]’s criticality index focus on practical aggregation methods but may lack granularity. For impact scope, Couce-Vieira et al. [
59] explores multidimensional impacts, including reputational damage and environmental consequences, and in contrast, Cavusoglu et al. [
58] and Goel et al. [
65] focus narrowly on financial and stock market repercussions, missing broader socio-economic or operational effects.
The recurring limitations across these research papers underscore key gaps in the field, including standardisation challenges, data/model constraints, integration/accessibility and multidimensional complexity. For standardisation challenges, Aldasoro et al. [
39], Eling et al. [
62], and Orlando et al. [
48] highlight the lack of standardised cyber event definitions, data collection methodologies, and maturity metrics, complicating cross-comparisons and consistency. For data and model constraints, limited datasets (e.g., Anderson et al. [
55]) and reliance on assumptions (e.g., Cavusoglu et al. [
58], and Kamiya at al., [
69]) restrict the accuracy of many models. Expanding data availability and incorporating real-time updates, as noted in Dongre at al., [
60], are critical future directions. For integration and accessibility, several techniques, such as Assen et al. [
40]’s QuantTM and Couce-Vieira et al.’s category-based modelling, face challenges in integration into organisational workflows and user-friendly applications. Bridging the gap between theoretical models and practical implementation is a pressing need. For multidimensional complexity, Bentley et al. [
56], and Orlando et al. [
48] tackle dependencies and multidimensional impacts; they simplify key components (e.g., fractional mitigation levels or asset interdependencies). More robust models are needed to capture cascading risks and systemic vulnerabilities.
The insights from these research papers suggest a path forward for cyber impact quantification by integrating pre and post-cyber attack impact quantification approaches, advancing data-driven models, improving standardisation/interoperability, expanding impact dimensions, and a renewed focus on risks and systemic impacts. Integrating pre and post-cyber attack impact quantification approaches will create a hybrid model that combines the foresight of PRE techniques with the precision of POST analyses. For instance, integrating threat modelling [
40] with real-world VaR metrics [
39], ref. [
42] could create robust frameworks for end-to-end risk management. For the advanced data-driven models, enhancing access to anonymised datasets [
62] and real-time data updates [
60] will be pivotal. As suggested in Franco et al. [
42], ML and AI techniques can improve factor calibration and dynamic risk modelling, addressing limitations in current methods. For standardisation and interoperability, establishing universal standards for cyber event definitions, risk metrics, and loss quantification (e.g., Aldasoro et al. [
39], Eling et al. [
62], and Eling et al. [
61]) will facilitate cross-industry benchmarking and the development of globally accepted frameworks. For expanding impact dimensions, Couce-Vieira et al. [
59], and Portela et al. [
72] highlight the importance of incorporating less tangible impacts into quantification models, such as reputational damage, regulatory penalties, and cascading risks. Tailoring impact prioritisation to organisational profiles will further enhance their relevance. For emerging risks and systemic impacts, Eling et al. [
61], and Thomas et al. [
78] underscore the need to address systemic risks, including cascading failures and supply chain vulnerabilities. Incorporating agent-based simulation and macro-level risk modelling can address these gaps.
RQ3: What are the top data sources used by academics and practitioners in quantifying the multidimensional impact of cyber attacks, how are they being used, and what are their limitations?
To answer this research question, this paper retrieved and analysed the top data sources used in academic research publications on quantifying the multidimensional impact of cyber security attacks. The results are presented in
Table 11. The analysis of data sources used in academic research on the multidimensional impact of cyber attacks reveals three prominent intertwined themes: data origin (internal vs. external), temporal applicability (PRE vs. POST attack), and impact type (physical/digital, economic, and reputational). The triangulation of these three themes is required for a complete understanding and empirical quantification of the complex multidimensional impact of cyber attacks.
A holistic computation necessitates the integration of both internal organisational data (e.g., Incident Response Metrics Database, Banking Transactions, Internal Asset Database, Security Control Inventory, and Risk Assessment Reports) and external organisational data (e.g., Operational Riskdata eXchange [
84], Privacy Rights Clearinghouse [
85], Industry Reports, Advisen Dataset [
87], Technology News, and LexisNexis [
90]). Internal data is unique to the organisations technnologies, and digital services. However, internal data still needs to be benchmarked with external data to provide a broader context and the necessary industry perspective. However, an apparent difficulty is accommodation and fixing data quality issues. To bring this to life, an organisation seeking to quantify the financial losses from a recent ransomware attack can use its internal data, such as servers and databases from the Internal Asset Database, to identify its crown jewels and vulnerable assets while using external data, such as associated fines from Privacy Rights Clearinghouse [
85] and benchmarks data such as the severity of similar cyber attack incidents from the Operational Riskdata eXchange [
84] to provide a broader industry perspective.
Crucially, PRE- and POST-attack data are essential for a complete understanding. PRE-attack data, typically sourced from threat models, vulnerability scan reports, and threat intelligence(e.g., Internal Asset Database, Security Control Inventory, Risk Assessment Reports, and potentially Banking Transactions for predictive modelling), facilitates proactive risk assessment and mitigation planning. POST-attack data, derived from incident response logs, financial records, and reputational analysis (e.g., Incident Response Metrics, Operational Riskdata eXchange [
84], Privacy Rights Clearinghouse [
85], Industry Reports, Advisen Dataset [
87], Technology News, and LexisNexis [
90]), provides empirical evidence of actual impacts, enabling model calibration and validation. However, PRE-attack data is fundamentally a forward look from the lens of an attacker and includes unverifiable data. On the other hand POST-attack data is based on verifiable data, hence they need to be correctly intertwined to improve the capability to anticipate the multidimensional impact of cyber attacks. To bring this to life, a financial company wanting to understand the reputational impact of a past data breach can model its PRE-attack data, such as emerging threats from its threat intelligence feeds, to model its potential breaches while using POST-attack data, such as articles and reports from Technology and News/Websites to assess the actual reputational damage.
Also, these data sources contribute to quantifying the various dimension of the multidimensional impacts of cyber attacks: physical/digital (e.g., Incident Response Metrics Database), economic (e.g., Operational Riskdata eXchange [
84], Privacy Rights Clearinghouse [
85], Industry Reports, Advisen Dataset), and reputational (e.g., Technology News/Websites, LexisNexis [
90], and potentially survey data—though not explicitly listed, it is often a key component). Integrating data across these three themes—origin, time, and impact—is crucial for comprehensively analysing the multidimensional impact of cyber attacks. To bring this to life, a building society that suffers a cyber attack that disrupts logistics operations and affects customer confidence can understand the multidimensional impact dimensions by analysing physical/digital impact data like downtime of critical systems, the number of affected machines from Incident Response Metrics Database, economic Impact data like average cost of ransomware attacks from Advisen Dataset [
87] and reputational impact data like media coverage, public sentiment, and stock price fluctuations post-incident from Technology News/Websites. The company combines data from all three impact categories to understand the attack’s consequences. The Incident Response Metrics Database quantifies the physical/digital disruption, the Advisen Dataset [
87] and internal records quantify the economic impact, and media/social media monitoring captures the reputational damage. This comprehensive assessment helps the company quantify the multidimensional impact of cyber attacks to the inform future security investments.
Similarly, these data sources diverge in their practical application, data deficiencies, and Computational intricacies. Internal data sources offer organisation contextual, and nuanced information, that lack external context, whereas external data sources provide industry benchmarks, but may have regionally prejudice or obsolete information.
RQ4: What are the state-of-the-art implementations of ML and DL in cyber security risk quantification ?
To answer this question, this paper retrieved and analysed research publications that implement ML and DL models in the quantification of cyber security risks. The results are presented in a chronological order, as shown in
Table 12, including the citation, use case, ML/DL algorithms and their evaluation criteria.
Table 12 provides a comprehensive overview, detailing how various ML and DL models are specifically applied, their identified use cases, the algorithms employed, and their respective evaluation criteria across the surveyed literature. This systematic presentation provides the foundational context for the subsequent discussion on prevalent algorithmic trends and their nuanced applications within cyber risk quantification.These academic research studies demonstrate a growing integration of ML and DL models for cyber security risk assessment across a wide range of industry sectors. Abdulsatar et al. [
50] explored a deep learning-based framework tailored to microservice architectures, leveraging LSTM models to assess risks in highly dynamic and containerised environments. Their work highlights the importance of adaptability and scalability in risk quantification. Similarly, Ahmadi-Assalemi et al. [
51] developed a Super Learner Ensemble method for anomaly detection and risk scoring in Industrial Control Systems (ICS), effectively combining multiple ML models to enhance detection accuracy and cyber risk prediction in critical infrastructure. Alagappan et al. [
52] focused on the Internet of Things (IoT), employing probabilistic ML techniques to account for the heterogeneity and limited resources typical of IoT devices, emphasising lightweight yet effective quantification models. In the energy domain, Kumar et al. [
47] applied deep learning to evaluate the economic impact of cyber threats on Virtual Power Plants (VPPs), illustrating how LSTM models can learn temporal attack patterns and predict potential financial consequences. Alsaadi et al. [
54] turned attention to the financial sector, applying predictive analytics through CNNs to model cyber risk trends and support proactive risk mitigation strategies. Also, Yao et al. [
81] addressed the construction industry, proposing a machine learning-based framework to assess cyber risks in Building Information Modeling (BIM) systems, underlining the growing relevance of cyber risk management in traditionally non-digital sectors. These studies reveal how ML and DL are increasingly tailored to domain-specific challenges—such as real-time data in ICS, resource constraints in IoT, temporal dependencies in VPPs, and complex data structures in finance and construction. Despite differences in focus, they share a common objective: enhancing cyber security risk quantification by moving beyond basic detection to assess threat likelihood, impact, and severity in contextually rich environments.
Within the analysed body of literature, two distinct trends emerge concerning algorithm selection for cyber security risk quantification: a clear preference for Support Vector Machines (SVM) and Naïve Bayes (NB) among traditional ML techniques, and a comparable emphasis on Convolutional Neural Networks (CNN) and Long Short-Term Memory (LSTM) models within DL approaches. This dual-track adoption highlights how algorithm selection is influenced by data characteristics, computational resources, and the intended application domain. SVM and NB feature prominently across several studies due to their relative interpretability, maturity, and proven effectiveness in classification tasks. SVM, in particular, has been extensively adopted due to its capacity to create optimal decision boundaries (hyperplanes) that separate data into distinct classes, making it well-suited for high-dimensional cyber security datasets. This strength is emphasised in the works of Ahmadi-Assalemi et al. [
51], Ali et al. [
53], and Biswas et al. [
57], who leverage SVM for accurate classification and risk scoring. Further studies by Franco et al. [
42] and Rafaiani et al. [
73] highlight SVM’s flexibility in handling non-linear data through kernel functions, which transform data into higher-dimensional spaces where linear separability becomes feasible. Such capability is particularly valuable in cyber security, where attack patterns often exhibit complex, non-linear relationships. However, the computational cost associated with training multiple binary classifiers, especially in large or multi-class datasets—poses a scalability challenge. As noted by Franco et al. [
64], the One-vs-Rest strategy used by SVM introduces inefficiencies that may hinder real-time applicability in dynamic threat environments.
Contrastingly, NB offers a lightweight, probabilistic alternative that excels in scenarios where rapid inference and uncertainty handling are critical. Studies by Alagappan et al. [
52] and Kumar et al. [
47] illustrate NB’s utility in modelling probabilistic relationships between features and outcomes, facilitating risk estimation in environments with incomplete or noisy data. NB’s low computational overhead makes it particularly suitable for resource-constrained contexts, such as IoT or edge devices, as supported by findings in Yao et al. [
81]. However, its core limitation lies in the assumption of feature independence, which is an unrealistic constraint in many cyber security datasets where features are often correlated. This assumption can compromise accuracy and limit the algorithm’s generalisability in complex threat detection tasks. DL models, specifically LSTMs and CNNs, have gained increasing traction due to their superior capacity for pattern recognition and sequence modelling, both critical for anticipating and quantifying cyber risks in real-time. LSTM models, as adopted by Ali et al. [
53], Alsaadi et al. [
54], and Goyal et al. [
67], are particularly effective in capturing temporal dependencies within time-series cyber data. Their internal architecture, comprising forget, input, and output gates, alongside memory cells, allows LSTMs to preserve long-range dependencies while mitigating vanishing gradient issues common in recurrent neural networks (RNNs). This capability is essential in cyber risk contexts where attack sequences evolve over time and exhibit delayed effects. Studies by Huang et al. [
66], Sangiorgio et al. [
75], and Yang et al. [
80] further confirm LSTM’s ability to detect nuanced patterns and forecast threat trajectories, enhancing both the accuracy and granularity of cyber risk assessments. However, computational intensity and susceptibility to overfitting, especially when training with limited or noisy data, remain key limitations. These factors can hinder the real-time responsiveness and robustness of LSTM-based systems in rapidly changing threat environments.
CNNs, by contrast, offer an alternative DL approach that emphasises spatial feature extraction. As documented by Ali et al. [
53] and Alsaadi et al. [
54], CNNs utilize convolutional layers to detect hierarchical feature patterns, followed by pooling layers to reduce dimensionality and mitigate overfitting. This architecture enables CNNs to learn abstract representations from raw input data, making them effective in analysing structured inputs such as network traffic matrices, binary code, or log files. In risk quantification contexts, this allows for more refined risk scoring through feature abstraction and anomaly detection. However, as with LSTMs, CNNs are resource-intensive, often requiring large labelled datasets and high-performance computing environments for effective training and deployment. Ali et al. [
53] highlight this challenge in their research on efficient training and model scalability, noting that the depth of CNN architectures directly correlates with computational burden and training duration. When comparing ML and DL approaches across the reviewed studies, a consistent pattern emerges: while SVM and NB are favoured for their simplicity, interpretability, and lower computational requirements, LSTM, and CNN models consistently outperform them in terms of predictive accuracy and adaptability to complex data. Ali et al. [
53] explicitly note the superior performance of DL models in capturing the intricate, non-linear relationships inherent in cyber security data, which traditional ML models often struggle to represent effectively. However, these performance gains come at a cost. DL models demand substantial training data, are often difficult to interpret, posing challenges for explainability in high-stakes environments, and are prone to overfitting, particularly when domain-specific data is scarce.
In summary, the reviewed literature reflects a progressive evolution from traditional ML to more sophisticated DL techniques in cyber risk quantification, driven by the need for models that can effectively process high-volume, high-velocity, and high-variety data. While ML algorithms like SVM and NB continue to offer valuable trade-offs between accuracy and efficiency in constrained environments, DL models such as LSTM and CNN are becoming increasingly essential for developing predictive, adaptive, and context-aware risk assessment frameworks. The selection of algorithms, therefore, hinges on a careful balancing of performance needs, computational capacity, data availability, and the specific threat landscape of the application domain.