Previous Article in Journal
A Lightweight Multimodal Framework for Misleading News Classification Using Linguistic and Behavioral Biometrics
Previous Article in Special Issue
The Rise of Hacking in Integrated EHR Systems: A Trend Analysis of U.S. Healthcare Data Breaches
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Game-Theoretic Approach for Quantification of Strategic Behaviors in Digital Forensic Readiness

by
Mehrnoush Vaseghipanah
1,
Sam Jabbehdari
1,* and
Hamidreza Navidi
2
1
Department of Computer, Tehran North Branch, Islamic Azad University, Tehran 1651153511, Iran
2
Department of Mathematics and Computer Sciences, Shahed University, Tehran 3319118651, Iran
*
Author to whom correspondence should be addressed.
J. Cybersecur. Priv. 2025, 5(4), 105; https://doi.org/10.3390/jcp5040105
Submission received: 8 October 2025 / Revised: 12 November 2025 / Accepted: 17 November 2025 / Published: 26 November 2025
(This article belongs to the Special Issue Cyber Security and Digital Forensics—2nd Edition)

Abstract

Small and Medium-sized Enterprises (SMEs) face disproportionately high risks from Advanced Persistent Threats (APTs), which often evade traditional cybersecurity measures. Existing frameworks catalogue adversary tactics and defensive solutions but provide limited quantitative guidance for allocating limited resources under uncertainty, a challenge amplified by the growing use of AI in both offensive operations and digital forensics. This paper proposes a game-theoretic model for improving digital forensic readiness (DFR) in SMEs. The approach integrates the MITRE ATT&CK and D3FEND frameworks to map APT behaviors to defensive countermeasures and defines 32 custom DFR metrics, weighted using the Analytic Hierarchy Process (AHP), to derive utility functions for both attackers and defenders. The main analysis considers a non-zero-sum attacker–defender bimatrix game and yields a single Nash equilibrium in which the attacker concentrates on Impact-oriented tactics and the defender on Detect-focused controls. In a synthetic calibration across ten organizational profiles, the framework achieves a median readiness improvement of 18.0% (95% confidence interval: 16.3% to 19.7%) relative to pre-framework baselines, with targeted improvements in logging and forensic preservation typically reducing key attacker utility components by around 15–30%. A zero-sum variant of the game is also analyzed as a robustness check and exhibits consistent tactical themes, but all policy conclusions are drawn from the empirical non-zero-sum model. Despite relying on expert-driven AHP weights and synthetic profiles, the framework offers SMEs actionable, equilibrium-informed guidance for strengthening forensic preparedness against advanced cyber threats.

1. Introduction

Digital forensic readiness (DFR) enables organizations to proactively collect and preserve admissible digital evidence, reducing legal risks and supporting business continuity. It is particularly valuable for Small and Medium-sized Businesses and Enterprises (SMBs/SMEs)—encompassing both the commercial/business context (SMB) and the broader organizational/industrial context (SME)—which often face resource constraints in cybersecurity operations. A robust DFR strategy ensures that significant cyber incidents can be addressed efficiently, lawfully, and professionally, conserving investigative resources, reducing costs, protecting organizational reputation, and maintaining compliance with applicable regulations.
Despite heavy investment in Computer Security Incident Response Teams (CSIRTs), digital forensics and incident response (DFIR) units, and advanced monitoring technologies—such as EDR, XDR, NDR, SIEM, and IDPS—organizations still struggle to achieve effective incident detection and response. Such limitations become especially pronounced against Advanced Persistent Threats (APTs), which are sophisticated, well-funded actors conducting prolonged cyber campaigns. APTs are stealthy, long-term cyberattacks by unauthorized entities to remain undetected in networks [1,2]. Recent threat intelligence reports indicate that, while global median dwell time has decreased to 11 days, this metric rises significantly when organizations rely on external notifications (median 26 days), highlighting the importance of internal detection capabilities [3]. Data breaches impose severe financial consequences on organizations; the global average cost reached USD 4.44 million in 2025, with financial sector breaches averaging USD 6.08 million [4,5]. Historical data suggest that cyber incidents can have devastating impacts on small enterprises, with earlier studies indicating that a substantial portion may face severe operational disruptions following major breaches [6]. For example, Baker [7] notes that, in the SolarWinds incident, threats persisted within networks for prolonged periods without detection.
The rapid proliferation of artificial intelligence (AI) has further complicated this landscape. AI-driven tools empower attackers with advanced automation, adaptive tactics, and the ability to launch more sophisticated and targeted attacks, thereby increasing the potency of APTs. Conversely, while AI offers defenders enhanced capabilities for faster and more accurate detection, it also introduces unprecedented forensic challenges. These include the complexity of analyzing AI-generated attacks, the potential for AI-based evidence manipulation, and the need for new techniques to handle AI-related incidents. For SMBs, these challenges are particularly acute due to resource constraints. These technical challenges are compounded by the broader organizational struggle to effectively govern AI systems and mitigate associated risks, a problem highlighted in recent literature [8].
Organizations often perceive this issue as primarily technical in nature. However, this challenge fundamentally encompasses the interplay of technology, human expertise, and processes. Without skilled personnel and planning, even the most advanced technology stack may fail against determined assailants. Situations where inadequate DFR hinders effective cyber security incident investigations—often due to poor data retention, ineffective log management, or compromised digital evidence integrity—exemplify what we term non-forensicability (see Section 5.1 for a detailed discussion). Wrightson [9] emphasizes that understanding an attacker’s motivations and capabilities, as well as knowing their past actions, helps investigators categorize and respond to diverse cyber threats.
Digital forensic investigators must know both defense and offense strategies, preempt emerging attack techniques, and collaborate closely with defense teams. Årnes [10] characterizes digital forensics, as a sub-discipline of forensic science, as encompassing scientifically validated methods for the management of digital evidence. These methods are essential for reconstructing criminal incidents or anticipating unauthorized activities.
To address the need for a formal strategic framework for DFR, we propose a game-theoretic approach to model the strategic interactions between cyber attackers and defenders. This approach helps organizations anticipate threats, optimize defense strategies, and make more informed decisions. We focus on the strategic behavior in digital forensics, drawing from Sun Tzu’s wisdom in The Art of War, which emphasizes the importance of understanding both one’s own abilities and the opponent’s strengths and strategies. As Tzu [11] states, “If you know the enemy and know yourself, you need not fear the result of a hundred battles. If you know yourself but not the enemy, for every victory gained, you will also suffer a defeat. You will succumb in every battle if you know neither the enemy nor yourself.” This highlights the importance of knowing the adversary’s motivations, methods, and goals, as well as the capabilities and limitations of one’s own tools and techniques.
Inspired by Sun Tzu’s philosophy, our game-theoretic model operationalizes this wisdom by quantifying how knowledge asymmetries between attacker and defender impact forensic readiness. We operationalize “know the enemy” via 16 attacker utilities and “know yourself” via 16 defender metrics that quantify organizational capabilities that model adversary behaviors. These 32 metrics are linked through an explicit ATT&CK↔D3FEND coupling Γ that yields measurable forensic readiness improvements and richer post-incident evidence. We formalize three strategic states: comprehensive knowledge (targeted defense), partial knowledge (vulnerable defense), and ignorance (minimal resilience). This approach is especially important in the AI era, where modeling emerging AI-powered attack surfaces and their forensic implications becomes essential for building resilient systems. Game theory provides a mathematical foundation for analyzing strategic interactions among rational decision-makers [12]. Its application in cybersecurity is growing, as it offers a structured approach to the following:
  • Model Strategic Decisions: Capture the objectives and constraints of both attackers and defenders [13].
  • Conduct Risk Analysis: Elucidate payoffs and tactics to identify critical vulnerabilities and optimal defensive strategies [14].
  • Enable Adaptive Defense: Capture the dynamic nature of cyber threats, including those augmented by AI, to inform adaptive countermeasures [15].
  • Optimize Resource Allocation: Evaluate strategy effectiveness to guide efficient investment of limited defensive resources [16].
To operationalize this game-theoretic approach, our methodology is grounded in established cybersecurity standards and formal decision-making processes. We build upon best practices from the National Institute of Standards and Technology (NIST) for metric development and forensic readiness [17]. Specifically, we integrate the MITRE ATT&CK framework to systematically model adversary behaviors and the complementary MITRE D3FEND framework to map defensive countermeasures. This integration provides a standardized taxonomy that bridges attacker tactics with defender responses. Based on these frameworks, we define 32 custom DFR metrics, weighted using the Analytic Hierarchy Process (AHP), to compute quantifiable utility functions for both attackers and defenders. This addresses a critical gap in the field: the absence of quantifiable payoffs in strategic DFR planning. Furthermore, we present an end-to-end algorithmic suite for scoring, classification, and gap analysis, moving beyond fragmented assessments towards a holistic readiness model.
This paper makes the following key contributions:
  • A game-theoretic readiness-planning framework for DFR that quantifies strategic attacker–defender interactions.
  • A practical integration of MITRE ATT&CK and D3FEND with AHP-weighted metrics to ground utilities in real-world tactics and techniques.
  • An equilibrium-based analysis that derives actionable, resource-constrained guidance for SMBs/SMEs.
  • An empirical evaluation on APT-inspired, multi-vector scenarios showing that the framework can improve readiness and reduce attacker success under realistic constraints.
The remainder of this paper is structured as follows: Section 2 reviews the related works in digital forensics investigation and readiness. Section 3 describes our game-theoretic approach and algorithms for DFR. Section 4 presents our experimental analysis and results. Section 5 concludes with our findings and future work.

2. Related Works

Enhancing cybersecurity and digital forensics has spurred a plethora of studies. These foundational works span technical defenses, strategic modeling, and simulation of cyber interactions. While appreciating their contributions, we identify areas for further exploration.

2.1. Game Theory in Digital Forensics

Alpcan et al. [18] provided a foundational contribution to the field of network security by presenting theoretical approaches for decision making in security from a game-theoretic perspective. Their work serves as a valuable reference not only for researchers and graduate students but also for practitioners such as system administrators and security officers seeking to apply quantitative models grounded in control, optimization, and decision theory. Casey [19] established the conceptual foundation for incorporating game theory into digital forensics, contextualizing how strategic analysis can enhance forensic practices.
Manshaei et al. [20] offered a comprehensive overview of game-theoretic methods in network security and privacy, highlighting their capability to model strategic interactions in complex adversarial environments. Their study provided in-depth insights into how game theory can strengthen computer and communication network security across multiple layers, including physical and MAC layers, self-organizing networks, intrusion detection systems, anonymity and privacy mechanisms, network security economics, and cryptography. The authors summarized key concepts such as equilibrium analysis and mechanism design, emphasizing the significance of addressing information limitations and learning factors in developing effective security solutions.
Several subsequent studies have built on this foundation to explore game-theoretic applications in digital forensics. Nisioti et al. [21] presented a Bayesian game model for analyzing interactions between a forensic investigator and a strategic attacker on a multi-host forensic investigation graph. Hasanabadi et al. [22] developed a model representing attacker–investigator dynamics involving rootkits and anti-rootkits, defining each player’s actions and profiling their characteristics. Extending these ideas, Karabiyik et al. [23] proposed a game-theoretic approach to optimize tool selection in digital forensics, particularly focusing on file carving tools and the strategic adaptation of selection decisions during investigations. Hasanabadi et al. [24] later introduced a memory-based mechanism to expand action spaces within forensic game models, reducing convergence iterations when new anti-forensic or counter-anti-forensic tools emerge. Caporusso et al. [25] further analyzed post-attack decision dynamics in human-controlled ransomware scenarios, modeling negotiation strategies and emphasizing the role of information availability, user education, and human factors in developing resilient defensive responses.

2.2. Digital Forensics Readiness and Techniques

Kebande et al. [26] introduced a technique for implementing DFR in cloud computing environments through a modified obfuscated Non-Malicious Botnet (NMB). Operating as a distributed forensic Agent-Based Solution (ABS), this method enables forensic logging for readiness purposes across cloud infrastructures. In a related effort, Kebande et al. [27] proposed the construction of a Digital Forensic Readiness Intelligence Repository (DFRIR) founded on knowledge-sharing principles. The repository cross-references potential evidence sources, aims to reduce the time required for forensic investigations, and supports sharing across multiple jurisdictions.
Englbrecht et al. [28] developed a DFR-specific Capability Maturity Model (CMM) to guide organizations in implementing readiness measures. The framework draws on COBIT 5 IT-Governance principles and incorporates the core characteristics necessary for effective DFR implementation. Reddy et al. [29] built a Digital Forensic Readiness Management System (DFRMS) tailored for large organizations. Based on requirements identified through a comprehensive literature review, the DFRMS architecture comprises five modules: event analysis, DFR information management, costing, access control, and user interface. A proof-of-concept prototype demonstrated the system’s practical feasibility and its potential to improve readiness in enterprise contexts.
Grobler et al. [30] positioned DFR as a means to strengthen organizational security strategies by preparing for incidents while minimizing disruptions to business processes. Their guidelines emphasize ensuring legal admissibility of evidence, detecting resource misuse, and demonstrating due diligence in protecting valuable company assets. The authors contend that revisions to current information systems architectures, strategies, and best practices are needed to enable successful prosecutions, pointing to deficiencies in admissible evidence and procedural rigor. Lakhdhar et al. [31] proposed a game-theoretic model for forensic-ready systems utilizing cognitive security concepts; however, this work lacks practical tools applicable to SMBs/SMEs.
Elyas et al. [32] designed and validated a DFR framework through expert focus groups. The framework assists organizations in assessing their forensic strategies by identifying critical factors in capacity development. It categorizes governance, top management support, and culture as organizational dimensions, while technology and architecture are grouped under forensic infrastructure. Baiquni and Amiruddin [33] applied the Digital Forensic Readiness Index (DiFRI) to quantitatively evaluate a cyber organization’s operational readiness, offering tailored improvement recommendations. Although informative, this methodology does not address strategic adversary behavior or optimal resource allocation—gaps targeted by our proposed game-theoretic approach.
Complementing DFR frameworks with an SME-focused perspective, Rawindaran et al. [34] introduce an enhanced ROHAN model integrated with the Cyber Guardian Framework (CGF) to improve cybersecurity resilience in resource-constrained organizations. Their mixed-method study emphasizes role-specific awareness, continuous improvement, and the use of AI-enabled decision support—principles aligned with readiness thinking. However, while ROHAN+CGF advance organizational practice, they do not explicitly model adversarial strategy or attacker–defender interdependence; our game-theoretic formulation targets precisely this gap by coupling readiness with strategic behavior and optimal resource allocation.
Trenwith et al. [35] advocated centralized logging as a cornerstone of effective DFR, enabling rapid acquisition of evidential data and accelerated investigative analysis. While centralized log management streamlines evidence collection, it does not account for the diverse evidence types necessary in investigations, particularly within cloud environments. Cloud systems present additional challenges due to the dynamic and distributed nature of data storage and processing, which demand solutions beyond efficient logging.
In the context of microservice architectures, Monteiro et al. [36] proposed “Adaptive Observability,” a game theory-driven method designed to address evidence challenges in ephemeral environments where traditional observability mechanisms fail after container termination. By dynamically adjusting observability based on user–service interactions, the approach enhances evidence retention while optimizing resource consumption. Comparative evaluations show performance improvements ranging from 3.1 % to 42.50 % over conventional techniques. The authors suggest future work should incorporate varying attacker risk preferences and extend into industrial case studies, with additional metrics covering cost-effectiveness and scalability.

2.3. Advancement in Cybersecurity Modeling

Xiong et al. [37] developed a threat modeling language for enterprise security based on the MITRE Enterprise ATT&CK Matrix and implemented using the Meta Attack Language framework. This language enables the simulation of cyberattacks on modeled system instances to analyze security configurations and assess potential architectural modifications aimed at improving system resilience.
Wang et al. [38] proposed a sequential Defend–Attack framework that integrates adversarial risk analysis. Their approach introduces a new class of influence diagram algorithms, termed hybrid Bayesian network inference, to identify optimal defensive strategies under adversarial conditions. This model enhances understanding of the interdependent decision processes between attackers and defenders in dynamic threat environments.
Usman et al. [39] presented a hybrid methodology for IP reputation prediction and zero-day attack categorization that fuses Dynamic Malware Analysis, Cyber Threat Intelligence, Machine Learning, and Data Forensics. This integrated system simultaneously evaluates severity, risk score, confidence, and threat lifespan using machine learning techniques, illustrating how data-driven analytics can support forensic and security objectives. The study also highlights persistent data forensic challenges when automating classification and reputation modeling for emerging cyber threats.

2.4. Innovative Tools and Methodologies

Li et al. [40] introduced LEChain, a blockchain-based lawful evidence management scheme for digital forensics designed to address security and privacy concerns often overlooked in cloud computing and blockchain-based evidence management. LEChain implements fine-grained access control through ciphertext-policy attribute-based encryption and employs brief randomizable signatures to protect witness privacy during evidence collection.
Soltani and Seno [41] presented a Software Signature Detection Engine (SSDE) for digital forensic triage. The SSDE architecture comprises two subsystems: signature construction and signature detection. Signatures are generated using a differential analysis model that compares file system states before and after execution of specific software. Their study evaluates multiple design parameters, resulting in the creation and assessment of 576 distinct SSDE models.
At the storage–firmware boundary, Rother and Chen [42] present ACRecovery, a flash-translation-layer (FTL) forensics mechanism that can roll back OS access-control metadata after an OS-level compromise by exploiting out-of-place updates in raw flash. Their prototype on EXT2/EXT3 and OpenNFM demonstrates efficient recovery with minimal performance impact, highlighting a promising post-compromise remediation path. While orthogonal to our strategic readiness modeling, such FTL-aware techniques complement DFR by preserving evidential integrity and enabling rapid restoration when preventive controls are bypassed.
Nikkle [43] described the Registration Data Access Protocol (RDAP) as a secure, standardized, and internationalized alternative to the legacy WHOIS system. While WHOIS and RDAP are expected to coexist for some time, RDAP offers enhanced security, automation capabilities, tool integration, and authoritative data sourcing—features that strengthen its utility in digital forensic investigations. Furthermore, Nikkle [44] introduced the concept of Fintech Forensics as a new sub-discipline, noting how the rise of digital transformation and financial technology has created novel avenues for criminal activity, necessitating dedicated forensic methodologies for financial transactions.

2.5. Digital Forensics in Emerging Domains

Seo et al. [45] proposed a Metaverse forensic framework structured around four phases derived from NIST’s digital forensic guidelines: data collection, examination and retrieval of evidence, analysis, and reporting. The study also outlines three procedures for data collection and examination distributed across user, service, and Metaverse platform domains, providing a systematic approach for investigating offenses occurring in virtual environments.
Malhotra [46] explored the intersection of digital forensics and artificial intelligence (AI), presenting current approaches and emerging trends. The author emphasized that, in today’s increasingly digital society, the rise in cybercrimes and financial frauds has made digital forensics indispensable. Integrating AI techniques into forensic analysis offers promising opportunities to address these challenges effectively. Malhotra further argued that AI-driven digital forensics could transform investigative efficiency, catalyzing the so-called Fourth Industrial Revolution. Consequently, continued investment in AI-enabled forensic technologies, specialized training, and advanced analytical tools is critical for ensuring preparedness against evolving cyber threats.
Tok and Chattopadhyay [47] examined cybersecurity challenges within Smart City Infrastructures (SCIs), proposing a unified definition and applying the STRIDE threat modeling methodology to identify potential offenses and evidence sources. Their study provides valuable guidance for investigators by mapping technical and legal aspects of digital forensics in SCI environments. However, the authors note that the applicability of their framework may depend on contextual variations in regulatory standards and implementation practices across jurisdictions.

2.6. Advanced Persistent Threats and Cybercrime

Han et al. [48] examined defensive strategies against long-term and stealthy cyberattacks, such as Advanced Persistent Threats (APTs). Their work underscores the necessity of strategic and proactive measures to counter increasingly sophisticated adversaries capable of prolonged network infiltration.
Chandra and Snowe [49] defined cybercrime as criminal activity involving computer technology and proposed a taxonomy built upon four foundational principles: mutual exclusivity, structural clarity, exhaustiveness, and well-defined categorization. This taxonomy facilitates the classification and differentiation of various cybercrime types and could be extended to organizational applications, metrics development, integration with traditional crime taxonomies, and automated classification for improved efficiency.
Collectively, these contributions highlight the potential of combining game theory with advanced technologies—such as artificial intelligence and blockchain—to enhance the effectiveness of digital forensic investigations. Casey et al. [50] introduced the Cyber-investigation Analysis Standard Expression (CASE), a community-driven specification language designed to improve interoperability and coordination among investigative tools. By building upon the Unified Cyber Ontology (UCO), CASE offers a standardized structure for representing and exchanging cyber-investigation data across multiple organizations and jurisdictions. Its versatility allows application in criminal, corporate, and intelligence contexts, supporting comprehensive analysis. Through illustrative examples and a proof-of-concept API, Casey et al. demonstrated how CASE enables structured data capture, facilitates sharing and collaboration, and incorporates data marking for controlled dissemination within the cyber-investigation community.
Despite notable progress in cybersecurity and digital forensics—particularly via the integration of game theory, enhanced readiness techniques, and diverse modeling tools—several critical challenges remain. Current approaches often struggle to represent the dynamic and asymmetric interactions between attackers and defenders in APT scenarios. Moreover, game-theoretic models frequently overlook nuanced decision-making processes inherent to forensic investigations and fail to fully account for the rapidly evolving tactics of modern cyber adversaries. Additionally, many DFR frameworks emphasize technical countermeasures while insufficiently addressing strategic adversary dynamics, leaving organizations vulnerable and less responsive to emerging threats.

2.7. Novelty

This work develops a quantitative game-theoretic model of digital forensic readiness for APT-focused SMEs. Its main contribution is the metricization of readiness via AHP-weighted utilities explicitly grounded in MITRE ATT&CK/D3FEND, as opposed to proposing a new game-theoretic solution concept.
To clarify the scope and positioning of our work relative to prior digital forensics and cybersecurity games, we compare our framework systematically with the closest prior work. Our approach integrates four key components that, to our knowledge, have not previously been combined in a single readiness-planning framework: (i) ATT&CK–D3FEND knowledge coupling with quantified utilities; (ii) AHP-weighted DFR metrics for payoff grounding; (iii) explicit PNE and MNE analysis with support conditions (one MNE for the non-zero-sum bimatrix ( A , D ) , and five equilibria—two pure-strategy and three mixed-strategy—for the zero-sum variant ( A , D ) as a robustness check); and (iv) actionable SME guidance under resource constraints.
Table 1 provides a systematic comparison across 12 evaluation dimensions, comparing our framework against five representative works in game-theoretic forensics: Nisioti et al. (Bayesian anti-forensics), Karabiyik et al. (tool selection games), Lakhdhar et al. (provability taxonomies), Wang et al. (adversarial risk analysis), and Monteiro et al. (microservice observability). This comparison demonstrates how our integration of standardized knowledge frameworks (ATT&CK/D3FEND), expert-driven metric weighting (AHP), and quantitative equilibrium analysis (MNE) addresses gaps in prior work, particularly the lack of quantitative payoffs, standardized taxonomies, and SME-focused guidance. Prior work has advanced game-theoretic forensics in specific domains: Nisioti et al. focus on Bayesian anti-forensics, Karabiyik et al. on tool selection games, Lakhdhar et al. on provability taxonomies, Wang et al. on adversarial risk analysis, and Monteiro et al. on microservice observability. Building on these foundations, our framework jointly (i) integrates both ATT&CK and D3FEND (whereas Nisioti et al. use ATT&CK without D3FEND, and others use neither); (ii) employs the AHP with expert panels for metric weighting (whereas others use CVSS mapping, rule-based, or implicit weighting); (iii) provides explicit MNE analysis with support conditions for non-zero-sum bimatrix games; and (iv) explicitly targets SME/SMB applicability with resource-constrained guidance.
This combination is, to our knowledge, not yet available in existing work, and is intended as a practical integration that addresses a gap in quantifying strategic payoffs in DFR planning using standardised knowledge frameworks; it does not introduce a new game-theoretic solution concept.

3. Materials and Methods

In this section, the problem statement is provided in Section 3.1. The methodology of the research is stated in Section 3.2. The fundamental concepts of game theory are presented in Appendix B.12. The proposed approach is detailed in Section 3.3, followed by the utility function discussion in Section 3.4. The prioritization of DFR is addressed in Section 3.5, respectively. The reevaluation of DFR is covered in Section 3.6.

3.1. Problem Statement

Let A represent the set of attackers and D represent the set of defenders in a cyber environment. The objective of this research is to model the strategic interactions between A and D during the DFR phase using game theory.
Let us define the following variables:
  • S A : Strategies available to attackers, corresponding to MITRE ATT&CK tactics (e.g., Reconnaissance, Resource Development, Initial Access, Execution, Persistence, etc.).
  • S D : Strategies available to defenders, corresponding to MITRE D3FEND countermeasures (e.g., Model, Detect, Harden, Isolate, Deceive, etc.)
  • P: Parameters influencing game models, such as attack severity, defense effectiveness, and forensic capability.
  • U A ( s A , s D ) : Utility function for attackers, representing the payoff based on their strategy s A and the defenders’ strategy s D .
  • U D ( s A , s D ) : Utility function for defenders, representing the payoff based on their strategy s D and the attackers’ strategy s A .
This research aims to solve the following problems:
  • Model Construction: Construct game models G ( A , D , S A , S D , P ) to represent the interactions between A and D.
  • Equilibrium Analysis: Identify Nash equilibria ( s A * , s D * ) such that
    U A ( s A * , s D * ) U A ( s A , s D * ) s A S A U D ( s A * , s D * ) U D ( s A * , s D ) s D S D
The goal is to derive optimal strategies ( s A * , s D * ) that enhance DFR, thereby informing the development of effective cybersecurity policies and strategies. This research contributes to the theoretical understanding of strategic interactions in cybersecurity, providing a foundation for future empirical studies and practical applications.

3.2. Methodology

We implemented a game-theoretic framework (Appendix B.12) integrating ATT&CK–D3FEND knowledge mapping with AHP-weighted DFR metrics. The framework consists of five components: (i) ATT&CK–D3FEND mapping (Section 4.1), (ii) DFR metric development, (iii) AHP weight determination (Section 3.5), (iv) payoff matrix construction (Section 3.3.3), and (v) equilibrium computation (Section 3.3.5).

Notation and Symbols

We employ standard mathematical notation throughout the paper, including sets ( S , T ), matrices (A, D), vectors (x, y), and scalar-valued functions ( A ( s , t ) , D ( s , t ) ). Key symbols are listed in Table 2, while the comprehensive notation table is provided in Appendix B.1. A summary of all strategy-related notation is presented in Table 3.

3.3. Proposed Approach

Inspired by Sun Tzu’s strategic principles, our approach models digital forensics as a normal-form game between two primary entities: attacker and defender. This game captures their strategy sets and resulting payoffs as follows:
  • Players:
    Attacker: 14 strategies ( s 1 , s 2 , , s 14 ).
    Defender: 6 strategies ( t 1 , t 2 , , t 6 ).
  • Payoff Matrices: Shown in Table 4 for the attacker and Table 5 for the defender, each matrix displays payoffs for every strategy combination.
  • Rationality: Both players are presumed rational, seeking to maximize their individual payoffs given knowledge of the opponent’s strategy. The game is simultaneous and non-zero-sum.
Attacker strategies include actions such as reconnaissance, execution, privilege escalation, and others. Defender strategies encompass modeling, detecting, deceiving, and additional controls. Modeling these interactions provides insight into the dynamic strategic landscape of digital forensics. As visualized in Figure 1, analysis of the payoff matrices reveals both outcomes and equilibrium points, highlighting the evolving nature of cyber threats. Darker matrix shades indicate higher attacker payoffs.

3.3.1. PNE Analysis

A pure-strategy Nash equilibrium (PNE) represents a stable outcome where neither player can improve their payoff by unilaterally changing strategy. Intuitively, this means that, (i) given the defender’s choice, the attacker’s strategy yields the highest possible payoff and (ii) given the attacker’s choice, the defender’s strategy yields the highest possible payoff. For the non-zero-sum bimatrix ( A , D ) , a strategy profile ( s * , t * ) is a PNE if
A ( s * , t * ) A ( s , t * ) s S , D ( s * , t * ) D ( s * , t ) t T ,
where S and T are the attacker and defender pure strategy sets, respectively. The first inequality ensures the attacker cannot gain by switching from s * to any other strategy s when the defender plays t * ; the second ensures the defender cannot gain by switching from t * to any other strategy t when the attacker plays s * .
For our game, we find that ( s 14 , t 3 ) = (‘Impact’, ‘Detect’) is a PNE, verified by checking the best-response conditions:
A ( s 14 , t 3 ) A ( s k , t 3 ) k { 1 , 2 , , 14 } ,
D ( s 14 , t 3 ) D ( s 14 , t l ) l { 1 , 2 , , 6 } .
Specifically, by inspecting Table 4 and Table 5, (i) A ( s 14 , t 3 ) = 41 is the maximum in column t 3 (attacker’s best response); (ii) D ( s 14 , t 3 ) = 29 is the maximum in row s 14 (defender’s best response). A full best-response scan over all 14 × 6 = 84 pure strategy pairs confirms this is the unique PNE (see Appendix B.7 for the verification algorithm). This PNE is highlighted in Figure 1.

3.3.2. MNE Analysis

Main Equilibrium (Non-Zero-Sum)
All results in the main text are based on the non-zero-sum bimatrix ( A , D ) constructed from independent attacker/defender utilities. Using nashpy’s vertex_enumeration on ( A , D ) , we obtained exactly one Nash equilibrium, which is pure at ( s 14 = Impact , t 3 = Detect ) . Support enumeration yields the same point, and the Karush–Kuhn–Tucker (KKT) conditions [51] are satisfied. The equilibrium is non-degenerate and stable under ε -perturbations up to ε 10 6 . This is the equilibrium reported in Table 4 and Table 5 and Figure 1. The zero-sum transform ( A , D ) yields exactly five equilibria under vertex enumeration: two pure equilibria at ( s 14 , t 1 ) = ( Impact , Model ) and ( s 12 , t 4 ) = ( Command   and   Control , Isolate ) , and three mixed equilibria with supports { s 12 , s 14 } × { t 1 , t 4 } , { s 9 , s 12 } × { t 4 , t 5 } , and { s 9 , s 11 } × { t 4 , t 5 } . All pass KKT verification, are non-degenerate, and are ε -stable for ε 10 6 (Note: Support enumeration reports only 3 equilibria on this instance; therefore, vertex enumeration is used as the primary method and ground truth). Complete support sets and probability distributions for these five equilibria are provided in Table A3 in the Appendix B.9, where they are presented as a robustness check.

3.3.3. Payoff Construction from ATT&CK→D3FEND Coverage

We construct defender-side coverage rates by aggregating many-to-many links from MITRE ATT&CK® Enterprise techniques (v13.1) to MITRE D3FEND techniques (v0.12.0-BETA-2) at the tactic × control-family level. Let S be the set of 14 ATT&CK tactics and T the set of six D3FEND control families. For each cell ( s , t ) we define
C ( s , t ) = # { k s : k t } # { k s } ,
where k t denotes the existence of at least one D3FEND technique in family t mitigating technique (or sub-technique) k. Sub-techniques were treated as first-class and were not rolled up into parent techniques. Counts were de-duplicated once per ( APT , technique , family ) when aggregating to tactics, as detailed in Section 4.2.
The attacker payoff matrix A is defined analogously from attacker-centric effectiveness against the same family t (selection rules unchanged).
For each strategy pair ( s , t ) (attacker tactic s vs. defender countermeasure t), we computed a weighted utility score that aggregates multiple DFR metrics. Intuitive explanation: We evaluated how well a defender countermeasure t addresses an attacker tactic s across 16 different dimensions (e.g., logging quality, evidence preservation, detection capability), weighted each dimension by its importance (determined via AHP expert elicitation), and summed the weighted scores to get an overall utility. Formal computation: Let w i denote the AHP weight for the i-th DFR metric (16 attacker metrics + 16 defender metrics = 32 total), and let M i ( s , t ) be the normalized metric score (0–1) for tactic s and countermeasure family t. The raw payoff value A ˜ ( s , t ) was computed as a weighted sum:
A ˜ ( s , t ) = i = 1 16 w i ( attacker ) · M i ( s , t ) ,
where the metric scores M i ( s , t ) are derived from coverage statistics (e.g., C ( s , t ) ) and attacker-centric effectiveness assessments. The defender-side matrix D ˜ ( s , t ) was computed analogously using the 16 defender metrics. These raw utility scores (typically in the range [ 0 , 1 ] ) were then linearly scaled and rounded to integers in the range [ 0 , 41 ] to produce the final payoff matrices shown in Table 4 and Table 5. The scaling transformation is as follows:
A ( s , t ) = 41 · A ˜ ( s , t ) + 0.5 , D ( s , t ) = 41 · D ˜ ( s , t ) + 0.5 ,
where · + 0.5 denotes rounding to the nearest integer. This scaling preserves the relative magnitudes of utility differences while mapping to a discrete payoff range suitable for equilibrium computation. Example: If A ˜ ( s , t ) = 0.85 , then A ( s , t ) = 41 × 0.85 + 0.5 = 35.35 = 35 . Because A and the defender-side matrix D are derived from distinct statistics, the game is non-zero-sum; in particular, we do not impose D = C . All main-text equilibria are computed on the non-zero-sum bimatrix ( A , D ) ; the zero-sum transform ( A , D ) is provided only as a robustness check in the Appendix B.9.
All scripts, versioning, and reproducibility information are provided in the Appendix B.

3.3.4. Payoff Matrices

The final payoff matrices for attacker and defender strategies are then shown in Table 4 and Table 5.

3.3.5. Mixed Nash Equilibrium Computation

We computed mixed Nash equilibria (MNE) using nashpy’s vertex_enumeration routine [52], which implements the vertex enumeration algorithm for bimatrix games (see [52] for implementation details). The method operates on the non-zero-sum bimatrix ( A , D ) , where A and D are attacker and defender payoff matrices derived independently from ATT&CK→D3FEND mappings (see Section 3.3.3).
Both A and D are utilities to be maximized. We passed the bimatrix ( A , D ) directly to the solver. When presenting defender costs C for interpretability, we converted to utilities via D = C for equilibrium computation and stated this explicitly where applicable. In our case, D was already constructed as a utility matrix (higher values are better for the defender), so no transformation was needed; the game was passed to nashpy as Game(A, D) without modification (where A and D are the matrix arrays). All code for equilibrium computation was archived in a public repository (see Appendix B).
For payoff matrices A and D, a mixed-strategy Nash equilibrium (MNE) is a pair of probability distributions ( x * , y * ) over strategies where neither player can improve their expected payoff by changing their probability distribution. Here, x * Δ 14 represents the attacker’s probability distribution over 14 strategies, and y * Δ 6 represents the defender’s probability distribution over 6 strategies.
Intuitive interpretation: In an MNE, players randomize over strategies such that (i) all strategies used with positive probability yield the same expected payoff (no strategy is better than another), and (ii) any unused strategy would yield a lower expected payoff (no incentive to switch). This creates strategic unpredictability while maintaining optimality.
Formal conditions: Let ( A y * ) i = j = 1 6 A ( s i , t j ) y j * denote the attacker’s expected payoff when playing pure strategy s i against the defender’s mixed strategy y * (i.e., the weighted average of payoffs across all defender strategies, weighted by their probabilities). Similarly, ( x * D ) j = i = 1 14 x i * D ( s i , t j ) denotes the defender’s expected payoff when playing pure strategy t j against the attacker’s mixed strategy x * . Then ( x * , y * ) is an MNE if and only if there exist constants v A and v D (the equilibrium values) such that
( A y * ) i = v A i supp ( x * ) , ( A y * ) i v A i supp ( x * ) ,
( x * D ) j = v D j supp ( y * ) , ( x * D ) j v D j supp ( y * ) ,
where supp ( x * ) = { i : x i * > 0 } and supp ( y * ) = { j : y j * > 0 } are the supports of the mixed strategies.
These are the standard KKT conditions for Nash equilibrium, ensuring that (i) all actions in support receive equal expected payoffs ( v A and v D ) and (ii) no excluded action yields a higher payoff (no profitable deviations).
For the non-zero-sum bimatrix ( A , D ) , vertex enumeration yields exactly one Nash equilibrium, which is pure at ( s 14 , t 3 ) = ( Impact , Detect ) . The vertex_enumeration method enumerates all vertices of the best-response polytopes, returning all equilibria of the game. Vertex enumeration can return equilibria that support enumeration misses in some numerical configurations, as it covers all polytope vertices rather than searching only over specific support sizes; for this reason, we use vertex enumeration as the primary method and ground truth.

3.3.6. Dynamics Illustration (Zero-Sum Variant)

The convergence trajectories shown in Figure 2 are based on the zero-sum variant ( A , D ) and illustrate attractor points under discrete-time best-response dynamics. For visualization, we ran best-response dynamics: starting from uniform random initial strategies, each player iteratively updates to a pure best response against the opponent’s current strategy. The attacker updates via s ( t + 1 ) = arg max s S A ( s , t ( t ) ) , and the defender updates via t ( t + 1 ) = arg max t T ( D ) ( s ( t ) , t ) (equivalently, arg min t T D ( s ( t ) , t ) for the zero-sum variant), where s ( t ) and t ( t ) denote pure strategies at iteration t. The process converges to attractor points corresponding to equilibria of ( A , D ) .
For the non-zero-sum bimatrix ( A , D ) , best-response dynamics converge to the unique PNE ( Impact , Detect ) . The trajectories shown in Figure 2 illustrate multiple attractor points under the zero-sum variant ( A , D ) , demonstrating different strategic patterns that emerge under the transformed game structure. These results are provided for exploratory purposes; all policy conclusions in this paper are drawn from the non-zero-sum bimatrix ( A , D ) .
Methodological Transparency Statement
Our main strategic conclusions are drawn from the empirical, non-zero-sum bimatrix ( A , D ) . To probe sensitivity to a worst-case, antagonistic setting, we also study a zero-sum variant ( A , D ) in Appendix B.9; this produces five equilibria (two pure-strategy and three mixed-strategy) and consistent tactical themes but is not used for headline results.

3.4. Utility Function

We modeled attacker–defender interactions using utility functions that quantify the payoff for each party. This is grounded in Multi-Criteria Decision Analysis (MCDA), an established framework for evaluating complex, conflicting criteria [13,52,53]. MCDA is well-suited for assessing the multifaceted nature of cybersecurity strategies.

3.4.1. Attacker Utility Function

The attacker’s utility was evaluated across 16 dimensions, such as Attack Success Rate, Resource Efficiency, and Stealthiness. Each metric was normalized between 0 (least favorable) and 1 (most favorable) and assigned a weight w i based on its relative importance. The attacker utility function is formulated as follows:
U Attacker = i = 1 16 w i M i
where M i is the normalized score for the i-th metric. This provides a granular view of attacker priorities and effectiveness (Table 6).

3.4.2. Defender Utility Function

Similarly, the defender’s utility evaluates 16 dimensions such as Logging Capabilities, Evidence Integrity, and Standards Compliance. The defender utility function is the following:
U Defender = j = 1 16 w j M j
where M j is the normalized score for the j-th metric. This reflects the organization’s forensic readiness (Table 7).
Ten DFIR practitioners independently completed two 16 × 16 pairwise comparison matrices (PCMs) covering attacker and defender metrics. We aggregated their Saaty-scale judgments via the geometric-mean consensus recommended in [54], normalized the resulting vectors to unit sum, and verified C R < 0.10 for both attacker and defender PCMs (details in Appendix B). Figure 3 summarizes the elicitation workflow; the full instrument, anonymized PCMs, and processing scripts are released with Appendix B. An expanded, step-by-step description of the elicitation protocol is provided in Appendix B.13.

3.4.3. Utility Calculation Algorithms

Utility scores follow the weighted sums in Equations (6) and (7). In practice this is a single O ( n ) pass that validates metric inputs, normalizes the AHP weights when necessary, and clamps floating-point drift to the [ 0 , 1 ] range. The exact pseudocode and input-checking logic are provided in Appendix B (Algorithm A1).
Readiness classification compares the resulting score u with a policy threshold T [ 0 , 1 ] (typically T 0.7 for “High DFR”). The decision rule is a constant-time check summarized in Algorithm A2 (Appendix B); we retain the notation here only to reference subsequent analyses that condition the { HighDFR , NeedsImprovement } outcome.
Given a readiness classification of NeedsImprovement , we flagged low-performing metrics by comparing each m i to the same threshold T. This scan produces an index set of candidate improvements that subsequently feeds the prioritisation step. The simple iteration is described in Algorithm A3 (Appendix B) and underpins the gap analysis reported in Section 4.7.

3.5. Prioritizing DFR Improvements

Enhancing DFR requires strategically targeting metrics within the utility function that have the greatest potential impact. Calibration with real-world experimental data ensures the validity of the model, aligning the results with operational realities [55].
To systematically determine improvement priorities, we apply the AHP, a structured multi-criteria decision framework that combines quantitative and qualitative assessments [54]. AHP provides a mathematical basis for ranking metrics, particularly highlighting low-scoring factors with high weight (Figure 4).

3.5.1. AHP Methodology for Weight Determination

We aggregate expert PCMs via the element-wise geometric mean, compute the principal eigenvector of the consensus matrix, and normalize the resulting weights to unit sum. The full eigenvector/LLSM workflow, together with numerical tolerances and fallback rules, is documented in Algorithm A4 (Appendix B). For both attacker and defender matrices the resulting C R values remain below 0.10 , confirming acceptable consistency. Figure 4 visualizes the prioritized weights, and Table 8 lists the values used throughout the analysis.
Expert Panel Procedures and Transparency
Recruitment and inclusion criteria. Ten domain experts were recruited based on the following criteria: (i) a minimum of 5 years of professional experience in digital forensics (DF), digital forensics and incident response (DFIR), or security operations; and/or (ii) peer-reviewed publications on game-theoretic security or digital forensic readiness. All participants provided written informed consent for participation and publication of anonymized, aggregated results. Participants declared any conflicts of interest and submitted domain-only email addresses for communication.
Data collection and independence. To limit anchoring bias and dominance effects, judgments were collected independently via an online instrument. Each expert completed two 16 × 16 PCMs, one for attacker metrics and one for defender metrics, without knowledge of other participants’ responses. Per-expert consistency ratios (CRs) were computed; participants had the option to revise judgments if C R > 0.10 . The released CSV files report per-expert CRs; no personally identifiable information is included.
Anonymization and data availability. Expert responses were anonymized prior to analysis. Anonymized demographics (years of experience, primary domain expertise, geographic region) and per-expert CR distributions are summarized in Appendix B (Figure A9) and provided in Appendix B. The full attacker/defender PCMs (six-decimal precision) and aggregated weights are released as CSV tables (Table A4) together with scripts to recompute eigenvector and LLSM priorities (available in the repository, Appendix B).
Institutional review and ethics. Under the Islamic Azad University Research Ethics policy, this expert-elicitation exercise—in which adult professionals provided non-sensitive technical judgments anonymously and no personally identifiable information was collected—does not constitute human-subject research requiring REC/IRB review. Electronic consent was obtained at the start of the instrument via an on-screen information sheet and an “I agree to participate” confirmation. No names, emails, IP addresses, or other identifiers were recorded; responses were stored only in anonymized, aggregate form (see Section 3 and the institutional review statement in the Acknowledgments Section).
Reporting Precision and Repeated Weights
Weights in Table 8 are shown to four decimals for readability. Because (i) judgments use a discrete 1–9 Saaty scale and (ii) we aggregate experts multiplicatively via geometric means, priority-vector components can legitimately cluster; rounding can therefore make nearby values appear equal (e.g., 0.0881 repeated). We provide six-decimal weights in Table A4; except where experts explicitly judged equal importance (yielding proportional rows/columns and thus equal eigenvector components), clustered entries separate at higher precision. Both aggregated PCMs satisfy the usual AHP criterion ( C R < 0.10 ).
Plausibility of Small and Similar CR Values
For each consensus PCM, we compute C I = ( λ max n ) / ( n 1 ) and C R = C I / R I with n = 16 and R I = 1.59 . Our consensus matrices yield λ max = 16.3200 and 16.3157 , hence C I = 0.02133 , 0.02105 and C R = 0.038 , 0.0132 . Low and similar CRs are expected under log-space geometric aggregation, which reduces dispersion and improves consistency across both PCMs produced by the same expert panel and protocol.
Additional AHP Diagnostics and Robustness
As robustness checks, we (i) recomputed priorities using the logarithmic least-squares (row geometric mean, LLSM) method and obtained cosine similarity > 0.999 with the eigenvector solution as well as identical top-k rankings; (ii) reported Koczkodaj’s triad inconsistency and the geometric consistency index (GCI) for the consensus PCMs (Table A5); (iii) performed a local perturbation study (1000 runs) that jitters entries by ± 1 Saaty step and applies ± 5 % multiplicative noise, observing median Spearman rank correlation ρ 0.95 and C R 0.10 (Figure A8); and (iv) summarized per-expert consistency via CR distributions, where aggregation reduces inconsistency (Figure A9).

3.5.2. Prioritization Process

The prioritization step ranks metrics by combining their AHP weight with the observed score gap and then screens the list against organizational feasibility constraints. We therefore concentrate remediation on high-weight, low-score metrics while treating the remaining steps (strategy design, deployment, monitoring) as operational details covered in Appendix B.

3.5.3. DFR Improvement Algorithm

We translate the priority list into an actionable program by sorting candidate metrics by their weight–gap product and selecting those that pass feasibility checks on cost, time, and staffing. Implementation details, including deterministic tie-breaking and monitoring hooks, are documented in Algorithm A5 (Appendix B). This keeps the main text focused on how the prioritized improvements feed the readiness evaluation reported in Section 4.7.
This process ensures high-impact improvements are implemented first, maximizing readiness gains within resource constraints.

3.6. Reevaluating the DFR

Following improvement implementation, the system’s forensic readiness was reevaluated by comparing updated utility scores to baseline values. An increased score confirms readiness enhancement, whereas stagnant or diminished scores indicate the need for further targeted measures.
This reevaluation provides a quantitative, evidence-based feedback loop, reinforcing decision making grounded in rigorous analysis. A comprehensive understanding of potential threats, combined with expertise in defensive and forensic techniques, enables organizations to continually strengthen preparedness and accelerate investigative processes.

4. Results

This section presents a detailed analysis of cyber threat dynamics, emphasizing the interplay between attacker tactics and defender strategies. It integrates empirical data, game-theoretic insights, and readiness evaluation to examine how different strategic behaviors influence DFR. Our findings illustrate the alignment between simulated outcomes and practical cybersecurity trends, providing a comprehensive understanding of real-world implications.

4.1. Data Collection and Methodology

We used MITRE ATT&CK® Enterprise v13.1 (9 May 2023) and MITRE D3FEND v0.12.0-BETA-2 (21 March 2023) via STIX 2.1 [56]. From the relationship path
intrusion-set uses attack-pattern
(Enterprise scope; direct edges only), we excluded objects/relations with revoked==true or x_mitre_deprecated==true. Technique IDs were normalized to uppercase; sub-techniques (e.g., T1027.013) were treated as distinct from their parents (e.g., T1027) and counted separately (no roll-up to parent techniques).
The final ATT&CK evidence set contains 260 technique assignments across ten intrusion sets: LeafMiner (17), Silent Librarian (13), OilRig (56), Ajax Security Team (6), Moses Staff (12), Cleaver (5), CopyKittens (8), APT33 (32), APT39 (52), MuddyWater (59).
Let S denote the 14 ATT&CK tactics and T the six D3FEND control families F = { Harden , Model , Evict , Isolate , Deceive , Detect } . For each cell ( s , t ) we aggregate many-to-many ATT&CK→D3FEND links and normalize by the number of (sub-)techniques under tactic s:
C ( s , t ) = # { k s : k t } # { k s } .
Here C is a tactic × family coverage rate. Game-theoretic payoff functions for attackers and defenders are defined later in Section 4.6; they are not constrained to satisfy D = C ; hence, the bimatrix game is non-zero-sum. Versioning, STIX scripts, and mapping CSVs are provided in the repository (Appendix B).

Extraction and Mapping Objects

Let A be the set of APT groups (intrusion sets), X the ATT&CK (sub-)techniques (Enterprise v13.1), and Y the D3FEND techniques (v0.12.0-BETA-2). We extract
E = { ( a , x ) A × X : a uses x } , M = { ( x , y ) X × Y : x mitigated_by y } ,
drop revoked/deprecated objects, and retain sub-techniques as first-class elements. D3FEND techniques are categorized by family ( y ) F .
Versioning, STIX scripts, and the robustness check are provided in the Appendix B.3.

4.2. Analysis of Tactics and Techniques

We clarify how the many-to-many ATT&CK↔D3FEND graph is aggregated at the tactic layer and how double counting is avoided.

4.2.1. Named Metrics

We use two de-duplicated, tactic-level metrics:
  • Family-coverage count (APT–technique–family incidence):
    Count ( τ , f ) = ( a , x ) : x τ 1 y : ( x , y ) M , family ( y ) = f ,
    which counts, for each ATT&CK tactic τ and D3FEND family f, the number of unique ( APT a , technique x ) instances with at least one mapped D3FEND technique in family f, de-duplicated once per ( a , x , f ) even if multiple y in the same family map to ( a , x ) .
  • Tactic recurrence (de-duplicated):
    Freq ( τ ) = a A 1 ( x τ : ( a , x ) E and x is the most specific technique observed for τ in a ) ,
i.e., for each APT a a tactic τ is credited at most once, preferring the most specific observed sub-technique (no observed sub-technique of x exists for a under τ ). Raw counts are shown in the figure; shares s ( τ ) = Freq ( τ ) / τ Freq ( τ ) and per-APT normalizations are reported in Appendix B.

4.2.2. How Figures Are Computed

Figure 5 reports Count ( τ , f ) as defined above (family coverage, not raw edge multiplicity). We also report per-APT normalizations p ^ ( τ , f ) = Count ( τ , f ) / | A | in Appendix B. Figure 6 reports Freq ( τ ) (one credit per APT per tactic, preferring the most specific sub-technique).

4.2.3. Notes and Limitations

Results inherit snapshot bias (versioning and reporting density). Mappings capture plausible mitigations, not guaranteed prevention. Full STIX extraction scripts and JSON snapshots are archived for reproducibility.

4.3. DFR Metrics Overview and Impact Quantification

Our analysis employs a set of 32 DFR metrics—16 attacker-centric and 16 defender-centric—detailed in Table 6 and Table 7. Each metric is normalized and weighted according to expert-driven AHP priorities.
The aggregate utility scores are computed as weighted sums of these metric values using Equations (6) and (7). Readiness is then computed as the difference between defender and attacker utility scores (Equation (9)).

4.3.1. Methods: Calibration-Based Synthetic Attacker Profiles

Notation Disambiguation
To avoid ambiguity, we use subscript notation consistently throughout: S T d for defender Staff Training (vs. S T for attacker Stealthiness) and P d for defender Preservation (vs. P for attacker Persistence). This notation is applied in tables, figures, and text wherever defender metrics are referenced.
To ensure coherence between defender improvements and adversary pressure, we co-generated attacker utility profiles under an explicit coupling prior. Let X ( d ) [ 0 , 1 ] 16 and X ( a ) [ 0 , 1 ] 16 denote defender/attacker metric vectors. We specified a sparse coupling matrix Γ (ATT&CK↔D3FEND-informed) that links defender capabilities (e.g., logging L, volatile capture V D C C , network forensics N F ) to reductions in adversarial stealth ( S T ), exfiltration effectiveness ( D E E ), and attribution resistance ( A R ), among others. For each case, attacker “before” profiles were drawn from weakly correlated Beta priors; “after” profiles were updated by the calibrated rule
X after ( a ) = clip [ 0 , 1 ] X before ( a ) λ Γ Δ d + ϵ , Δ d = X after ( d ) X before ( d ) ,
with case-wise λ [ 0.8 , 1.2 ] and small noise ϵ [ 0.02 , 0.02 ] . This construction avoids unrealistic across-the-board gains, preserves heterogeneity across cases, and operationalizes the ATT&CK↔D3FEND mapping. The AHP weights w ( d ) , w ( a ) are then applied to quantify readiness as
Readiness = k = 1 16 w k ( d ) x k ( d ) = 1 16 w ( a ) x ( a ) .
Limitations
These attacker profiles are synthetic, calibration-based for illustrative evaluation rather than field measurements; inter-rater reliability is not applicable to this section. For per-expert consistency ratios (CRs) and AHP validation, see Figure A9 (Appendix B).
In our synthetic calibration, profiles with limited logging exhibit higher attacker utility in the readiness balance; targeted improvements centered on logging and forensic preservation typically reduce the attacker utility component by approximately 15–30% under the specified Γ settings (see manifest and coupling files in Appendix B.5), contributing to higher net readiness. Specifically, linked attacker metrics (e.g., S T , DEE, AR) show reductions in the 15–30% range, while the overall weighted attacker utility component (computed via Equation (6)) shows a median reduction of approximately 8.0% across the C = 10 cases. Across the C = 10 synthetic cases, median readiness improvement ( Δ Readiness) is 18.0% (95% CI: [16.3%, 19.7%]).
This explicit linkage confirms the abstract’s key quantitative claims, grounded in our comprehensive DFR metric framework and empirical simulations.

4.4. Comparative Overview of Prior Game-Theoretic Models

To clarify how our contribution fits within the existing game-theoretic literature on digital forensic readiness, Table 9 contrasts the problem focus, game formulation, and reported quantitative effects of three representative approaches. The goal is not a head-to-head benchmark—the objectives, datasets, and evaluation protocols differ substantially—but rather to highlight complementary emphases and outputs.
These approaches address different operational contexts and are complementary. Our contribution adds an ATT&CK/D3FEND-grounded readiness score for APT-focused SMEs, analyzes both equilibrium and evolutionary dynamics, and also releases reproducible artifacts that practitioners can adapt to their environments.

Connection to Sun Tzu’s Strategic Wisdom

Our results quantitatively validate the strategic principles introduced in Section 3.3, operationalizing Sun Tzu’s dictum: “If you know the enemy and know yourself, you need not fear the result of a hundred battles.” “Know yourself” is operationalized through the 16 defender metrics (Table 7) that quantify organizational capabilities. Our results demonstrate that organizations with limited logging (incomplete self-knowledge) exhibit higher attacker utility, whereas strategic improvements centered on logging and forensic preservation (enhanced self-knowledge) reduce attacker utility by 15–30%, validating the importance of self-awareness. Know the enemy is operationalized through the 16 attacker utilities (Table 6) derived from empirical ATT&CK data on real APT groups. The Nash equilibrium at ( s 14 = Impact , t 3 = Detect ) reveals that understanding attacker priorities (Impact tactics) enables optimal defensive strategy (Detect), demonstrating the practical value of threat intelligence. We formalize three strategic states from Sun Tzu’s wisdom: comprehensive knowledge (targeted defense, high readiness), partial knowledge (vulnerable defense, moderate readiness), and ignorance (minimal resilience, low readiness). Our quantitative results show that moving from ignorance to comprehensive knowledge yields 18.0% median readiness improvement, providing concrete evidence that Sun Tzu’s strategic principles translate into measurable forensic readiness gains.

4.5. Attackers vs. Defenders: A Comparative Study

We analyzed how defensive techniques correspond to attacker strategies in frequency and efficacy. Figure 7 shows the distribution of D3FEND methods, such as Detect, Harden, Model, Evict, Isolate, and Deceive.
Our results indicate that attackers most frequently employ the Credential Access technique, with Impact-related tactics demonstrating the highest success rates. On the defense side, Detect emerged as the most frequently employed strategy, albeit with data limitations for the Impact category within the MITRE frameworks.

4.6. Game Dynamics and Strategy Analysis

4.6.1. Non-Zero-Sum ( A , D )

Using vertex enumeration, we obtain exactly one Nash equilibrium, which is pure at ( s 14 = Impact , t 3 = Detect ) . Support enumeration yields the same point, and the KKT conditions are satisfied. The equilibrium is non-degenerate and stable under ε -perturbations up to ε 10 6 .

4.6.2. Zero-Sum ( A , D )

Vertex enumeration returns exactly five Nash equilibria: two pure at ( s 14 , t 1 ) = ( Impact , Model ) and ( s 12 , t 4 ) = ( Command   and   Control , Isolate ) and three mixed with supports { s 12 , s 14 } × { t 1 , t 4 } , { s 9 , s 12 } × { t 4 , t 5 } , and { s 9 , s 11 } × { t 4 , t 5 } . All pass KKT verification, are non-degenerate, and are ε -stable for ε 10 6 (Note: Support enumeration reports only three equilibria on this instance; therefore, vertex enumeration is used as the primary method and ground truth). These results, detailed in Appendix B.9 and Table A3, demonstrate different strategic patterns under the transformed game structure and illustrate that attackers diversify tactics in response to defender adaptations, while defenders strategically redistribute effort based on attack probability.
All computed equilibria satisfy the KKT optimality conditions, are non-degenerate in the game-theoretic sense, and remain invariant under small payoff perturbations up to ε 10 6 . For each equilibrium we numerically verified KKT feasibility, dual feasibility, and complementarity, and we report best-response residuals < 10 14 (see Appendix B.8 for details).
Both analyses align with empirical evidence, showing that strategic flexibility—not rigid planning—enhances readiness. Convergence between theoretical modeling and real-world data reveals interdependencies between adaptive behaviors, informing more resilient DFR optimization frameworks.
While support enumeration formally identifies the PNE at the Attacker strategy ‘Impact’ paired with the Defender strategy ‘Detect’, the dynamic convergence analysis reveals that early trajectory states—starting from uniform or neutral mixed strategies—tend to gravitate toward the ‘Command_and_Control’ strategy for the attacker paired with ‘Detect’ for the defender. This suggests that, during the learning or adaptation phase, the system often stabilizes near this local attractor before potentially progressing to the PNE or possibly remaining trapped depending on the adaptation dynamics and information of the players. Therefore, both states are significant: the PNE represents the theoretically stable solution assuming full rationality and optimal play, whereas the observed convergence behavior reflects realistic intermediate strategic positioning players may occupy during actual cybersecurity engagements. Recognizing this duality informs defenders that, while ‘Impact/Detect’ is a strategic target equilibrium, adaptive defense must also address the commonly emerging patterns around ‘Command_and_Control/Detect’ to guide attackers toward less damaging behaviors.

4.7. Synthetic, Calibration-Based Case Profiles

To validate the effectiveness of our proposed framework, we generated synthetic, calibration-based case profiles that simulate forensic readiness scenarios before and after implementing the framework. These profiles are illustrative and calibration-based rather than field measurements; they operationalize the ATT&CK↔D3FEND mapping through an explicit coupling mechanism (see Section 4.3.1 and Appendix B). Ten case profiles are presented in Table 10 and Table 11.
Notation: To avoid ambiguity, defender metrics use a subscript d (e.g., S T d = Staff Training, P d = Preservation), while attacker metrics keep bare symbols (e.g., S T = Stealthiness, P = Persistence).
A comparative visualization (Figure 8) shows measurable improvement in post-implementation readiness scores for most metrics, validating the framework’s effectiveness. The 32 DFR metrics (16 defender + 16 attacker) serve as quantitative indicators of both forensic readiness (process capability) and forensicability (system capability to support investigations). Higher metric scores indicate improved readiness and enhanced forensicability, enabling organizations to transition from non-forensicability to forensicable states. Quality control metrics (detailed in Appendix B.5) confirm that all cases show positive defender and readiness improvements, with realistic heterogeneity: mean readiness improvement of approximately 18.0% (95% CI: [ 16.3 % , 19.7 % ]), with 34% of attacker metrics remaining unchanged per case, demonstrating selective suppression rather than global collapse. Across the 10 synthetic cases, organizations showed measurable improvements in forensic readiness, with systems becoming more forensicable as evidenced by enhanced logging, volatile data capture, and network forensics capabilities. The scores in Table 10 and Table 11 are generated from the YAML configuration and RNG seed (42) listed in the manifest (Appendix B.5), enabling exact regeneration of these values.

4.8. Sensitivity Analysis

4.8.1. Local Perturbation Sensitivity

We assess ranking robustness for both attacker and defender criteria using local perturbations of the aggregated AHP pairwise comparison matrices. For each metric i in turn, all entries in the i-th row/column (i.e., all comparisons involving i) are shifted by exactly one step on the Saaty 1–9 scale (up or down with equal probability), reciprocity is re-enforced, and a multiplicative uniform noise of ± 5 % is applied. We repeat this R = 200 times per metric and recompute the principal-eigenvector weights after each perturbation. The stability of metric i is quantified as
Stability i = 1 n j = 1 n rank j ( i ) rank j ( orig ) ,
where rank ( orig ) are ranks under the unperturbed matrix and rank ( i ) are ranks after perturbing metric i. Lower values indicate higher rank stability. The combined results for attacker (orange) and defender (blue) metrics are shown in Figure 9. In our data several metrics (e.g., ASR on the attacker side and L (Logging) on the defender side) exhibit relatively low average rank changes.

4.8.2. Monte Carlo Simulation

To examine how uncertainty in metric levels affects overall readiness, we run a Monte Carlo simulation with N = 20 , 000 draws. For each run we sample attacker and defender metric values independently from [ 0 , 1 ] and compute weighted scores using the AHP-derived weights. Readiness is computed as in Equation (9). We quantify each metric’s global sensitivity as the absolute Pearson correlation between the metric value and the readiness score. Figure 10 reports these correlations (higher bars indicate stronger influence). Figure 11 and Figure 12 visualize the bivariate relationships for each side.
Parameters including Collaboration (CB), Reputation and Prestige (RP), and Volatile Data Capture Capabilities (VDCC) had lower sensitivities (Figure 11), but their presence contributes to broader defense stability.
Overall, a few high-sensitivity metrics drive most of the variability in readiness, while the remaining ones provide a complementary signal that stabilizes performance.

4.9. Distribution of Readiness Score

The histogram in Figure 13 displays the standardized (z-scored) readiness values, z = ( x μ ) / σ , centered at zero. Because readiness is defined as defender minus attacker score, the raw values lie approximately in [ 1 , 1 ] ; standardization clarifies relative deviations from the mean, hence, the presence of both negative and positive values. The near-symmetric shape indicates a balanced spread around the average level of preparedness, with low-frequency tails representing unusually weak or unusually strong cases.
Key observations include the following:
  • Central Peak at 0.0: A high frequency around 0.0 indicates balanced readiness in most systems.
  • Symmetrical Spread: Even tapering on both sides suggests system stability across environments.
  • Low-Frequency Extremes: Outliers at the tails (−0.3 and +0.3) denote rare but critical deviations requiring targeted intervention.
This symmetrical distribution implies consistent readiness performance with occasional exceptional cases—either highly prepared or notably weak systems. When combined with sensitivity outcomes, this distribution reinforces the importance of continuous evaluation, adaptive planning, and targeted investment in high-impact metrics to sustain forensic readiness.

5. Discussion

Applying the proposed game-theoretic framework within an organizational cybersecurity context entails multiple phases and distinct challenges. Figure 14 could visualize these steps, which are summarized as follows:
  • Implementation Challenges: Real-world adoption may encounter barriers such as limited resources, integration costs, and the need for game theory expertise. Organizational resistance to change and adaptation to new analytical frameworks are additional challenges.
  • Integration with Existing Tools: The framework can align synergistically with existing platforms such as threat intelligence systems, SIEM, and EDR tools. These integrations can enhance decision making and optimize forensic investigation response times.
  • Decision Support Systems: Game-theoretic models can augment decision support processes by helping security teams prioritize investments, allocate resources, and optimize incident response based on adaptive risk modeling.
  • Training and Awareness Programs: Building internal capability is crucial. Training programs integrating game-theoretic principles into cybersecurity curricula can strengthen decision making under adversarial uncertainty.
  • Collaborative Defense Strategies: The framework supports collective defense through shared intelligence and coordinated responses. Collaborative action can improve deterrence and resilience against complex, multi-organizational threats.
  • Policy Implications: Incorporating game theory into cybersecurity has policy ramifications, including regulatory alignment, responsible behavior standards, and ethical considerations regarding autonomous or strategic decision models.
  • Case Studies and Use Cases: Documented implementations of game-theoretic approaches demonstrate measurable improvements in risk response and forensic readiness. Future research can expand these to varied industry sectors.
  • Future Directions: Continued innovation in game model development, integration with AI-driven threat analysis, and tackling emerging cyber challenges remain promising directions.
While adoption may face organizational or technical barriers, the approach remains adaptable. Incorporation with SIEM, EDR, and threat intelligence workflows allows for effective deployment, while targeted training mitigates skill gaps. Ultimately, these methods can significantly enhance decision support and defense coordination across security ecosystems.

5.1. Forensicability and Non-Forensicability

The dual concepts of forensicability and non-forensicability capture the degree to which digital systems are prepared to support forensic investigation and incident response.
Non-forensicability refers to an environment’s inability to effectively preserve or provide forensic evidence, typically arising from poor data retention, weak logging, or compromised evidence integrity. It represents a subjective assessment grounded in measurable deficiencies of DFR. Quantitatively, this can be evaluated via parameters such as log resolution, retention time, or audit trail completeness.
Conversely, forensicability characterizes systems that exhibit the structural and procedural maturity necessary for reliable forensic investigations. Hallmarks of forensicable systems include secure log management, redundancy in evidence capture, and adherence to recognized forensic standards. These factors not only strengthen internal visibility but also ensure evidence admissibility in legal contexts.
For organizations, enhancing forensicability means institutionalizing proactive DFR practices—ensuring data capture, protection, and retrieval mechanisms are integral to operations. Continuous assessment through forensic readiness metrics helps organizations transition from fragile, reactive postures to resilient, evidence-supported defenses.

5.2. Evolutionary Game Theory Analysis

Using Evolutionary Game Theory (EGT) enables modeling of how attacker and defender strategies evolve concurrently over time. This approach captures adaptation cycles that traditional static game models overlook.
The simulation results in Table 12 and Figure 15 illustrate how strategy populations change across generations. Attackers and defenders adjust probabilistically based on observed payoffs, with defender readiness influencing long-term stability.
Key insights derived from EGT include the following:
  • Evolutionary Dynamics: Attackers and defenders co-adapt in continuous feedback cycles; the success of one influences the next strategic shift in the other.
  • Replication and Mutation: Successful tactics replicate, while mutations introduce strategic diversity critical for both exploration and adaptation.
  • Equilibrium and Stability: Evolutionary Stable Strategies (ESSs) represent steady states where neither party benefits from deviation.
  • Co-evolutionary Context: The model exposes the perpetual nature of cyber escalation, showing that proactive defense and continuous readiness optimization are essential to remain resilient.

5.3. Attack Impact on Readiness and Investigation Phases

The simulation represented in Figure 16 demonstrates how attacks influence DFR through overlapping utility functions between attackers and defenders during investigation phases. Each incident reveals opportunities for defenders to improve readiness, forming a feedback mechanism between preparedness and investigative learning.
Observed overlaps indicate that investigation phases contribute directly to capability growth—highlighting that post-incident analysis enriches strategic defense planning and improves future preparedness.

5.4. Readiness and Training Level of the Defender

Simulations comparing varying defender experience levels (Junior, Mid-level, Senior) reveal a direct correlation between training maturity and overall forensic readiness (Figure 17). Higher training levels correlate with improved detection accuracy and evidence capture, illustrating that defensive effectiveness is both strategic and skill-dependent.

5.5. Attack Success and Evidence Collection Rates

Monte Carlo simulations of attack outcomes (Table 13) show that higher attacker capability increases success rates, while robust forensic processes substantially raise evidence collection probability across scenarios.

5.6. Comparative Analysis in SMB and SME Organizations

Recognizing that SMEs and SMBs differ in resource availability and defensive maturity, a comparative simulation was conducted (Table 14 and Table 15). Results show that SMBs typically exhibit higher resilience, yet both types face elevated risks under “irrational” attacker behaviors.
Figure 18 visualizes how SME/SMB impact metrics shift between baseline and irrational attacker settings for the SQLi and DDoS scenarios described above.

Irrational Attacker Behavior Analysis

By modeling partial randomness in adversarial decision making, “irrational behavior” introduces deviations from expected attacks, thus reflecting real-world unpredictability. Figure 19 and Figure 20 illustrate the expanded range of outcomes.
This model highlights the necessity for robust intrusion detection, endpoint monitoring, and anomaly-based analytics to counteract unpredictable threats and enhance resilience in both small- and mid-scale organizations.

5.7. Limitations and Future Work

While this research offers a structured quantitative contribution to DFR and security strategy development, certain limitations acknowledge the boundaries of current modeling:
  • Model Complexity: Real-world human elements and deep organizational dynamics may extend beyond current model parameters.
  • Data Availability: Reliance on open-source ATT&CK and D3FEND datasets limits coverage of emerging threat behaviors.
  • Computational Needs: Evolutionary modeling and large-scale simulations require high-performance computing resources.
  • Expert Bias: AHP-based weighting depends on expert judgment, introducing potential subjective bias despite structured controls.
Future research could pursue the following:
  • Real-time Adaptive Models: Integrating continuous learning to instantly adapt to threat changes.
  • AI/ML Integration: Employing predictive modeling for attacker intent recognition and defense automation.
  • Cross-Organizational Collaboration: Expanding to cooperative game structures for shared threat response.
  • Empirical Validation: Conducting large-scale quantitative studies to reinforce and generalize model applicability.

6. Conclusions

This study presents a comprehensive game-theoretic framework that formalizes classical strategic principles, notably those of Sun Tzu, into a structured model applicable to contemporary cyber conflict analysis. By modeling the strategic interplay between attackers and defenders, the framework bridges traditional strategic insight and modern decision-theoretic planning. It integrates MITRE ATT&CK–D3FEND mappings, incorporates readiness scoring across simulated organizational scenarios, and aligns these insights with quantitative game-theoretical analyses.
Our results on the non-zero-sum bimatrix ( A , D ) identify exactly one Nash equilibrium, which is pure at ( s 14 = Impact , t 3 = Detect ) . The PNE emphasizes the defender’s Detect strategy as a robust counter to attackers’ Impact-focused operations. Analysis of the zero-sum variant ( A , D ) (see Appendix B.9) yields exactly five equilibria: two pure at ( s 14 , t 1 ) = ( Impact , Model ) and ( s 12 , t 4 ) = ( Command   and   Control , Isolate ) and three mixed with supports { s 12 , s 14 } × { t 1 , t 4 } , { s 9 , s 12 } × { t 4 , t 5 } , and { s 9 , s 11 } × { t 4 , t 5 } . These results suggest that, under the transformed game structure, defenders should allocate approximately 90–95% of their forensic effort toward modeling controls, preserving a smaller fraction for real-time detection. This balance introduces useful strategic unpredictability, increases the attacker’s required effort, and diminishes overall intrusion success probabilities. However, all policy conclusions in this paper are drawn from the non-zero-sum bimatrix ( A , D ) .
Operationally, these insights were translated into a four-phase assessment process encompassing readiness scoring, maturity classification, gap identification, and roadmap prioritization. Through this practical translation, our model enables measurable digital forensic improvements. In our synthetic calibration across 10 organizational profiles, limited logging increases the attacker utility component of readiness, whereas strategic improvements centered on logging and forensic preservation typically reduce it. Specifically, linked attacker metrics (e.g., S T , DEE, AR) show reductions in the 15–30% range, while the overall weighted attacker utility component (computed via Equation (6)) shows a median reduction of approximately 8.0% across cases, contributing to higher net readiness. Most importantly, our quantitative results demonstrate that moving from ignorance to comprehensive knowledge yields a median readiness improvement of 18.0% (95% CI: [16.3%, 19.7%]) across N = 20 , 000 Monte Carlo trials, providing concrete evidence that Sun Tzu’s strategic principles translate into measurable forensic readiness gains (see manifest hyperparameters in Appendix B.5).
The contribution of the framework is best seen through the systematic comparison with prior game-theoretic forensics approaches (Table 1). While prior work has advanced game-theoretic forensics in specific domains (Bayesian modelling, tool selection, cognitive security, risk analysis, adaptive observability), our framework brings these ideas together by the following: (i) systematically integrating ATT&CK and D3FEND into a unified game-theoretic model; (ii) grounding quantitative utilities in AHP-weighted DFR metrics; (iii) providing explicit MNE analysis with support conditions; and (iv) delivering actionable SME guidance under resource constraints.
Our goal is to complement prior theoretical work by showing how existing game-theoretic concepts and knowledge graphs can be instantiated for quantitative forensic-readiness planning. The novelty lies in the metricisation and integration choices, not in proposing a new game-theoretic foundation.

6.1. Limitations

The framework’s accuracy depends on the quality and granularity of metric data as well as expert input for AHP weighting. Factors such as organizational diversity, resource variability, and evolving adversary behaviors could influence transferability across different organizational contexts. Additionally, the assumption of static utility parameters (derived from ATT&CK–D3FEND mappings and AHP weights) in our normal-form bimatrix game model simplifies real-world dynamics, which are inherently fluid and adaptive. The simultaneous-move structure does not capture temporal evolution or learning behaviors that may occur in extended adversarial interactions.

6.2. Future Research Directions

Building upon this foundation, several research extensions are envisaged:
  • Extended Environmental Applications: Adapting the framework to cloud-native, IoT, and blockchain ecosystems where architectural differences create distinct forensic challenges.
  • Dynamic Threat Intelligence Integration: Employing real-time data feeds and AI-based analytics to enable adaptive recalibration of utilities and strategy distributions.
  • Standardized Readiness Benchmarks: Developing comparative industry baselines for forensic maturity that support cross-organizational evaluation and improvement.
  • Automated Response Coupling: Integrating automated incident response and orchestration tools to bridge the gap between detection and remediation.
  • Enhanced Evolutionary Models: Expanding evolutionary game formulations to capture longer-term strategic co-adaptations between attackers and defenders.
  • Large-Scale Empirical Validation: Conducting multi-sector, empirical measurement campaigns to statistically validate and refine equilibrium predictions.
In conclusion, the proposed game-theoretic approach provides a mathematically grounded, strategically informed basis for advancing DFR. By linking equilibrium analysis with empirical readiness metrics, the framework offers a repeatable methodology for optimizing resource allocation, reducing attacker advantage, and fostering systemic resilience against persistent and adaptive cyber threats.

Supplementary Materials

The following supporting information can be downloaded from https://www.mdpi.com/article/10.3390/jcp5040105/s1.

Author Contributions

Conceptualization, M.V., S.J. and H.N.; methodology, M.V. and H.N.; software, M.V.; validation, M.V., S.J. and H.N.; formal analysis, M.V. and H.N.; investigation, M.V.; resources, S.J.; data curation, M.V.; writing—original draft preparation, M.V.; writing—review and editing, S.J. and H.N.; visualization, M.V.; supervision, S.J. and H.N.; project administration, S.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding. The APC was funded by the authors.

Institutional Review Board Statement

Under the Islamic Azad University Research Ethics policy, this expert-elicitation study—adult professionals providing non-sensitive technical judgments, with no collection of personally identifiable information—does not constitute human-subject research and therefore did not require REC/IRB review. Participation was anonymous by design; no names, emails, IP addresses, or other identifiers were recorded. Procedures and safeguards are summarized in Section 3 (“Expert panel procedures and transparency”) and detailed in Appendix B (Figure A9 and associated data files).

Informed Consent Statement

Electronic consent was obtained at the start of the online instrument via an on-screen information page and an “I agree to participate” confirmation; participation was voluntary and anonymous.

Data Availability Statement

The original contributions presented in this study are included in the article material. Further inquiries can be directed to the corresponding author.

Acknowledgments

This work received no external financial or in-kind support beyond the authors’ personal resources.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

    The following abbreviations are used in this manuscript:
AHPAnalytic Hierarchy Process
APTAdvanced Persistent Threat
ATT&CKMITRE Adversarial Tactics, Techniques, and Common Knowledge
CASECyber-investigation Analysis Standard Expression
CIAConfidentiality, Integrity, Availability (triad)
CSIRTComputer Security Incident Response Team
CVSSCommon Vulnerability Scoring System
D3FENDMITRE Defensive Countermeasures Knowledge Graph
DFIRDigital Forensics and Incident Response
DFRDigital Forensic Readiness
DDoSDistributed Denial of Service
EDREndpoint Detection and Response
EGTEvolutionary Game Theory
ESSEvolutionarily Stable Strategy
IDPSIntrusion Detection and Prevention System
JCPJournal of Cybersecurity and Privacy
MCDAMulti-Criteria Decision Analysis
MNEMixed Nash Equilibrium
NDRNetwork Detection and Response
NENash Equilibrium
PNEPure Nash Equilibrium
SIEMSecurity Information and Event Management
SMBSmall and Medium Business
SMESmall and Medium Enterprise
SQLiStructured Query Language injection
TTPTactics, Techniques, and Procedures
UCOUnified Cyber Ontology
XDRExtended Detection and Response

Appendix A. Simulation Model and Settings

Appendix A.1. Readiness Components

Let T [ 0 , 1 ] (training), E [ 0 , 1 ] (experience), and V [ 0 , 1 ] (attacker capability; larger is stronger). We define
C = w T T + w Aw Aw , F = w E E + w P P ,
where Aw is security awareness, and P denotes forensics procedures. We use ( w T , w Aw ) = ( 0.7 , 0.3 ) and ( w E , w P ) = ( 0.8 , 0.2 ) , with C , F [ 0 , 1 ] .

Appendix A.2. Outcome Probabilities

For attacker strength s { Low , Med , High } ,
p attack ( s C ) = clip b s [ 1 α ( C μ C ) ] , 0 , 1 ,
p collect ( s F , κ ) = clip e s + β ( F μ F ) γ ( κ μ κ ) , 0 , 1 ,
where ( b Low , b Med , b High ) = ( 0.25 , 0.53 , 0.75 ) , ( e Low , e Med , e High ) = ( 0.93 , 0.96 , 0.94 ) , α = 0.5 , β = 0.20 , γ = 0.25 , and μ C = μ F = 0.75 , μ κ = 0.60 are centering constants. Evidence complexity κ U ( 0.30 , 0.90 ) . The function clip ( x , 0 , 1 ) truncates to [ 0 , 1 ] .

Appendix A.3. Sampling and Maturity Regimes

For each trial we draw T , A , E , P from regime-specific ranges:
  • Junior+Mid+Senior: T U ( 0.40 , 0.90 ) , E U ( 0.40 , 1.00 ) ;
  • Mid+Senior: T U ( 0.60 , 0.90 ) , E U ( 0.60 , 1.00 ) ;
  • Senior: T U ( 0.70 , 0.90 ) , E U ( 0.70 , 1.00 ) .
Attacker capability V used in Figure 17 is sampled per point to shape the green curve.

Appendix A.4. Experiment Size and Uncertainty

We run N = 50 , 000 trials per attacker strength with seed 42. Rates are reported as p ^ ± 1.96 p ^ ( 1 p ^ ) / N (95% CI).

Appendix B. Additional Materials

Appendix B.1. Notation Table

To ensure clarity and avoid confusion between similar-looking symbols, we provide a notation table that systematically lists all symbols used throughout this manuscript. The notation follows standard mathematical conventions: sets use calligraphic font (e.g., S , T ), matrices use bold uppercase (e.g., A, D), vectors use bold lowercase (e.g., x, y), and scalars use italic font (e.g., A ( s , t ) , D ( s , t ) ).
Notation disambiguation: To avoid confusion between similar symbols:
  • A ( s , t ) vs. A: A ( s , t ) is a scalar (single entry in the matrix), while A is the entire matrix (bold uppercase). Similarly, D ( s , t ) is a scalar, D is the matrix.
  • x, y vs. x, y: Lowercase x, y denote individual elements or indices; bold x, y denote mixed-strategy vectors (probability distributions over pure strategies).
  • Strategy indices: Pure strategies s S and t T are elements (not vectors); mixed strategies x and y are probability vectors over these sets.
Table A1. Notation table for game-theoretic DFR framework.
Table A1. Notation table for game-theoretic DFR framework.
SymbolTypeDescription
Game Structure
S SetAttacker strategy set: S = { s 1 , s 2 , , s 14 } where | S | = 14 ATT&CK tactics (see Table 3)
T SetDefender strategy set: T = { t 1 , t 2 , , t 6 } where | T | = 6 D3FEND control families (see Table 3)
sElementPure attacker strategy: s S (e.g., s 1 = Reconnaissance , s 14 = Impact )
tElementPure defender strategy: t T (e.g., t 1 = Model , t 3 = Detect )
AMatrixAttacker payoff matrix: A R 14 × 6 , where entry A ( s , t ) is the attacker’s utility for strategy pair ( s , t )
DMatrixDefender payoff matrix: D R 14 × 6 , where entry D ( s , t ) is the defender’s utility for strategy pair ( s , t )
A ( s , t ) ScalarAttacker scalar payoff: A ( s , t ) [ 0 , 41 ] (unitless utility, higher is better for attacker)
D ( s , t ) ScalarDefender scalar payoff: D ( s , t ) [ 0 , 41 ] (unitless utility, higher is better for defender)
( s * , t * ) PairPure Nash equilibrium: strategy profile where s * S and t * T are mutual best responses
Mixed Strategies and Equilibria
xVectorAttacker mixed strategy: x Δ 14 where Δ 14 = { x R 0 14 : i = 1 14 x i = 1 } (probability distribution over S )
yVectorDefender mixed strategy: y Δ 6 where Δ 6 = { y R 0 6 : j = 1 6 y j = 1 } (probability distribution over T )
x * VectorAttacker equilibrium mixed strategy: x * Δ 14 (optimal probability vector)
y * VectorDefender equilibrium mixed strategy: y * Δ 6 (optimal probability vector)
x i Scalari-th component of x: x i [ 0 , 1 ] (probability that attacker plays strategy s i )
y j Scalarj-th component of y: y j [ 0 , 1 ] (probability that defender plays strategy t j )
Δ n SetProbability simplex: Δ n = { p R 0 n : i = 1 n p i = 1 }
supp ( x * ) SetSupport of x * : supp ( x * ) = { i : x i * > 0 } (indices of strategies played with positive probability)
supp ( y * ) SetSupport of y * : supp ( y * ) = { j : y j * > 0 }
( A y * ) i ScalarAttacker expected payoff: ( A y * ) i = j = 1 6 A ( s i , t j ) y j * (expected utility when attacker plays s i and defender plays y * )
( x * D ) j ScalarDefender expected payoff: ( x * D ) j = i = 1 14 x i * D ( s i , t j ) (expected utility when defender plays t j and attacker plays x * )
v A ScalarAttacker equilibrium value: v A = ( A y * ) i for all i supp ( x * ) (constant expected payoff at equilibrium)
v D ScalarDefender equilibrium value: v D = ( x * D ) j for all j supp ( y * ) (constant expected payoff at equilibrium)
DFR Metrics and Utilities
M i ( s , t ) ScalarNormalized DFR metric i: M i ( s , t ) [ 0 , 1 ] for strategy pair ( s , t ) (see Section 3.3.3)
w i ( A ) ScalarAHP weight for attacker metric i: w i ( A ) [ 0 , 1 ] , i = 1 16 w i ( A ) = 1 (see Table 8)
w j ( D ) ScalarAHP weight for defender metric j: w j ( D ) [ 0 , 1 ] , j = 1 16 w j ( D ) = 1 (see Table 8)
A ˜ ( s , t ) ScalarNormalized attacker utility: A ˜ ( s , t ) = i = 1 16 w i ( A ) M i ( s , t ) [ 0 , 1 ]
D ˜ ( s , t ) ScalarNormalized defender utility: D ˜ ( s , t ) = j = 1 16 w j ( D ) M j ( D ) ( s , t ) [ 0 , 1 ]
x k ( d ) ScalarDefender metric k value: x k ( d ) [ 0 , 1 ] (see Section 4.3.1)
x ( a ) ScalarAttacker metric value: x ( a ) [ 0 , 1 ] (see Section 4.3.1)
Γ MatrixCoupling matrix: Γ R 16 × 16 linking defender capabilities to attacker metric suppression (see Section 4.3.1)

Appendix B.2. Repository and Data Access

All code, scripts, data snapshots, and additional resources supporting this study are available through the following repositories:
  • Main repository (reproducibility package): https://github.com/Mehrn0ush/gtDFR (accessed on 12 November 2025). This repository contains all Python 3.10 scripts, configuration files (config/), processed datasets (data/), analysis scripts for equilibrium computation and profile generation (scripts/), and supplementary artifacts (Supplementary Materials, including AHP weight tables, expert consistency reports, equilibrium exports, and manifest files). The repository is version e, d and a tagged release matching this manuscript is provided.
  • Historical mapping snapshots (archival only): https://github.com/Mehrn0ush/gtDFR/tree/main/aptcsv/archive (accessed on 12 November 2025). These gists archive earlier CSV/JSON exports used during initial ATT&CK/D3FEND data exploration and MITRE STIX parsing. They are not required to reproduce the main results but are provided for transparency.

Framework Versions and Snapshots

This study used the following framework versions: (i) MITRE ATT&CK® Enterprise: v13.1 (released 9 May 2023); (ii) MITRE D3FEND: v0.12.0-BETA-2 (released 21 March 2023). For robustness validation, we compared our baseline (ATT&CK v13.1, D3FEND v0.12.0-BETA-2) against MITRE ATT&CK® Enterprise v14.1 (14 November 2023) and v18.0 (28 October 2025) paired with MITRE D3FEND v0.21.0 (August 2025); detailed comparison results are documented in Appendix B.3.

Appendix B.3. ATT&CK/D3FEND Snapshot Comparison Report (Robustness Check)

Appendix B.3.1. Purpose

To assess robustness of our findings to framework updates, we re-ran the exact STIX 2.1 pipeline used in the main text on later public ATT&CK snapshots and compared per-APT technique sets across three versions.

Appendix B.3.2. Pipeline

The extraction pipeline follows the same methodology described in Appendix B.4. In this robustness run, the D3FEND mapping CSV was not available; therefore, only ATT&CK technique-set deltas are reported (coverage-by-family deltas are omitted).
Versions Compared
Table A2. ATT&CK bundle snapshots used for robustness checks.
Table A2. ATT&CK bundle snapshots used for robustness checks.
SnapshotATT&CK VersionBundle Modified (UTC)
Baseline (closest to analysis window)v13.12023-05-09T14:00:00.188Z
Intermediatev14.12023-11-14T14:00:00.188Z
Latestv18.02025-10-28T14:00:00.188Z

Appendix B.3.3. Executive Summary

From v13.1 → v18.0 across our ten intrusion sets: 26 techniques were added and 7 removed overall. Five of ten groups show any change; the largest net change is for OilRig (+21, −1). Qualitative conclusions in the main text (tactic × family coverage trends and tactic recurrence patterns) remain substantively unchanged.

Appendix B.3.4. Per-APT Deltas (v13.1 → v18.0)

APT Groupv13.1v14.1v18.0AddedRemoved
APT3332323112
APT3952525321
AjaxSecurityTeam66600
Cleaver55500
CopyKittens88800
LeafMiner17171700
MosesStaff12121211
MuddyWater59595812
OilRig565676211
SilentLibrarian13131300

Appendix B.3.5. Provenance (Console Excerpt)

ATT\&CK/D3FEND Robustness Check - Three-Version Comparison: v13.1 ’ v14.1 ’ v18.0
 
Loaded March 2023 bundle: v13.1 (Modified: 2023-05-09T14:00:00.188Z)
Loaded v14.1 bundle      (Modified: 2023-11-14T14:00:00.188Z)
Loaded latest bundle     (Modified: 2025-10-28T14:00:00.188Z)
 
D3FEND CSV not found (coverage deltas omitted this run).
 
Per-APT counts (v13.1, v14.1, v18.0):
APT33 32,32,31 | APT39 52,52,53 | AjaxSecurityTeam 6,6,6 | Cleaver 5,5,5 |
CopyKittens 8,8,8 | LeafMiner 17,17,17 | MosesStaff 12,12,12 |
MuddyWater 59,59,58 | OilRig 56,56,76 | SilentLibrarian 13,13,13
 
Summary: groups with any change (v13.1’v18.0): 5/10; largest change: OilRig (+21, -1)
Totals (v13.1’v18.0): Added 26, Removed 7

Appendix B.3.6. Note on D3FEND Coverage in This Run

The D3FEND mapping CSV was not available on this machine at execution time; therefore, only ATT&CK technique-set deltas are reported above. When the mapping CSV is provided, the same script emits family-coverage deltas as well; our qualitative conclusions were unchanged in test runs with the mapping present.

Appendix B.3.7. Reproducibility

The robustness analysis report (attack_d3fend_robustness_report.md) and the exact STIX bundle versions used in this comparison are archived in Appendix B.

Appendix B.4. STIX Data Extraction Methodology

Data extraction from STIX 2.1 bundles followed this process:
  • STIX object types: Intrusion-set (APT groups), relationship (with relationship_type="uses"), attack-pattern (techniques).
  • Scope: Enterprise ATT&CK only (excluded Mobile, ICS, PRE-ATT&CK, which is historical and no longer maintained).
  • Relationship path: Intrusion-set.id uses attack-pattern.id, using only direct relationships (not transitive via malware/tools); we extracted external_references with source_name="mitre-attack" to obtain technique IDs.
  • Filtering: Excluded all objects and relationships with revoked==true or x_mitre_ deprecated==true.
  • Normalization: ATT&CK technique IDs normalized to uppercase (e.g., t1110.003T1110.003).
  • Sub-technique handling: Counted exact sub-techniques (e.g., T1027.013) as distinct from parent techniques (e.g., T1027); no rollup performed.
Extraction scripts implementing this methodology are archived in the repository. A robustness check comparing our v13.1 baseline with MITRE ATT&CK® Enterprise v14.1 (November 2023) and v18.0 (October 2025) confirms technique assignment stability across versions; the robustness check script and detailed comparison report are available in the repository (see Appendix B.3).

Appendix B.5. Synthetic Profile Generation for Case Studies

Each profile is generated by a scripted pipeline that couples empirical ATT&CK→ D3FEND coverage with stochastic sampling. We draw “before” defender and attacker metric vectors from Beta distributions parameterized by the observed coverage statistics, apply targeted uplifts to defender metrics flagged by the framework, and update attacker metrics through the coupled rule in Equation (8) using the sparse Γ matrix to encode cross-framework suppression effects. Every run records the random seed, hyperparameters, and emitted CSV/manifest files, enabling exact regeneration of the ten-case bundle reported in Section 4.7.
The synthetic, calibration-based case profiles used in Section 4.7 (Table 10 and Table 11) are generated using a reproducible pipeline documented in the repository. Key components are listed below.

Appendix B.9.1. Configuration Files

  • Metric definition configuration: Canonical definitions of all 32 DFR metrics (16 defender + 16 attacker), including metric IDs, names, descriptions, and framework targeting flags. This configuration serves as the single source of truth for metric ordering and naming across all scripts, tables, and visualizations (available in the repository, Appendix B).
  • Coupling matrix configuration: Sparse coupling matrix Γ (16 × 16) encoding ATT&CK↔D3FEND linkages. Nonzero entries (coefficients typically 0.05–0.20) specify which defender metrics suppress which attacker utilities, operationalizing the game-theoretic framework (available in the repository, Appendix B).

Appendix B.5.2. Generator Script

The profile generation script (Section 4.3.1, available in the repository) generates defender and attacker profiles using the following:
  • Defender before profiles: Beta-distributed samples with mild correlation structure (L↔D↔LR, IR↔ ST d ↔NF, I↔ P d ).
  • Defender after profiles: For metrics with targeted_by_framework: true, add uplift in [0.10, 0.30]; for others, allow small drift [−0.02, +0.05].
  • Attacker before profiles: Weakly correlated Beta priors with semantic blocks (ASR–IA, ST–AR–DN, RE–RT–FG–RP).
  • Attacker after profiles: Coupled update via Equation (8) with case-wise scaling λ [ 0.8 , 1.2 ] and noise ϵ [ 0.02 , 0.02 ] .

Appendix B.5.3. Output Files

The generator produces CSV files containing defender and attacker profiles (before and after intervention) and a manifest file containing the following:
  • RNG seed and hyperparameters.
  • Nonzero pattern and values of Γ .
  • SHA-256 hashes of all output files.
  • QC metrics: mean Δ Defender, Δ Attacker, Δ Readiness with 95% confidence intervals; fraction of unchanged attacker metrics per case; median decrease for linked metrics; and verification that attacker mean does not collapse to zero.

Appendix B.5.4. Visualizations

Four diagnostic plots (generated by the visualization script available in the repository) demonstrate the defender→attacker linkage:
  • Coupling heatmap (Figure A1): 16 × 16 heatmap of Γ showing nonzero entries (attacker rows × defender columns). Darker cells denote stronger suppression coefficients.
  • Before/after ridges (Figure A2): Density distributions for selected attacker metrics ( S T , DEE, AR as linked; RE, CB as controls) across C cases, comparing before vs. after profiles. Linked metrics shift modestly downward; controls remain stable.
  • Bipartite linkage graph (Figure A3): Bipartite graph with attacker metrics (left) and defender metrics (right); edge set { ( a , k ) : Γ [ a , k ] > 0 } with widths proportional to Γ [ a , k ] . This visualizes the structural prior behind the calibration.
  • Readiness waterfall (Figure A4): Per-case stacked bars showing Δ D c (defender uplift), Δ A c (attacker suppression), and net Δ Readiness from Equation (9). This clarifies how defender uplift and attacker suppression combine.
All files are archived in the repository with full documentation (see Supplementary Materials README_PROFILES.md).
Figure A1. Coupling matrix Γ (attacker rows, defender columns). Darker cells denote stronger suppression coefficients. Zero cells (pale) indicate no assumed linkage under Γ . Defender columns use subscript notation ( S T d for Staff Training, P d for Preservation) to distinguish from attacker metrics ( S T for Stealthiness, P for Persistence). This matrix operationalizes the ATT&CK↔D3FEND mapping used in Equation (8).
Figure A1. Coupling matrix Γ (attacker rows, defender columns). Darker cells denote stronger suppression coefficients. Zero cells (pale) indicate no assumed linkage under Γ . Defender columns use subscript notation ( S T d for Staff Training, P d for Preservation) to distinguish from attacker metrics ( S T for Stealthiness, P for Persistence). This matrix operationalizes the ATT&CK↔D3FEND mapping used in Equation (8).
Jcp 05 00105 g0a1
Figure A2. Before/after distributions for selected attacker metrics across C = 10 cases. Linked metrics ( S T , DEE, AR) shift modestly downward; controls (RE, CB) remain stable. This demonstrates selective, realistic shifts where Γ is nonzero, rather than global collapse. Zero cells (pale) in the coupling heatmap indicate no assumed linkage under Γ . In the heatmap (Figure A1), defender metrics use subscript notation ( S T d , P d ) to distinguish from attacker metrics ( S T , P).
Figure A2. Before/after distributions for selected attacker metrics across C = 10 cases. Linked metrics ( S T , DEE, AR) shift modestly downward; controls (RE, CB) remain stable. This demonstrates selective, realistic shifts where Γ is nonzero, rather than global collapse. Zero cells (pale) in the coupling heatmap indicate no assumed linkage under Γ . In the heatmap (Figure A1), defender metrics use subscript notation ( S T d , P d ) to distinguish from attacker metrics ( S T , P).
Jcp 05 00105 g0a2
Figure A3. Bipartite linkage graph: attacker metrics (left) and defender metrics (right). Edge widths proportional to Γ [ a , k ] . This visualizes the structural prior behind the calibration, making explicit which defender capabilities suppress which attacker utilities.
Figure A3. Bipartite linkage graph: attacker metrics (left) and defender metrics (right). Edge widths proportional to Γ [ a , k ] . This visualizes the structural prior behind the calibration, making explicit which defender capabilities suppress which attacker utilities.
Jcp 05 00105 g0a3
Figure A4. Per-case waterfall: defender uplift ( Δ D ), attacker change ( Δ A ), and net Δ Readiness. The dotted horizontal line at 0 marks the baseline. Net Δ Readiness equals the difference between defender uplift (green) and attacker suppression (red). This clarifies how defender improvements and attacker suppression combine to yield net readiness gains. All cases show positive readiness improvement.
Figure A4. Per-case waterfall: defender uplift ( Δ D ), attacker change ( Δ A ), and net Δ Readiness. The dotted horizontal line at 0 marks the baseline. Net Δ Readiness equals the difference between defender uplift (green) and attacker suppression (red). This clarifies how defender improvements and attacker suppression combine to yield net readiness gains. All cases show positive readiness improvement.
Jcp 05 00105 g0a4

Appendix B.6. Optional Fuzzy Robustness (Not Used for Table 4 and Table 5)

Appendix B.6.1. Scope

This appendix provides a compact, reproducible fuzzy formulation (Mamdani min–max, centroid over [ 0 , 1 ] ) that can be used to stress-test the readiness scoring, but it was not used to compute the main payoff matrices in Table 4 and Table 5.

Appendix B.6.2. Specification

Inputs (normalized to [ 0 , 1 ] ): EvidenceQuality, DetectionCoverage, ResponseLatency; Output: ForensicReadinessScore. Triangular membership functions and labels (Low/Medium/High for Evidence Quality and Detection Coverage; Fast/Moderate/Slow for latency) are provided in Table A3, while the output Low/Medium/High sets appear in Table A4. The full 3 × 3 × 3 rule base is listed in Table A5, and Algorithm A6 summarizes the pipeline. A ± 10 % perturbation on ( a , c ) shows small variation (MAD < 0.03 ), indicating qualitative robustness of the readiness ranking.

Appendix B.7. Pure Nash Equilibrium Verification

A pure-strategy Nash equilibrium ( s * , t * ) requires that s * is a best response to t * and t * is a best response to s * . We verify PNE by a full best-response scan: for each defender action t j ( j { 1 , , 6 } ), we compute the (possibly set-valued) attacker best-response set BR A ( t j ) = arg max i A ( i , j ) ; for each attacker action s i ( i { 1 , , 14 } ), we compute the defender set BR D ( s i ) = arg max j D ( i , j ) . A profile ( s i , t j ) is a PNE iff i BR A ( t j ) and j BR D ( s i ) . Exhaustive enumeration over all 14 × 6 = 84 pairs confirms that ( s 14 , t 3 ) = ( Impact , Detect ) is the unique PNE.

Appendix B.8. Equilibrium Computation Details

Mixed Nash equilibria (MNE) were computed using the nashpy vertex_enumeration method [52] on the bimatrix game ( A , D ) (non-zero-sum), where A and D are attacker and defender utility matrices constructed independently from ATT&CK→D3FEND mappings (Section 3.3.3).

Appendix B.8.1. Sign Convention

Both A and D are utilities to be maximized. We therefore pass Game(A, D) directly to the solver. When presenting defender costs C for interpretability, we convert to utilities via D = C . In our implementation, D is already a utility matrix (higher is better), so no transformation is needed.

Appendix B.8.2. Main Results (Non-Zero-Sum)

For ( A , D ) , vertex enumeration yields exactly one equilibrium, which is pure at ( s 14 , t 3 ) = ( Impact , Detect ) . Support enumeration on ( A , D ) returns the same single equilibrium.

Appendix B.8.3. Zero-Sum Variant

For the zero-sum transform ( A , D ) , vertex enumeration returns exactly five equilibria: two pure at ( s 14 , t 1 ) = ( Impact , Model ) and ( s 12 , t 4 ) = ( Command   and   Control , Isolate ) , and three mixed with supports { s 12 , s 14 } × { t 1 , t 4 } , { s 9 , s 12 } × { t 4 , t 5 } , and { s 9 , s 11 } × { t 4 , t 5 } . Support enumeration reports only three equilibria on this instance; therefore, we treat vertex enumeration as the primary method and ground truth. All five equilibria satisfy first-order optimality (KKT) conditions—we numerically verified KKT feasibility, dual feasibility, and complementarity—and have tiny best-response residuals (on the order of 10 15 for all equilibria, computed as the difference between maximum and minimum expected payoffs within support). These results are provided as a robustness check; policy conclusions in the main text are based on the non-zero-sum bimatrix ( A , D ) .
All code—including exact payoff matrices—is archived in the repository (Appendix B). We provide two driver scripts for equilibrium computation: one for the non-zero-sum bimatrix ( A , D ) and one for the zero-sum variant ( A , D ) . Solver versioning and environment details are documented in the repository. No ad-hoc tie-breaking was required; all reported solutions are bona fide Nash equilibria.

Appendix B.8.4. Non-Degeneracy and ε-Perturbation

The scaling from [ 0 , 1 ] to [ 0 , 41 ] with integer rounding could, in principle, introduce degeneracy (ties). Empirically, all reported equilibria are non-degenerate: (i) the unique PNE in ( A , D ) has distinct payoffs at equilibrium; (ii) all equilibria in ( A , D ) and ( A , D ) have well-defined supports with equal payoffs only within support. We also performed ε -perturbation tests (uniform noise with ε { 10 6 , 10 7 , 10 8 } ): the number and type of equilibria remained unchanged for ( A , D ) and remained five for ( A , D ) at ε = 10 6 , with coordinate differences effectively zero for ( A , D ) . This indicates numerical stability to small perturbations.

Appendix B.9. Zero-Sum Sensitivity Variant (A,-D)

For exploratory purposes, we analyze the zero-sum sensitivity variant G = ( A , D ) . Unlike the main non-zero-sum bimatrix ( A , D ) , G yields five equilibria under vertex enumeration: two pure-strategy equilibria (MNE 2: ( Impact , Model ) and MNE 4: ( Command   and   Control , Isolate ) ) and three mixed-strategy equilibria (MNE 1, MNE 3, MNE 5). These results are provided only as a robustness check; all policy conclusions are drawn from ( A , D ) .
Figure A5 shows the equilibrium probability distributions for all five equilibria; Figure A6 contrasts the unique PNE from ( A , D ) with the two pure equilibria from ( A , D ) . Figure A7 provides a grid view of all five equilibria.
Figure A5. All five equilibria in the zero-sum variant ( A , D ) : 2 pure + 3 mixed. Bars show equilibrium probabilities over strategies. MNE 2 and MNE 4 are pure-strategy equilibria; MNE 1, MNE 3, and MNE 5 are mixed-strategy equilibria. Note: Main results use the non-zero-sum bimatrix ( A , D ) , which yields a single equilibrium at ( Impact , Detect ) .
Figure A5. All five equilibria in the zero-sum variant ( A , D ) : 2 pure + 3 mixed. Bars show equilibrium probabilities over strategies. MNE 2 and MNE 4 are pure-strategy equilibria; MNE 1, MNE 3, and MNE 5 are mixed-strategy equilibria. Note: Main results use the non-zero-sum bimatrix ( A , D ) , which yields a single equilibrium at ( Impact , Detect ) .
Jcp 05 00105 g0a5
Table A3. Zero-sum sensitivity variant ( A , D ) : Complete support sets and probability distributions for all five Mixed Nash Equilibria (MNE). Only strategies with non-zero probabilities are shown. Probabilities rounded to four decimal places. MNE 2 and MNE 4 are pure-strategy equilibria; MNE 1, MNE 3, and MNE 5 are mixed-strategy equilibria. Note: These equilibria are from the zero-sum transform ( A , D ) ; main results use the non-zero-sum bimatrix ( A , D ) .
Table A3. Zero-sum sensitivity variant ( A , D ) : Complete support sets and probability distributions for all five Mixed Nash Equilibria (MNE). Only strategies with non-zero probabilities are shown. Probabilities rounded to four decimal places. MNE 2 and MNE 4 are pure-strategy equilibria; MNE 1, MNE 3, and MNE 5 are mixed-strategy equilibria. Note: These equilibria are from the zero-sum transform ( A , D ) ; main results use the non-zero-sum bimatrix ( A , D ) .
MNEAttacker StrategyDefender Strategy
MNE 1 s 12 : Command and Control (0.5714) t 1 : Model (0.9512)
(mixed) s 14 : Impact (0.4286) t 4 : Isolate (0.0488)
MNE 2 s 14 : Impact (1.0000) t 1 : Model (1.0000)
(pure)
MNE 3 s 9 : Discovery (0.2000) t 4 : Isolate (0.7857)
(mixed) s 12 : Command and Control (0.8000) t 5 : Deceive (0.2143)
MNE 4 s 12 : Command and Control (1.0000) t 4 : Isolate (1.0000)
(pure)
MNE 5 s 9 : Discovery (0.1111) t 4 : Isolate (0.0769)
(mixed) s 11 : Collection (0.8889) t 5 : Deceive (0.9231)
Figure A6. Left: unique PNE for ( A , D ) at ( s 14 , t 3 ) = ( Impact , Detect ) . Middle/Right: the two pure equilibria in ( A , D ) : MNE 2 at ( s 14 , t 1 ) = ( Impact , Model ) and MNE 4 at ( s 12 , t 4 ) = ( Command   and   Control , Isolate ) . Note: Main results use the non-zero-sum bimatrix ( A , D ) .
Figure A6. Left: unique PNE for ( A , D ) at ( s 14 , t 3 ) = ( Impact , Detect ) . Middle/Right: the two pure equilibria in ( A , D ) : MNE 2 at ( s 14 , t 1 ) = ( Impact , Model ) and MNE 4 at ( s 12 , t 4 ) = ( Command   and   Control , Isolate ) . Note: Main results use the non-zero-sum bimatrix ( A , D ) .
Jcp 05 00105 g0a6
Figure A7. Grid view of attacker/defender equilibrium distributions for each of the five equilibria in the zero-sum variant ( A , D ) . MNE 2 and MNE 4 are pure-strategy equilibria; MNE 1, MNE 3, and MNE 5 are mixed-strategy equilibria. See Table A3 for exact probabilities and support sets. Note: Main results use the non-zero-sum bimatrix ( A , D ) , which yields a single equilibrium at ( Impact , Detect ) .
Figure A7. Grid view of attacker/defender equilibrium distributions for each of the five equilibria in the zero-sum variant ( A , D ) . MNE 2 and MNE 4 are pure-strategy equilibria; MNE 1, MNE 3, and MNE 5 are mixed-strategy equilibria. See Table A3 for exact probabilities and support sets. Note: Main results use the non-zero-sum bimatrix ( A , D ) , which yields a single equilibrium at ( Impact , Detect ) .
Jcp 05 00105 g0a7
Table A4. Six-decimal AHP-derived metric weights (attacker/defender).
Table A4. Six-decimal AHP-derived metric weights (attacker/defender).
AttackerDefender
Metric Weight Metric Weight
ASR0.109 421L0.088 138
RE0.047 640I0.088 138
ST0.092 165D0.042 359
DEE0.088 782VDCC0.064 290
TTE0.047 640E0.046 193
EC0.088 782IR0.088 138
AR0.081 419DR0.048 163
RT0.047 640NF0.081 980
IA0.092 165ST0.081 980
P0.081 419LR0.048 163
AD0.057 193A0.055 784
DN0.026 406C0.046 039
LG0.043 317T0.069 305
CB0.026 246R0.053 188
FG0.021 048V0.042 359
RP0.048 717P0.055 784
Table A5. Consistency diagnostics for aggregated AHP matrices.
Table A5. Consistency diagnostics for aggregated AHP matrices.
Matrixn λ max CICRGCIKoczkodaj
Attacker1616.452 8000.030 2000.019 0000.066 6000.875 000
Defender1616.315 7000.021 0000.013 2000.047 7000.500 000
Figure A8. AHP rank stability under local perturbations. For each metric i, all pairwise comparisons in row/column i are shifted by one 1–9 Saaty step, and multiplicative noise U [ 0.95 , 1.05 ] is applied; reciprocity is restored, and priorities are recomputed. Bars show mean absolute rank change (lower = more stable). Left: attacker; right: defender.
Figure A8. AHP rank stability under local perturbations. For each metric i, all pairwise comparisons in row/column i are shifted by one 1–9 Saaty step, and multiplicative noise U [ 0.95 , 1.05 ] is applied; reciprocity is restored, and priorities are recomputed. Bars show mean absolute rank change (lower = more stable). Left: attacker; right: defender.
Jcp 05 00105 g0a8
Figure A9. Per-expert Consistency Ratio (CR) and Demographics. Violin plots for attacker and defender experts showing CR distributions. Dashed line indicates C R = 0.10 . Anonymized demographics (years of experience, primary domain expertise, geographic region) and per-expert CR values are provided in Appendix B.
Figure A9. Per-expert Consistency Ratio (CR) and Demographics. Violin plots for attacker and defender experts showing CR distributions. Dashed line indicates C R = 0.10 . Anonymized demographics (years of experience, primary domain expertise, geographic region) and per-expert CR values are provided in Appendix B.
Jcp 05 00105 g0a9
Table A6. ATT&CK tactic × D3FEND control-family coverage counts (family-coverage). Each cell shows Count ( τ , f ) : the number of unique ( APT a , technique x ) instances with at least one mapped D3FEND technique in family f, de-duplicated once per ( APT , technique , family ) (see Section 4.2.1 and Section 4.2). Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023). Raw data extracted from STIX 2.1 JSON files and mapping CSVs (see Appendix B).
Table A6. ATT&CK tactic × D3FEND control-family coverage counts (family-coverage). Each cell shows Count ( τ , f ) : the number of unique ( APT a , technique x ) instances with at least one mapped D3FEND technique in family f, de-duplicated once per ( APT , technique , family ) (see Section 4.2.1 and Section 4.2). Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023). Raw data extracted from STIX 2.1 JSON files and mapping CSVs (see Appendix B).
ATT&CK TacticModelHardenDetectIsolateDeceiveEvict
Collection312128151818
Command and Control411541151541
Credential Access585258434949
Defense Evasion474245333340
Discovery271824131518
Execution302227181921
Exfiltration231520121316
Initial Access352632212326
Lateral Movement362833222427
Persistence382935242628
Privilege Escalation393036252729
Reconnaissance000000
Resource Development000000
Impact292226181921
Total371274360253228288
Table A7. ATT&CK tactic frequency counts across APT groups (de-duplicated). Each tactic τ is credited at most once per APT (preferring the most specific sub-technique evidence; see Section 4.2.1 and Section 4.2). Raw counts shown; shares s ( τ ) = Freq ( τ ) / τ Freq ( τ ) and per-APT normalizations p ^ ( τ ) = Freq ( τ ) / | A | are available in the repository CSV exports. Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023).
Table A7. ATT&CK tactic frequency counts across APT groups (de-duplicated). Each tactic τ is credited at most once per APT (preferring the most specific sub-technique evidence; see Section 4.2.1 and Section 4.2). Raw counts shown; shares s ( τ ) = Freq ( τ ) / τ Freq ( τ ) and per-APT normalizations p ^ ( τ ) = Freq ( τ ) / | A | are available in the repository CSV exports. Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023).
ATT&CK TacticFrequency
Collection8
Command and Control9
Credential Access10
Defense Evasion10
Discovery8
Execution9
Exfiltration7
Initial Access9
Lateral Movement9
Persistence9
Privilege Escalation9
Reconnaissance0
Resource Development0
Impact8
Total104
Table A8. D3FEND defensive control family frequency counts across mapped APT–technique instances. Counts represent the number of unique APT–technique pairs mapped to each D3FEND family, aggregated across all ATT&CK tactics (see Section 4.2). Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023).
Table A8. D3FEND defensive control family frequency counts across mapped APT–technique instances. Counts represent the number of unique APT–technique pairs mapped to each D3FEND family, aggregated across all ATT&CK tactics (see Section 4.2). Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023).
D3FEND Control FamilyFrequency
Model371
Harden274
Detect360
Isolate253
Deceive228
Evict288
Total1774

Appendix B.10. Attacker Utility Metrics and Scoring Preferences (Full Details)

The complete detailed table for Attacker Utility Metrics and Scoring Preferences, originally referenced as Table 6 in the main text, is provided below. This table includes qualitative descriptions for each metric level (0, 0.1–0.3, 0.4–0.6, 0.7–0.9, 1) to guide metric evaluation. Each metric is evaluated on a continuous scale from 0 (least favorable) to 1 (most favorable).
Table A9. Attacker utility metrics and scoring preferences (full details).
Table A9. Attacker utility metrics and scoring preferences (full details).
MetricDescriptionScore
Attack Success Rate (ASR)Attack success rate is nearly nonexistent0
Attacks are occasionally successful0.1–0.3
Attacks are successful about half of the time0.4–0.6
Attacks are usually successful0.7–0.9
Attacks are always successful1
Resource Efficiency (RE)Attacks require considerable resources with low payoff0
Attacks require significant resources but have a moderate payoff0.1–0.3
Attacks are somewhat resource efficient0.4–0.6
Attacks are quite resource efficient0.7–0.9
Attacks are exceptionally resource efficient1
Stealthiness (ST)Attacks are always detected and attributed0
Attacks are usually detected and often attributed0.1–0.3
Attacks are sometimes detected and occasionally attributed0.4–0.6
Attacks are seldom detected and rarely attributed0.7–0.9
Attacks are never detected nor attributed1
Data Exfiltration Effectiveness (DEE)Data exfiltration attempts always fail0
Data exfiltration attempts succeed only occasionally0.1–0.3
Data exfiltration attempts often succeed0.4–0.6
Data exfiltration attempts usually succeed0.7–0.9
Data exfiltration attempts always succeed1
Time-to-Exploit (TTE)Vulnerabilities are never successfully exploited before patching0
Vulnerabilities are exploited before patching only occasionally0.1–0.3
Vulnerabilities are often exploited before patching0.4–0.6
Vulnerabilities are usually exploited before patching0.7–0.9
Vulnerabilities are always exploited before patching1
Evasion of Countermeasures (EC)Countermeasures always successfully thwart attacks0
Countermeasures often successfully thwart attacks0.1–0.3
Countermeasures sometimes fail to thwart attacks0.4–0.6
Countermeasures often fail to thwart attacks0.7–0.9
Countermeasures never successfully thwart attacks1
Attribution Resistance (AR)The attacker is always accurately identified0
The attacker is often accurately identified0.1–0.3
The attacker is sometimes accurately identified0.4–0.6
The attacker is seldom accurately identified0.7–0.9
The attacker is never accurately identified1
Reusability of Attack Techniques (RT)Attack techniques are always one-off, never reusable0
Attack techniques are occasionally reusable0.1–0.3
Attack techniques are often reusable0.4–0.6
Attack techniques are usually reusable0.7–0.9
Attack techniques are always reusable1
Impact of Attacks (IA)Attacks cause no notable disruption or loss0
Attacks cause minor disruption or loss0.1–0.3
Attacks cause moderate disruption or loss0.4–0.6
Attacks cause major disruption or loss0.7–0.9
Attacks cause catastrophic disruption or loss1
Persistence (P)The attacker cannot maintain control over compromised systems0
The attacker occasionally maintains control over compromised systems0.1–0.3
The attacker often maintains control over compromised systems0.4–0.6
The attacker usually maintains control over compromised systems0.7–0.9
The attacker always maintains control over compromised systems1
Adaptability (AD)The attacker is unable to adjust strategies in response to changing defenses0
The attacker occasionally adjusts strategies in response to changing defenses0.1–0.3
The attacker often adjusts strategies in response to changing defenses0.4–0.6
The attacker usually adjusts strategies in response to changing defenses0.7–0.9
The attacker always adjusts strategies in response to changing defenses1
Deniability (DN)The attacker cannot deny involvement in attacks0
The attacker can occasionally deny involvement in attacks0.1–0.3
The attacker can often deny involvement in attacks0.4–0.6
The attacker can usually deny involvement in attacks0.7–0.9
The attacker can always deny involvement in attacks1
Longevity (LG)The attacker’s operations are quickly disrupted0
The attacker’s operations are often disrupted0.1–0.3
The attacker’s operations are occasionally disrupted0.4–0.6
The attacker’s operations are rarely disrupted0.7–0.9
The attacker’s operations are never disrupted1
Collaboration (CB)The attacker never collaborates with others0
The attacker occasionally collaborates with others0.1–0.3
The attacker often collaborates with others0.4–0.6
The attacker usually collaborates with others0.7–0.9
The attacker always collaborates with others1
Financial Gain (FG)The attacker never profits from attacks0
The attacker occasionally profits from attacks0.1–0.3
The attacker often profits from attacks0.4–0.6
The attacker usually profits from attacks0.7–0.9
The attacker always profits from attacks1
Reputation and Prestige (RP)The attacker gains no reputation or prestige from attacks0
The attacker gains little reputation or prestige from attacks0.1–0.3
The attacker gains some reputation or prestige from attacks0.4–0.6
The attacker gains considerable reputation or prestige from attacks0.7–0.9
The attacker’s reputation or prestige is greatly enhanced by each attack1

Appendix B.11. Defender Utility Metrics and Scoring Preferences (Full Details)

The complete detailed table for Defender Utility Metrics and Scoring Preferences, originally referenced as Table 7 in the main text, is provided below. This table includes qualitative descriptions for each metric level (0, 0.1–0.3, 0.4–0.6, 0.7–0.9, 1) to guide metric evaluation. Each metric is evaluated on a continuous scale from 0 (least favorable) to 1 (most favorable).
Table A10. Defender utility metrics and scoring preferences (full details).
Table A10. Defender utility metrics and scoring preferences (full details).
MetricDescriptionScore
Logging and Audit Trail Capabilities (L)No logging or audit trail capabilities0
Minimal or ineffective logging and audit trail capabilities0.1–0.3
Moderate logging and audit trail capabilities0.4–0.6
Robust logging and audit trail capabilities with some limitations0.7–0.9
Comprehensive and highly effective logging and audit trail capabilities1
Integrity and Preservation of Digital Evidence (I)Complete loss of all digital evidence, including backups0
Severe damage or compromised backups with limited recoverability0.1–0.3
Partial loss of digital evidence, with some recoverable data0.4–0.6
Reasonable integrity and preservation of digital evidence, with recoverable backups0.7–0.9
Full integrity and preservation of all digital evidence, including secure and accessible backups1
Documentation and Compliance with Digital Forensic Standards (D)No documentation or non-compliance with digital forensic standards0
Incomplete or inadequate documentation and limited adherence to digital forensic standards0.1–0.3
Basic documentation and partial compliance with digital forensic standards0.4–0.6
Well-documented processes and good adherence to digital forensic standards0.7–0.9
Comprehensive documentation and strict compliance with recognized digital forensic standards1
Volatile Data Capture Capabilities (VDCC)No volatile data capture capabilities0
Limited or unreliable volatile data capture capabilities0.1–0.3
Moderate volatile data capture capabilities0.4–0.6
Effective volatile data capture capabilities with some limitations0.7–0.9
Robust and reliable volatile data capture capabilities1
Encryption and Decryption Capabilities (E)No encryption or decryption capabilities0
Weak or limited encryption and decryption capabilities0.1–0.3
Moderate encryption and decryption capabilities0.4–0.6
Strong encryption and decryption capabilities with some limitations0.7–0.9
Highly secure encryption and decryption capabilities1
Incident Response Preparedness (IR)No incident response plan or team in place0
Initial incident response plan, not regularly tested or updated, with limited team capability0.1–0.3
Developed incident response plan, periodically tested, with trained team0.4–0.6
Comprehensive incident response plan, regularly tested and updated, with a well-coordinated team0.7–0.9
Advanced incident response plan, continuously tested and optimized, with a dedicated, experienced team1
Data Recovery Capabilities (DR)No data recovery processes or tools in place0
Basic data recovery tools, with limited effectiveness0.1–0.3
Advanced data recovery tools, with some limitations in terms of capabilities0.4–0.6
Sophisticated data recovery tools, with high success rates0.7–0.9
Comprehensive data recovery tools and processes, with excellent success rates1
Network Forensics Capabilities (NF)No network forensic capabilities0
Basic network forensic capabilities, limited to capturing packets or logs0.1–0.3
Developed network forensic capabilities, with ability to analyze traffic and detect anomalies0.4–0.6
Advanced network forensic capabilities, with proactive threat detection0.7–0.9
Comprehensive network forensic capabilities, with full spectrum threat detection and automated responses1
Staff Training and Expertise ( ST d )No trained staff or expertise in digital forensics0
Few staff members with basic training in digital forensics0.1–0.3
Several staff members with intermediate-level training, some with certifications0.4–0.6
Most staff members with advanced-level training, many with certifications0.7–0.9
All staff members are experts in digital forensics, with relevant certifications1
Legal & Regulatory Compliance (LR)Non-compliance with applicable legal and regulatory requirements0
Partial compliance with significant shortcomings0.1–0.3
Compliance with most requirements, some minor issues0.4–0.6
High compliance with only minor issues0.7–0.9
Full compliance with all relevant legal and regulatory requirements1
Accuracy (A)No consistency in results, many errors and inaccuracies in digital forensic analysis0
Frequent errors in analysis, high level of inaccuracy0.1–0.3
Some inaccuracies in results, needs further improvement0.4–0.6
High level of accuracy, few inconsistencies or errors0.7–0.9
Extremely accurate, consistent results with virtually no errors1
Completeness (C)Significant data overlooked, very incomplete analysis0
Some relevant data collected, but analysis remains substantially incomplete0.1–0.3
Most of the relevant data collected and analyzed, but some gaps remain0.4–0.6
High degree of completeness in data collection and analysis, minor gaps0.7–0.9
Comprehensive data collection and analysis, virtually no information overlooked1
Timeliness (T)Extensive delays in digital forensic investigation process, no urgency0
Frequent delays, slow response time0.1–0.3
Reasonable response time, occasional delays0.4–0.6
Quick response time, infrequent delays0.7–0.9
Immediate response, efficient process, no delays1
Reliability (R)Unreliable techniques, inconsistent and unrepeatable results0
Some reliability in techniques, but results are often inconsistent0.1–0.3
Mostly reliable techniques, occasional inconsistencies in results0.4–0.6
High reliability in techniques, few inconsistencies0.7–0.9
Highly reliable and consistent techniques, results are dependable and repeatable1
Validity (V)No adherence to standards, methods not legally or scientifically acceptable0
Minimal adherence to standards, many methods not acceptable0.1–0.3
Moderate adherence to standards, some methods not acceptable0.4–0.6
High adherence to standards, majority of methods are acceptable0.7–0.9
Strict adherence to standards, all methods used are legally and scientifically acceptable1
Preservation ( P d )No procedures in place for evidence preservation, evidence frequently damaged or lost0
Minimal preservation procedures, evidence sometimes damaged or lost0.1–0.3
Moderate preservation procedures, occasional evidence damage or loss0.4–0.6
Robust preservation procedures, rare instances of evidence damage or loss0.7–0.9
Comprehensive preservation procedures, virtually no damage or loss of evidence1

Appendix B.12. Game Theory Background

This appendix provides detailed game-theoretic preliminaries, including formal definitions of players, actions, payoffs, and equilibrium concepts, that complement the concise overview in Section 3.

Appendix B.12.1. Players and Actions

We consider finite games with player index set N = { 1 , 2 , , n } . Each player i N has a finite strategy (action) set A i , and the set of all pure-strategy profiles is the Cartesian product
A = A 1 × A 2 × × A n .

Appendix B.12.2. Payoff Functions and Utility

Each player has a payoff function u i : A R that maps a profile a = ( a 1 , a 2 , , a n ) to a real-valued utility. These utilities encode the players’ preferences given their own choices and those of their opponents.
In the digital-forensic readiness context, the two primary players are the following:
  • Defender (D): The organization allocating forensic and defensive resources.
  • Attacker (A): The adversary selecting tactics to compromise digital assets.

Appendix B.12.3. Illustrative Scenario

To ground the notation, consider a stylized two-player defender–attacker interaction where the defender chooses between high or low forensic-tool investment (HI FT vs. LI FT), and the attacker chooses between a sophisticated attack (SA) and a simpler attack (SI). The payoff structure can be represented as shown below.
Table A11. Payoff matrix for a toy defender–attacker interaction (defender payoff first).
Table A11. Payoff matrix for a toy defender–attacker interaction (defender payoff first).
Attack (SA)Attack (SI)
Defender (HI FT) ( u D ( HI FT , SA ) , u A ( HI FT , SA ) ) ( u D ( HI FT , SI ) , u A ( HI FT , SI ) )
Defender (LI FT) ( u D ( LI FT , SA ) , u A ( LI FT , SA ) ) ( u D ( LI FT , SI ) , u A ( LI FT , SI ) )

Appendix B.12.4. Payoff Interpretation

The qualitative interpretation of Table A11 is as follows:
  • Defender payoffs: High investment improves readiness for sophisticated attacks at the expense of potential over-provisioning when the attacker chooses simpler tactics; low investment reduces costs but increases risk against sophisticated attacks.
  • Attacker payoffs: Sophisticated attacks demand more resources yet yield higher gains if defences are weak; simpler attacks are cheaper but produce lower returns, especially against well-prepared defenders.

Appendix B.12.5. Equilibrium Concepts

This toy model highlights standard equilibrium concepts:
  • Pure Nash Equilibrium (PNE): A profile ( s * , t * ) where neither player can improve utility by deviating unilaterally.
  • Mixed Nash Equilibrium (MNE): Players randomize over strategies to prevent opponents from gaining by deviation, a common representation for adversaries facing uncertainty (e.g., APTs varying attack vectors).
The full empirical analysis in the main text replaces this toy payoff matrix with the ATT&CK/D3FEND-derived utilities A and D.

Appendix B.13. Expert Weight Elicitation Workflow (Extended)

The step-by-step protocol underlying the AHP weight estimation referenced in Section 3.5 is summarized below.
  • Expert identification: Recruit DF/DFIR experts with ATT&CK domain knowledge and documented experience (minimum of 5 years professional experience and/or peer-reviewed publications).
  • Threat landscape briefing: Provide each expert with the current ATT&CK tactics, associated TTPs, and D3FEND control families relevant to SME/SMB environments.
  • Evaluation criteria: Present weighting criteria (Likelihood, Impact, Detectability, Effort) used to reason about attacker tactics and defender capabilities.
  • Independent judgments: Collect pairwise comparison matrices (PCMs) independently using the Saaty 1–9 scale for both attacker and defender metric hierarchies.
  • Aggregation and normalization: Combine judgments via the element-wise geometric mean, normalize the resulting eigenvectors, and check C R < 0.10 for each consensus PCM (Algorithm A4).
  • Weight publication: Publish the aggregated weights, per-expert consistency diagnostics, and supporting scripts (see Table 8 and Appendix B.2).
These expanded instructions mirror the data-collection workflow shown in Figure 3 and the artefacts archived in the supplementary repository bundle.

Appendix B.14. Supplementary Algorithms

Algorithms A1–A5 provide the detailed pseudocode referenced in Section 3. They implement (i) validated utility aggregation, (ii) readiness classification, (iii) low-score identification, (iv) AHP weight derivation from expert PCMs, and (v) the prioritized improvement plan. Each routine mirrors the implementation released with the supplementary code bundle.
Algorithm A1 Utility computation with validation
Input:
       M = [ m 1 , , m n ] : Normalised metric scores, m i [ 0 , 1 ]
       W = [ w 1 , , w n ] : Metric weights, w i 0 , i w i = 1
Output:    u [ 0 , 1 ] : Weighted utility score
  1:
n length ( M )
  2:
if  n length ( W )   then
  3:
    abort with “Mismatch in array lengths.”
  4:
if  n = 0   then
  5:
    abort with “Empty metric array.”
  6:
if  i : ¬ isfinite ( M [ i ] ) ¬ isfinite ( W [ i ] )   then
  7:
    abort with “Non-finite input values.”
  8:
if  i : W [ i ] < 0   then
  9:
    abort with “Weights must be non-negative.”
10:
sumW i = 1 n W [ i ]
11:
if  | sumW 1 | > 10 9   then
12:
     W W / sumW            ▹ Renormalise weights
13:
u 0
14:
for  i 1   to  n  do
15:
     u u + M [ i ] · W [ i ]
16:
u min ( 1 , max ( 0 , u ) )            ▹ Clamp to [ 0 , 1 ]
17:
return  u
Algorithm A2 Readiness classification rule
Input:
       u [ 0 , 1 ] : Utility score (from Algorithm 1)
       T [ 0 , 1 ] : Readiness threshold (e.g., T = 0.7 )
Output:    status { HighDFR , NeedsImprovement }
  1:
if  u [ 0 , 1 ]   then
  2:
    abort with “Utility u must be in [ 0 , 1 ] .”
  3:
if  T [ 0 , 1 ]   then
  4:
    abort with “Threshold T must be in [ 0 , 1 ] .”
  5:
if  u T   then
  6:
    return  HighDFR
  7:
else
  8:
    return  NeedsImprovement
Algorithm A3 Low-score metric identification
Input:
       M = [ m 1 , , m n ] : Normalised metric scores
       T [ 0 , 1 ] : Threshold used in Algorithm A2
Output:     L { 1 , , n } : Indices of metrics requiring improvement
  1:
n length ( M )
  2:
if  n = 0   then
  3:
    return 
  4:
if  i : ¬ isfinite ( M [ i ] )   then
  5:
    abort with “Non-finite metric values.”
  6:
if  T [ 0 , 1 ]   then
  7:
    abort with “Threshold T must be in [ 0 , 1 ] .”
  8:
L
  9:
for  i 1   to  n  do
10:
    if  M [ i ] < T  then
11:
         L L { i }
12:
return  L
Algorithm A4 Eigenvector-based AHP weight derivation
Input:
      Expert PCMs A ( k ) = [ a i j ( k ) ] n × n , k = 1 , , K , with a i j ( k ) { 1 / 9 , , 9 } and a j i ( k ) = 1 / a i j ( k )
Output:    ( w , λ max , C I , C R ) : Normalised weights and consistency diagnostics
  1:
A ¯ [ a ¯ i j ] where a ¯ i j = k = 1 K a i j ( k ) 1 / K    ▹ Element-wise geometric aggregation
  2:
Enforce reciprocity: a ¯ j i 1 / a ¯ i j for i < j
  3:
Solve A ¯ w = λ max w with tolerance 10 12 ; set w w / i w i
  4:
Optional LLSM cross-check: w i LLSM j = 1 n a ¯ i j 1 / n , normalise, compare cosine similarity with w
  5:
C I ( λ max n ) / ( n 1 ) ; R I 1.59 (Saaty index for n = 16 ); C R C I / R I
  6:
if  C R 0.10   then
  7:
    warn “Consensus PCM inconsistent; consider judgment revision.”
  8:
return  ( w , λ max , C I , C R )
Algorithm A5 Prioritised DFR improvement plan
Input:
       M = [ m 1 , , m n ] , W = [ w 1 , , w n ] : Metrics and weights
       P { 1 , , n } : Candidate improvement indices (from Algorithm A3)
       feas , dev , mon : Feasibility, strategy, and monitoring oracles
Output:   Ordered action list A = [ ( i , strategy i , monitor i ) ]
  1:
if  length ( M ) length ( W ) or length ( M ) = 0  then
  2:
    abort with “Invalid inputs.”
  3:
if  i : ¬ isfinite ( M [ i ] ) ¬ isfinite ( W [ i ] )   then
  4:
    abort with “Non-finite inputs.”
  5:
if  i W [ i ] 1 > 10 9   then
  6:
     W W / i W [ i ]
  7:
Build list L = { ( i , m i , w i , m i · w i ) } i = 1 n
  8:
Sort L ascending by m i · w i ; break ties by larger w i , then index i
  9:
A [ ]
10:
for each ( i , m i , w i , _ ) in L  do
11:
    if  i P and feas ( i ) = true  then
12:
         strategy i dev ( i , m i , w i )
13:
         monitor i mon ( i )
14:
        Append ( i , strategy i , monitor i ) to A
15:
return  A
Algorithm 6 Fuzzy readiness scoring pipeline
Input:
      Normalised inputs ( EvidenceQuality , DetectionCoverage , ResponseLatency ) [ 0 , 1 ] 3
      Triangular membership functions ( a , b , c ) for each input/output label (Table A3 and Table A4)
      Rule table R mapping input labels to output labels (Table A5)
Output:    ForensicReadinessScore [ 0 , 1 ]
1:
Fuzzify inputs: for each input x and label , compute μ x [ ] triangular ( x ; a , b , c )
2:
Initialise outputs:  β [ Low ] 0 , β [ Medium ] 0 , β [ High ] 0
3:
for each rule r R  do
4:
     α r min ( μ EQ [ r . EQ ] , μ DC [ r . DC ] , μ RL [ r . Latency ] )         ▹ Mamdani min
5:
     β [ r . Output ] max ( β [ r . Output ] , α r )
6:
Aggregate outputs: discretise Y = { 0 , 0.01 , , 1 } and for each y Y set μ agg ( y ) max { Low , Medium , High } β [ ] triangular ( y ; a , b , c )
7:
Defuzzify (centroid):  score y Y y · μ agg ( y ) y Y μ agg ( y )
8:
return min ( 1 , max ( 0 , score ) )

References

  1. Chen, P.; Desmet, L.; Huygens, C. A study on advanced persistent threats. In Proceedings of the Communications and Multimedia Security: 15th IFIP TC 6/TC 11 International Conference, CMS 2014, Aveiro, Portugal, 25–26 September 2014; Proceedings 15. Springer: Berlin/Heidelberg, Germany, 2014; pp. 63–72. [Google Scholar]
  2. Rowlingson, R. A ten step process for forensic readiness. Int. J. Digit. Evid. 2004, 2, 1–28. [Google Scholar]
  3. Google Mandiant. M-Trends 2025: Executive Edition; Google LLC: Mountain View, CA, USA, 2025; Based on Mandiant Consulting Investigations of Targeted Attack Activity Conducted Between 1 January 2024 and 31 December 2024; Available online: https://services.google.com/fh/files/misc/m-trends-2025-executive-edition-en.pdf (accessed on 10 November 2025).
  4. IBM. Cost of a Data Breach Report 2025: The AI Oversight Gap; IBM Corporation: Armonk, NY, USA, 2025; Based on IBM analysis of research data independently compiled by Ponemon Institute. [Google Scholar]
  5. Bonderud, D. Cost of a Data Breach 2024: Financial Industry. Available online: https://www.ibm.com/think/insights/cost-of-a-data-breach-2024-financial-industry (accessed on 10 November 2025).
  6. Johnson, R. 60 Percent of Small Companies Close Within 6 Months of Being Hacked. 2019. Available online: https://cybersecurityventures.com/60-percent-of-small-companies-close-within-6-months-of-being-hacked (accessed on 10 November 2025).
  7. Baker, P. The SolarWinds Hack Timeline: Who Knew What, and When? 2021. Available online: https://www.csoonline.com/article/570537/the-solarwinds-hack-timeline-who-knew-what-and-when.html (accessed on 10 November 2025).
  8. Batool, A.; Zowghi, D.; Bano, M. AI governance: A systematic literature review. AI Ethics 2025, 5, 3265–3279. [Google Scholar] [CrossRef]
  9. Wrightson, T. Advanced Persistent Threat Hacking: The Art and Science of Hacking Any Organization; McGraw-Hill Education Group: New York, NY, USA, 2014; ISBN 978-0-07-182836-9. [Google Scholar]
  10. Årnes, A. Digital Forensics; John Wiley & Sons: Hoboken, NJ, USA, 2017; ISBN 978-1-119-26238-4. [Google Scholar]
  11. Griffith, S.B. Sun Tzu: The Art of War; Oxford University Press: London, UK, 1963; Volume 39, ISBN 9780195014761. [Google Scholar]
  12. Myerson, R.B. Game Theory; Harvard University Press: Cambridge, MA, USA, 2013; ISBN 978-0-674-72861-5. [Google Scholar]
  13. Belton, V.; Stewart, T. Multiple Criteria Decision Analysis: An Integrated Approach; Springer Science & Business Mediar: Cham, Switzerland, 2002; ISBN 978-0-7923-7505-0. [Google Scholar]
  14. Lye, K.W.; Wing, J.M. Game strategies in network security. Int. J. Inf. Secur. 2005, 4, 71–86. [Google Scholar] [CrossRef]
  15. Roy, S.; Ellis, C.; Shiva, S.; Dasgupta, D.; Shandilya, V.; Wu, Q. A survey of game theory as applied to network security. In Proceedings of the 2010 43rd Hawaii International Conference on System Sciences (HICSS), Honolulu, HI, USA, 5–8 January 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 1–10. [Google Scholar]
  16. Zhu, Q.; Basar, T. Game-theoretic methods for robustness, security, and resilience of cyberphysical control systems: Games-in-games principle for optimal cross-layer resilient control systems. IEEE Control Syst. Mag. 2015, 35, 46–65. [Google Scholar]
  17. Kent, K.; Chevalier, S.; Grance, T. Guide to Integrating Forensic Techniques into Incident Response; NIST Special Publication 800-86; National Institute of Standards and Technology: Gaithersburg, MD, USA, 2006. [Google Scholar] [CrossRef]
  18. Alpcan, T.; Başar, T. Network Security: A Decision and Game-Theoretic Approach; Cambridge University Press: Cambridge, UK, 2010. [Google Scholar]
  19. Casey, E. Digital Evidence and Computer Crime: Forensic Science, Computers, and the Internet; Academic Press: Cambridge, MA, USA, 2011. [Google Scholar]
  20. Manshaei, M.H.; Zhu, Q.; Alpcan, T.; Bacşar, T.; Hubaux, J.P. Game theory meets network security and privacy. ACM Comput. Surv. (CSUR) 2013, 45, 1–39. [Google Scholar] [CrossRef]
  21. Nisioti, A.; Loukas, G.; Rass, S.; Panaousis, E. Game-theoretic decision support for cyber forensic investigations. Sensors 2021, 21, 5300. [Google Scholar] [CrossRef]
  22. Hasanabadi, S.S.; Lashkari, A.H.; Ghorbani, A.A. A game-theoretic defensive approach for forensic investigators against rootkits. Forensic Sci. Int. Digit. Investig. 2020, 33, 200909. [Google Scholar] [CrossRef]
  23. Karabiyik, U.; Karabiyik, T. A game theoretic approach for digital forensic tool selection. Mathematics 2020, 8, 774. [Google Scholar] [CrossRef]
  24. Hasanabadi, S.S.; Lashkari, A.H.; Ghorbani, A.A. A memory-based game-theoretic defensive approach for digital forensic investigators. Forensic Sci. Int. Digit. Investig. 2021, 38, 301214. [Google Scholar] [CrossRef]
  25. Caporusso, N.; Chea, S.; Abukhaled, R. A game-theoretical model of ransomware. In Proceedings of the Advances in Human Factors in Cybersecurity: Proceedings of the AHFE 2018 International Conference on Human Factors in Cybersecurity, Orlando, FL, USA, 21–25 July 2018; Loews Sapphire Falls Resort at Universal Studios. Springer: Cham, Switzerland, 2018; pp. 69–78. [Google Scholar]
  26. Kebande, V.R.; Venter, H.S. Novel digital forensic readiness technique in the cloud environment. Aust. J. Forensic Sci. 2018, 50, 552–591. [Google Scholar] [CrossRef]
  27. Kebande, V.R.; Karie, N.M.; Choo, K.R.; Alawadi, S. Digital forensic readiness intelligence crime repository. Secur. Priv. 2021, 4, e151. [Google Scholar] [CrossRef]
  28. Englbrecht, L.; Meier, S.; Pernul, G. Towards a capability maturity model for digital forensic readiness. Wirel. Netw. 2020, 26, 4895–4907. [Google Scholar] [CrossRef]
  29. Reddy, K.; Venter, H.S. The architecture of a digital forensic readiness management system. Comput. Secur. 2013, 32, 73–89. [Google Scholar] [CrossRef]
  30. Grobler, C.P.; Louwrens, C. Digital forensic readiness as a component of information security best practice. In Proceedings of the IFIP International Information Security Conference; Springer: Boston, MA, USA, 2007; pp. 13–24. [Google Scholar]
  31. Lakhdhar, Y.; Rekhis, S.; Sabir, E. A Game Theoretic Approach For Deploying Forensic Ready Systems. In Proceedings of the 2020 International Conference on Software, Telecommunications and Computer Networks (SoftCOM), Split, Croatia, 17–19 September 2020; pp. 1–6. [Google Scholar]
  32. Elyas, M.; Ahmad, A.; Maynard, S.B.; Lonie, A. Digital forensic readiness: Expert perspectives on a theoretical framework. Comput. Secur. 2015, 52, 70–89. [Google Scholar] [CrossRef]
  33. Baiquni, I.Z.; Amiruddin, A. A case study of digital forensic readiness level measurement using DiFRI model. In Proceedings of the 2022 International Conference on Informatics, Multimedia, Cyber and Information System (ICIMCIS), Jakarta, Indonesia, 16–17 November 2022; pp. 184–189. [Google Scholar]
  34. Rawindaran, N.; Jayal, A.; Prakash, E. Cybersecurity Framework: Addressing Resiliency in Welsh SMEs for Digital Transformation and Industry 5.0. J. Cybersecur. Priv. 2025, 5, 17. [Google Scholar] [CrossRef]
  35. Trenwith, P.M.; Venter, H.S. Digital forensic readiness in the cloud. In Proceedings of the 2013 Information Security for South Africa, Johannesburg, South Africa, 14–16 August 2013; pp. 1–5. [Google Scholar]
  36. Monteiro, D.; Yu, Y.; Zisman, A.; Nuseibeh, B. Adaptive Observability for Forensic-Ready Microservice Systems. IEEE Trans. Serv. Comput. 2023, 16, 3196–3209. [Google Scholar] [CrossRef]
  37. Xiong, W.; Legrand, E.; Åberg, O.; Lagerström, R. Cyber security threat modeling based on the MITRE Enterprise ATT&CK Matrix. Softw. Syst. Model. 2022, 21, 157–177. [Google Scholar]
  38. Wang, J.; Neil, M. A Bayesian-network-based cybersecurity adversarial risk analysis framework with numerical examples. arXiv 2021, arXiv:2106.00471. [Google Scholar]
  39. Usman, N.; Usman, S.; Khan, F.; Jan, M.A.; Sajid, A.; Alazab, M.; Watters, P. Intelligent dynamic malware detection using machine learning in IP reputation for forensics data analytics. Future Gener. Comput. Syst. 2021, 118, 124–141. [Google Scholar] [CrossRef]
  40. Li, M.; Lal, C.; Conti, M.; Hu, D. LEChain: A blockchain-based lawful evidence management scheme for digital forensics. Future Gener. Comput. Syst. 2021, 115, 406–420. [Google Scholar] [CrossRef]
  41. Soltani, S.; Seno, S.A.H. Detecting the software usage on a compromised system: A triage solution for digital forensics. Forensic Sci. Int. Digit. Investig. 2023, 44, 301484. [Google Scholar] [CrossRef]
  42. Rother, C.; Chen, B. Reversing File Access Control Using Disk Forensics on Low-Level Flash Memory. J. Cybersecur. Priv. 2024, 4, 805–822. [Google Scholar] [CrossRef]
  43. Nikkel, B. Registration Data Access Protocol (RDAP) for digital forensic investigators. Digit. Investig. 2017, 22, 133–141. [Google Scholar] [CrossRef]
  44. Nikkel, B. Fintech forensics: Criminal investigation and digital evidence in financial technologies. Forensic Sci. Int. Digit. Investig. 2020, 33, 200908. [Google Scholar] [CrossRef]
  45. Seo, S.; Seok, B.; Lee, C. Digital forensic investigation framework for the metaverse. J. Supercomput. 2023, 79, 9467–9485. [Google Scholar] [CrossRef]
  46. Malhotra, S. Digital forensics meets ai: A game-changer for the 4th industrial revolution. In Artificial Intelligence and Blockchain in Digital Forensics; River Publishers: Aalborg, Denmark, 2023; pp. 1–20. [Google Scholar]
  47. Tok, Y.C.; Chattopadhyay, S. Identifying threats, cybercrime and digital forensic opportunities in Smart City Infrastructure via threat modeling. Forensic Sci. Int. Digit. Investig. 2023, 45, 301540. [Google Scholar] [CrossRef]
  48. Han, K.; Choi, J.H.; Choi, Y.; Lee, G.M.; Whinston, A.B. Security defense against long-term and stealthy cyberattacks. Decis. Support Syst. 2023, 166, 113912. [Google Scholar] [CrossRef]
  49. Chandra, A.; Snowe, M.J. A taxonomy of cybercrime: Theory and design. Int. J. Account. Inf. Syst. 2020, 38, 100467. [Google Scholar] [CrossRef]
  50. Casey, E.; Barnum, S.; Griffith, R.; Snyder, J.; van Beek, H.; Nelson, A. Advancing coordinated cyber-investigations and tool interoperability using a community developed specification language. Digit. Investig. 2017, 22, 14–45. [Google Scholar] [CrossRef]
  51. Boyd, S.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  52. Knight, V.; Campbell, J. Nashpy: A Python library for the computation of Nash equilibria. J. Open Source Softw. 2018, 3, 904. [Google Scholar] [CrossRef]
  53. Zopounidis, C.; Pardalos, P.M. Handbook of Multicriteria Analysis; Springer Science & Business Media: Cham, Switzerland, 2010; Volume 103. [Google Scholar]
  54. Saaty, T.L. Analytic hierarchy process. In Encyclopedia of Operations Research and Management Science; Springer: Cham, Switzerland, 2013; pp. 52–64. [Google Scholar]
  55. Joint Committee for Guides in Metrology (JCGM). Evaluation of Measurement Data—Supplement 1: Propagation of Distributions Using a Monte Carlo Method; JCGM 101:2008; JCGM: Sèvres, France, 2008; Available online: https://www.bipm.org/documents/20126/2071204/JCGM_101_2008_E.pdf (accessed on 10 November 2025)JCGM 101:2008.
  56. The MITRE Corporation. MITRE ATT&CK STIX Data. Structured Threat Information Expression (STIX 2.1) Datasets for Enterprise, Mobile, and ICS ATT&CK. 2024. Available online: https://github.com/mitre-attack/attack-stix-data (accessed on 10 November 2025).
Figure 1. Attacker (A) and Defender (D) payoff matrices. The unique pure PNE at ( s 14 , t 3 ) = ( Impact , Detect ) is highlighted.
Figure 1. Attacker (A) and Defender (D) payoff matrices. The unique pure PNE at ( s 14 , t 3 ) = ( Impact , Detect ) is highlighted.
Jcp 05 00105 g001
Figure 2. Zero-sum variant ( A , D ) : convergence under discrete-time best-response dynamics from a grid of initial conditions (seed = 42, step size = 1, convergence tolerance = 10 6 ). The system converges to two pure attractors: ( s 12 = Command   and   Control , t 4 = Isolate ) and ( s 14 = Impact , t 1 = Model ) ; mixed equilibria are saddle/unstable under this dynamic. Note: Main results use the non-zero-sum bimatrix ( A , D ) , which yields a single PNE at ( s 14 = Impact , t 3 = Detect ) . See Appendix B.9 for details.
Figure 2. Zero-sum variant ( A , D ) : convergence under discrete-time best-response dynamics from a grid of initial conditions (seed = 42, step size = 1, convergence tolerance = 10 6 ). The system converges to two pure attractors: ( s 12 = Command   and   Control , t 4 = Isolate ) and ( s 14 = Impact , t 1 = Model ) ; mixed equilibria are saddle/unstable under this dynamic. Note: Main results use the non-zero-sum bimatrix ( A , D ) , which yields a single PNE at ( s 14 = Impact , t 3 = Detect ) . See Appendix B.9 for details.
Jcp 05 00105 g002
Figure 3. Expert-driven weight calculation workflow for MITRE ATT&CK tactics.
Figure 3. Expert-driven weight calculation workflow for MITRE ATT&CK tactics.
Jcp 05 00105 g003
Figure 4. Attacker and defender metric weights derived via AHP.
Figure 4. Attacker and defender metric weights derived via AHP.
Jcp 05 00105 g004
Figure 5. Empirical counts of ATT&CK tactic × D3FEND control-family coverage (family-coverage) derived from real-world APT group data: for each tactic τ and family f, we count APT–technique instances with at least one mapped D3FEND technique in family f, de-duplicated once per ( APT , technique , family ) . Data extracted from MITRE STIX bundles for ten APT groups (LeafMiner, Silent Librarian, OilRig, Ajax Security Team, Moses Staff, Cleaver, CopyKittens, APT33, APT39, MuddyWater). Y-axis: count (unitless); X-axis: ATT&CK tactic (rows) × D3FEND control family (columns: Model, Harden, Detect, Isolate, Deceive, Evict). Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023). Raw counts are provided in Appendix B (Table A6).
Figure 5. Empirical counts of ATT&CK tactic × D3FEND control-family coverage (family-coverage) derived from real-world APT group data: for each tactic τ and family f, we count APT–technique instances with at least one mapped D3FEND technique in family f, de-duplicated once per ( APT , technique , family ) . Data extracted from MITRE STIX bundles for ten APT groups (LeafMiner, Silent Librarian, OilRig, Ajax Security Team, Moses Staff, Cleaver, CopyKittens, APT33, APT39, MuddyWater). Y-axis: count (unitless); X-axis: ATT&CK tactic (rows) × D3FEND control family (columns: Model, Harden, Detect, Isolate, Deceive, Evict). Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023). Raw counts are provided in Appendix B (Table A6).
Jcp 05 00105 g005
Figure 6. Empirical frequency of ATT&CK tactics across real-world APT groups with parent/sub-technique de-duplication: each tactic is credited at most once per APT (prefer the most specific sub-technique evidence). Data extracted from MITRE STIX bundles for ten APT groups (LeafMiner, Silent Librarian, OilRig, Ajax Security Team, Moses Staff, Cleaver, CopyKittens, APT33, APT39, MuddyWater). Raw counts shown (y-axis: count; x-axis: ATT&CK tactic). Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023). Shares and per-APT normalizations are provided in Appendix B (Table A7).
Figure 6. Empirical frequency of ATT&CK tactics across real-world APT groups with parent/sub-technique de-duplication: each tactic is credited at most once per APT (prefer the most specific sub-technique evidence). Data extracted from MITRE STIX bundles for ten APT groups (LeafMiner, Silent Librarian, OilRig, Ajax Security Team, Moses Staff, Cleaver, CopyKittens, APT33, APT39, MuddyWater). Raw counts shown (y-axis: count; x-axis: ATT&CK tactic). Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023). Shares and per-APT normalizations are provided in Appendix B (Table A7).
Jcp 05 00105 g006
Figure 7. Empirical frequency of D3FEND defensive control families (tactics) across mapped APT–technique instances derived from real-world data. Counts represent the number of unique APT–technique pairs mapped to each D3FEND family, aggregated across all ATT&CK tactics. Data extracted from MITRE STIX bundles for ten real APT groups (LeafMiner, Silent Librarian, OilRig, Ajax Security Team, Moses Staff, Cleaver, CopyKittens, APT33, APT39, MuddyWater). Y-axis: count; X-axis: D3FEND control family (Model, Harden, Detect, Isolate, Deceive, Evict). Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023). Raw data are provided in Appendix B Table A8).
Figure 7. Empirical frequency of D3FEND defensive control families (tactics) across mapped APT–technique instances derived from real-world data. Counts represent the number of unique APT–technique pairs mapped to each D3FEND family, aggregated across all ATT&CK tactics. Data extracted from MITRE STIX bundles for ten real APT groups (LeafMiner, Silent Librarian, OilRig, Ajax Security Team, Moses Staff, Cleaver, CopyKittens, APT33, APT39, MuddyWater). Y-axis: count; X-axis: D3FEND control family (Model, Harden, Detect, Isolate, Deceive, Evict). Versions: ATT&CK Enterprise v13.1; D3FEND v0.12.0-BETA-2 (snapshot Mar–May 2023). Raw data are provided in Appendix B Table A8).
Jcp 05 00105 g007
Figure 8. Mean readiness score before and after implementation.
Figure 8. Mean readiness score before and after implementation.
Jcp 05 00105 g008
Figure 9. AHP-based rank-stability under ± 1 Saaty step plus ± 5 % noise (lower bars = more stable).
Figure 9. AHP-based rank-stability under ± 1 Saaty step plus ± 5 % noise (lower bars = more stable).
Jcp 05 00105 g009
Figure 10. Global sensitivity analysis of DFR metrics to readiness scores. Y-axis: absolute Pearson correlation coefficient ( | r | , unitless, range [0,1]); X-axis: metric ID (attacker: ASR, RE, ST, DEE, TTE, EC, AR, RT, IA, P, AD, DN, LG, CB, FG, RP; defender: L, I, D, VDCC, E, IR, DR, NF, ST d , LR, A, C, T, R, V, P d ). Sensitivity computed from N = 20 , 000 Monte Carlo draws sampling metric values independently from [ 0 , 1 ] ; readiness computed via Equation (9) using AHP-derived weights (Table 8). Higher bars indicate stronger influence on readiness variability.
Figure 10. Global sensitivity analysis of DFR metrics to readiness scores. Y-axis: absolute Pearson correlation coefficient ( | r | , unitless, range [0,1]); X-axis: metric ID (attacker: ASR, RE, ST, DEE, TTE, EC, AR, RT, IA, P, AD, DN, LG, CB, FG, RP; defender: L, I, D, VDCC, E, IR, DR, NF, ST d , LR, A, C, T, R, V, P d ). Sensitivity computed from N = 20 , 000 Monte Carlo draws sampling metric values independently from [ 0 , 1 ] ; readiness computed via Equation (9) using AHP-derived weights (Table 8). Higher bars indicate stronger influence on readiness variability.
Jcp 05 00105 g010
Figure 11. Bivariaterelationships between attacker metrics and readiness scores. Y-axis: readiness score (computed via Equation (9), unitless, range approximately [ 1 , 1 ] ); X-axis: attacker metric value (normalized [ 0 , 1 ] ). Each panel shows one of the 16 attacker metrics (ASR, RE, ST, DEE, TTE, EC, AR, RT, IA, P, AD, DN, LG, CB, FG, RP); scatter points sampled from N = 20 , 000 Monte Carlo draws with metric values independently sampled from [ 0 , 1 ] and readiness computed using AHP-derived weights (Table 8).
Figure 11. Bivariaterelationships between attacker metrics and readiness scores. Y-axis: readiness score (computed via Equation (9), unitless, range approximately [ 1 , 1 ] ); X-axis: attacker metric value (normalized [ 0 , 1 ] ). Each panel shows one of the 16 attacker metrics (ASR, RE, ST, DEE, TTE, EC, AR, RT, IA, P, AD, DN, LG, CB, FG, RP); scatter points sampled from N = 20 , 000 Monte Carlo draws with metric values independently sampled from [ 0 , 1 ] and readiness computed using AHP-derived weights (Table 8).
Jcp 05 00105 g011
Figure 12. Bivariate relationships between defender metrics and readiness scores. Y-axis: readiness score (computed via Equation (9), unitless, range approximately [ 1 , 1 ] ); X-axis: defender metric value (normalized [ 0 , 1 ] ). Each panel shows one of the 16 defender metrics (L, I, D, VDCC, E, IR, DR, NF, ST d , LR, A, C, T, R, V, P d ); scatter points sampled from N = 20 , 000 Monte Carlo draws with metric values independently sampled from [ 0 , 1 ] and readiness computed using AHP-derived weights (Table 8).
Figure 12. Bivariate relationships between defender metrics and readiness scores. Y-axis: readiness score (computed via Equation (9), unitless, range approximately [ 1 , 1 ] ); X-axis: defender metric value (normalized [ 0 , 1 ] ). Each panel shows one of the 16 defender metrics (L, I, D, VDCC, E, IR, DR, NF, ST d , LR, A, C, T, R, V, P d ); scatter points sampled from N = 20 , 000 Monte Carlo draws with metric values independently sampled from [ 0 , 1 ] and readiness computed using AHP-derived weights (Table 8).
Jcp 05 00105 g012
Figure 13. Distribution of standardized (z-scored) readiness values across Monte Carlo simulation ( N = 20 , 000 draws). Y-axis: frequency (count); X-axis: standardized readiness score z = ( x μ ) / σ (unitless, centered at zero). Readiness computed via Equation (9) using AHP-derived weights (Table 8); raw readiness values approximately in [ 1 , 1 ] (defender utility minus attacker utility). The near-symmetric distribution indicates balanced variability around the mean.
Figure 13. Distribution of standardized (z-scored) readiness values across Monte Carlo simulation ( N = 20 , 000 draws). Y-axis: frequency (count); X-axis: standardized readiness score z = ( x μ ) / σ (unitless, centered at zero). Readiness computed via Equation (9) using AHP-derived weights (Table 8); raw readiness values approximately in [ 1 , 1 ] (defender utility minus attacker utility). The near-symmetric distribution indicates balanced variability around the mean.
Jcp 05 00105 g013
Figure 14. Implementation roadmap and outcomes. Left: key adoption challenges. Center: phased workflow for deploying the game-theoretic DFR framework. Right: expected outcomes. Bottom band: cross-cutting enablers (policy, collaboration, upskilling, and measurement).
Figure 14. Implementation roadmap and outcomes. Left: key adoption challenges. Center: phased workflow for deploying the game-theoretic DFR framework. Right: expected outcomes. Bottom band: cross-cutting enablers (policy, collaboration, upskilling, and measurement).
Jcp 05 00105 g014
Figure 15. Evolution of attacker and defender strategies in EGT simulation.Panels (ai) represent the nine combinations of resource budgets (rows: 1, 3, 5 units) and defender population sizes (columns: 10, 15, 25 agents), with the attacker population fixed at five agents. Each panel shows the replicator-dynamics trajectories of attacker (solid line) and defender (dashed line) strategy proportions over 250 generations. The partial overlap between the solid and dashed curves indicates convergence toward the same equilibrium and does not hinder interpretation, as the trajectories remain visually distinguishable through distinct line styles and colors; no data are obscured.
Figure 15. Evolution of attacker and defender strategies in EGT simulation.Panels (ai) represent the nine combinations of resource budgets (rows: 1, 3, 5 units) and defender population sizes (columns: 10, 15, 25 agents), with the attacker population fixed at five agents. Each panel shows the replicator-dynamics trajectories of attacker (solid line) and defender (dashed line) strategy proportions over 250 generations. The partial overlap between the solid and dashed curves indicates convergence toward the same equilibrium and does not hinder interpretation, as the trajectories remain visually distinguishable through distinct line styles and colors; no data are obscured.
Jcp 05 00105 g015
Figure 16. Effect of attacks on investigation phases: (a) Pareto chart; (b) attacker and defender utility; (c) utility coordination visualization.
Figure 16. Effect of attacks on investigation phases: (a) Pareto chart; (b) attacker and defender utility; (c) utility coordination visualization.
Jcp 05 00105 g016
Figure 17. Defender readiness vs. training level in three maturity regimes: (a) Junior + Mid + Senior, (b) Mid + Senior, and (c) Senior.
Figure 17. Defender readiness vs. training level in three maturity regimes: (a) Junior + Mid + Senior, (b) Mid + Senior, and (c) Senior.
Jcp 05 00105 g017
Figure 18. Impact comparison for SMEs and SMBs under SQLi and DDoS scenarios (baseline vs. irrational).
Figure 18. Impact comparison for SMEs and SMBs under SQLi and DDoS scenarios (baseline vs. irrational).
Jcp 05 00105 g018
Figure 19. Impact of irrational attacker behavior on SQLi and DDoS for SME and SMB simulations.
Figure 19. Impact of irrational attacker behavior on SQLi and DDoS for SME and SMB simulations.
Jcp 05 00105 g019
Figure 20. Behavioral distribution under irrational attack scenarios for SMEs and SMBs.
Figure 20. Behavioral distribution under irrational attack scenarios for SMEs and SMBs.
Jcp 05 00105 g020
Table 1. Comparative analysis of game-theoretic approaches to digital forensic readiness.
Table 1. Comparative analysis of game-theoretic approaches to digital forensic readiness.
DimensionOur ApproachNisioti
et al. [21]
Karabiyik
et al. [23]
Lakhdhar
et al. [31]
Wang
et al. [38]
Monteiro
et al. [36]
Game ModelNon-zero-sum
Bimatrix
(MNE/PNE)
Bayesian
(BNE)
2 × 2
Normal-Form
Non-cooperativeARA
(Influence
Diagram)
Bayesian
(BNE)
ATT&CKExplicit (14
Tactics)
Explicit××××
D3FENDExplicit (6
Families)
×××××
Knowledge Coupling
ATT&CK↔D3FEND
ATT&CK +
CVSS
Empirical
(ForGe)
Internal
(CSLib)
Probabilistic
(HBN)
CVSS +
OpenTelemetry
Weighting MethodAHP (10
Experts)
CVSS +
SME
Rule-based Parametric
( α )
Implicit Scalar
Parameters
Quantitative Utilities32 AHP
Metrics
Payoff
Functions
Payoff
Matrix
ParametricUtility
Nodes

Closed-Form
EquilibriumPNE &
MNE
BNEPure/Mixed
NE
Pure/Mixed
NE
× ARABNE
DFR FocusDFR Post-mortem Investigation EfficiencyForensic
Readiness
× Cyber RiskForensic
Readiness
SME/SMBExplicitly
Targeted

Domain-Agnostic
Potential Applicable Feasible Implicit
Standardization ATT&CK,
D3FEND,
STIX
ATT&CK,
STIX, CVSS

Open-Source

CVE/US-CERT

Self-Contained
CVSS,
OpenTelemetry
ReproducibilityCode, Data,
Seeds
Public
Inputs, No
Code
Code on
Request
No Public
Code/Data
No Code,
Commercial
Benchmark,
Repo
Key DifferentiatorIntegrated
DFR
Bayesian
Anti-Forensics
Tool SelectionProvability
Taxonomy
Adversarial
Risk Analysis
Microservice
Observability
Legend:  = fully addressed; = partially addressed; and × = not addressed. Abbreviations: AHP (Analytic Hierarchy Process), MNE (Mixed Nash Equilibrium), PNE (Pure Nash Equilibrium), BNE (Bayesian Nash Equilibrium), ARA (Adversarial Risk Analysis), HBN (Hybrid Bayesian Network). Our framework uniquely integrates ATT&CK–D3FEND knowledge with AHP-weighted utilities and explicit SME/SMB targeting—a combination not found in prior work.
Table 2. Notation reference for key symbols (see Table A1 in Appendix B for complete listing).
Table 2. Notation reference for key symbols (see Table A1 in Appendix B for complete listing).
SymbolDescription
Game Structure
S Attacker strategy set: S = { s 1 , , s 14 } , | S | = 14 ATT&CK tactics
T Defender strategy set: T = { t 1 , , t 6 } , | T | = 6 D3FEND control families
AAttacker payoff matrix: A R 14 × 6 , entry A ( s , t ) [ 0 , 41 ]
DDefender payoff matrix: D R 14 × 6 , entry D ( s , t ) [ 0 , 41 ]
Strategies
s S Attacker pure strategy (ATT&CK tactic)
t T Defender pure strategy (D3FEND control family)
xAttacker mixed strategy: x Δ 13 , probability vector over S
yDefender mixed strategy: y Δ 5 , probability vector over T
x * , y * Nash equilibrium mixed strategies
Utilities and Metrics
A ˜ ( s , t ) Normalized attacker utility: A ˜ ( s , t ) = i = 1 16 w i ( A ) M i ( s , t ) [ 0 , 1 ]
D ˜ ( s , t ) Normalized defender utility: D ˜ ( s , t ) = j = 1 16 w j ( D ) M j ( D ) ( s , t ) [ 0 , 1 ]
w i ( A ) AHP weight for attacker metric i: w i ( A ) [ 0 , 1 ] , i = 1 16 w i ( A ) = 1
w j ( D ) AHP weight for defender metric j: w j ( D ) [ 0 , 1 ] , j = 1 16 w j ( D ) = 1
M i ( s , t ) Attacker DFR metric i value: M i ( s , t ) [ 0 , 1 ]
M j ( D ) ( s , t ) Defender DFR metric j value: M j ( D ) ( s , t ) [ 0 , 1 ]
Notation Disambiguation
A ( s , t ) vs. A A ( s , t ) is a scalar (single entry); A is the entire matrix
x, y vs. x*, y*x, y are elements/indices; x*, y* are mixed-strategy vectors
Table 3. Strategy notation reference.
Table 3. Strategy notation reference.
Defender StrategiesATT&CK Tactics
t 1 = Model s 1 = Reconnaissance
t 2 = Harden s 2 = Resource Development
t 3 = Detect s 3 = Initial Access
t 4 = Isolate s 4 = Execution
t 5 = Deceive s 5 = Persistence
t 6 = Evict s 6 = Privilege Escalation
s 7 = Defense Evasion
s 8 = Credential Access
s 9 = Discovery
s 10 = Lateral Movement
s 11 = Collection
s 12 = Command and Control
s 13 = Exfiltration
s 14 = Impact
Table 4. Attacker’s Payoff Matrix A ( s , t ) (utility values; higher is better). Column labels: t 1 = Model , t 2 = Harden , t 3 = Detect , t 4 = Isolate , t 5 = Deceive , and t 6 = Evict .
Table 4. Attacker’s Payoff Matrix A ( s , t ) (utility values; higher is better). Column labels: t 1 = Model , t 2 = Harden , t 3 = Detect , t 4 = Isolate , t 5 = Deceive , and t 6 = Evict .
t1t2t3t4t5t6
s15678910
s2001234
s31413121100
s41617181800
s51920201800
s6232221765
s7242526242526
s8322829303127
s9333435303332
s10323536675
s1136373863530
s123738393900
s13383940000
s14394041000
Table 5. Defender’s Payoff Matrix D ( s , t ) (utility values; higher is better). Column labels: t 1 = Model , t 2 = Harden , t 3 = Detect , t 4 = Isolate , t 5 = Deceive and, t 6 = Evict .
Table 5. Defender’s Payoff Matrix D ( s , t ) (utility values; higher is better). Column labels: t 1 = Model , t 2 = Harden , t 3 = Detect , t 4 = Isolate , t 5 = Deceive and, t 6 = Evict .
t1t2t3t4t5t6
s1571175
s26810266
s379115811
s48102525912
s59112481013
s610122481110
s711212010127
s81814259525
s91315231248
s1014162211149
s11151720121314
s12161821131525
s13172020101617
s14121929161716
Table 6. Attacker Utility Metrics (Summary). Each metric is evaluated on a continuous scale from 0 (least favorable) to 1 (most favorable). Detailed scoring preferences with qualitative descriptions for each metric level are provided in Appendix B (Table A9).
Table 6. Attacker Utility Metrics (Summary). Each metric is evaluated on a continuous scale from 0 (least favorable) to 1 (most favorable). Detailed scoring preferences with qualitative descriptions for each metric level are provided in Appendix B (Table A9).
MetricDescription
Attack Success Rate (ASR)Likelihood of successful attack execution
Resource Efficiency (RE)Ratio of attack payoff to resource expenditure
Stealthiness (ST)Ability to avoid detection and attribution
Data Exfiltration Effectiveness (DEE)Success rate of data exfiltration attempts
Time-to-Exploit (TTE)Speed of vulnerability exploitation before patching
Evasion of Countermeasures (EC)Ability to bypass defensive measures
Attribution Resistance (AR)Difficulty in identifying the attacker
Reusability of Attack Techniques (RT)Extent to which attack techniques can be reused
Impact of Attacks (IA)Magnitude of disruption or loss caused
Persistence (P)Ability to maintain control over compromised systems
Adaptability (AD)Capacity to adjust strategies in response to defenses
Deniability (DN)Ability to deny involvement in attacks
Longevity (LG)Duration of operations before disruption
Collaboration (CB)Extent of collaboration with other attackers
Financial Gain (FG)Monetary profit from attacks
Reputation and Prestige (RP)Enhancement of attacker reputation
Table 7. Defender Utility Metrics (Summary). Each metric is evaluated on a continuous scale from 0 (least favorable) to 1 (most favorable). Detailed scoring preferences with qualitative descriptions for each metric level are provided in Appendix B (Table A10).
Table 7. Defender Utility Metrics (Summary). Each metric is evaluated on a continuous scale from 0 (least favorable) to 1 (most favorable). Detailed scoring preferences with qualitative descriptions for each metric level are provided in Appendix B (Table A10).
MetricDescription
Logging and Audit Trail Capabilities (L)Extent of logging and audit trail coverage
Integrity and Preservation of Digital Evidence (I)Ability to preserve evidence integrity and backups
Documentation and Compliance with Digital Forensic Standards (D)Adherence to forensic standards and documentation quality
Volatile Data Capture Capabilities (VDCC)Effectiveness of volatile data capture
Encryption and Decryption Capabilities (E)Strength of encryption/decryption capabilities
Incident Response Preparedness (IR)Quality of incident response plans and team readiness
Data Recovery Capabilities (DR)Effectiveness of data recovery tools and processes
Network Forensics Capabilities (NF)Sophistication of network forensic analysis
Staff Training and Expertise ( ST d )Level of staff training and certifications
Legal & Regulatory Compliance (LR)Compliance with legal and regulatory requirements
Accuracy (A)Consistency and correctness of forensic analysis
Completeness (C)Extent of comprehensive data collection and analysis
Timeliness (T)Speed and efficiency of forensic investigation process
Reliability (R)Consistency and repeatability of forensic techniques
Validity (V)Adherence to legal and scientific standards
Preservation ( P d )Effectiveness of evidence preservation procedures
Table 8. AHP-derived metric weights for attacker and defender utility functions. Notation: Defender metrics use subscript d ( S T d = Staff Training, P d = Preservation); attacker metrics use bare symbols ( S T = Stealthiness, P = Persistence).
Table 8. AHP-derived metric weights for attacker and defender utility functions. Notation: Defender metrics use subscript d ( S T d = Staff Training, P d = Preservation); attacker metrics use bare symbols ( S T = Stealthiness, P = Persistence).
Metric (Attacker)WeightMetric (Defender)Weight
ASR0.1094L0.0881
RE0.0476I0.0881
ST0.0921D0.0423
DEE0.0887VDCC0.0642
TTE0.0476E0.0461
EC0.0887IR0.0881
AR0.0814DR0.0481
RT0.0476NF0.0819
IA0.0921 ST d 0.0819
P0.0814LR0.0481
AD0.0571A0.0557
DN0.0264C0.0460
LG0.0433T0.0693
CB0.0262R0.0531
FG0.0210V0.0423
RP0.0487 P d 0.0557
Precision note. Values are rounded to four decimals for readability. Six-decimal weights are provided in Table A4; apparent duplicates at four decimals are either rounding artifacts or reflect intended equal-importance judgments.
Table 9. Representative game-theoretic approaches to digital forensic readiness and observability.
Table 9. Representative game-theoretic approaches to digital forensic readiness and observability.
Lakhdhar et al. [31]Monteiro et al. [36]This Work
Forensic objectiveInvestigation-ready infrastructure with cognitive defenceAdaptive observability and evidence collection for microservicesATT&CK/D3FEND-grounded readiness planning for APT-focused organisations
Game formulation2-player non-cooperative game; pure/mixed Nash equilibria2-player Bayesian game; Bayesian Nash equilibriaNon-zero-sum bimatrix game, zero-sum variant, and evolutionary dynamics
Reported quantitative effectsSecurity cost gain of ≈45– 75 % vs. static defence; false decisions reduced by ≈70– 90 % F 1 improvements of ≈ 3.1 42.5 % vs. full/sampling observabilityMedian readiness improvement of 18.0 % (95% CI [16.3, 19.7]); linked attacker metrics reduced by ≈15– 30 %
Evaluation assetsSymbolic scenario library (CSLib); parameter sweeps over ( w 1 , w 2 , w 3 , γ ) TrainTicket benchmark (41 microservices, 10 2 10 5 concurrent users)Empirical ATT&CK→D3FEND mappings (10 APT groups); 10 calibrated synthetic readiness profiles
Released artefactsNot publicly documentedPrototype and evaluation scripts (GitHub:no public version available)STIX extraction scripts, mapping CSVs, and YAML manifests for profile generation (GitHub; version: v1.0)
Reported percentages are taken directly from the original studies; because objectives, datasets, and metrics differ (cost, F 1 , readiness), these figures are not directly comparable across columns.
Table 10. Synthetic DFR profiles (before implementation). Illustrative, calibration-based scenarios; not field measurements.
Table 10. Synthetic DFR profiles (before implementation). Illustrative, calibration-based scenarios; not field measurements.
CaseLIDVDCCEIRDRNF ST d LRACTRV P d
10.50.60.30.40.50.60.20.50.20.60.70.20.60.10.20.4
20.10.20.70.60.10.20.60.10.60.40.20.60.20.10.60.5
30.60.10.60.50.60.40.20.20.60.10.60.10.20.60.10.6
40.70.20.20.70.20.60.40.60.20.10.20.60.10.20.60.2
50.70.60.30.50.60.70.40.20.60.30.60.20.10.60.20.3
60.50.70.50.70.50.40.60.60.30.20.60.10.60.20.40.6
70.40.60.30.60.70.60.20.20.70.60.20.70.60.20.50.4
80.10.20.60.50.60.20.50.40.20.60.10.20.60.70.60.2
90.60.30.20.60.20.30.60.60.40.20.60.30.20.60.20.5
100.50.60.30.20.60.20.70.20.50.60.20.40.20.60.50.2
Table 11. Synthetic DFR profiles (after implementation). Illustrative, calibration-based scenarios; not field measurements. Profiles generated using targeted framework uplift with coupling constraints (see Section 4.3.1).
Table 11. Synthetic DFR profiles (after implementation). Illustrative, calibration-based scenarios; not field measurements. Profiles generated using targeted framework uplift with coupling constraints (see Section 4.3.1).
CaseLIDVDCCEIRDRNF ST d LRACTRV P d
10.80.80.70.90.80.80.70.90.70.60.80.70.80.70.70.7
20.90.80.90.80.70.90.70.80.60.70.70.80.70.60.60.8
30.80.70.80.90.80.90.80.90.70.80.80.70.70.70.80.7
40.80.90.90.80.70.90.90.80.70.70.70.80.70.70.60.8
50.70.70.90.70.80.90.70.90.80.80.70.70.60.80.70.7
60.70.80.80.90.70.80.60.90.60.70.60.80.70.90.70.7
70.80.70.90.70.60.90.80.90.70.80.70.70.80.70.80.8
80.70.60.90.80.80.90.80.80.80.70.70.80.70.60.80.7
90.90.70.80.70.70.90.70.80.70.80.80.70.60.70.70.7
100.80.80.90.70.70.90.80.70.70.80.70.70.80.80.60.8
Table 12. Simulation results based on evolutionary game theory.
Table 12. Simulation results based on evolutionary game theory.
Res.Def.Att.Scen.Final Val.Avg. Att.Avg. Def.Avg. Read.
1105a0.560.840.00
1155b0.520.940.00
1255c0.610.690.00
3105d0.960.580.00
3255f1.001.000.00
5155h0.910.750.03
Table 13. Simulation results (mean ± 95% CI; N = 50 , 000 trials per setting).
Table 13. Simulation results (mean ± 95% CI; N = 50 , 000 trials per setting).
LowMediumHigh
Attack success rate0.25 ± 0.00380.53 ± 0.00440.75 ± 0.0038
Evidence collection rate0.93 ± 0.00220.96 ± 0.00170.94 ± 0.0021
Table 14. Simulation results of attack success rate for SME and SMB organizations.
Table 14. Simulation results of attack success rate for SME and SMB organizations.
IDSMESMBImpact Metrics
TypeMalic.Str.ImpactCVSSTypeMalic.Str.ImpactCVSSWorkloadAvail.Conf.Integ.
0DDoS0.751.12High7DDoS0.751.12High71.1250.800
1SQLI0.751.12High9SQLI0.751.12High92.72.587.27.2
2DDoS0.751.12Med0DDoS0.751.12Med01.1250.9600
3SQLI0.751.12High9SQLI0.751.12High91.1251.0057.27.2
4DDoS0.751.12Low0DDoS0.751.12Low01.1250.9600
5SQLI0.751.12Med7SQLI0.751.12Med72.72.582.82.8
Table 15. Simulation result of attack success rate—irrational behavior.
Table 15. Simulation result of attack success rate—irrational behavior.
IDSMESMBImpact Metrics
TypeMalic.Str.ImpactCVSSTypeMalic.Str.ImpactCVSSWorkloadAvail.Conf.Integ.
0SQLI0.490.73Med7SQLI0.490.73Med70.730.612.82.8
1DDoS0.751.12High7DDoS0.751.12High71.120.8000
2DDoS0.801.21High7DDoS0.801.21High71.210.8000
3SQLI0.160.24High9SQLI0.160.24High90.240.127.27.2
4SQLI0.580.87High9SQLI0.580.87High92.452.337.27.2
5DDoS0.841.26High7DDoS0.841.26High72.840.8000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Vaseghipanah, M.; Jabbehdari, S.; Navidi, H. A Game-Theoretic Approach for Quantification of Strategic Behaviors in Digital Forensic Readiness. J. Cybersecur. Priv. 2025, 5, 105. https://doi.org/10.3390/jcp5040105

AMA Style

Vaseghipanah M, Jabbehdari S, Navidi H. A Game-Theoretic Approach for Quantification of Strategic Behaviors in Digital Forensic Readiness. Journal of Cybersecurity and Privacy. 2025; 5(4):105. https://doi.org/10.3390/jcp5040105

Chicago/Turabian Style

Vaseghipanah, Mehrnoush, Sam Jabbehdari, and Hamidreza Navidi. 2025. "A Game-Theoretic Approach for Quantification of Strategic Behaviors in Digital Forensic Readiness" Journal of Cybersecurity and Privacy 5, no. 4: 105. https://doi.org/10.3390/jcp5040105

APA Style

Vaseghipanah, M., Jabbehdari, S., & Navidi, H. (2025). A Game-Theoretic Approach for Quantification of Strategic Behaviors in Digital Forensic Readiness. Journal of Cybersecurity and Privacy, 5(4), 105. https://doi.org/10.3390/jcp5040105

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop