Next Article in Journal
Synthesis of Chitosan and Ferric-Ion (Fe3+)-Doped Brushite Mineral Cancellous Bone Scaffolds
Next Article in Special Issue
An Enhanced Tree-Seed Algorithm for Function Optimization and Production Optimization
Previous Article in Journal
Advancement in Cancer Vasculogenesis Modeling through 3D Bioprinting Technology
Previous Article in Special Issue
A Parameter Reduction-Based Decision-Making Method with Interval-Valued Neutrosophic Soft Sets for the Selection of Bionic Thin-Wall Structures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing the Efficiency of a Cybersecurity Operations Center Using Biomimetic Algorithms Empowered by Deep Q-Learning

1
Escuela de Ingeniería Informática, Universidad de Valparaíso, Valparaíso 2362905, Chile
2
Escuela de Ingeniería y Negocios, Universidad Viña del Mar, Viña del Mar 2572007, Chile
3
Escuela de Ingeniería Informática, Pontificia Universidad Católica de Valparaíso, Valparaíso 2362807, Chile
*
Author to whom correspondence should be addressed.
Biomimetics 2024, 9(6), 307; https://doi.org/10.3390/biomimetics9060307
Submission received: 10 April 2024 / Revised: 8 May 2024 / Accepted: 15 May 2024 / Published: 21 May 2024
(This article belongs to the Special Issue Computer-Aided Biomimetics: 2nd Edition)

Abstract

:
In the complex and dynamic landscape of cyber threats, organizations require sophisticated strategies for managing Cybersecurity Operations Centers and deploying Security Information and Event Management systems. Our study enhances these strategies by integrating the precision of well-known biomimetic optimization algorithms—namely Particle Swarm Optimization, the Bat Algorithm, the Gray Wolf Optimizer, and the Orca Predator Algorithm—with the adaptability of Deep Q-Learning, a reinforcement learning technique that leverages deep neural networks to teach algorithms optimal actions through trial and error in complex environments. This hybrid methodology targets the efficient allocation and deployment of network intrusion detection sensors while balancing cost-effectiveness with essential network security imperatives. Comprehensive computational tests show that versions enhanced with Deep Q-Learning significantly outperform their native counterparts, especially in complex infrastructures. These results highlight the efficacy of integrating metaheuristics with reinforcement learning to tackle complex optimization challenges, underscoring Deep Q-Learning’s potential to boost cybersecurity measures in rapidly evolving threat environments.

1. Introduction

In the digital age, the landscape of the contemporary world is increasingly shaped by technological advancements, where threats in the cyber realm pose significant challenges to enterprises. Recognizing these threats necessitates a nuanced understanding of cybersecurity culture, the development of robust cyber-risk management strategies, and the adoption of a proactive, collaborative approach tailored to each organization’s unique context [1,2]. In response, this paper introduces a novel, adaptable cybersecurity risk management framework designed to seamlessly integrate with the evolving threat landscape, leveraging technological progress and aligning with specific organizational needs.
The advancement of optimization techniques, driven by an expansion in scientific knowledge, has led to notable breakthroughs in various fields, including cybersecurity [3]. Artificial intelligence (AI) plays a pivotal role in this evolution, especially through the development of bio-inspired optimization algorithms. These algorithms, inspired by natural processes, have been instrumental in enhancing cyber-risk management strategies by offering innovative solutions and efficiencies [4]. Despite their effectiveness in solving complex problems, these algorithms can encounter limitations, such as stagnation at local optima, which poses a challenge to achieving global optimization [5]. Nevertheless, this challenge also presents an opportunity for a strategic focus on diversification within the search domain, facilitating significant improvements in cyber-risk management efficacy.
Bio-inspired algorithms often struggle to achieve global optimization due to their inherent design, which tends to favor convergence towards local optima based on immediate environmental information. This can lead to the premature acceptance of suboptimal solutions [6,7]. Addressing this issue is crucial and involves promoting a balanced approach to exploration and exploitation, encouraging the exploration of previously uncharted territories and the pursuit of untapped opportunities, thereby enhancing the identification and mitigation of cyber risks [8].
This research proposes a cutting-edge hybrid algorithm that combines metaheuristic algorithms with reinforcement learning to efficiently search for and identify optimal solutions in global optimization tasks. This approach aims to strike a delicate balance between exploration and exploitation, gradually offering more advantageous solutions over time, while avoiding the pitfalls of premature convergence [9]. By leveraging the strengths of bio-inspired algorithms such as Particle Swarm Optimization (PSO), the Bat Algorithm (BAT), the Gray Wolf Optimizer (GWO), and the Orca Predator Algorithm (OPA) for initial detection, and subsequently optimizing the search process with Deep Q-Learning (DQL), this study seeks to address and overcome the challenges of the exploration–exploitation balance and computational complexity, especially in high-dimensional search spaces [10,11].
Enhancing the methodology outlined in [12], this paper extends the integration of bio-inspired algorithms with Deep Q-Learning to optimize the implementation of Cybersecurity Operations Centers (Cyber SOCs). It focuses on a comprehensive risk and requirement evaluation, the establishment of clear objectives, and the creation of a robust technological infrastructure, featuring key tools such as Security Information and Event Management (SIEM) and Network Intrusion Detection Systems (NIDSs) for effective real-time monitoring and threat mitigation [13,14].
Structured to provide a thorough investigation, this paper is organized as follows: Section 2 offers a detailed review of recent integrations of machine learning with metaheuristics in cybersecurity, highlighting multi-objective optimization challenges. Section 3 delves into preliminary concepts of bio-inspired algorithms, emphasizing the principles of PSO, BAT, GWO, and OPA, alongside a formal introduction to DQL and advancements in Cyber SOC and SIEM technologies. Section 4 outlines the development of the proposed solution, with Section 5 detailing the experimental design methodology. Section 6 analyzes the results, discussing the hybridization’s effectiveness in generating efficient solutions. Finally, Section 7 concludes the study, summarizing key findings and suggesting directions for future research.

2. Related Work

The rising frequency and severity of cyberattacks underscore the essential role of cybersecurity in protecting organizational assets. Research such as the study by [15] introduces a groundbreaking multi-objective optimization approach for cybersecurity countermeasures using Genetic Algorithms. Their methodology aims to fine-tune Artificial Immune System parameters to achieve an ideal balance between minimizing risk and optimizing execution time. The robustness of the model is demonstrated through comprehensive testing across a broad spectrum of inputs, showcasing its capacity for a swift and effective cybersecurity response.
In the realm of machine learning (ML), techniques are being increasingly applied across diverse domains, including the creation of advanced machine learning models, enhancing physics simulations, and tackling complex linear programming challenges. The research conducted by [16] delves into the significant impact of machine learning on the domain knowledge of metaheuristics, leading to enhanced problem-solving methodologies. Furthermore, the integration of machine learning with metaheuristics, as explored in studies [17,18], opens up promising avenues for cyber-risk management, showcasing the transformative potential of ML in developing new strategies and enhancing existing cybersecurity mitigation efforts.
The synergy between advanced machine learning techniques and metaheuristics is pivotal in crafting solutions that effectively address the sophisticated and ever-evolving landscape of cyber threats. Notably, research such as [19] emphasizes the utility of integrating Q-Learning with Particle Swarm Optimization for the resolution of combinatorial problems, marking a significant advancement over traditional PSO methodologies. The approach not only enhances solution quality but also exemplifies the effectiveness of learning-based hybridizations in the broader context of swarm intelligence algorithms, providing a novel and adaptable methodology for tackling optimization challenges.
Innovative algorithmic design further underscores the progress in optimization techniques, with the introduction of the self-adaptive virus optimization algorithm by [20]. The novel algorithm improves upon the conventional virus optimization algorithm by minimizing the reliance on user-defined parameters, thus facilitating a broader application across various problem domains. The dynamic adaptation of its parameters significantly elevates the algorithm’s performance on benchmark functions, showcasing its superiority, particularly in scenarios where the traditional algorithm exhibited limitations. The advancement is achieved by streamlining the algorithm, reducing controllable parameters to a singular one, thereby enhancing its efficiency and versatility for continuous domain optimization challenges.
The discourse on metaheuristic algorithms for solving complex optimization problems is enriched by [21], which addresses the manual design of these algorithms without a cohesive framework. Proposing a General Search Framework to amalgamate diverse metaheuristic strategies, the method introduces a systematic approach for the selection of algorithmic components, facilitating the automated design of sophisticated algorithms. The framework enables the development of novel, population-based algorithms through reinforcement learning, marking a pivotal step towards the automation of algorithm design supported by effective machine learning techniques.
In the domain of intrusion detection, Ref. [22] introduces an innovative technique, a metaheuristic with a deep learning-enabled intrusion detection system for a secured smart environment (MDLIDS–SSE), which combines metaheuristics with deep learning to secure intelligent environments. Employing Z-score normalization for data preprocessing via the improved arithmetic optimization algorithm-based feature selection (IAOA–FS), the method achieves high precision in intrusion classification, surpassing recent methodologies. Experimental validation underscores its potential in safeguarding smart cities, buildings, and healthcare systems, demonstrating promising results in accuracy, recall, and detection rates.
Additionally, the Q-Learning Vegetation Evolution algorithm, as presented in [23], exemplifies the integration of Q-Learning for optimizing coverage in numerical and wireless sensor networks. The approach, featuring a mix of exploitation and exploration strategies and the use of online Q-Learning for dynamic adaptation, demonstrates significant improvements over conventional methods through rigorous testing on CEC2020 benchmark functions and real-world engineering challenges. The research contributes a sophisticated approach to solving complex optimization problems, highlighting the efficacy of hybrid strategies in the field.
In the sphere of cyber-risk management, particularly from the perspective of the Cyber SOC and SIEM, research efforts focus on strategic optimization, automated responses, and adaptive methodologies to navigate the dynamic cyber-threat landscape. Works such as [12,24] explore efficient strategies for designing network topologies and optimizing cybersecurity incident responses within SIEM systems. These studies leverage multi-objective optimization approaches and advanced machine learning models, like Deep Q neural networks, to enhance decision-making processes, showcasing significant advancements in the automation and efficiency of cybersecurity responses.
Emerging strategies in intrusion detection and network security, highlighted by [25,26], emphasize the integration of reinforcement learning with oversampling and undersampling algorithms, and the combination of Particle Swarm Optimization–Genetic Algorithm with the LSTM–GRU of deep learning that fused the GRU (gated recurrent unit) and LSTM (long short-term memory). These approaches demonstrate a significant leap forward in detecting various types of attacks within Internet of Things (IoT) networks, showcasing the power of combining machine learning and optimization techniques for IoT security. The model’s accuracy in classifying different attack types, as tested on the CICIDS-2017 dataset, outperforms existing methods and suggests a promising direction for future research in this domain.
Furthermore, Ref. [27] introduces a semi-supervised alert filtering scheme that leverages semi-supervised learning and clustering techniques to efficiently distinguish between false and true alerts in network security monitoring. The method’s effectiveness, as evidenced by its superior performance over traditional approaches, offers a fresh perspective on alert filtering, significantly contributing to the improvement of network security management by reducing alert fatigue.
The exploration of machine learning’s effectiveness and cost-efficiency in NIDS for small and medium enterprises (SMEs) in the UK is presented in [28]. The study assesses various intrusion detection and prevention devices, focusing on their ability to manage zero-day attacks and related costs. The research, conducted during the COVID-19 pandemic, investigates both commercial and open-source NIDS solutions, highlighting the balance between cost, required expertise, and the effectiveness of machine learning-enhanced NIDS in safeguarding SMEs against cyber threats.
From the perspective of Cyber SOCs, Ref. [29] addresses the increasing complexity of cyberattacks and their implications for public sector organizations. The study proposes a ‘Wide–Scope CyberSOC’ model as a unique outsourced solution to enhance cybersecurity awareness and implementation across various operational domains, tackling the challenges faced by public institutions in building a skilled cybersecurity team and managing the blend of internal and external teams amidst the prevailing outsourcing trend.
Lastly, Ref. [30] offers a comprehensive analysis of the bio-inspired Internet of Things, underscoring the synergy between biomimetics and advanced technologies. The research evaluates the current state of Bio-IoT, focusing on its benefits, challenges, and future potential. The integration of natural principles with IoT technology promises to create more efficient and adaptable solutions, addressing key challenges such as data security and privacy, interoperability, scalability, energy management, and data handling.

3. Preliminaries

In this study, we integrated bio-inspired algorithms with an advanced machine learning technique to tackle a complex optimization problem. Specifically, we utilized Particle Swarm Optimization, the Bat Algorithm, the Grey Wolf Optimizer and the Orca Predator Algorithm, which are inspired by the intricate processes and behaviors observed in nature and among various animal species. These algorithms were improved by incorporating reinforcement learning through Deep Q-Learning into the search process of bio-inspired methods.

3.1. Particle Swarm Optimization

Particle Swarm Optimization is a computational method that simulates the social behavior observed in nature, such as birds flocking or fish schooling, to solve optimization problems [31]. This technique is grounded in the concept of collective intelligence, where simple agents interact locally with one another and with their environment to produce complex global behaviors.
In PSO, a swarm of particles moves through the solution space of an optimization problem, with each particle representing a potential solution. The movement of these particles is guided by their own best-known positions in the space, as well as the overall best-known positions discovered by any particle in the swarm. This mechanism encourages both individual exploration of the search space and social learning from the success of other particles. The position of each particle is updated according to Equations (1) and (2):
v i ( t + 1 ) = w v i ( t ) + c 1 r a n d 1 ( p b e s t i x i ( t ) ) + c 2 r a n d 2 ( g b e s t x i ( t ) )
x i ( t + 1 ) = x i ( t ) + v i ( t + 1 )
where v i ( t + 1 ) is the velocity of particle i at iteration t + 1 . w is the weight of inertia that helps balance exploration and exploitation. c 1 and c 1 are coefficients representing self-confidence and social trust, respectively. r a n d 1 and r a n d 1 are random numbers between 0 and 1. p b e s t i is the best known position for the particle i and g b e s t is the best position known of the entire population. Finally, x i ( t ) and x i ( t + 1 ) represent the current position of the particle i and the next one, respectively.
The algorithm iterates these updates, allowing particles to explore the solution space, with the aim of converging towards the global optimum. The parameters w, c 1 , and c 2 play crucial roles in the behavior of the swarm, affecting the convergence speed and the algorithm’s ability to escape local optima.
PSO is extensively employed due to its simplicity, efficiency, and versatility, enabling its application across a broad spectrum of optimization problems. Its ability to discover solutions without requiring gradient information renders it especially valuable for problems characterized by complex, nonlinear, or discontinuous objective functions.

3.2. Bat Algorithm

The Bat Algorithm is an optimization technique inspired by the echolocation behavior of bats. It simulates the natural echolocation mechanism that bats use for navigation and foraging. This algorithm captures the essence of bats’ sophisticated biological sonar systems, translating the dynamics of echolocation and flight into a computational algorithm capable of searching for global optima in complex optimization problems [6].
In the Bat Algorithm, a population of virtual bats navigates the solution space, where each bat represents a potential solution. The bats use a combination of echolocation and a random walk to explore and exploit the solution space effectively. They adjust their echolocation parameters, such as frequency, pulse rate, and loudness, to locate prey, analogous to finding the optimal solutions in a given problem space. The algorithm employs the Equations (3)–(5), for updating the bats’ positions and velocities:
f i = f m i n + ( f m a x f m i n ) β
v i ( t + 1 ) = v i ( t ) + ( x i ( t ) g b e s t ) f i
x i ( t + 1 ) = x i ( t ) + v i ( t + 1 )
where f i is the frequency of the bat i, ranging from f m i n to f m a x with β being a random number between 0 and 1. v i ( t + 1 ) represents the velocity of bat i at iteration t + 1 , and g b e s t signifies the global best solution found by any bat. x i ( t + 1 ) denotes the position of bat i for the next iteration.
Additionally, to model the bats’ local search and exploitation capability, a random walk is incorporated into the best solution found so far (see Equation (6)). This is achieved by modifying a bat’s position using the average loudness A of all the bats and the pulse emission rate r, guiding the search towards the optimum:
x n e w = x g b e s t + ϵ A
where x n e w represents a new solution generated by local search around the global best position x g b e s t , and ϵ is a random number drawn from a uniform distribution. The values of A and r decrease and increase, respectively, over the course of iterations, fine-tuning the balance between exploration and exploitation based on the proximity to the prey, i.e., the optimal solution.
The Bat Algorithm’s efficiency stems from its dual approach of global search, facilitated by echolocation-inspired movement and local search, enhanced by the random walk based on pulse rate and loudness. This combination allows the algorithm to explore vast areas of the search space while also intensively searching areas near the current best solutions.

3.3. Gray Wolf Optimizer

The Gray Wolf Optimizer is an optimization algorithm inspired by the social hierarchy and hunting behavior of gray wolves in nature. This algorithm mimics the leadership and team dynamics of wolves in packs to identify and converge on optimal solutions in multidimensional search spaces [32]. The core concept behind GWO is the emulation of the way gray wolves organize themselves into a social hierarchy and collaborate during hunting, applying these behaviors to solve optimization problems.
In a gray wolf pack, there are four types of wolves: alpha ( α ), beta ( β ), delta ( δ ), and omega ( ω ), representing the leadership hierarchy. The alpha wolves lead the pack, followed by beta and delta wolves, with omega wolves being at the bottom of the hierarchy. This social structure is translated into the algorithm where the best solution is considered the alpha, the second-best the beta, and the third-best the delta. The rest of the candidate solutions are considered omega wolves, and they follow the lead of the alpha, beta, and delta wolves towards the prey (optimal solution).
The positions of the wolves are updated based on the positions of the alpha, beta, and delta wolves, simulating the hunting strategy and encircling of prey. The mathematical models for updating the positions of the gray wolves are given by Equations (7) and (8):
D = | C · x p ( t ) x ( t ) |
x ( t + 1 ) = x p ( t ) A · D
where x p ( t ) represents the position vector of the prey (or the best solution found so far), x is the position vector of a wolf, A and C are coefficient vectors, and t indicates the current iteration. The vectors A and C are calculated by Equations (8) and (10):
A = 2 · a · r 1 a
C = 2 · r 2
where a linearly decreases from 2 to 0 over the course of iterations, and r 1 , r 2 are random vectors in [ 0 , 1 ] .
The hunting (optimization) is guided mainly by the alpha, beta, and delta wolves, with omega wolves following their lead. The algorithm effectively simulates the wolves’ approach and encircling of prey, exploration of the search area, and exploitation of promising solutions.

3.4. Orca Predator Algorithm

The Orca Predator Algorithm draws inspiration from the sophisticated hunting techniques of orcas, known for their strategic and cooperative behaviors [33]. Orca societies are characterized by complex structures and collaborative efforts in predation, employing echolocation for navigation and prey detection in their aquatic environments. OPA models solutions as n-dimensional vectors within a solution space x = [ x 1 , x 2 , , x n ] T , mimicking these marine predators’ approaches to tracking and capturing prey.
OPA’s methodology encompasses two main phases reflective of orca predation: the chase, involving herding and encircling tactics, and the attack, focusing on the actual capture of prey. During the chase phase, the algorithm alternates between herding prey towards the surface and encircling it to limit escape opportunities, with the decision based on a parameter p and a random number r within [ 0 , 1 ] . The attack phase simulates the final assault on the prey, highlighting the importance of coordination and precision.
x c h a s e , 1 , i t = a × ( d × x b e s t t F × ( b × M t + c × x i t ) )
x c h a s e , 2 , i t = e × x b e s t t x i t
M = i = 1 N x i t N , c = 1 b
x n e w = x c h a s e , 1 , i t = x i t + x c h a s e , 1 , i t when q > r a n d x c h a s e , 2 , i t = x i t + x c h a s e , 2 , i t when q r a n d
Equations (11)–(14) detail the algorithm’s dynamics, modeling velocity and spatial adjustments reflective of orca hunting behaviors. x i t represents the position of the i-th orca at time t, with x b e s t t denoting the optimal solution’s position. Parameters a, b, d, and e are random coefficients that influence the algorithm’s exploration and exploitation mechanisms, with F indicating the attraction force between agents.
After herding prey to the surface, orcas coordinate to finalize the hunt, using their positions and the positions of randomly chosen peers to strategize their attack. This collective behavior is encapsulated in Equations (15) and (16), illustrating the algorithm’s mimicry of orca hunting techniques:
x c h a s e , 3 , i , k t = x j 1 , k t + u × ( x j 2 , k t x j 3 , k t )
u = 2 × ( r a n d 0.5 ) × M a x I t e r t M a x I t e r
These formulations demonstrate how orcas adapt their positions based on the dynamics of their surroundings and the behaviors of their pod members, optimizing their strategies to efficiently capture prey. Through this algorithm, the intricate and collaborative nature of orca predation is leveraged as a metaphor for solving complex optimization problems, with a focus on enhancing solution accuracy and efficiency.

3.5. Reinforcement Learning

Reinforcement learning revolves around the concept of agents operating autonomously to optimize rewards through their decisions, as outlined in comprehensive studies [34]. These agents navigate their learning journey via a trial and error mechanism, pinpointing behaviors that accrue maximum rewards, both immediately and in the future, a hallmark trait of reinforcement learning [35].
During the reinforcement learning journey, agents are in constant interaction with their surroundings, engaging with essential elements like the policy, value function, and, at times, a simulated representation of the environment [36,37,38,39]. The value function assesses the potential success of the actions taken by the agent within its environment, while adjustments in the agent’s policy are influenced by the rewards received.
One pivotal reinforcement learning method, Q-Learning, aims to define a function that evaluates the potential success of an action a t in a certain state s t at time t [40,41]. This evaluation function, or Q function, undergoes updates as per Equation (17):
Q s t , a t Q s t , a t + α r t + 1 + γ max a Q s t + 1 , a Q s t , a t
Here, α symbolizes the learning rate, and γ represents the discount factor, with r t + 1 being the reward after executing action a t .
Deep Q-Learning (DQL) merges the robust capabilities of deep learning with the adaptive mechanisms of reinforcement learning, offering significant advancements in handling complex, high-dimensional environments [42]. This integration allows agents to learn and refine strategies through direct interaction with their environment, optimizing their decision-making processes over time. DQL has proven particularly effective in applications such as autonomous vehicles, robotics, and complex game environments, where agents must make real-time decisions based on incomplete information. Despite its strengths, DQL faces challenges such as the need for extensive training data and potential overfitting, underscoring the importance of ongoing research to enhance its stability and efficiency in diverse applications.
Within Deep Q-Learning, the Q function is articulated as Q ( s t , a t ; θ ) , where s t denotes the present state, a t the action undertaken by the agent at time t, and θ the network’s weights [43]. The Q function’s update mechanism is guided by Equation (18):
Q ( s t , a t ; θ ) Q ( s t , a t ; θ ) + α r t + 1 + γ max a t + 1 Q ( s t + 1 , a t + 1 ; θ ) Q ( s t , a t ; θ )
Here, s t + 1 and a t + 1 indicate the subsequent state and action at time t + 1 , respectively. The learning rate α influences the extent of Q value function updates at each learning step. A higher α facilitates rapid adjustment to environmental changes, beneficial during the learning phase’s early stages or in highly variable settings. Conversely, a lower α ensures a more gradual and steady learning curve but might extend the convergence period. The discount factor γ prioritizes the future over immediate rewards, promoting strategies focused on long-term gain. In contrast, a lower γ favors immediate rewards, suitable for less predictable futures or scenarios necessitating quick policy development. The reward r t + 1 is received post-action a t execution in state s t , with θ denoting the parameters of a secondary neural network that periodically synchronizes with θ to enhance training stability.
A hallmark of Deep Q-Learning is the incorporation of replay memory, a pivotal component of its learning framework [44,45]. Replay memory archives the agent’s experiences as tuples s t , a t , r t + 1 , s t + 1 , with each tuple capturing a distinct experience involving the current state s t , the executed action a t , the obtained reward r t + 1 , and the ensuing state s t + 1 . This methodology of preserving and revisiting past experiences significantly improves learning efficiency and efficacy, enabling the agent to draw from a broader spectrum of experiences. It also diminishes the sequential dependency of learning events, a crucial strategy for mitigating the risk of over-reliance on recent data and fostering a more expansive learning approach. Furthermore, DQL employs the mini-batch strategy for extracting experiences from replay memory throughout the training phase [46]. Rather than progressing from individual experiences one by one, the algorithm opts for a random selection of mini-batches of experiences. This technique of batch sampling bolsters learning stability by ensuring sample independence and optimizes computational resource utilization.
DQL has proven highly effective across a range of applications, demonstrating its versatility and robust decision-making capabilities [47,48]. In autonomous vehicles [49], DQL developed systems that process vast sensory data to make real-time navigational decisions, significantly enhancing road safety and efficiency. In the realm of robotics, an application of a Deep Q-Learning algorithm to enhance the positioning accuracy of an industrial robot was studied in [50]. Additionally, DQL’s strategic decision-making prowess was applied in healthcare to optimize the security and privacy of healthcare data in IoT systems, focusing on authentication, malware, and DDoS attack mitigation, and evaluating performance through metrics like energy consumption and accuracy [51]. In the financial sector, the study [52] introduced an automated trading system that combines reinforcement learning with a deep neural network to predict share quantities and employs transfer learning to overcome data limitations. This approach significantly boosts profits across various stock indices, outperforming traditional systems in volatile markets. Moreover, in supply chain and logistics, the manuscript [53] reviewed the increasing deep reinforcement learning to address challenges stemming from evolving business operations and E-commerce growth, discussing methodologies, applications, and future research directions.
DQL is governed by a loss function according to Equation (19), which measures the discrepancy between the estimated Q and target values:
Loss ( θ t ) = E × y Q ( s t , a t ; θ ) 2
where y is the target value, calculated by Equation (20):
y = r t + 1 + γ × max a t + 1 Q ( s t + 1 , a t + 1 ; θ )
Here, r t + 1 is the reward received after taking action a t in the state s t , and γ is the discount factor, which balances the importance of short-term and long-term rewards. The formulation max a Q ( s t + 1 , a t + 1 ; θ ) represents the maximum estimated value for the next state s t + 1 , according to the target network with parameters θ . Q ( s t , a t ; θ t ) is the Q value estimated by the evaluation network for the current state s t and action a t , using the current parameters θ t . In each training step in DQL, the evaluation network receives a loss function, which is backpropagated based on a batch of experiences randomly selected from the experience replay memory. The evaluation network’s parameter, θ , is then updated by minimizing the loss function through the Stochastic Gradient Descent (SGD) function. After several steps, the target network’s parameter, θ , is updated by assigning the latest parameter θ to θ . After a period of training, the two neural networks are trained stably.

3.6. Cybersecurity Operations Centers

Recent years have seen many organizations establish Cyber SOCs in response to escalating security concerns, necessitating substantial investments in technology and complex setup processes [54]. These centralized hubs enhance incident detection, investigation, and response capabilities by analyzing data from various sources, thereby increasing organizational situational awareness and improving security issue management [55]. The proliferation of the Internet and its integral role in organizations brings heightened security risks, emphasizing the need for continuous monitoring and the implementation of optimization methods to tackle challenges like intrusion detection and prevention effectively [56].
Security Information and Event Management systems have become essential for Cyber SOCs, playing a critical role in safeguarding the IT infrastructure by enhancing cyber-threat detection and response, thereby improving operational efficiency and mitigating security incident impacts [57]. The efficient allocation of centralized NIDS sensors through an SIEM system is crucial for optimizing detection coverage and operational efficiency, considering the organization’s specific security needs [58]. This strategic approach allows for cohesive management and comprehensive security data analysis, leading to a faster and more effective response to security incidents [59]. SIEM systems, widely deployed to manage cyber risks, have evolved into comprehensive solutions that offer broad visibility into high-risk areas, focusing on proactive mitigation strategies to reduce incident response costs and times [60]. Figure 1 illustrates the functional characteristics of a Cyber SOC.
Today’s computer systems are universally vulnerable to cyberattacks, necessitating continuous and comprehensive security measures to mitigate risks [61]. Modern technology infrastructures incorporate various security components, including firewalls, intrusion detection and prevention systems, and security software on devices, to fortify against threats [62]. However, the autonomous operation of these measures requires the integration and analysis of data from different security elements for a complete threat overview, highlighting the importance of Security Information and Event Management systems [63]. As the core of Cyber SOCs, SIEM systems aggregate data from diverse sources, enabling effective threat management and security reporting [64].
SIEM architectures consist of key components such as source device integration, log collection, and event monitoring, with a central engine performing log analysis, filtering, and alert generation [65,66]. These elements work together to provide real-time insights into network activities, as depicted in Figure 2.
NIDS sensors, often based on cost-effective Raspberry Pi units, serve as adaptable and scalable modules for network security, requiring dual Ethernet ports for effective integration into the SIEM ecosystem [67]. This study aims to enhance the assignment and management of NIDS sensors within a centralized network via SIEM, improving the optimization of sensor deployment through the application of Deep Q-Learning to metaheuristics, advancing upon previous work [12].
In this context, cybersecurity risk management is essential for organizations to navigate the evolving threat landscape and implement appropriate controls [68]. It aims to balance securing networks and minimizing losses from vulnerabilities [69], requiring continuous model updates and the strategic deployment of security measures [70]. Cyber-risk management strategies, including the adoption of SIEM systems, are vital for monitoring security events and managing incidents [69].
The optimization problem focuses on deploying NIDS sensors effectively, considering cost, benefits, and indirect costs of non-installation. This involves formulations to minimize sensor costs (Equation (22)), maximize benefits (Equation (23)), and minimize indirect costs (Equation (24)), with constraints ensuring sufficient sensor coverage (Equation (25)) and network reliability (Equation (26)):
F ( x ) = f 1 ( x ) , f 2 ( x ) , f 3 ( x )
f 1 ( x ) : min x i j X i = 1 s j = 1 n x i j c i j
f 2 ( x ) : max x i j X j = 1 n x i j d i j , i
f 3 ( x ) : min x i j X j = 1 n ( 1 x i j ) i i j , i
j = 1 n x i j 1 , i
j = 1 n p j ( 1 x i j ) j = 1 n p j ( 1 u ) , i
This streamlined approach extends the model to larger networks and emphasizes the importance of regular updates and expert collaboration to improve cybersecurity outcomes [12,71].
Expanding on research [12] which optimized NIDS sensor allocation in medium-sized networks, this study extends the approach to larger networks. By analyzing a case study, this research first tackled instance zero with ten VLANs, assigning qualitative variables to each based on operational importance and failure susceptibility for strategic NIDS sensor placement. This formulation led to an efficient allocation of NIDS sensors for the foundational instance zero, as depicted in Figure 3. The study scaled up to forty additional instances, providing a robust examination of NIDS sensor deployment strategies in varied network configurations.

4. Developed Solution

This solution advances the integration of bio-inspired algorithms—Particle Swarm Optimization, the Bat Algorithm, the Grey Wolf Optimizer, and the Orca Predator Algorithm—with Deep Q-Learning to dynamically fine-tune the parameters of these algorithms. Utilizing the collective and adaptive behaviors of PSO, BAT, GWO, and OPA alongside the capabilities of DQL to handle extensive state and action spaces, we enhanced the efficiency of feature selection. Inspired by the natural strategies of their respective biological counterparts and combined with DQL’s proficiency in managing high-dimensional challenges [33,72], this approach innovates optimization tactics while effectively addressing complex combinatorial issues.
DQL is pivotal for shifting towards exploitation, particularly in later optimization phases. As PSO, BAT, GWO, and OPA explore the solution space, DQL focuses the exploration on the most promising regions through an epsilon-greedy policy, optimizing action selection as the algorithm progresses and learns [73].
Each algorithm functions as a metaheuristic with agents (particles, bats, wolves, or agents) representing search agents within the binary vector solution space.
DQL’s reinforcement learning strategy fine-tunes the operational parameters of these algorithms, learning from their performance outcomes to enhance exploration and exploitation balance. Through replay memory, DQL benefits from historical data, incrementally improving NIDS sensor mapping for an SIEM system.
Figure 4 displays the collaborative workflow between the bio-inspired algorithms and DQL, showcasing an efficient and effective optimization methodology that merges nature-inspired exploration with DQL’s adaptive learning.
The essence of our methodology is captured in the pseudocode of Algorithm 1, beginning with dataset input and leading to the global best solution identification. This process involves initializing agents, adjusting their positions and velocities, and employing a training phase to compute fitness, followed by DQL’s refinement of exploration and exploitation strategies.
The core loop iterates until reaching a specified limit, with each agent’s position and velocity updated and fitness evaluated for refining the search strategy.
Finally, the computational complexity of our metaheuristic is O ( k n ) , where n is the problem dimension and k is the number of iterations or population size, reflecting total function evaluations during the algorithm’s execution. The complexity of the Deep Q-Learning (DQL) algorithm is typically O ( M N ) [74,75], with M as the sample size and N as the number of network parameters, crucial for comprehensive dataset analysis. Although M and N can be large, their fixed nature justifies the increased computational demand for enhanced results. Furthermore, ongoing advancements in computing technology help offset the impact of this increased complexity.
Algorithm 1: Enhanced bio-inspired optimization method
Biomimetics 09 00307 i001

5. Experimental Setup

Forty instances were proposed for the experimental stage. These instances entailed random operating parameters, which are detailed below for each instance: the number of operational VLANs, the types of sensors used, the range of sensor costs, the range of benefits associated with sensor installation in a particular VLAN, the range of indirect costs incurred when a sensor is not installed in a given VLAN, and the probability of non-operation for a given VLAN.
Additionally, it is important to note that, as per the mathematical modeling formulation outlined earlier, one of its constraints mandates a minimum operational or uptime availability of ninety percent for the organization’s network. The specific values for each instance are provided in detail in Table 1.
Once the solution vector has been altered, it becomes necessary to implement a binarization step for the usage of continuous metaheuristics in a binary domain [76]. This involves comparing the sigmoid function to a randomly uniform value δ that falls within the range of 0 and 1. Subsequently, a conversion function, for instance, [ 1 / ( 1 + e x i j ) ] > δ , is employed as a method of discretization. In this scenario, if the statement holds true, then x i j 1 . Conversely, if it does not hold true, then x i j 0 .
Our objective was to devise plans and offer recommendations for the trial phase, thereby demonstrating that the recommended strategy is a feasible solution for determining the location of the NIDS sensor. The time taken to solve was calculated to gauge the duration of metaheuristics required to achieve efficient solutions. We used the highest value as a critical measure to evaluate subsequent outcomes, which the Equation determines (27).
( p , q ) p q K f p ( x ) e p ( x b e s t ) ω p m a x + c ^ f q ( x ) c ^ e q ( x b e s t ) ω q m i n , ω ( p , q ) 0
where ω ( p , q ) represents weight of objective functions and ω ( p , q ) = 1 must be satisfied. Values of ω ( p , q ) are defined by analogous estimating. f ( p , q ) ( x ) is the single-objective function and e ( p , q ) ( x b e s t ) stores the best value met independently. Finally, c ^ is an upper bound of minimization single-objective functions.
Following this, we employed ordinal examination to assess the adequacy of the strategy. Subsequently, we elaborate on the hardware and software utilized to duplicate computational experiments. Outcomes are depicted through tables and graphics.
We highlight that test scenarios were developed using standard simulated networks designed to mimic the behavior and characteristics of real networks. These simulations represent the operational characteristics of networks in organizations of various sizes, from minor to medium and large. Depending on its scale and extent, defined by the number of VLANs, each VLAN consists of multiple devices, such as computers, switches, and server farms, along with their related connections. The research evaluated test networks that varied in size, starting from networks with ten VLANs, moving to networks with twenty-five VLANs, and extending to more extensive networks with up to fifty VLANs. The simulation considered limitations such as bandwidth capacity, latency, packet loss, and network congestion by replicating the test networks and considering their functional and working properties. These aspects, along with other factors, were critical in defining the uptime of each VLAN. Network availability is defined as the time or percentage during which the network remains operational and accessible, without experiencing significant downtime. For this study, it was essential that networks maintained a minimum availability of 90%, as interruptions and periods of downtime may occur due to equipment failure, network congestion, or connectivity problems. Implementing proactive monitoring through the SIEM will ensure high availability on the network.

5.1. Methodology

In this study, we employed a multi-phase research approach to ensure thorough analysis and robust findings. Initially, we identified the key variables influencing our research question through an extensive literature review. Following this, we designed our experimental setup to test these variables under controlled conditions, detailing the participant selection criteria, data collection methods, and the statistical tools used for analysis. This meticulous approach enabled us to confirm our data’s reliability and guarantee that our results could be replicated and validated by future studies.
The forty instances representing networks of various sizes and complexities, described in Table 1, were used to evaluate the performance between the native and enhanced metaheuristics following the principles established in [77]. The methodological proposal consisted of an analytical comparison between the hybridization results, that is, the results obtained from the original form of the algorithm and the results obtained with the application of Deep Q-Learning. To achieve this, we implemented the following methodological approach:
  • Preparation and planning: In this phase, network instances that emulated real-world cases, from medium-sized networks to large networks, were generated, randomly covering the various operational and functional scenarios of modern networks. Subsequently, the objectives to achieve were defined as having a secure, operational, and highly available network. These objectives were to minimize the number of NIDS sensors assigned to the network, maximize the installation benefits, and minimize the indirect costs of non-installation. Experiments were designed to systematically evaluate hybridization improvements in a controlled manner, ensuring balanced optimization of the criteria described above.
  • Execution and assessment: We carried out a comprehensive evaluation of both native and improved metaheuristics, analyzing the quality of the solutions obtained and the efficiency in terms of calculation and convergence characteristics. We implement comprehensive tests to perform performance comparisons with descriptive statistical methods and performed the Mann–Whitney–Wilcoxon test for comparative analysis. This method involves determining the appropriateness of each execution for each given instance.
  • Analysis and validation: We performed a comprehensive and in-depth analysis to understand the influence of Deep Q-Learning and the behavior of the PSO, BAT, GWO, and OPA metaheuristics in generating efficient solutions for the corresponding instances. To do this, comparative tables and graphs of the solutions generated by the native and improved metaheuristics were built.

5.2. Implementation Aspects

To ensure clarity and transparency of our experimental design, we present our experimental parameters and platforms in Table 2, below. This table outlines all crucial aspects of our setup, allowing for the easy replication of our methods and verification of our results.

6. Results and Discussion

Table 3, Table 4, Table 5, Table 6 and Table 7 shows the main findings corresponding to the execution of the native metaheuristics and the metaheuristics improved with Deep Q-Learning. The tables are structured into forty sections (one per instance), each consisting of six rows that statistically describe the value of the metric corresponding to the scalarization of objectives, considering the best value obtained as the minimum value and the worst value obtained as the maximum value. The median represents the middle value, and the mean denotes the average of the results, while the standard deviation (STD) and the interquartile range (IQR) quantify the variability in the findings. Concerning columnar representation, PSO, BAT, GWO, and OPA detail results for bio-inspired optimizers lacking a learning component. PSODQL, BATDQL, GWODQL, and OPADQL represent our enhanced versions of biomimetic algorithms.
When analyzing instances one to nine, it is evident that both the native metaheuristics and the metaheuristics improved with Deep Q-Learning produce identical solutions and metrics, given the low complexity of the IT infrastructure of these first instances; however, despite generating exact values regarding the best scalarization value, instances six, eight, and nine show variations in their generation, which can be seen in the variation of the standard deviations of PSO, BAT, and BATDQL.
From instances ten to sixteen, there are slight variations in the solutions obtained by each metaheuristic, although the value of the best solution remains identical in most instances. As for the worst value generated, variations begin to develop, causing variations to appear in the average, standard deviation, median, and interquartile range. In metaheuristics improved with Deep Q-Learning, specifically PSODQL, BATDQL, and OPADQL, it is verified that the standard deviation is lower compared to their corresponding native metaheuristics. This exciting finding demonstrates that the solution values are very similar to the average; in other words, there is little variability among the solution results, suggesting that the results are consistent and stable. Moreover, experiments with Deep Q-Learning metaheuristics indicate that the experiments are reliable and that random errors have a minimal impact on the outcomes.
Subsequently, in instance seventeen, a great variety is observed in the solutions generated, with PSODQL providing the best solution and OPADQL in second place, maintaining the previous finding with respect to the standard deviation.
For instance, from eighteen to twenty, there is a wide variety of solutions, highlighting PSODQL, BATDQL, and OPADQL. It is interesting to verify that BAT, GWO, and OPA, both native and improved, generate the exact value of the best solution. However, the standard deviation in the improved metaheuristics is lower than that obtained in the native metaheuristics, which reaffirms the consistency and stability of the results.
From instance twenty-one to instance thirty-two, the PSODQL, BATDQL, and OPADQL metaheuristics generate better solution values concerning their corresponding native metaheuristics, and regarding their corresponding standard deviations, they are lower concerning the native metaheuristics; PSODQL’s performance stands out as it produces the best solution values.
In instances thirty-three and thirty-four, the performance of the metaheuristics PSODQL, BATDQL, and OPADQL is maintained, highlighting the excellent performance of BATDQL in instance thirty-three and OPADQL in instance thirty-four.
Concluding with instances thirty-five to forty, we can observe that PSODQL, BATDQL, and OPADQL continue to obtain the best solution values; the standard deviations maintain a small value compared to their native counterparts, and PSODQL, which generated the best solution value, is highlighted.
In the application of metaheuristics with Deep Q-Learning, specifically PSODQL, BATDQL, and OPADQL, in addition to generating better solution values, observing a low standard deviation is beneficial as it indicates that the generated solutions are efficiently clustered around optimal values, thus reflecting the high precision and consistency of the results. This pattern suggests the algorithms’ notable effectiveness in identifying optimal or near-optimal solutions, with minimal variation across multiple executions, a crucial aspect for effectively resolving complex problems. Furthermore, a reduced interquartile range reaffirms the concentration of solutions around the median, decreasing data dispersion and refining the search towards regions of the solution space with high potential, which improves precision in reaching efficient solutions.
To present the results graphically, we faced the challenge of analyzing and comparing samples generated from non-parametric underlying processes, that is, processes whose data do not assume a normal distribution. Given this, it became essential to use a visualization tool such as the violin diagram, which adequately handles the non-parametric nature of the data and provides a clear and detailed view of their corresponding distributions. Visualizing these graphs allows us to consolidate the previously analyzed results, corresponding to the evaluation metric and, later in this section, the Wilcoxon–Mann–Whitney test.
Figure 5, Figure 6, Figure 7 and Figure 8 enrich our comprehension of the effectiveness of biomimetic algorithms (left side) and their enhanced version (right side). These graphical illustrations reveal the data’s distribution; highlighting that the learning component provides a real improvement for each optimization algorithm. The violin diagram is an analytical tool that combines box plots and kernel density diagrams to compare data distribution between two samples; it was used to visualize the results. It shows summarized statistics, such as medians and quartiles, and the data density along its range. It helps identify and analyze significant differences between two samples, offering insights into patterns and the data structure [78]. This way, we can appreciate that in instances fifteen and sixteen, the standard deviation is small in the metaheuristics with DQL compared to native metaheuristics, especially PSODQL, BATDQL, and OPADQL. Furthermore, the median in PSODQL in instance fifteen is much lower than in native PSO. For instances seventeen to twenty, in addition to noting the minor standard deviation in the metaheuristics with Q-Learning, the medians for PSODQL and OPADQL are significantly lower than their native counterparts. From twenty to twenty-six, the previous results for the metaheuristics with DQL are maintained, and the distributions and medians for PSODQL and OPADQL move to lower values. For instances twenty-seven and twenty-eight, the standard deviation is small in the metaheuristics with DQL compared to the native metaheuristics. For instance, we can verify that the distribution and the median in PSODQL reach lower values in twenty-nine. For instances thirty and thirty-one, the distributions and medians in PSODQL, BATDQL, and OPADQL reach lower values. For instance thirty-two, both PSODQL and OPADQL distributions and medians reach lower values, and from thirty-three to forty, we can verify that in most cases, PSODQL, BATDQL, and OPA’s medians tend to lower values. From the above, we can confirm that the visualizations of the solutions for the instances allow us to reaffirm the findings and results of the substantial improvement of the metaheuristics with DQL compared to the native metaheuristics, highlighting PSODQL as the one that generates the best solutions throughout the experimentation phase.
It is worth mentioning that the visualization of the solutions from instances one to fourteen is impossible to graph, since they mainly generate the same statistical values.
In the context of this research, we conducted the Wilcoxon–Mann–Whitney test, a non-parametric statistical test used to compare two independent samples [79]. It was used to determine if there were significant differences in two groups of samples that may not have the same distribution, which were generated from native metaheuristics and DQL. The significance level was previously set at p = 0.05 to conduct the test.
The results are detailed in Table 8, describing the following findings. It is verified that from instances fifteen and sixteen, there are significant differences between the samples generated by PSODQL and native PSO, concluding that there is an improvement in the results obtained by PSODQL. For BAT and BATDQL, there are no significant differences between the samples; the same is true for GWO and GWODQL. However, for OPA and OPADQL, there is a substantial difference between the samples. PSODQL shows a more remarkable improvement, as it has a more significant difference than OPADQL since the obtained p-value is lower, as verified in the table. In instances seventeen, for the samples generated by PSO, there is a significant difference between the samples, resulting in a better performance of PSODQL; the same happens with BAT, resulting in a better BATDQL; for native GWO, it is better than GWODQL, and for the samples generated by OPA, there is a significant difference, resulting in a better OPADQL. In the eighteenth and nineteenth instances, it is confirmed that PSODQL is better than PSO. For BAT and BATDQL, there are no significant differences between the samples, just as for GWO and GWODQL. Moreover, OPADQL is better for OPA, as there are substantial differences between the samples. In both cases, PSODQL is better since it has the lowest p value. In instance twenty, PSODQL and OPADQL show significant differences between their samples; however, OPADQL is better since it has the lowest p value. In instance twenty-one, given the obtained results, PSODQL is better than native PSO, and the same applies to BAT; for GWO, native GWO is better, and for OPADQL, there are no significant differences between the samples. For this instance, PSODQL is better since it has the lowest p value.
For instances from twenty-three to twenty-eight, significant differences are verified between the samples generated by PSO, BAT, and OPA, the result being that the samples generated by DQL are better. For GWO, there are cases of significant differences between their samples. For the twenty-ninth instance, significant differences exist for the samples generated by PSO and OPA, resulting in better PSODQL and OPADQL. For instances thirty to thirty-two, PSODQL, BATDQL, and OPADQL are better. For the thirty-third instances, PSODQL and OPADQL turned out to be better. For instances thirty-four to thirty-six, PSODQL, BATDQL, and OPADQL are better. For the thirty-seventh instance, PSODQL and OPADQL are better. PSODQL, BATDQL, and OPADQL are the best for the thirty-eighth cases. For instances thirty-nine and forty, PSODQL and OPADQL are the best.
The central objective of this study was to evaluate the impact of integrating the Deep Q-Learning technique into traditional metaheuristics to improve their effectiveness in optimization tasks. The results demonstrate that the Deep Q-Learning-enhanced versions, specifically PSODQL, BATDQL, and OPADQL, exhibited superior performance compared to their native counterparts. Notably, PSODQL stood out significantly, outperforming native PSO in one hundred percent of the cases during the experimental phase. These findings highlight the potential of reinforcement learning through Deep Q-Learning as an effective strategy to enhance the performance of metaheuristics in optimization problems.

7. Conclusions

The presented research tackles the challenge of enhancing the efficiency of Cybersecurity Operations Centers through the integration of biomimetic algorithms and Deep Q-Learning, a reinforcement learning technique. This approach is proposed to improve the deployment of sensors across network infrastructures, balancing security imperatives against deployment costs. The research is grounded in the premise that the dynamic nature of cyber threats necessitates adaptive and efficient solutions for cybersecurity management.
The study demonstrated that incorporating DQL into biomimetic algorithms significantly improves the effectiveness of these algorithms, enabling optimal resource allocation and efficient intrusion detection. Experimental results validated the hypothesis that combining biomimetic optimization techniques with deep reinforcement learning leads to superior solutions compared to conventional strategies.
A comparative analysis between native biomimetic algorithms and those enhanced with DQL revealed a notable improvement in the accuracy and consistency of the solutions obtained. This enhancement is attributed to the ability of DQL to dynamically adapt and fine-tune the algorithms’ parameters, focusing the search towards the most promising regions of the solution space. Moreover, the implementation of replay memory and the mini-batch strategy in DQL contributed to learning efficiency and training stability.
The study underscores the importance of integrating machine learning techniques with optimization algorithms to address complex problems in cybersecurity. The adaptability and improved performance of biomimetic algorithms enhanced with DQL offer a promising approach to efficient Cyber SOC management, highlighting the potential of these advanced techniques in the cybersecurity domain.
Future works could pivot towards creating adaptive defense mechanisms by integrating biomimetic algorithms with Deep Q-Learning, focusing on real-time threat responses and evolutionary security frameworks. This would entail embedding ethical AI principles to ensure that these advanced systems operate without bias and respect privacy. Additionally, exploring federated learning for collaborative defense across Cyber SOCs could revolutionize how threat intelligence is shared, fostering a unified global response to cyber threats without compromising sensitive data. These directions promise to significantly elevate the cybersecurity landscape, making it more resilient, ethical, and collaborative.

Author Contributions

Formal analysis, R.O., O.S., C.R., R.S. and B.C.; investigation, R.O., O.S., C.R., R.S. and B.C.; methodology, R.O. and R.S.; Resources, R.O.; software, O.S. and C.R.; validation, R.O., R.S. and B.C.; writing—original draft, O.S., C.R. and R.O.; writing—review and editing, R.O., O.S., C.R., R.S. and B.C. All the authors of this paper hold responsibility for every part of this manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

R.O. is supported by grant ANID/FONDECYT/INICIACION/11231016. B.C. is supported by grant ANID/FONDECYT/REGULAR/1210810 and the Spanish Ministry of Science and Innovation Project PID2019-109891RB-I00, under the European Regional Development Fund (FEDER).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Acknowledgments

Omar Salinas wishes to thank the Doctorado en Ingeniería Informática Aplicada at the Universidad de Valparaíso for their support under Grant 101.016/2020.

Conflicts of Interest

The authors declare no conflicts of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Yıldırım, İ. Cyber Risk Management in Banks: Cyber Risk İnsurance. In Global Cybersecurity Labor Shortage and International Business Risk; IGI Global: Hershey, PA, USA, 2019; pp. 38–50. [Google Scholar]
  2. Melaku, H.M. Context-Based and Adaptive Cybersecurity Risk Management Framework. Risks 2023, 11, 101. [Google Scholar] [CrossRef]
  3. Darwish, S.M.; Farhan, D.A.; Elzoghabi, A.A. Building an Effective Classifier for Phishing Web Pages Detection: A Quantum-Inspired Biomimetic Paradigm Suitable for Big Data Analytics of Cyber Attacks. Biomimetics 2023, 8, 197. [Google Scholar] [CrossRef] [PubMed]
  4. Broeckhoven, C.; Winters, S. Biomimethics: A critical perspective on the ethical implications of biomimetics in technological innovation. Bioinspir. Biomimetics 2023, 18, 053001. [Google Scholar] [CrossRef]
  5. Ding, H.; Liu, Y.; Wang, Z.; Jin, G.; Hu, P.; Dhiman, G. Adaptive Guided Equilibrium Optimizer with Spiral Search Mechanism to Solve Global Optimization Problems. Biomimetics 2023, 8, 383. [Google Scholar] [CrossRef]
  6. Yang, X.; Li, H. Evolutionary-state-driven Multi-swarm Cooperation Particle Swarm Optimization for Complex Optimization Problem. Inf. Sci. 2023, 646, 119302. [Google Scholar] [CrossRef]
  7. Li, W.; Liang, P.; Sun, B.; Sun, Y.; Huang, Y. Reinforcement learning-based particle swarm optimization with neighborhood differential mutation strategy. Swarm Evol. Comput. 2023, 78, 101274. [Google Scholar] [CrossRef]
  8. Nama, S.; Saha, A.K.; Chakraborty, S.; Gandomi, A.H.; Abualigah, L. Boosting particle swarm optimization by backtracking search algorithm for optimization problems. Swarm Evol. Comput. 2023, 79, 101304. [Google Scholar] [CrossRef]
  9. Seyyedabbasi, A. A reinforcement learning-based metaheuristic algorithm for solving global optimization problems. Adv. Eng. Softw. 2023, 178, 103411. [Google Scholar] [CrossRef]
  10. Taye, M.M. Understanding of Machine Learning with Deep Learning: Architectures, Workflow, Applications and Future Directions. Computers 2023, 12, 91. [Google Scholar] [CrossRef]
  11. Peres, F.; Castelli, M. Combinatorial optimization problems and metaheuristics: Review, challenges, design, and development. Appl. Sci. 2021, 11, 6449. [Google Scholar] [CrossRef]
  12. Salinas, O.; Soto, R.; Crawford, B.; Olivares, R. An integral cybersecurity approach using a many-objective optimization strategy. IEEE Access 2023, 11, 91913–91936. [Google Scholar] [CrossRef]
  13. Wawrowski, Ł.; Białas, A.; Kajzer, A.; Kozłowski, A.; Kurianowicz, R.; Sikora, M.; Szymańska-Kwiecień, A.; Uchroński, M.; Białczak, M.; Olejnik, M.; et al. Anomaly detection module for network traffic monitoring in public institutions. Sensors 2023, 23, 2974. [Google Scholar] [CrossRef] [PubMed]
  14. Kaur, G.; Lashkari, A.H. An introduction to security operations. In Advances in Cybersecurity Management; Springer: Berlin/Heidelberg, Germany, 2021; pp. 463–481. [Google Scholar]
  15. Nespoli, P.; Gomez Marmol, F.; Kambourakis, G. AISGA: Multi-objective parameters optimization for countermeasures selection through genetic algorithm. In Proceedings of the 16th International Conference on Availability, Reliability and Security, Vienna, Austria, 17–20 August 2021; pp. 1–8. [Google Scholar]
  16. da Costa Oliveira, A.L.; Britto, A.; Gusmão, R. Machine learning enhancing metaheuristics: A systematic review. Soft Comput. 2023, 27, 15971–15998. [Google Scholar] [CrossRef]
  17. Almasoud, A.S. Enhanced Metaheuristics with Machine Learning Enabled Cyberattack Detection Model. Intell. Autom. Soft Comput. 2023, 37, 2849–2863. [Google Scholar] [CrossRef]
  18. Albahri, O.; AlAmoodi, A. Cybersecurity and Artificial Intelligence Applications: A Bibliometric Analysis Based on Scopus Database. Mesopotamian J. Cybersecur. 2023, 2023, 158–169. [Google Scholar] [CrossRef] [PubMed]
  19. Olivares, R.; Soto, R.; Crawford, B.; Ríos, V.; Olivares, P.; Ravelo, C.; Medina, S.; Nauduan, D. A learning–based particle swarm optimizer for solving mathematical combinatorial problems. Axioms 2023, 12, 643. [Google Scholar] [CrossRef]
  20. Liang, Y.C.; Cuevas Juarez, J.R. A self-adaptive virus optimization algorithm for continuous optimization problems. Soft Comput. 2020, 24, 13147–13166. [Google Scholar] [CrossRef]
  21. Yi, W.; Qu, R.; Jiao, L.; Niu, B. Automated design of metaheuristics using reinforcement learning within a novel general search framework. IEEE Trans. Evol. Comput. 2022, 27, 1072–1084. [Google Scholar] [CrossRef]
  22. Malibari, A.A.; Alotaibi, S.S.; Alshahrani, R.; Dhahbi, S.; Alabdan, R.; Al-wesabi, F.N.; Hilal, A.M. A novel metaheuristics with deep learning enabled intrusion detection system for secured smart environment. Sustain. Energy Technol. Assess. 2022, 52, 102312. [Google Scholar] [CrossRef]
  23. Zhong, R.; Peng, F.; Yu, J.; Munetomo, M. Q-learning based vegetation evolution for numerical optimization and wireless sensor network coverage optimization. Alex. Eng. J. 2024, 87, 148–163. [Google Scholar] [CrossRef]
  24. Alturkistani, H.; El-Affendi, M.A. Optimizing cybersecurity incident response decisions using deep reinforcement learning. Int. J. Electr. Comput. Eng. 2022, 12, 6768. [Google Scholar] [CrossRef]
  25. Abedzadeh, N.; Jacobs, M. A Reinforcement Learning Framework with Oversampling and Undersampling Algorithms for Intrusion Detection System. Appl. Sci. 2023, 13, 11275. [Google Scholar] [CrossRef]
  26. Al-kahtani, M.S.; Mehmood, Z.; Sadad, T.; Zada, I.; Ali, G.; ElAffendi, M. Intrusion detection in the Internet of Things using fusion of GRU-LSTM deep learning model. Intell. Autom. Soft Comput. 2023, 37, 2283. [Google Scholar] [CrossRef]
  27. Shon, H.G.; Lee, Y.; Yoon, M. Semi-Supervised Alert Filtering for Network Security. Electronics 2023, 12, 4755. [Google Scholar] [CrossRef]
  28. Rawindaran, N.; Jayal, A.; Prakash, E.; Hewage, C. Cost benefits of using machine learning features in NIDS for cyber security in UK small medium enterprises (SME). Future Internet 2021, 13, 186. [Google Scholar] [CrossRef]
  29. Domínguez-Dorado, M.; Rodríguez-Pérez, F.J.; Carmona-Murillo, J.; Cortés-Polo, D.; Calle-Cancho, J. Boosting holistic cybersecurity awareness with outsourced wide-scope CyberSOC: A generalization from a spanish public organization study. Information 2023, 14, 586. [Google Scholar] [CrossRef]
  30. Alabdulatif, A.; Thilakarathne, N.N. Bio-inspired internet of things: Current status, benefits, challenges, and future directions. Biomimetics 2023, 8, 373. [Google Scholar] [CrossRef]
  31. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995. [Google Scholar] [CrossRef]
  32. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  33. Jiang, Y.; Wu, Q.; Zhu, S.; Zhang, L. Orca predation algorithm: A novel bio-inspired algorithm for global optimization problems. Expert Syst. Appl. 2022, 188, 116026. [Google Scholar] [CrossRef]
  34. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  35. Wang, L.; Pan, Z.; Wang, J. A review of reinforcement learning based intelligent optimization for manufacturing scheduling. Complex Syst. Model. Simul. 2021, 1, 257–270. [Google Scholar] [CrossRef]
  36. Sun, H.; Yang, L.; Gu, Y.; Pan, J.; Wan, F.; Song, C. Bridging locomotion and manipulation using reconfigurable robotic limbs via reinforcement learning. Biomimetics 2023, 8, 364. [Google Scholar] [CrossRef] [PubMed]
  37. Zhu, K.; Zhang, T. Deep reinforcement learning based mobile robot navigation: A review. Tsinghua Sci. Technol. 2021, 26, 674–691. [Google Scholar] [CrossRef]
  38. Azar, A.T.; Koubaa, A.; Ali Mohamed, N.; Ibrahim, H.A.; Ibrahim, Z.F.; Kazim, M.; Ammar, A.; Benjdira, B.; Khamis, A.M.; Hameed, I.A.; et al. Drone deep reinforcement learning: A review. Electronics 2021, 10, 999. [Google Scholar] [CrossRef]
  39. Alavizadeh, H.; Alavizadeh, H.; Jang-Jaccard, J. Deep Q-learning based reinforcement learning approach for network intrusion detection. Computers 2022, 11, 41. [Google Scholar] [CrossRef]
  40. Zhang, L.; Tang, L.; Zhang, S.; Wang, Z.; Shen, X.; Zhang, Z. A Self-Adaptive Reinforcement-Exploration Q-Learning Algorithm. Symmetry 2021, 13, 1057. [Google Scholar] [CrossRef]
  41. Jang, B.; Kim, M.; Harerimana, G.; Kim, J.W. Q-learning algorithms: A comprehensive classification and applications. IEEE Access 2019, 7, 133653–133667. [Google Scholar] [CrossRef]
  42. Wang, H.N.; Liu, N.; Zhang, Y.Y.; Feng, D.W.; Huang, F.; Li, D.S.; Zhang, Y.M. Deep reinforcement learning: A survey. Front. Inf. Technol. Electron. Eng. 2020, 21, 1726–1744. [Google Scholar] [CrossRef]
  43. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef]
  44. Diekmann, N.; Walther, T.; Vijayabaskaran, S.; Cheng, S. Deep reinforcement learning in a spatial navigation task: Multiple contexts and their representation. In Proceedings of the 2019 Conference on Cognitive Computational Neuroscience, Berlin, Germany, 13–16 September 2019. [Google Scholar]
  45. Schaul, T.; Quan, J.; Antonoglou, I.; Silver, D. Prioritized Experience Replay. arXiv 2015, arXiv:1511.05952. [Google Scholar]
  46. Ramicic, M.; Bonarini, A. Correlation minimizing replay memory in temporal-difference reinforcement learning. Neurocomputing 2020, 393, 91–100. [Google Scholar] [CrossRef]
  47. Ji, Z.; Xiao, W. Improving decision-making efficiency of image game based on deep Q-learning. Soft Comput. 2020, 24, 8313–8322. [Google Scholar] [CrossRef]
  48. Yavas, U.; Kumbasar, T.; Ure, N.K. A New Approach for Tactical Decision Making in Lane Changing: Sample Efficient Deep Q Learning with a Safety Feedback Reward. In Proceedings of the 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020. [Google Scholar] [CrossRef]
  49. Cai, P.; Wang, H.; Sun, Y.; Liu, M. DQ-GAT: Towards Safe and Efficient Autonomous Driving with Deep Q-Learning and Graph Attention Networks. IEEE Trans. Intell. Transp. Syst. 2022, 23, 21102–21112. [Google Scholar] [CrossRef]
  50. Sumanas, M.; Petronis, A.; Bucinskas, V.; Dzedzickis, A.; Virzonis, D.; Morkvenaite-Vilkonciene, I. Deep Q-Learning in Robotics: Improvement of Accuracy and Repeatability. Sensors 2022, 22, 3911. [Google Scholar] [CrossRef] [PubMed]
  51. Roy, P.P.; Teju, V.; Kandula, S.R.; Sowmya, K.V.; Stan, A.I.; Stan, O.P. Secure Healthcare Model Using Multi-Step Deep Q Learning Network in Internet of Things. Electronics 2024, 13, 669. [Google Scholar] [CrossRef]
  52. Jeong, G.; Kim, H.Y. Improving financial trading decisions using deep Q-learning: Predicting the number of shares, action strategies, and transfer learning. Expert Syst. Appl. 2019, 117, 125–138. [Google Scholar] [CrossRef]
  53. Yan, Y.; Chow, A.H.; Ho, C.P.; Kuo, Y.H.; Wu, Q.; Ying, C. Reinforcement learning for logistics and supply chain management: Methodologies, state of the art, and future opportunities. Transp. Res. Part E Logist. Transp. Rev. 2022, 162, 102712. [Google Scholar] [CrossRef]
  54. Vaarandi, R.; Mäses, S. How to Build a SOC on a Budget. In Proceedings of the 2022 IEEE International Conference on Cyber Security and Resilience (CSR), Rhodes, Greece, 27–29 July 2022; pp. 171–177. [Google Scholar]
  55. János, F.D.; Dai, N.H.P. Security concerns towards security operations centers. In Proceedings of the 2018 IEEE 12th International Symposium on Applied Computational Intelligence and Informatics (SACI), Timisoara, Romania, 17–19 May 2018; pp. 000273–000278. [Google Scholar]
  56. Alterazi, H.A.; Kshirsagar, P.R.; Manoharan, H.; Selvarajan, S.; Alhebaishi, N.; Srivastava, G.; Lin, J.C.W. Prevention of cybersecurity with the internet of things using particle swarm optimization. Sensors 2022, 22, 6117. [Google Scholar] [CrossRef] [PubMed]
  57. Menges, F.; Latzo, T.; Vielberth, M.; Sobola, S.; Pöhls, H.C.; Taubmann, B.; Köstler, J.; Puchta, A.; Freiling, F.; Reiser, H.P.; et al. Towards GDPR-compliant data processing in modern SIEM systems. Comput. Secur. 2021, 103, 102165. [Google Scholar] [CrossRef]
  58. Kotecha, K.; Verma, R.; Rao, P.V.; Prasad, P.; Mishra, V.K.; Badal, T.; Jain, D.; Garg, D.; Sharma, S. Enhanced network intrusion detection system. Sensors 2021, 21, 7835. [Google Scholar] [CrossRef]
  59. Aghmadi, A.; Hussein, H.; Polara, K.H.; Mohammed, O. A Comprehensive Review of Architecture, Communication, and Cybersecurity in Networked Microgrid Systems. Inventions 2023, 8, 84. [Google Scholar] [CrossRef]
  60. González-Granadillo, G.; González-Zarzosa, S.; Diaz, R. Security information and event management (SIEM): Analysis, trends, and usage in critical infrastructures. Sensors 2021, 21, 4759. [Google Scholar] [CrossRef] [PubMed]
  61. Ghiasi, M.; Niknam, T.; Wang, Z.; Mehrandezh, M.; Dehghani, M.; Ghadimi, N. A comprehensive review of cyber-attacks and defense mechanisms for improving security in smart grid energy systems: Past, present and future. Electr. Power Syst. Res. 2023, 215, 108975. [Google Scholar] [CrossRef]
  62. Wanjau, S.K.; Wambugu, G.M.; Oirere, A.M.; Muketha, G.M. Discriminative spatial-temporal feature learning for modeling network intrusion detection systems. J. Comput. Secur. 2023, 32, 1–30. [Google Scholar] [CrossRef]
  63. Younus, Z.; Alanezi, M. A Survey on Network Security Monitoring: Tools and Functionalities. Mustansiriyah J. Pure Appl. Sci. 2023, 1, 55–86. [Google Scholar]
  64. Tuyishime, E.; Balan, T.C.; Cotfas, P.A.; Cotfas, D.T.; Rekeraho, A. Enhancing Cloud Security—Proactive Threat Monitoring and Detection Using a SIEM-Based Approach. Appl. Sci. 2023, 13, 12359. [Google Scholar] [CrossRef]
  65. Bezas, K.; Filippidou, F. Comparative Analysis of Open Source Security Information & Event Management Systems (SIEMs). Indones. J. Comput. Sci. 2023, 12, 443–468. [Google Scholar]
  66. Muhammad, A.R.; Sukarno, P.; Wardana, A.A. Integrated Security Information and Event Management (SIEM) with Intrusion Detection System (IDS) for Live Analysis based on Machine Learning. Procedia Comput. Sci. 2023, 217, 1406–1415. [Google Scholar] [CrossRef]
  67. Awajan, A. A novel deep learning-based intrusion detection system for IOT networks. Computers 2023, 12, 34. [Google Scholar] [CrossRef]
  68. Kure, H.I.; Islam, S.; Mouratidis, H. An integrated cybersecurity risk management framework and risk predication for the critical infrastructure protection. Neural Comput. Appl. 2022, 34, 15241–15271. [Google Scholar] [CrossRef]
  69. Safitra, M.F.; Lubis, M.; Fakhrurroja, H. Counterattacking cyber threats: A framework for the future of cybersecurity. Sustainability 2023, 15, 13369. [Google Scholar] [CrossRef]
  70. Oyedokun, G.E.; Campbell, O. Imperatives of Risk Analysis and Asset Management on Cybersecurity in a Technology-Driven Economy. In Effective Cybersecurity Operations for Enterprise-Wide Systems; IGI Global: Hershey, PA, USA, 2023; pp. 147–168. [Google Scholar]
  71. Zhang, Y.; Malacaria, P. Optimization-time analysis for cybersecurity. IEEE Trans. Dependable Secur. Comput. 2021, 19, 2365–2383. [Google Scholar] [CrossRef]
  72. Tan, F.; Yan, P.; Guan, X. Deep reinforcement learning: From Q-learning to deep Q-learning. In Proceedings of the Neural Information Processing: 24th International Conference, ICONIP 2017, Guangzhou, China, 14–18 November 2017; Proceedings, Part IV 24. Springer: Berlin/Heidelberg, Germany, 2017; pp. 475–483. [Google Scholar]
  73. Fotouhi, A.; Ding, M.; Hassan, M. Deep q-learning for two-hop communications of drone base stations. Sensors 2021, 21, 1960. [Google Scholar] [CrossRef]
  74. Hu, X.; Chu, L.; Pei, J.; Liu, W.; Bian, J. Model complexity of deep learning: A survey. Knowl. Inf. Syst. 2021, 63, 2585–2619. [Google Scholar] [CrossRef]
  75. Fan, J.; Wang, Z.; Xie, Y.; Yang, Z. A theoretical analysis of deep Q-learning. In Proceedings of the Learning for Dynamics and Control, Virtual, 10–11 June 2020; pp. 486–489. [Google Scholar]
  76. Crawford, B.; Soto, R.; Astorga, G.; García, J.; Castro, C.; Paredes, F. Putting Continuous Metaheuristics to Work in Binary Search Spaces. Complexity 2017, 2017, 8404231. [Google Scholar] [CrossRef]
  77. Bartz-Beielstein, T.; Preuss, M. Experimental research in evolutionary computation. In Proceedings of the 9th Annual Conference Companion on Genetic and Evolutionary Computation, London, UK, 7–11 July 2007; pp. 3001–3020. [Google Scholar]
  78. Hund, A.K.; Stretch, E.; Smirnoff, D.; Roehrig, G.H.; Snell-Rood, E.C. Broadening the taxonomic breadth of organisms in the bio-inspired design process. Biomimetics 2023, 8, 48. [Google Scholar] [CrossRef]
  79. Wilcox, R. A Heteroscedastic Analog of the Wilcoxon–Mann–Whitney Test When There Is a Covariate. Int. J. Stat. Probab. 2023, 12. [Google Scholar] [CrossRef]
Figure 1. Cyber SOC functional characteristics.
Figure 1. Cyber SOC functional characteristics.
Biomimetics 09 00307 g001
Figure 2. SIEM functional characteristics.
Figure 2. SIEM functional characteristics.
Biomimetics 09 00307 g002
Figure 3. Network topology. Instance-zero solution.
Figure 3. Network topology. Instance-zero solution.
Biomimetics 09 00307 g003
Figure 4. Solution developed using four metaheuristics with Deep Q-Learning.
Figure 4. Solution developed using four metaheuristics with Deep Q-Learning.
Biomimetics 09 00307 g004
Figure 5. Computational result distributions between improved biomimetic algorithms against their native versions. Hardest instances from 15 to 22.
Figure 5. Computational result distributions between improved biomimetic algorithms against their native versions. Hardest instances from 15 to 22.
Biomimetics 09 00307 g005
Figure 6. Computational result distributions between improved biomimetic algorithms against their native versions. Hardest instances from 23 to 30.
Figure 6. Computational result distributions between improved biomimetic algorithms against their native versions. Hardest instances from 23 to 30.
Biomimetics 09 00307 g006
Figure 7. Computational result distributions between improved biomimetic algorithms against their native versions. Hardest instances from 31 to 38.
Figure 7. Computational result distributions between improved biomimetic algorithms against their native versions. Hardest instances from 31 to 38.
Biomimetics 09 00307 g007
Figure 8. Computational result distributions between improved biomimetic algorithms against their native versions. Hardest instances from 39 to 40.
Figure 8. Computational result distributions between improved biomimetic algorithms against their native versions. Hardest instances from 39 to 40.
Biomimetics 09 00307 g008
Table 1. Specification of the operational parameters for the forty instances.
Table 1. Specification of the operational parameters for the forty instances.
InstanceNumber of VLANsType of SensorsUptimeRange of Direct CostsQualitative Profit RangeRange of Indirect CostsPerformance of Subnets
110290%[100–150][1–20][1–7][0.39–0.80]
210290%[100–150][5–20][1–7][0.10–0.80]
310290%[100–150][1–20][1–7][0.02–0.80]
410290%[100–150][1–20][1–5][0.11–0.80]
515290%[100–150][1–20][1–7][0.14–0.85]
615290%[100–150][1–20][1–7][0.01–0.94]
715290%[100–150][1–20][1–7][0.01–0.94]
815290%[100–150][1–20][1–7][0.07–0.96]
915290%[100–150][1–20][3–7][0.07–0.96]
1020290%[100–150][1–20][1–7][0.04–0.61]
1120290%[100–150][1–20][1–7][0.07–0.56]
1220290%[100–150][1–20][1–7][0.10–0.91]
1320290%[100–150][1–20][1–7][0.01–0.99]
1420290%[100–150][1–20][1–7][0.05–0.88]
1525290%[100–150][1–20][1–7][0.07–0.96]
1625290%[100–150][1–20][1–7][0.07–0.96]
1725290%[100–150][1–20][1–7][0.07–0.89]
1825290%[100–150][1–20][1–5][0.08–0.97]
1925290%[100–150][1–20][1–7][0.06–0.99]
2030290%[100–150][10–20][1–7][0.50–0.89]
2130290%[100–150][1–20][1–7][0.22–0.89]
2230290%[100–150][1–20][1–7][0.07–0.96]
2330290%[100–150][1–20][1–7][0.08–0.97]
2430290%[100–150][1–20][1–7][0.05–0.98]
2535290%[100–150][1–20][1–7][0.10–0.96]
2635290%[100–150][1–20][1–7][0.07–0.94]
2735290%[100–150][1–20][1–7][0.07–0.94]
2835290%[100–150][1–20][1–7][0.03–0.98]
2935290%[100–150][1–20][1–7][0.08–0.98]
3040290%[100–150][1–20][1–7][0.06–0.98]
3140290%[100–150][1–20][1–7][0.05–0.98]
3240290%[100–150][1–20][1–7][0.04–0.97]
3340290%[100–150][1–20][1–7][0.16–0.93]
3440290%[100–150][1–20][1–7][0.09–0.95]
3545290%[100–150][1–20][1–7][0.01–0.95]
3645290%[100–150][1–20][1–7][0.07–0.97]
3745290%[100–150][1–20][1–7][0.01–0.95]
3845290%[100–150][1–20][1–7][0.03–0.97]
3945290%[100–150][1–20][1–7][0.07–0.96]
4050290%[100–150][1–20][1–7][0.02–0.84]
Table 2. Experimental parameters and platform details.
Table 2. Experimental parameters and platform details.
ParameterValue
Particle Swarm Optimization
Inertia weight (w) 0.6
Cognitive acceleration ( c 1 ) 0.6
Social acceleration ( c 2 ) 0.6
Number of particles ( p s )10
Maximum iterations (T)100
Bat Algorithm
Larger search space jumps for broader exploration ( f m i n ) 0.75
Finer adjustments for more detailed exploitation ( f m a x ) 1.25
Modulating the decay rate of loudness over time ( α ) 0.9
Modulating the pulse rate’s decay over time ( γ ) 0.9
Adding randomness to the bat’s movement. ( ϵ ) 0.9
Number of virtual bats ( p s )10
Maximum iterations (T)100
Gray Wolf Optimization
Number of wolves ( p s )10
Maximum iterations (T)100
Orca Predator Algorithm
Explorative and exploitative behaviors (p) 0.5
Influence of leading orcas on the group’s movement (q) 0.75
Attraction force (F)2
Number of orcas ( p s )10
Maximum iterations (T)100
Deep Q-Learning
Action size40
Neurons per layer20
ActivationReLU (layers), Linear (final layer)
Loss functionHuber
OptimizerRMSprop with a learning rate of 0.001
Epsilon-greedyStarts at 1.0, decays to 0.01
Network updateEvery 50 training steps
Platform details
Operating systemmacOS 14.2.1 Darwin Kernel v23
Programming languagePython 3.10
Hardware specificationsUltra M2 chip, 64 GB RAM
Table 3. Comparison between improved biomimetic algorithms against their native versions. Instances from 1 to 8.
Table 3. Comparison between improved biomimetic algorithms against their native versions. Instances from 1 to 8.
InstancesMetricsNative AlgorithmsImproved Algorithms
PSOBATGWOOPAPSODQLBATDQLGWODQLOPADQL
1Best950950950950950950950950
Worst950950950950950950950950
Mean950950950950950950950950
Std00000000
Median950950950950950950950950
Iqr00000000
2Best773773773773773773773773
Worst773773773773773773773773
Mean773773773773773773773773
Std00000000
Median773773773773773773773773
Iqr00000000
3Best822822822822822822822822
Worst822822822822822822822822
Mean822822822822822822822822
Std00000000
Median822822822822822822822822
Iqr00000000
4Best872872872872872872872872
Worst872872872872872872872872
Mean872872872872872872872872
Std00000000
Median872872872872872872872872
Iqr00000000
>5Best12151215121512151215121512151215
Worst12151253121512151215121512151215
Mean12151244.2121512151215121512151215
Std014.1000000
Median12150121512151215121512151215
Iqr025000000
6Best12351235123512351235123512351235
Worst13181397123512351235131812351235
Mean1238.501282.801235123512351238.8712351235
Std15.2763.0200015.3200
Median12351235123512351235123512351235
Iqr093000000
7Best12871287128712871287128712871287
Worst12871287128712871287128712871287
Mean12871287128712871287128712871287
Std00000000
Median12871287128712871287128712871287
Iqr00000000
8Best12691269126912691269126912691269
Worst12841303126912691269126912691269
Mean1270.301278.33126912691269126912691269
Std3.9914.47000000
Median12691269126912691269126912691269
Iqr019.75000000
Table 4. Comparison between improved biomimetic algorithms against their native versions. Instances from 9 to 16.
Table 4. Comparison between improved biomimetic algorithms against their native versions. Instances from 9 to 16.
InstancesMetricsNative AlgorithmsImproved Algorithms
PSOBATGWOOPAPSODQLBATDQLGWODQLOPADQL
9Best13031303130313031303130313031303
Worst13051303130313031303130313031303
Mean1303.071303130313031303130313031303
Std0.370000000
Median13031303130313031303130313031303
Iqr00000000
10Best15361536153615361536153615361535
Worst16361737159215961547167915961547
Mean1560.731584.531543.071551.101536.371569.571548.771536.70
Std27.5456.9314.8721.862.0135.2019.792.81
Median1547154715361536153615641541.501536
Iqr4592.751145056110
11Best15931593159315931593159315931593
Worst16871690160716501593168116411599
Mean1606.171607.171593.871597.2015931600.901595.401593.20
Std23.7630.522.9113.59019.568.861.10
Median15961593159315931593159315931593
Iqr146000600
12Best16081608160816081608160816081608
Worst16891703164216581642168916421615
Mean16291628.8716111616.671611.631633.0316111608
Std21.7326.636.6414.6010.3725.236.641.78
Median16151615161116081608164216081608
Iqr40407704270
13Best15301530153015301530153015301530
Worst16321626153715681531163315661537
Mean1547.471548.701530.931537.531530.031545.531535.131530.37
Std27.2128.312.1511.360.1827.368.181.30
Median15351535153015351530153115331530
Iqr35.25361.00702970
14Best14491449144914491449144914491449
Worst15881609150715491497155915401508
Mean1498.431495.031462.201486.431452.271490.531488.571452.30
Std45.6446.6721.5135.079.2539.6233.3111.07
Median14971478144914971449149714971449
Iqr90822058089700
15Best19101920191019201910191019101910
Worst20892105201820622020202021802018
Mean1997.971984.871972.101989.971931.271980.831985.501956.30
Std46.7256.3332.6047.7534.6732.7854.7232.59
Median20081965196619991915199919801956
Iqr49.259754.2573.5021.503854.7553.75
16Best19941994195519551955195519551955
Worst21122112203720872001208720492012
Mean2028.272028.272005.232009.501982.032025.302006.131986.60
Std37.2837.2816.3225.4519.5030.6718.2819.73
Median20052005200120051994201220011996
Iqr56.505712.252041511242
Table 5. Comparison between improved biomimetic algorithms against their native versions. Instances from 17 to 24.
Table 5. Comparison between improved biomimetic algorithms against their native versions. Instances from 17 to 24.
InstancesMetricsNative AlgorithmsImproved Algorithms
PSOBATGWOOPAPSODQLBATDQLGWODQLOPADQL
17Best675651683749461698781609
Worst120312561089101191112841098878
Mean971.831054.53890.77893.57744.37948.27950.03723.30
Std126.54144.8597.8361.44100.04125.3577.9760.57
Median9691075.50897.50901.50759929958724
Iqr175.75150159.5074.75170.50182.25139.5098
18Best18321801183218011801180118321801
Worst20052044195019581889200919491936
Mean1915.901930.031881.801909.271837.601938.901897.131853.30
Std47.2860.5344.6442.1620.7950.1338.8031.52
Median1918.50193718871927183519451898.501849
Iqr51.5077.258660.258.752686.5025.25
19Best19351930193019301905193019301930
Worst20742075202420691978204220362032
Mean1984.131979.431962.271976.071936.131981.301960.231947.33
Std38.1938.9025.7031.5114.5537.8726.7121.64
Median198119781962.271978.5019351979.5019471942
Iqr80.5051.504942128242.7512
20Best23372331233423392293233723342293
Worst24702507245024592426253224372428
Mean2413.332416.102383.202408.902364.772410.202390.632367.38
Std33.9247.4835.0230.3430.6041.5827.8728.53
Median2419.50242323742413.502366241623842365
Iqr50.5076.7568.2546.2536.7555.254138.25
21Best973978915868702805896745
Worst1419164812771289980145513211196
Mean1161.901303.631114.031070.77829.531119.201196.101061.37
Std86.07173.08100103.8273.28156.6491.1793.70
Median116112911108.501081.5082911021207.501076
Iqr101312.75124172.25110154105119
22Best24002423234924232323240023842349
Worst25962557252025382473252425292467
Mean2484.102486.432462.1324782372.202450.772468.932427.70
Std46.4336.8234.0929.0843.3933.1134.7028.82
Median24772480.502456.5024772357.5024472464.502431.50
Iqr5869.7543.5041.25845563.2524.50
23Best22952319232323232244229523262297
Worst24782489243024692390247424362399
Mean2390.502398.032372.032393.8023222400.832383.502349.60
Std45.8451.3336.5430.2934.6642.4132.6627.03
Median23912390.502373.50239423282410.502390.502338
Iqr7893.75703937.7555.2557.2548.25
24Best22322279222822482193227522382238
Worst24392502238624912337243424232411
Mean2354.502391.402317.872352.632269.772363.1323502315.57
Std52.0752.944254.6942.3951.7541.0836.85
Median2363.50239723272342227923682342.502320
Iqr62.259957.755276925039.50
Table 6. Comparison between improved biomimetic algorithms against their native versions. Instances from 25 to 32.
Table 6. Comparison between improved biomimetic algorithms against their native versions. Instances from 25 to 32.
InstancesMetricsNative AlgorithmsImproved Algorithms
PSOBATGWOOPAPSODQLBATDQLGWODQLOPADQL
25Best28262872282128172782279128422805
Worst30083176298429752922297629802946
Mean2931.902964.932910.472921.672863.032872.332919.372887.87
Std44.7855.3336.0443.4435.7843.9732.8134.57
Median2936.5029592915.5029312870287429242887.50
Iqr51.755752.5064.5063.256731.7546.50
26Best30223018305630222968301630243037
Worst32083220315231443106317231553141
Mean3123.803133.203112.173101.273048.203114.203081.703083.27
Std44.7051.7624.7233.4634.1331.6432.3629.32
Median312131403112.5031093048.503119.5030833081
Iqr72.5078.753549.7537.75364445.50
27Best27242721271627752711271427142717
Worst29483052288428862848289128892851
Mean2840.902882.732803.472841.562765.872827.272813.272787.63
Std55.3281.3051.3329.3341.0748.7739.2832.25
Median28422858.502811.5028512767.502840.502821.502794
Iqr64.259675.7533.2578.756461.2532.25
28Best27762790281727522714272727782717
Worst30243037296030052915304529632911
Mean2911.202949.432894.972935.372825.772905.102883.172827.50
Std52.2265.1242.5953.774879.3148.0450.01
Median2919.502954.502903.502945.5028312894.502895.502832
Iqr73.7590.7563.7566.5067.25114.5073.5059
29Best28502870285928592770285928202813
Worst30513108300731512913306030262973
Mean29702968.832935.232984.432868.072965.232947.902914.27
Std47.166346.5870.0546.7251.6444.5937.74
Median29662957.5029532971.5028622971.5029562910
Iqr49.7592.5065.759753.5083.5075.7543.50
30Best33143310323133183218325732473258
Worst34873579345735033367347134703409
Mean3403.633438.473368.173417.773293.033388.1733803344.23
Std51.3863.3850.0443.6145.1252.4254.3943.35
Median3417343533683413328534.1333803356.50
Iqr9610558.755874.258170.2562
31Best31793224310932743044311931333116
Worst33993447332834753232335733193341
Mean3281.733325.803226.703375.373160.533262.173246.503258.23
Std58.1069.5260.4159.7551.7368.1247.3957.50
Median32763307.5032303380.50316232613250.503272
Iqr73.50105103.5010261.5096.5071.2564
32Best30343062297330632946305730262974
Worst32763317322033293145328332183206
Mean3185.633168.503138.903200.733071.103177.273148.173113.77
Std57.7262.2259.7765.9353.0659.7844.1561.49
Median3195.503167.50315232003074.503182.5031453129.50
Iqr88.5010982.2585651026271.50
Table 7. Comparison between improved biomimetic algorithms against their native versions. Instances from 33 to 40.
Table 7. Comparison between improved biomimetic algorithms against their native versions. Instances from 33 to 40.
InstancesMetricsNative AlgorithmsImproved Algorithms
PSOBATGWOOPAPSODQLBATDQLGWODQLOPADQL
33Best29042897289029382838283028962838
Worst31193188307631053017316731123017
Mean3019.313020.732980.833034.272932.402984.073014.402932.40
Std58.9574.0049.4747.7341.5858.7357.0341.58
Median30233008.502983.503034.502944298530162944
Iqr91.50123.255973565582.7556
34Best30193035293929792929293829892918
Worst31833357312533703102320331573108
Mean3091.603153.333065.603107.802999.933074.673076.633032
Std44.1774.8943.8774.7241.2258.8344.2150.01
Median309931343070.5030993008.503067.5030873028.50
Iqr7814473.505638.2567.2576.5056.25
35Best32093317321932173110326032513160
Worst34963647349734633328353834513437
Mean3394.873470.573343.533359.833236.833386.903350.433301.43
Std77.4291.8057.9069.6560.8285.7152.2867.38
Median34073474.503338.503387.503247.50337633543309.50
Iqr105.2515973.75119.75118160.7583.7592.75
36Best34753537341634483331350934383456
Worst37253743366336543602372836493625
Mean35923640.403570.203564.373480.733597.033567.603545.17
Std68.6661.4447.4149.9264.0157.9249.7543.33
Median358336433572.503559.503477.5035863573.503556.50
Iqr9285.505573.5070.5097.257250.50
37Best35163478347134633408350934623473
Worst36943752365336793581372236753615
Mean3604.633621.403569.673600.673511.903630.133595.973553.27
Std48.0173.9743.7653.0147.4558.4147.4338.00
Median36053624.503574.50361235193640.5036013566
Iqr82.75111.5062.2563.257898.2556.2559.75
38Best32483307310932563114321131343211
Worst35043703344534913331355934163392
Mean3372.873443.803321.233371.873208.333367.833334.703302.03
Std65.7490.0168.0368.5057.9696.9871.8147.40
Median3370342933313355.50321033653351.503310
Iqr999673105.5081.50146.25100.2576
39Best35193434345235023346347234953449
Worst37383811368036903590375136693635
Mean3613.673617.333592.203617.933495.603642.903579.273543.67
Std63.4495.1145.8254.3558.8469.7847.0544.80
Median36093623359836343512.503654.5035753546.50
Iqr93.2514146.7577.5085.2598.5084.2560
40Best38133940385039193758382638423820
Worst41714203409442253999404741064065
Mean4025.904052.073993.174031.603886.573963.473982.403972.60
Std74.3171.9252.1162.1759.7259.8961.0261.75
Median4025.5040463991402239064001.503976.503982
Iqr123.75104.7562.25829610778.7576
Table 8. p-values obtained from Wilcoxon–Mann–Whitney Test.
Table 8. p-values obtained from Wilcoxon–Mann–Whitney Test.
InstancesPSO
v/s
PSODQL
PSODQL
v/s
PSO
BAT
v/s
BATDQL
BATDQL
v/s
BAT
GWO
v/s
GWODQL
GWODQL
v/s
GWO
OPA
v/s
OPADQL
OPADQL
v/s
OPA
15 1.4 × 10 12 1.5 × 10 3
16 6.5 × 10 15 1.5 × 10 12
17 4.2 × 10 16 7.1 × 10 4 1.1 × 10 2 2.4 × 10 15
18 4.8 × 10 16 1.1 × 10 13
19 3.7 × 10 15 4.6 × 10 12
20 1.8 × 10 13 2.6 × 10 13
21 4.2 × 10 15 4.5 × 10 3 7.1 × 10 4
22 4.1 × 10 17 3.5 × 10 4 1.8 × 10 15
23 2.3 × 10 14 7.1 × 10 14
24 7.6 × 10 15 3.9 × 10 2 2.3 × 10 3 1.1 × 10 3
25 5.7 × 10 15 1.2 × 10 15 5.1 × 10 4
26 1.9 × 10 15 3.6 × 10 2 1.1 × 10 12 8.6 × 10 3
27 4.4 × 10 13 3.5 × 10 3 6.7 × 10 15
28 7.7 × 10 15 9.1 × 10 3 3.1 × 10 16
29 2.3 × 10 16 3.1 × 10 13
30 2.4 × 10 16 2.7 × 10 3 7.6 × 10 14
31 1.7 × 10 16 8.4 × 10 4 3.2 × 10 14
32 1.1 × 10 15 9.4 × 10 12 4.6 × 10 13
33 2.1 × 10 14 3.1 × 10 12 7.7 × 10 3 6.6 × 10 12
34 1.7 × 10 16 9.3 × 10 13 7.4 × 10 13
35 4.6 × 10 16 8.6 × 10 4 9.1 × 10 4
36 8.6 × 10 14 5.1 × 10 3
37 9.6 × 10 15 1.2 × 10 2 4.1 × 10 12
38 1.9 × 10 17 2.1 × 10 3 1.6 × 10 4
39 3.7 × 10 16 3.7 × 10 13
40 1.9 × 10 16 2.1 × 10 12 2.5 × 10 10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Olivares, R.; Salinas, O.; Ravelo, C.; Soto, R.; Crawford, B. Enhancing the Efficiency of a Cybersecurity Operations Center Using Biomimetic Algorithms Empowered by Deep Q-Learning. Biomimetics 2024, 9, 307. https://doi.org/10.3390/biomimetics9060307

AMA Style

Olivares R, Salinas O, Ravelo C, Soto R, Crawford B. Enhancing the Efficiency of a Cybersecurity Operations Center Using Biomimetic Algorithms Empowered by Deep Q-Learning. Biomimetics. 2024; 9(6):307. https://doi.org/10.3390/biomimetics9060307

Chicago/Turabian Style

Olivares, Rodrigo, Omar Salinas, Camilo Ravelo, Ricardo Soto, and Broderick Crawford. 2024. "Enhancing the Efficiency of a Cybersecurity Operations Center Using Biomimetic Algorithms Empowered by Deep Q-Learning" Biomimetics 9, no. 6: 307. https://doi.org/10.3390/biomimetics9060307

Article Metrics

Back to TopTop