Improving Safety Awareness Campaigns Through the Use of Graph Neural Networks
Abstract
:1. Introduction
2. Network Communities
- Ring of cliques: This is a set of cliques of the same size connected to each other forming a ring. That is, each clique is connected to two cliques with a single link. To generate them, it is enough to detail only the number of cliques (n) and the size of them .
- Relaxed caveman graphs: These are formed by rewriting the links of a ring of cliques with a certain probability. To generate them, it is necessary to detail the number of cliques (n), the size of them (k) and the probability of rewriting a link (p).
- Planted partition graphs: This is a set of groups of the same size that are highly connected inside and to a low degree connected with other groups. To generate them, it is necessary to detail the number of groups (n), the size of them (k), the probability of forming links within each group and the probability of forming links outside the group .
- Random partition graphs: These can be considered as planted partition graphs where the groups are of different sizes. To generate them, it is necessary to detail the size of each of the groups , the probability of forming links within each group and the probability of forming links outside the group .
- Gaussian random partition graphs: These can also be considered as planted partition graphs where the size of the groups has a certain mean and variance. To generate them, it is necessary to detail the number of nodes , the mean of the group , a parameter that determines the variance of the size of the groups , the probability of forming links within each group and the probability of forming links outside the group .
- Stochastic Block Models: These can be considered as random partition graphs where it is also necessary to detail the link probabilities with each of the groups (including oneself). To generate them, it is necessary to detail the size of each of the groups and the probability of forming links of a group i with a group j with .
3. Belief-Propagation Models
3.1. Sznajd Model Variant
- A node and a neighbor of this node are chosen at random.
- Based on their opinions, the opinions of neighboring nodes are modified:
- If , then the neighbors of and , and become of the same opinion as these two nodes:
- If then no opinion changes.
- Steps 1 and 2 are repeated for a number of iterations.
3.2. Hegselmann–Krause Model
- For each node , we obtain the group of neighbors that share an opinion close to its:
- The opinion of each node is updated as the average of the opinions of this group:
- Steps 1 and 2 are repeated for a number of iterations or until the beliefs converge.
3.3. Model Convergence
4. Graph Neural Networks
4.1. Layers
- GraphSAGE [24,25]: In this article, we use the variant of the original GraphSAGE layer that pytorch goemetric (a library Python) defined. Given that the representation of a node is and the set of neighbors of node is , the representation of each node is updated as follows:
- Linear [26]: Given that the input dataset is x, we can obtain the output set as follows:
4.2. Functions
- Sigmoid: If the input dataset is x, the output dataset is as follows:It should be noted that these last two functions have no parameters for learning.
4.3. Loss Functions
- Binary cross entropy with logits: Considering that is the logit according to the model associated with node and is the label of node , we have that:
- Mean squared error: Considering that is the predicted value according to the model for node and is the label of node , we have that:
4.4. Optimizer
- Adam [28]: This algorithm takes into account the initial parameters, , , , , and the loss function with learning parameters at time , . Then, the learning parameters are updated to minimize the loss function following the following algorithm:
5. The Mathematical Problem
- The security awareness of each person can change over time based on the security awareness of their peers. In this article, the relationships of an individual i will be represented with their peers through a graph where V is the set of vertices and E is the set of edges. In this way, each person i will be represented by a node and the person’s security awareness i will be influenced by the neighboring nodes of the node , :
- It will be considered that the Markov property holds:
- Taking into account the two previous hypotheses, it will also be considered that the propagation of security awareness evolves according to the Sznajd model variant, the Hegselmann–Krause model or a combination of both:
6. Belief-Propagation Simulation
6.1. Social Network
6.2. Belief-Propagation Models
- Initially, all nodes are considered to have high awareness: for all .
- In the following, we consider the three largest communities where the majority of nodes () have low awareness, . In Sznajd’s model, the nodes with low awareness verify for all , while in the Hegselmann–Krause model is a random number belonging to the interval for all .
- Subsequently, we consider that there are some nodes with low awareness, , uniformly distributed throughout the network. In the Sznajd model, these nodes verify for all . In the Hegselmann–Krause model, it is verified that is a random number in the interval for all .
- Next, three communities are considered where most of the nodes (), , have high awareness. For the Sznajd model, for all and for the Hegselmann–Krause model is a random number in the interval for all .
- Finally, some nodes are considered with a medium awareness in the Hegselmann–Krause model (). Then, each node of these nodes verifies that is a random number in the interval .
6.3. Graph Neural Networks
- Prediction of the Sznajd model: As the opinions of each of the nodes is 0 or 1, binary cross entropy with logits (BCEWL) was used as a loss function. Adam with learning rate 0.001 was used as the optimizer. It was considered 100 epochs. Depending on the random assignment of initial beliefs of the nodes, different results were obtained. An accuracy between and was obtained in the test set. Figure A9 shows an example of the confusion matrix.
- Prediction of the Hegselmann–Krause model: As the opinion of each node is in the interval , the Mean Squared Error (MSE) was used as a loss function. The Adam optimizer with learning rate 0.001 was also used in this case. It was considered 300 epochs. We obtained an error between and in the test set. The mean absolute error is between and in the test set.
- Prediction of a combination of the Sznajd and Hegselmann–Krause models: In this case, the following combination of the two models was considered:
7. Application in Companies
- If you only have one security-awareness value, then you can apply Sznajd’s model and the Hegselmann–Krause model. These models are able to predict the evolution of these beliefs through a specific algorithm. Therefore, the main limitation of these models is that it only makes sense to apply them assuming that the beliefs evolve according to these algorithms. This means that these models do not always resemble the actual evolution of beliefs.
- If you have more than a single measure of safety awareness, you can use GNNs. The main advantage of these models is that they learn how security awareness evolves from the data and not by following a specific algorithm. In fact, as seen in the previous section, GNNs are able to learn the way of evolution from the two previous models (even a combination of both) with a considerably low error. Therefore, we consider that these types of deep learning models are more flexible and can better adapt to the actual evolution of the security awareness. In addition, these models can take into account characteristics of individuals that may influence propagation (age, position, etc.). However, these models would not be effective when people’s beliefs do not depend on their environment or it is excessively complicated to capture the effect that one person’s belief has on another. Depending on the information given, we propose the use of two different types of GNNs:
- If there are few values over time, a simple model similar to the one used in this article can be used.
- If there are many values over time, temporal GNNs can be used.
- A few randomly distributed nodes have medium future security awareness. To overcome this, it is sufficient for the people representing these nodes to learn from their colleagues. Then, this problem can be solved by holding meetings and group activities/games to exchange opinions about security with their colleagues.
- A few randomly distributed nodes have low future security awareness. For these people, it is recommended that they learn about the current state of malware. A simple way to overcome this is by taking individual security courses. These courses have to be effective and include practical activities.
- There are groups with low future security awareness. In this case, several security-learning talks and group activities can be conducted for groups with these weaknesses. Security experts must organize learning. In this way, they can learn about current malware threats.
8. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Appendix A
References
- Zwilling, M.; Klien, G.; Lesjak, D.; Åukasz, W.; Cetin, F.; Basim, H.N. Cyber Security Awareness, Knowledge and Behavior: A Comparative Study. J. Comput. Inf. Syst. 2022, 62, 82–97. [Google Scholar] [CrossRef]
- Alkhalil, Z.; Hewage, C.; Nawaf, L.; Khan, I. Phishing attacks: A recent comprehensive study and a new anatomy. Front. Comput. Sci. 2021, 3, 563060. [Google Scholar] [CrossRef]
- Bada, M.; Sasse, A.M.; Nurse, J.R.C. Cyber Security Awareness Campaigns: Why do they fail to change behaviour? arXiv 2019, arXiv:1901.02672. [Google Scholar]
- Siponen, M.T. A conceptual foundation for organizational information security awareness. Inf. Manag. Comput. Secur. 2000, 8, 31–41. [Google Scholar] [CrossRef]
- Kruger, H.; Kearney, W. A prototype for assessing information security awareness. Comput. Secur. 2006, 25, 289–296. [Google Scholar] [CrossRef]
- Hart, S.; Margheri, A.; Paci, F.; Sassone, V. Riskio: A Serious Game for Cyber Security Awareness and Education. Comput. Secur. 2020, 95, 101827. [Google Scholar] [CrossRef]
- Sharif, K.H.; Ameen, S.Y. A Review of Security Awareness Approaches With Special Emphasis on Gamification. In Proceedings of the 2020 International Conference on Advanced Science and Engineering (ICOASE), Duhok, Iraq, 23–24 December 2020; pp. 151–156. [Google Scholar] [CrossRef]
- Peltier, T.R. Implementing an Information Security Awareness Program. Inf. Syst. Secur. 2005, 14, 37–49. [Google Scholar] [CrossRef]
- Hansche, S. Designing a Security Awareness Program: Part 1. Inf. Syst. Secur. 2001, 9, 1–9. [Google Scholar] [CrossRef]
- Rezgui, Y.; Marks, A. Information security awareness in higher education: An exploratory study. Comput. Secur. 2008, 27, 241–253. [Google Scholar] [CrossRef]
- McCormac, A.; Zwaans, T.; Parsons, K.; Calic, D.; Butavicius, M.; Pattinson, M. Individual differences and Information Security Awareness. Comput. Hum. Behav. 2017, 69, 151–156. [Google Scholar] [CrossRef]
- Stefaniuk, T. Training in shaping employee information security awareness. Entrep. Sustain. Issues 2020, 7, 1832. [Google Scholar] [CrossRef] [PubMed]
- Spears, R. Social Influence and Group Identity. Annu. Rev. Psychol. 2021, 72, 367–390. [Google Scholar] [CrossRef] [PubMed]
- Hollebeek, L.D.; Sprott, D.E.; Sigurdsson, V.; Clark, M.K. Social influence and stakeholder engagement behavior conformity, compliance, and reactance. Psychol. Mark. 2022, 39, 90–100. [Google Scholar] [CrossRef]
- Hu, X.; Chen, X.; Davison, R.M. Social Support, Source Credibility, Social Influence, and Impulsive Purchase Behavior in Social Commerce. Int. J. Electron. Commer. 2019, 23, 297–327. [Google Scholar] [CrossRef]
- Sznajd-Weron, K. Sznajd model and its applications. Acta Phys. Pol. Ser. 2005, 36, 2537–2547. [Google Scholar]
- Douven, I.; Riegler, A. Extending the Hegselmann–Krause Model I. Log. J. IGPL 2009, 18, 323–335. [Google Scholar] [CrossRef]
- Hagberg, A.A.; Schult, D.A.; Swart, P.J. Exploring Network Structure, Dynamics, and Function using NetworkX. In Proceedings of the 7th Python in Science Conference, Pasadena, CA, USA, 19–24 August 2008; Varoquaux, G., Vaught, T., Millman, J., Eds.; pp. 11–15. [Google Scholar] [CrossRef]
- Sznajd-Weron, K.; Sznajd, J.; Weron, T. A review on the Sznajd model—20 years after. Phys. A Stat. Mech. Its Appl. 2021, 565, 125537. [Google Scholar] [CrossRef]
- Rossetti, G.; Milli, L.; Rinzivillo, S.; Sîrbu, A.; Pedreschi, D.; Giannotti, F. NDlib: A python library to model and analyze diffusion processes over complex networks. Int. J. Data Sci. Anal. 2017, 5, 61–79. [Google Scholar] [CrossRef]
- Rossetti, G.; Milli, L.; Rinzivillo, S.; Sirbu, A.; Pedreschi, D.; Giannotti, F. NDlib: Studying Network Diffusion Dynamics. In Proceedings of the 2017 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Tokyo, Japan, 19–21 October 2017; pp. 155–164. [Google Scholar] [CrossRef]
- Hegselmann, R.; Krause, U. Opinion Dynamics and Bounded Confidence Models, Analysis and Simulation. J. Artif. Soc. Soc. Simul. 2002, 5. [Google Scholar]
- Lanchier, N.; Li, H.L. Consensus in the Hegselmann–Krause Model. J. Stat. Phys. 2022, 187, 20. [Google Scholar] [CrossRef]
- Hamilton, W.L.; Ying, R.; Leskovec, J. Inductive Representation Learning on Large Graphs. arXiv 2018, arXiv:1706.02216. [Google Scholar]
- Fey, M.; Lenssen, J.E. Fast Graph Representation Learning with PyTorch Geometric. In Proceedings of the ICLR Workshop on Representation Learning on Graphs and Manifolds, 2019. Available online: https://arxiv.org/abs/1903.02428 (accessed on 23 April 2025).
- Salas, R. Redes neuronales artificiales. Univ. Valparaıso Dep. Computación 2004, 1, 1–7. [Google Scholar]
- He, J.; Lin Li, J.X.; Zheng, C. Relu Deep Neural Networks and Linear Finite Elements. J. Comput. Math. 2020, 38, 502–527. [Google Scholar] [CrossRef]
- Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2017, arXiv:1412.6980. [Google Scholar]
- Yang, S.; Yu, X.; Zhou, Y. LSTM and GRU Neural Network Performance Comparison Study: Taking Yelp Review Dataset as an Example. In Proceedings of the 2020 International Workshop on Electronic Communication and Artificial Intelligence (IWECAI), Shanghai, China, 12–14 June 2020; pp. 98–101. [Google Scholar] [CrossRef]
- Masuko, T. Computational cost reduction of long short-term memory based on simultaneous compression of input and hidden state. In Proceedings of the 2017 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), Okinawa, Japan, 16–20 December 2017; pp. 126–133. [Google Scholar] [CrossRef]
- Seo, Y.; Defferrard, M.; Vandergheynst, P.; Bresson, X. Structured Sequence Modeling with Graph Convolutional Recurrent Networks. arXiv 2016, arXiv:1612.07659. [Google Scholar]
- Chen, J.; Wang, X.; Xu, X. GC-LSTM: Graph Convolution Embedded LSTM for Dynamic Link Prediction. arXiv 2021, arXiv:1812.04206. [Google Scholar] [CrossRef]
- Leskovec, J.; Krevl, A. SNAP Datasets: Stanford Large Network Dataset Collection. 2014. Available online: http://snap.stanford.edu/data (accessed on 23 April 2025).
Case 1 | Case 2 | Case 3 | |
---|---|---|---|
Neurons (1° layer) | 32 | 32 | 32 |
Neurons (2° layer) | 32 | 32 | 32 |
Neurons (3° layer) | 32 | 32 | 32 |
Neurons (4° layer) | 1 | 1 | 1 |
Learning rate | 0.001 | 0.001 | 0.001 |
Epochs | 100 | 300 | 300 |
Loss function | BCEWL | MSE | MSE |
Activation function | ReLU | ReLU | ReLU |
Last activation function | - | Sigmoid | Sigmoid |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Guillén, J.D.H.; del Rey, A.M. Improving Safety Awareness Campaigns Through the Use of Graph Neural Networks. Axioms 2025, 14, 328. https://doi.org/10.3390/axioms14050328
Guillén JDH, del Rey AM. Improving Safety Awareness Campaigns Through the Use of Graph Neural Networks. Axioms. 2025; 14(5):328. https://doi.org/10.3390/axioms14050328
Chicago/Turabian StyleGuillén, Jose D. Hernández, and Angel Martín del Rey. 2025. "Improving Safety Awareness Campaigns Through the Use of Graph Neural Networks" Axioms 14, no. 5: 328. https://doi.org/10.3390/axioms14050328
APA StyleGuillén, J. D. H., & del Rey, A. M. (2025). Improving Safety Awareness Campaigns Through the Use of Graph Neural Networks. Axioms, 14(5), 328. https://doi.org/10.3390/axioms14050328