How to Conduct AI-Assisted (Large Language Model-Assisted) Content Analysis in Information Science and Cyber Security Research
Abstract
1. Introduction
1.1. Rationale
1.2. Differences Between CA and Thematic Analysis (TA) and Discourse Analysis (DA)
1.3. Current Study
2. Defining Content Analysis
Similarities and Differences Between Content, Thematic, and Discourse Analyses
3. The Utility of CA for Information Science and Cyber Security Research
3.1. Steps Involved in a Traditional CA
3.1.1. Reliability and Validity
3.1.2. Research Questions, Coding Frames and Unit of Analysis
3.1.3. Coding Scheme
3.1.4. Analysis and Interpretation
4. Employing AI-Assisted CA
4.1. Responsible AI
4.2. Hallucination Phenomenon and Error Mitigation
4.2.1. Verification Playbook
4.2.2. Errors Related to Cyber Security and Information Science
4.3. Hybrid Approach
5. Stages Recommended for an AI-Assisted CA
5.1. Deductive AI-Assisted CA
5.1.1. Research Questions, Population Texts, Samples and Units of Analysis
5.1.2. Coding and Reliability
5.1.3. Synthesising or Quantifying and Interpreting
5.2. Inductive AI-Assisted CA
5.2.1. Immersion, Code Development and Definitions
5.2.2. Developing a Codebook
5.2.3. Reliability and Reflexivity
5.3. Practical Implications
5.4. Apply in Information Science and Cyber Security Research
5.5. Limitations
6. Practical Example
7. Conclusions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Jiayimei, W.; Tao, N.; Wei-Bin, L.; Qingchuan, Z. A Contemporary Survey of Large Language Model Assisted Program Analysis. Trans. Artif. Intell. 2025, 1, 105–129. [Google Scholar] [CrossRef]
- Kharma, M.; Choi, S.; AlKhanafseh, M.; Mohaisen, D. Security and Quality in LLM-Generated Code: A Multi-Language, Multi-Model Analysis. arXiv 2025, arXiv:2502.01853. [Google Scholar]
- Du, X.; Du, M.; Zhou, Z.; Bai, Y. Facilitator or hindrance? The impact of AI on university students’ higher-order thinking skills in complex problem solving. Int. J. Educ. Technol. High. Educ. 2025, 22, 39. [Google Scholar] [CrossRef]
- Jie, A.L.X.; Kamrozzaman, N.A. The Challenges of Higher Education Students Face in Using Artificial Intelligence (AI) against Their Learning Experiences. Open J. Soc. Sci. 2024, 12, 362–387. [Google Scholar] [CrossRef]
- Ergene, O.; Ergene, B.C. AI ChatBots’ solutions to mathematical problems in interactive e-textbooks: Affordances and constraints from the eyes of students and teachers. Educ. Inf. Technol. 2025, 30, 509–545. [Google Scholar] [CrossRef]
- Sijing, L.; Lan, W. Artificial Intelligence Education Ethical Problems and Solutions. In Proceedings of the 2018 13th International Conference on Computer Science & Education (ICCSE), Colombo, Sri Lanka, 8–11 August 2018; pp. 1–5. [Google Scholar] [CrossRef]
- Chiu, T.K.F.; Moorhouse, B.L.; Chai, C.S.; Ismailov, M. Teacher support and student motivation to learn with Artificial Intelligence (AI) based chatbot. Interact. Learn. Environ. 2024, 32, 3240–3256. [Google Scholar] [CrossRef]
- Kim, J.; Lee, H.; Cho, Y.H. Learning design to support student-AI collaboration: Perspectives of leading teachers for AI in education. Educ. Inf. Technol. 2022, 27, 6069–6104. [Google Scholar] [CrossRef]
- Guo, K.; Zhang, E.D.; Li, D.; Yu, S. Using AI-supported peer review to enhance feedback literacy: An investigation of students’ revision of feedback on peers’ essays. Br. J. Educ. Technol. 2025, 56, 1612–1639. [Google Scholar] [CrossRef]
- Guo, K.; Pan, M.; Li, Y.; Lai, C. Effects of an AI-supported approach to peer feedback on university EFL students’ feedback quality and writing ability. Internet High. Educ. 2024, 63, 100962. [Google Scholar] [CrossRef]
- Whitfield, S.; Hofmann, M.A. Elicit: AI literature review research assistant. Public Serv. Q. 2023, 19, 201–207. [Google Scholar] [CrossRef]
- Al-Zahrani, A.M. The impact of generative AI tools on researchers and research: Implications for academia in higher education. Innov. Educ. Teach. Int. 2024, 61, 1029–1043. [Google Scholar] [CrossRef]
- Khalifa, M.; Albadawy, M. Using artificial intelligence in academic writing and research: An essential productivity tool. Comput. Methods Programs Biomed. Update 2024, 5, 100145. [Google Scholar] [CrossRef]
- Whitty, M.T. Leveraging Behaviour Sequence Analysis (BSA) in Information Systems: Examining malicious insider threat as an example. In Proceedings of the International Conferences IADIS Information Systems 2025 and e-Society 2025, Madeira Island, Portugal, 1–3 March 2025; pp. 517–521. [Google Scholar]
- Whitty, M.T. How the voice lost its voice: Applying the Dual Processing Theory to explain how mis/disinformation can deceive and persuade voting decision-making. J. Inf. Warf. 2025, 24, 93–108. [Google Scholar]
- Whitty, M.T.; Abdulgalimov, D.; Oliver, P.; Ruddy, C.; Seguin, J.; Young, G. Inside the Threat Matrix: Using Hybrid Computer Simulations to Educate Adults on Malicious Insider Threat and Technology Misuse. In HCI for Cybersecurity, Privacy and Trust; Moallem, A., Ed.; Springer Nature: Cham, Switzerland, 2025; pp. 298–312. [Google Scholar] [CrossRef]
- Whitty, M.T.; Ruddy, C.; Keatley, D.; Butavicius, M.; Grobler, M. The prince of insiders: A multiple pathway approach to understanding IP theft insider attacks. Inf. Comput. Secur. 2024, 32, 509–522. [Google Scholar] [CrossRef]
- Valizadeh, A.; Moassefi, M.; Nakhostin-Ansari, A.; Hosseini Asl, S.H.; Saghab Torbati, M.; Aghajani, R.; Maleki Ghorbani, Z.; Faghani, S. Abstract screening using the automated tool Rayyan: Results of effectiveness in three diagnostic test accuracy systematic reviews. BMC Med. Res. Methodol. 2022, 22, 160. [Google Scholar] [CrossRef] [PubMed]
- Gill, M.; Ng, J.; Szydlowski, N.; Fusco, N.; Ruiz, K. MSR95 Artificial Intelligence for Targeted Literature Review Screening within the Rayyan Platform. Value Health 2024, 27, S278. [Google Scholar] [CrossRef]
- Noy, S.; Zhang, W. Experimental evidence on the productivity effects of generative artificial intelligence. Science 2023, 381, 187–192. [Google Scholar] [CrossRef]
- Pack, A.; Barrett, A.; Escalante, J. Large language models and automated essay scoring of English language learner writing: Insights into validity and reliability. Comput. Educ. Artif. Intell. 2024, 6, 100234. [Google Scholar] [CrossRef]
- Neuendorf, K.A. The Content Analysis Guidebook; Sage: Los Angeles, CA, USA, 2017. [Google Scholar]
- Van Atteveldt, W.; Trilling, D.; Calderon, C.A. Computational Analysis of Communication; John Wiley & Sons: Hoboken, NJ, USA, 2022. [Google Scholar]
- Mingers, J. Combining IS research methods: Towards a pluralist methodology. Inf. Syst. Res. 2001, 12, 240–259. [Google Scholar] [CrossRef]
- Nicmanis, M.; Spurrier, H. Getting Started with Artificial Intelligence Assisted Qualitative Analysis: An Introductory Guide to Qualitative Research Approaches with Exploratory Examples from Reflexive Content Analysis. Int. J. Qual. Methods 2025, 24, 16094069251354863. [Google Scholar] [CrossRef]
- Krippendorff, K. Content Analysis: An Introduction to Its Methodology; Sage publications: Los Angeles, CA, USA, 2018. [Google Scholar]
- Gebru, T.; Morgenstern, J.; Vecchione, B.; Vaughan, J.W.; Wallach, H.; Daumé, H., III; Crawford, K. Datasheets for datasets. Commun. ACM 2021, 64, 86–92. [Google Scholar] [CrossRef]
- Whitty, M. Possible Selves: An Exploration of the Utility of a Narrative Approach. Identity 2002, 2, 211–228. [Google Scholar] [CrossRef]
- Badzinski, D.M.; Woods, R.H.; Nelson, C.M. Content analysis. In The Routledge Handbook of Research Methods in the Study of Religion; Routledge: Oxfordshire, UK, 2021; pp. 180–193. [Google Scholar]
- Baker, L.M.; King, A.E. Let’s get theoretical: A quantitative content analysis of theories and models used in the Journal of Applied Communications. J. Appl. Commun. 2016, 100, 5. [Google Scholar] [CrossRef]
- Braun, V.; Clarke, V. Supporting best practice in reflexive thematic analysis reporting in Palliative Medicine: A review of published research and introduction to the Reflexive Thematic Analysis Reporting Guidelines (RTARG). Palliat. Med. 2024, 38, 608–616. [Google Scholar] [CrossRef] [PubMed]
- Potter, J.; Wetherell, M. Discourse and social psychology. In The Discourse Studies Reader; John Benjamins Publishing Company: Amsterdam, The Netherlands, 2014; pp. 244–255. [Google Scholar]
- Wong, M.-F.; Guo, S.; Hang, C.-N.; Ho, S.-W.; Tan, C.-W. Natural Language Generation and Understanding of Big Code for AI-Assisted Programming: A Review. Entropy 2023, 25, 888. [Google Scholar] [CrossRef] [PubMed]
- Hassani, H.; Silva, E.S. The Role of ChatGPT in Data Science: How AI-Assisted Conversational Interfaces Are Revolutionizing the Field. Big Data Cogn. Comput. 2023, 7, 62. [Google Scholar] [CrossRef]
- Fazeli, S.; Sabetti, J.; Ferrari, M. Performing qualitative content analysis of video data in social sciences and medicine: The visual-verbal video analysis method. Int. J. Qual. Methods 2023, 22, 16094069231185452. [Google Scholar] [CrossRef]
- McCashin, D.; Murphy, C.M. Using TikTok for public and youth mental health—A systematic review and content analysis. Clin. Child Psychol. Psychiatry 2023, 28, 279–306. [Google Scholar] [CrossRef]
- Benjamin, S.; Bottone, E.; Lee, M. Beyond accessibility: Exploring the representation of people with disabilities in tourism promotional materials. In Justice and Tourism; Routledge: Oxfordshire, UK, 2021; pp. 153–171. [Google Scholar] [CrossRef]
- Drisko, J.W.; Maschi, T. Content Analysis; Oxford University Press: Oxford, UK, 2016. [Google Scholar]
- Riffe, D.; Lacy, S.; Watson, B.R.; Lovejoy, J. Analyzing Media Messages: Using Quantitative Content Analysis in Research; Routledge: Oxfordshire, UK, 2023. [Google Scholar]
- Chu, H. Research methods in library and information science: A content analysis. Libr. Inf. Sci. Res. 2015, 37, 36–41. [Google Scholar] [CrossRef]
- Choi, W.; Zhang, Y.; Stvilia, B. Exploring applications and user experience with generative AI tools: A content analysis of reddit posts on ChatGPT. Proc. Assoc. Inf. Sci. Technol. 2023, 60, 543–546. [Google Scholar] [CrossRef]
- Eriksson, J.; Giacomello, G. International relations, cybersecurity, and content analysis: A constructivist approach. In The Global Politics of Science and Technology-Vol. 2: Perspectives, Cases and Methods; Springer: Berlin/Heidelberg, Germany, 2014; pp. 205–219. [Google Scholar] [CrossRef]
- Walker, J.T. Mixed Method Content Analysis of the Cybersecurity Corpus of the US National Security System; Capitol Technology University: Laurel, MD, USA, 2024. [Google Scholar]
- Nicmanis, M. Reflexive Content Analysis: An Approach to Qualitative Data Analysis, Reduction, and Description. Int. J. Qual. Methods 2024, 23, 16094069241236603. [Google Scholar] [CrossRef]
- Vears, D.F.; Gillam, L. Inductive content analysis: A guide for beginning qualitative researchers. Focus Health Prof. Educ. Multi-Prof. J. 2022, 23, 111–127. [Google Scholar] [CrossRef]
- Russman, U.; Flick, U. Designing Qualitative Research for Working with Facebook Data. In The SAGE Handbook of Qualitative Research Design; Sage: Los Angeles, CA, USA, 2022; pp. 851–885. [Google Scholar]
- Harwood, T.G.; Garry, T. An overview of content analysis. Mark. Rev. 2003, 3, 479–498. [Google Scholar] [CrossRef]
- Bretl, D.J.; Cantor, J. The portrayal of men and women in US television commercials: A recent content analysis and trends over 15 years. Sex Roles 1988, 18, 595–609. [Google Scholar] [CrossRef]
- Stemler, S. An overview of content analysis. Pract. Assess. Res. Eval. 2000, 7, 17. [Google Scholar]
- Moretti, F.; van Vliet, L.; Bensing, J.; Deledda, G.; Mazzi, M.; Rimondini, M.; Zimmermann, C.; Fletcher, I. A standardized approach to qualitative content analysis of focus group discussions from different countries. Patient Educ. Couns. 2011, 82, 420–428. [Google Scholar] [CrossRef]
- De Wever, B.; Schellens, T.; Valcke, M.; Van Keer, H. Content analysis schemes to analyze transcripts of online asynchronous discussion groups: A review. Comput. Educ. 2006, 46, 6–28. [Google Scholar] [CrossRef]
- Oleinik, A. Mixing quantitative and qualitative content analysis: Triangulation at work. Qual. Quant. 2011, 45, 859–873. [Google Scholar] [CrossRef]
- Schreier, M. Qualitative Content Analysis in Practice; Sage Publications: Los Angeles, CA, USA, 2012. [Google Scholar]
- Forman, J.; Damschroder, L. Qualitative Content Analysis. In Empirical Methods for Bioethics: A Primer; Jacoby, L., Siminoff, L.A., Eds.; Emerald Group Publishing Limited: Leeds, UK, 2007; Volume 11, pp. 39–62. [Google Scholar]
- Hsieh, H.-F.; Shannon, S.E. Three Approaches to Qualitative Content Analysis. Qual. Health Res. 2005, 15, 1277–1288. [Google Scholar] [CrossRef] [PubMed]
- Mayring, P. Qualitative Content Analysis: Theoretical Background and Procedures. In Approaches to Qualitative Research in Mathematics Education: Examples of Methodology and Methods; Bikner-Ahsbahs, A., Knipping, C., Presmeg, N., Eds.; Springer: Amsterdam, The Netherlands, 2015; pp. 365–380. [Google Scholar]
- Kleinheksel, A.J.; Rockich-Winston, N.; Tawfik, H.; Wyatt, T.R. Demystifying Content Analysis. Am. J. Pharm. Educ. 2020, 84, 7113. [Google Scholar] [CrossRef]
- Dooley, K.J. Using manifest content analysis in purchasing and supply management research. J. Purch. Supply Manag. 2016, 22, 244–246. [Google Scholar] [CrossRef]
- Lee, J.-H.; Kim, Y.-G. A stage model of organizational knowledge management: A latent content analysis. Expert Syst. Appl. 2001, 20, 299–311. [Google Scholar] [CrossRef]
- Egberg Thyme, K.; Wiberg, B.; Lundman, B.; Graneheim, U.H. Qualitative content analysis in art psychotherapy research: Concepts, procedures, and measures to reveal the latent meaning in pictures and the words attached to the pictures. Arts Psychother. 2013, 40, 101–107. [Google Scholar] [CrossRef]
- Denzin, N.K. The new paradigm dialogs and qualitative inquiry. Int. J. Qual. Stud. Educ. 2008, 21, 315–325. [Google Scholar] [CrossRef]
- Sławecki, B. Paradigms in Qualitative Research. In Qualitative Methodologies in Organization Studies: Volume I: Theories and New Approaches; Ciesielska, M., Jemielniak, D., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 7–26. [Google Scholar]
- Jaspal, R. Content analysis, thematic analysis and discourse analysis. Res. Methods Psychol. 2020, 1, 285–312. [Google Scholar]
- Lytvyn, V.; Vysotska, V.; Peleshchak, I.; Basyuk, T.; Kovalchuk, V.; Kubinska, S.; Rusyn, B.; Pohreliuk, L.; Chyrun, L.; Salo, T. Identifying Textual Content Based on Thematic Analysis of Similar Texts in Big Data. In Proceedings of the 2019 IEEE 14th International Conference on Computer Sciences and Information Technologies (CSIT), Lviv, Ukraine, 17–20 September 2019; Volume 2, pp. 84–91. [Google Scholar] [CrossRef]
- Clarke, V.; Braun, V. Teaching thematic analysis: Overcoming challenges and developing strategies for effective learning. Psychologist 2013, 26, 120–123. [Google Scholar]
- Braun, V.; Clarke, V. Reflecting on reflexive thematic analysis. Qual. Res. Sport Exerc. Health 2019, 11, 589–597. [Google Scholar] [CrossRef]
- Ekstrom, J.J.; Lunt, B.M.; Parrish, A.; Raj, R.K.; Sobiesk, E. Information Technology as a Cyber Science. In Proceedings of the 18th Annual Conference on Information Technology Education, Rochester, NY, USA, 4–7 October 2017. [Google Scholar]
- Kaur, J.; Ramkumar, K.R. The recent trends in cyber security: A review. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34 Pt B, 5766–5781. [Google Scholar] [CrossRef]
- von Solms, R.; van Niekerk, J. From information security to cyber security. Comput. Secur. 2013, 38, 97–102. [Google Scholar] [CrossRef]
- Julien, H.; Pecoskie, J.; Reed, K. Trends in information behavior research, 1999–2008: A content analysis. Libr. Inf. Sci. Res. 2011, 33, 19–24. [Google Scholar] [CrossRef]
- Romanosky, S.; Ablon, L.; Kuehn, A.; Jones, T. Content analysis of cyber insurance policies: How do carriers price cyber risk? J. Cybersecur. 2019, 5, tyz002. [Google Scholar] [CrossRef]
- Blythe, J.M.; Sombatruang, N.; Johnson, S.D. What security features and crime prevention advice is communicated in consumer IoT device manuals and support pages? J. Cybersecur. 2019, 5, tyz005. [Google Scholar] [CrossRef]
- Boholm, M. Twenty-five years of cyber threats in the news: A study of Swedish newspaper coverage (1995–2019). J. Cybersecur. 2021, 7, tyab016. [Google Scholar] [CrossRef]
- Jackson, S.; Vanteeva, N.; Fearon, C. An investigation of the impact of data breach severity on the readability of mandatory data breach notification letters: Evidence from US firms. J. Assoc. Inf. Sci. Technol. 2019, 70, 1277–1289. [Google Scholar] [CrossRef]
- Morrow, E. Scamming higher ed: An analysis of phishing content and trends. Comput. Hum. Behav. 2024, 158, 108274. [Google Scholar] [CrossRef]
- Brown, S.A.; Upchurch, S.L.; Acton, G.J. A Framework for Developing a Coding Scheme for Meta-Analysis. West. J. Nurs. Res. 2003, 25, 205–222. [Google Scholar] [CrossRef] [PubMed]
- Lombard, M.; Snyder-Duch, J.; Bracken, C.C. Content Analysis in Mass Communication: Assessment and Reporting of Intercoder Reliability. Hum. Commun. Res. 2006, 28, 587–604. [Google Scholar] [CrossRef]
- Bojić, L.; Zagovora, O.; Zelenkauskaite, A.; Vuković, V.; Čabarkapa, M.; Veseljević Jerković, S.; Jovančević, A. Comparing large Language models and human annotators in latent content analysis of sentiment, political leaning, emotional intensity and sarcasm. Sci. Rep. 2025, 15, 11477. [Google Scholar] [CrossRef]
- Chew, R.; Bollenbacher, J.; Wenger, M.; Speer, J.; Kim, A. LLM-assisted content analysis: Using large language models to support deductive coding. arXiv 2023, arXiv:2306.14924. [Google Scholar]
- Morgan, D.L. Exploring the Use of Artificial Intelligence for Qualitative Data Analysis: The Case of ChatGPT. Int. J. Qual. Methods 2023, 22, 16094069231211248. [Google Scholar] [CrossRef]
- Fuller, K.A.; Morbitzer, K.A.; Zeeman, J.M.; Persky, A.M.; Savage, A.C.; McLaughlin, J.E. Exploring the use of ChatGPT to analyze student course evaluation comments. BMC Med. Educ. 2024, 24, 423. [Google Scholar] [CrossRef]
- Jenner, S.; Raidos, D.; Anderson, E.; Fleetwood, S.; Ainsworth, B.; Fox, K.; Kreppner, J.; Barker, M. Using large language models for narrative analysis: A novel application of generative AI. Methods Psychol. 2025, 12, 100183. [Google Scholar] [CrossRef]
- Bijker, R.; Merkouris, S.S.; Dowling, N.A.; Rodda, S.N. ChatGPT for Automated Qualitative Research: Content Analysis. J. Med. Internet Res. 2024, 26, e59050. [Google Scholar] [CrossRef]
- Theelen, H.; Vreuls, J.; Rutten, J. Doing Research with Help from ChatGPT: Promising Examples for Coding and Inter-Rater Reliability. Int. J. Technol. Educ. 2024, 7, 1–18. [Google Scholar] [CrossRef]
- Hitch, D. Artificial Intelligence Augmented Qualitative Analysis: The Way of the Future? Qual. Health Res. 2024, 34, 595–606. [Google Scholar] [CrossRef]
- Yan, L.; Echeverria, V.; Fernandez-Nieto, G.M.; Jin, Y.; Swiecki, Z.; Zhao, L.; Gašević, D.; Martinez-Maldonado, R. Human-ai collaboration in thematic analysis using chatgpt: A user study and design recommendations. In Proceedings of the Extended Abstracts of the CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 11–16 May 2024; pp. 1–7. [Google Scholar]
- Turobov, A.; Coyle, D.; Harding, V. Using ChatGPT for thematic analysis. arXiv 2024, arXiv:2405.08828. [Google Scholar]
- Doropoulos, S.; Karapalidou, E.; Charitidis, P.; Karakeva, S.; Vologiannidis, S. Beyond manual media coding: Evaluating large language models and agents for news content analysis. Appl. Sci. 2025, 15, 8059. [Google Scholar] [CrossRef]
- Adeseye, A.; Isoaho, J.; Mohammad, T. LLM-Assisted Qualitative Data Analysis: Security and Privacy Concerns in Gamified Workforce Studies. Procedia Comput. Sci. 2025, 257, 60–67. [Google Scholar] [CrossRef]
- Fan, J.; Ai, Y.; Liu, X.; Deng, Y.; Li, Y. Coding latent concepts: A human and LLM-coordinated content analysis procedure. Commun. Res. Rep. 2024, 41, 324–334. [Google Scholar] [CrossRef]
- Brumen, B.; Göllner, S.; Tropmann-Frick, M. Aspects and views on responsible artificial intelligence. In International Conference on Machine Learning, Optimization, and Data Science; Springer: Berlin/Heidelberg, Germany, 2022; pp. 384–398. [Google Scholar]
- Akbarighatar, P. Operationalizing responsible AI principles through responsible AI capabilities. AI Ethics 2025, 5, 1787–1801. [Google Scholar] [CrossRef]
- Buolamwini, J.; Gebru, T. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Proceedings of the Conference on Fairness, Accountability and Transparency, New York, NY, USA, 23–24 February 2018; pp. 77–91. [Google Scholar]
- Richardson, R.; Schultz, J.M.; Crawford, K. Dirty data, bad predictions: How civil rights violations impact police data, predictive policing systems, and justice. NYUL Rev. Online 2019, 94, 15–55. [Google Scholar]
- Nissenbaum, H. Privacy in context: Technology, policy, and the integrity of social life. In Privacy in Context; Stanford University Press: Redwood, CA, USA, 2009. [Google Scholar]
- Ahmadi, A. Unravelling the mysteries of hallucination in large language models: Strategies for precision in artificial intelligence language generation. Asian J. Comput. Sci. Technol. 2024, 13, 1–10. [Google Scholar] [CrossRef]
- Huang, L.; Yu, W.; Ma, W.; Zhong, W.; Feng, Z.; Wang, H.; Chen, Q.; Peng, W.; Feng, X.; Qin, B.; et al. A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions. ACM Trans. Inf. Syst. 2025, 43, 42. [Google Scholar] [CrossRef]
- Banerjee, S.; Agarwal, A.; Singla, S. Llms will always hallucinate, and we need to live with this. arXiv 2024, arXiv:2409.05746. [Google Scholar]
- Ji, Z.; Lee, N.; Frieske, R.; Yu, T.; Su, D.; Xu, Y.; Ishii, E.; Bang, Y.J.; Madotto, A.; Fung, P. Survey of Hallucination in Natural Language Generation. ACM Comput. Surv. 2023, 55, 248. [Google Scholar] [CrossRef]
- Williamson, S.M.; Prybutok, V. The Era of Artificial Intelligence Deception: Unraveling the Complexities of False Realities and Emerging Threats of Misinformation. Information 2024, 15, 299. [Google Scholar] [CrossRef]
- Clairoux-Trepanier, V.; Beauchamp, I.-M.; Ruellan, E.; Paquet-Clouston, M.; Paquette, S.-O.; Clay, E. The use of large language models (llm) for cyber threat intelligence (cti) in cybercrime forums. arXiv 2024, arXiv:2408.03354. [Google Scholar]
- Whitty, M.T.; Moustafa, N.; Grobler, M. Cybersecurity when working from home during COVID-19: Considering the human factors. J. Cybersecur. 2024, 10, tyae001. [Google Scholar] [CrossRef]
- Whitty, M.T. Mass-Marketing Fraud: A Growing Concern. IEEE Secur. Priv. 2015, 13, 84–87. [Google Scholar] [CrossRef]
- Jobin, A.; Ienca, M.; Vayena, E. The global landscape of AI ethics guidelines. Nat. Mach. Intell. 2019, 1, 389–399. [Google Scholar] [CrossRef]
- Ruddy, C.; Whitty, M.; Doherty, S. Fake and True Pandemic News Database; Zenodo: Honolulu, HI, USA, 2025. [Google Scholar] [CrossRef]
- Whitty, M.T.; Ruddy, C. COVID-19 lies and truths: Employing the Elaboration Likelihood Model (ELM) and Linguistic Inquiry and Word Count (LIWC) to gain insights into the persuasive techniques evident in disinformation (fake news). Comput. Hum. Behav. Rep. 2025, 20, 100797. [Google Scholar] [CrossRef]
Dimension | Content Analysis | Thematic Analysis | Discourse Analysis |
---|---|---|---|
Epistemological orientation | Positivist/post-positivist. | Flexible—positivist/constructionist/reflexive. | Constructionist/post-modernist/critical. |
Focus of Analysis | What is present in the text, frequency and distribution of categories, manifest and latent features. | Patterns of meaning—identifying, developing, and interpreting themes across the dataset. | How language works—discursive practices, rhetorical strategies, ideological functions. |
Coding process | Structured codebook with explicit rules; intercoder reliability checks. | Iterative coding and theme generation; phases include familiarisation, coding, theme development and refinement. | Close reading of extracts; interpretative identification of discursive repertories; metaphors and strategies. |
Reliability and Validity | Intercoder reliability measured by Krippendorff’s α or Cohen’s Kappa; Validity addressed through clear coding scheme that measure constructs and includes criterion checks. | Reliability conceptualised as transparency and reflexivity, can use interrater reliability or member checking and peer debriefing; requires an audit trail. | Reliability achieved through rigorous, systematic methods, clear research questions and intersubjectivity shared by other researchers; validity concerns include the accuracy of the reflection. |
Outputs | Quantitative summaries; descriptive statistics; cross-tabulations; visualisations of patterns. | Thematic maps, rich narrative accounts of themes, supported by illustrative extracts; tabulations; visualisations of patterns. | Critical interpretations of discourse practices; accounts of ideology, power, and identity construction. |
Strengths | Scalability, replicability, ability to combine with statistical analysis. | Flexibility, accessibility, capacity to capture patterned meanings. | Deep contextual insight; capacity to interrogate power, identity and ideology. |
Limitations | Can overlook nuance; risk of reductionism, especially if categories are poorly defined. | Risk of superficiality if rigour is not applied; themes may reflect researcher bias. | Limited replicability; findings are typically not generalisable. |
Study | Materials | Methods | Key Findings |
---|---|---|---|
Trends in information behaviour research 1999–2008: A content analysis [70] | 749 articles. | CA on human information literature using the searchers ‘information needs’ and ‘information uses’. | Overtime, scholarly researchers increased and practitioners decreased. |
Content analysis of cyber insurance policies: how do carriers price cyber risk? [71] | 235 documents. | CA on the losses covered by cyber insurance policies and those excluded; the questions carriers pose to applicants in order to assess risk; and how cyber insurance premiums determined. | Most important firm characteristics used to computer insurance premiums were the firm’s asset value base rate, rather than specific technology or governance controls. |
What security features and crime prevention advice is communicated in consumer IoT device manuals and support pages? [72] | Manuals and associated support pages for 270 consumer IoT devices produced by 220 different manufacturers. | CA on examined security features. | Manufacturers do not provide enough information about the security features of their devices. |
Twenty-five years of cyber threats in the news: A study of Swedish newspaper coverage (1995–2019) [73] | 1269 newspaper articles. | CA examined threats along several dimensions: modality, ambiguous themes, how threat has changed over time and event orientation. | Swedish papers cover multiple threats, hacking has multiple meanings, coverage has changed over time. |
An investigation of the impact of data breach severity on the readability of mandatory data breach notification letters: Evidence from U.S. firms [74] | 512 data breach incidents from 281 U.S. firms across 2012–2015. | CA examined data breach severity attributes and readability measures. | Data breach severity has a positive impact on reading complexity, and a negative impact on numerical terms. |
Scamming higher ed: An analysis of phishing content and trends [75] | 2300 phishing emails from 2010 to 2023. | CA examined persuasive and emotional appeals, topics, linguistic features and metadata | There has been a shift in the topics in phishing emails; persuaded using sources of authority, scarcity and fear appeals. |
Step | Description |
---|---|
| Specify focused questions that the analysis will address. |
| Identify the corpus and choose an appropriate sampling strategy. |
| A sample of the corpus. |
| Decide the textual unit (e.g., word, sentence, paragraph, document). |
| Construct clear, mutually exclusive and exhaustive categories (deductive and/or inductive). |
| Provide coder training; apply coding rules systematically and consistently. |
| Evaluate intercoder agreement (e.g., Cohen’s K) and refine the scheme if needed. |
| Conduct descriptive/quantitative summaries and/or qualitative pattern interpretation. |
| Link results to questions and theory, discuss implications, limitations, and transparency of procedures. |
Study | Materials | AI Role in CA | Findings & Conclusions | Type AI |
---|---|---|---|---|
Beyond manual media coding: Evaluating large language models and news agents for news content analysis [88] | 200 news articles on U.S. tariff policies. | 7 LLMs assessed under a unified zero-shot prompt. |
| not stated |
Comparing large language models and human annotators in latent content analysis of sentiment, political leaning, emotional intensity and sarcasm [78] | 100 curated textual items. | 33 human annotators; 8 LLM variants |
| GPT-4, Gemini, Llama 3.1-70B and Mixtral 8 x7B |
Using large language models for narrative analysis: A novel application of generative AI [82] | 138 short stories written by young people about social media, identity formation, and food choices. | Analysed by humans researchers and two different LLMs (Claude and GPT-01). |
| ChatGPT (version not stated) |
LLM- Assisted Content Analysis: Using large language models to support deductive coding. [79] | 4 publicly available datasets. | Compared GPT 3.5 with human researchers. |
| GPT 3.5 |
LLM-Assisted qualitative data analysis: Security and privacy concerns in gamified workforce studies [89] | 23 Interview transcripts. | Compared LLaMA, Gemma and Phi. |
| LLaMA, Gemma, and Phi. |
Coding latent concepts: A human and LLM-coordinated content analysis procedure [90] | 1000 public comments. | Compared human researchers with GPT-4o and GPT-3.5-turbo-1106. | Fine tuned GPT 3.5-turbo with smaller datasets can surpass GPT 4o’s performance. | GPT3.5 and GPT-4o |
Step1: Define research questions Who: Human researcher/s Process: Develop research questions based on theory and previous empirical research. | |
Step 2: Select content Who: Human researcher/s Process: Decide on the population of text or data that will be analysed (e.g., zines, books, newspaper articles, social media posts). | |
Step 3: Sample Who: Human researcher/s Process: Decide on a sample from the corpus. | |
Step 4: Determine units of analysis Who: Human researcher/s Process: decide on whether to code words, sentences, paragraphs, visuals, memes, documents, etc. | |
Step 5: Develop coding scheme and codebook Who: Human researcher/s and LLM/s Process: Human researcher/s first generate categories or codes with definitions informed from research questions, previous research and theory. They also outline the inclusion and exclusion criteria. LLM/s generate own set of categories or codes after being feed research questions, relevant theory, and key findings from the literature. Also outlines the inclusion and exclusion criteria. Human researcher/s make the final decision on the coding scheme considering also the LLMs output. Human researcher/s set up the codebook with categories/codes, definitions and inclusion/exclusion criteria (ideally with examples and counterexamples)—typically in a spreadsheet. | |
Step 6: Train coders and code and verification playbook Who: Human researcher/s and LLM/s Process: Human researcher/s and LLMs code independently with the same category/code definitions. Start with a sample and compare results. Human oversight of AI codes. Consider refining definitions. Repeat process until a final set of definitions is obtained. Humans have oversight of the final codes. Both human researcher/s and LLMs code full dataset. Employ a verification playbook to record LLMs failure types, hallucinated references and misclassification of nuanced categories. Can compare with other LLMs to increase reliability. | |
Step 7: Assess reliability and verification playbook Who: Human researcher/s Process: Carry out a statistical analysis comparing the human researcher/s and LLMs findings (e.g., using Krippendorff’s α or Cohen’s Kappa). | |
Step 8: Analyse coded data Who: Human researcher/s Process: Carry out the qualitative or statistical analysis on the codes. Qualitive analysis involves a pattern summary and quantitative involves carrying out appropriate statistical analysis. | |
Step 9: Interpret and report Who: Human researcher/s and can also use LLM/s (with caution) Process: Independently interpret the data. Support LLMs interpretation by feeding research questions, theory, previous key findings, and results. Compare interpretations. Check LLMs interpretation by finding evidence in the academic literature and be mindful of potential hallucination bias. Human researcher/s to make the final decision on the interpretation and write the report. |
Step1: Define research questions Who: Human researcher/s Process: Develop research questions based on gaps in the literature. | |
Step 2: Select content Who: Human researcher/s Process: Decide on the population of text or data that will be analysed (e.g., zines, books, newspaper articles, social media posts). | |
Step 3: Sample Who: Human researcher/s Process: Decide on a sample from the corpus. | |
Step 4: Determine units of analysis Who: Human researcher/s Process: decide on whether to code words, sentences, paragraphs, visuals, memes, documents, etc. | |
Step 5: Develop coding scheme and codebook Who: Human researcher/s and LLM/s Process: Human researcher/s read and re-read the data to gain a deep understanding of its context and nuances. An initial set of codes and definitions are developed that address the research questions. LLMs are feed the research questions and determine a first set of codes and definitions. Human researcher/s compare findings, re-think the categories or codes and cluster where appropriate. Human researcher’s and LLMs then independently undertake a second round of coding to develop subcategories if appropriate and more fine-grained codes (considering inclusion and exclusion criteria). Again, human researcher/s examine sets of findings and make the final decision on the coding scheme. They also set up the codebook with categories/codes, definitions and inclusion/exclusion criteria (ideally with examples and counterexamples)—typically in a spreadsheet. | |
Step 6: Train coders and code Who: Human researcher/s and LLM/s Process: Human researcher/s and LLMs code independently with the same category/code definitions. Start with a sample and compare results. Consider refining definitions. Repeat process until a final set of definitions is obtained. Both human researcher/s and LLMs code full dataset. Employ a verification playbook to record LLMs failure types, hallucinated references and misclassification of nuanced categories. Can compare with other LLMs to increase reliability. | |
Step 7: Assess reliability Who: Human researcher/s Process: Carry out a statistical analysis comparing the human researcher/s and LLMs findings (e.g., using Krippendorff’s α or Cohen’s Kappa). Alternatively, reliability could be strengthened through coder triangulation, reflexivity, and iterative consensus-building. | |
Step 8: Analyse coded data Who: Human researcher/s Process: Carry out the qualitative or statistical analysis on the codes. Qualitive analysis involves a pattern summary and quantitative involves carrying out appropriate statistical analysis. | |
Step 9: Interpret and report Who: Human researcher/s and can also use LLM/s (with caution) Process: Independently interpret the data. Support LLMs interpretation by feeding research questions, theory, previous key findings, and results. Compare interpretations. Check LLMs interpretation by finding evidence in the academic literature and be mindful of potential hallucination bias. Human researcher/s to make the final decision on the interpretation and write the report. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Whitty, M.T. How to Conduct AI-Assisted (Large Language Model-Assisted) Content Analysis in Information Science and Cyber Security Research. Electronics 2025, 14, 4104. https://doi.org/10.3390/electronics14204104
Whitty MT. How to Conduct AI-Assisted (Large Language Model-Assisted) Content Analysis in Information Science and Cyber Security Research. Electronics. 2025; 14(20):4104. https://doi.org/10.3390/electronics14204104
Chicago/Turabian StyleWhitty, Monica Therese. 2025. "How to Conduct AI-Assisted (Large Language Model-Assisted) Content Analysis in Information Science and Cyber Security Research" Electronics 14, no. 20: 4104. https://doi.org/10.3390/electronics14204104
APA StyleWhitty, M. T. (2025). How to Conduct AI-Assisted (Large Language Model-Assisted) Content Analysis in Information Science and Cyber Security Research. Electronics, 14(20), 4104. https://doi.org/10.3390/electronics14204104