Temporal Dynamics of Harmful Speech in Chatbot–User Dialogues: A Comparative Study of LLM and Chit-Chat Systems
Abstract
1. Introduction
2. Related Work
2.1. Addressing Harmful Expressions in Chatbots
2.2. Harmful Speech Types and Incidence in LLM Chatbots
2.3. Temporal Variation in Harmful Expressions
3. Methods
3.1. Dataset
3.2. Analysis Methods
3.3. Ethical Consideration
4. Results
5. Discussion
6. Limitations and Future Work
7. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Si, W.M.; Backes, M.; Blackburn, J.; De Cristofaro, E.; Stringhini, G.; Zannettou, S.; Zhang, Y. Why So Toxic? Measuring and Triggering Toxic Behavior in Open-Domain Chatbots. In Proceedings of the CCS ’22: 2022 ACM SIGSAC Conference on Computer and Communications Security, Los Angeles, CA, USA, 7–11 November 2022; pp. 2659–2673. [Google Scholar] [CrossRef]
- Chin, H.; Molefi, L.W.; Yi, M.Y. Empathy Is All You Need: How a Conversational Agent Should Respond to Verbal Abuse. In Proceedings of the CHI ’20: 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; pp. 1–13. [Google Scholar] [CrossRef]
- Song, H.; Hong, J.; Jung, C.; Chin, H.; Shin, M.; Choi, Y.; Choi, J.; Cha, M. Detecting offensive language in an open chatbot platform. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), Torino, Italy, 20–25 May 2024; pp. 4760–4771. [Google Scholar]
- Bender, E.M.; Gebru, T.; McMillan-Major, A.; Shmitchell, S. On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? In Proceedings of the FAccT ’21: 2021 ACM Conference on Fairness, Accountability, and Transparency, Virtual, 3–10 March 2021; pp. 610–623. [Google Scholar] [CrossRef]
- Shumailov, I.; Shumaylov, Z.; Zhao, Y.; Papernot, N.; Anderson, R.; Gal, Y. AI models collapse when trained on recursively generated data. Nature 2024, 631, 755–759. [Google Scholar] [CrossRef] [PubMed]
- Fortuna, P.; Nunes, S. A survey on automatic detection of hate speech in text. Acm Comput. Surv. (CSUR) 2018, 51, 1–30. [Google Scholar] [CrossRef]
- Jahan, M.S.; Oussalah, M. A systematic review of hate speech automatic detection using natural language processing. Neurocomputing 2023, 546, 126232. [Google Scholar] [CrossRef]
- Chin, H.; Song, H.; Baek, G.; Shin, M.; Jung, C.; Cha, M.; Choi, J.; Cha, C. The potential of chatbots for emotional support and promoting mental well-being in different cultures: Mixed methods study. J. Med. Internet Res. 2023, 25, e51712. [Google Scholar] [CrossRef]
- Lin, Z.; Wang, Z.; Tong, Y.; Wang, Y.; Guo, Y.; Wang, Y.; Shang, J. Toxicchat: Unveiling hidden challenges of toxicity detection in real-world user-ai conversation. arXiv 2023, arXiv:2310.17389. [Google Scholar]
- Ni, J.; Young, T.; Pandelea, V.; Xue, F.; Cambria, E. Recent advances in deep learning based dialogue systems: A systematic survey. Artif. Intell. Rev. 2023, 56, 3055–3155. [Google Scholar]
- Roller, S.; Boureau, Y.L.; Weston, J.; Bordes, A.; Dinan, E.; Fan, A.; Gunning, D.; Ju, D.; Li, M.; Poff, S.; et al. Open-domain conversational agents: Current progress, open problems, and future directions. arXiv 2020, arXiv:2006.12442. [Google Scholar] [CrossRef]
- Replika CEO Eugenia Kuyda Launches Wabi. Business Insider. 2025. Available online: https://www.businessinsider.com/replika-ceo-eugenia-kuyda-launch-wabi-2025-10 (accessed on 4 December 2025).
- The New York Times. Character.ai Abandons Making AI Models after $2.7bn Google Deal. 2024. Reports 20 Million Monthly Active Users. Available online: https://www.ft.com/content/f2a9b5d4-05fe-4134-b4fe-c24727b85bba (accessed on 4 December 2025).
- SimSimi Inc. SimSimi Official Website. 2025. Available online: https://simsimi.kr/ (accessed on 4 December 2025).
- Anthropic. How People Use Claude for Support, Advice, and Companionship. 2025. Available online: https://www.anthropic.com/news/how-people-use-claude-for-support-advice-and-companionship (accessed on 4 December 2025).
- Chatterji, A.; Cunningham, T.; Deming, D.J.; Hitzig, Z.; Ong, C.; Shan, C.Y.; Wadman, K. How People Use Chatgpt; Technical Report; National Bureau of Economic Research: Cambridge, MA, USA, 2025. [Google Scholar]
- Gehman, S.; Gururangan, S.; Sap, M.; Choi, Y.; Smith, N.A. RealToxicityPrompts: Evaluating Neural Toxic Degeneration in Language Models. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online Event, 16–20 November 2020; Cohn, T., He, Y., Liu, Y., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 3356–3369. [Google Scholar] [CrossRef]
- Shen, X.; Chen, Z.; Backes, M.; Shen, Y.; Zhang, Y. “Do Anything Now”: Characterizing and Evaluating In-The-Wild Jailbreak Prompts on Large Language Models. In Proceedings of the CCS ’24: 2024 on ACM SIGSAC Conference on Computer and Communications Security, Salt Lake City, UT, USA, 14–18 October 2024; pp. 1671–1685. [Google Scholar] [CrossRef]
- Zhou, W.; Zhu, X.; Han, Q.L.; Li, L.; Chen, X.; Wen, S.; Xiang, Y. The Security of Using Large Language Models: A Survey with Emphasis on ChatGPT. IEEE/CAA J. Autom. Sin. 2025, 12, 1–26. [Google Scholar] [CrossRef]
- Kim, Y.; Kim, J.H. The impact of ethical issues on public understanding of artificial intelligence. In Proceedings of the 23rd HCI International Conference, HCII 2021, Virtual Event, 24–29 July 2021; Springer: Cham, Switzerland, 2021; pp. 500–507. [Google Scholar]
- Deshpande, A.; Murahari, V.; Rajpurohit, T.; Kalyan, A.; Narasimhan, K. Toxicity in chatgpt: Analyzing persona-assigned language models. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, Singapore, 6–10 December 2023; Bouamor, H., Pino, J., Bali, K., Eds.; Association for Computational Linguistics: Singapore, 2023; pp. 1236–1270. [Google Scholar] [CrossRef]
- Weeks, C.; Cheruvu, A.; Abdullah, S.M.; Kanchi, S.; Yao, D.; Viswanath, B. A first look at toxicity injection attacks on open-domain chatbots. In Proceedings of the 39th Annual Computer Security Applications Conference, Austin, TX, USA, 4–8 December 2023; pp. 521–534. [Google Scholar]
- Golder, S.A.; Macy, M.W. Diurnal and seasonal mood vary with work, sleep, and daylength across diverse cultures. Science 2011, 333, 1878–1881. [Google Scholar] [CrossRef]
- Grinberg, N.; Naaman, M.; Shaw, B.; Lotan, G. Extracting diurnal patterns of real world activity from social media. In Proceedings of the International AAAI Conference on Web and Social Media, Cambridge, MA, USA, 8–11 July 2013; Volume 7, pp. 205–214. [Google Scholar]
- Anderson, C.; Platten, C.R. Sleep deprivation lowers inhibition and enhances impulsivity to negative stimuli. Behav. Brain Res. 2011, 217, 463–466. [Google Scholar] [CrossRef]
- Suler, J. The online disinhibition effect. Cyberpsychol. Behav. 2004, 7, 321–326. [Google Scholar] [PubMed]
- Joinson, D.; Haworth, C.M.; Simpson, E.; Cristianini, N.; Di Cara, N.H.; Davis, O.S. Active night-time tweeting is associated with meaningfully lower mental wellbeing in a UK birth cohort study. Sci. Rep. 2025, 15, 34301. [Google Scholar] [CrossRef] [PubMed]
- Scott, H.; Biello, S.M.; Woods, H.C. Social media use and adolescent sleep patterns: Cross-sectional findings from the UK millennium cohort study. BMJ Open 2019, 9, e031161. [Google Scholar] [CrossRef] [PubMed]
- Kumar, D.; Hancock, J.; Thomas, K.; Durumeric, Z. Understanding the behaviors of toxic accounts on reddit. In Proceedings of the ACM Web Conference 2023, Austin, TX, USA, 30 April–4 May 2023; pp. 2797–2807. [Google Scholar]
- Zhao, W.; Ren, X.; Hessel, J.; Cardie, C.; Choi, Y.; Deng, Y. Wildchat: 1m chatgpt interaction logs in the wild. arXiv 2024, arXiv:2405.01470. [Google Scholar]
- Bell, S.; Meglioli, M.C.; Richards, M.; Sánchez, E.; Ropers, C.; Wang, S.; Williams, A.; Sagun, L.; Costa-jussà, M.R. On the Role of Speech Data in Reducing Toxicity Detection Bias. In Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), Albuquerque, NM, USA, 29 April–4 May 2025; Chiruzzo, L., Ritter, A., Wang, L., Eds.; Association for Computational Linguistics: Albuquerque, NM, USA, 2025; pp. 1454–1468. [Google Scholar] [CrossRef]
- IBM. Context Window (Think Blog Topic Page). 2025. Available online: https://www.ibm.com/think/topics/context-window (accessed on 4 December 2025).
- Shin, M.; Chin, H.; Song, H.; Choi, Y.; Choi, J.; Cha, M. Context-Aware Offensive Language Detection in Human-Chatbot Conversations. In Proceedings of the 2024 IEEE International Conference on Big Data and Smart Computing (BigComp), Bangkok, Thailand, 18–21 February 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 270–277. [Google Scholar]
- Xu, J.; Ju, D.; Li, M.; Boureau, Y.L.; Weston, J.; Dinan, E. Bot-Adversarial Dialogue for Safe Conversational Agents. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; Toutanova, K., Rumshisky, A., Zettlemoyer, L., Hakkani-Tur, D., Beltagy, I., Bethard, S., Cotterell, R., Chakraborty, T., Zhou, Y., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 2950–2968. [Google Scholar] [CrossRef]
- Greenberg, A. Now Anyone Can Deploy Google’s Troll-Fighting AI. WIRED (Security). 2017. Available online: https://www.wired.com/2017/02/googles-troll-fighting-ai-now-belongs-world/ (accessed on 4 December 2025).
- Lima, L.H.Q.; Pagano, A.S.; da Silva, A.P.C. Toxic Content Detection in online social networks: A new dataset from Brazilian Reddit Communities. In Proceedings of the 16th International Conference on Computational Processing of Portuguese—Volume 1, Santiago de Compostela, Spain, 14–15 March 2024; pp. 472–482. [Google Scholar]
- Saveski, M.; Roy, B.; Roy, D. The structure of toxic conversations on Twitter. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 1086–1097. [Google Scholar]
- Fan, L.; Li, L.; Hemphill, L. Toxicity on Social Media During the 2022 Mpox Public Health Emergency: Quantitative Study of Topical and Network Dynamics. J. Med. Internet Res. 2024, 26, e52997. [Google Scholar] [CrossRef]
- Shen, Z.; Paik, I. Temporal Modeling of Social Media for Depression Forecasting: Deep Learning Approaches with Pretrained Embeddings. Appl. Sci. 2025, 15, 11274. [Google Scholar] [CrossRef]
- Andrews, S.; Ellis, D.A.; Shaw, H.; Piwek, L. Beyond self-report: Tools to compare estimated and real-world smartphone use. PLoS ONE 2015, 10, e0139004. [Google Scholar] [CrossRef]
- Guo, S.; He, Z.; Rao, A.; Morstatter, F.; Brantingham, J.; Lerman, K. The pulse of mood online: Unveiling emotional reactions in a dynamic social media landscape. ACM Trans. Web 2025, 19, 1–22. [Google Scholar] [CrossRef]
- Chin, H.; Lima, G.; Shin, M.; Zhunis, A.; Cha, C.; Choi, J.; Cha, M. User-chatbot conversations during the COVID-19 pandemic: Study based on topic modeling and sentiment analysis. J. Med. Internet Res. 2023, 25, e40922. [Google Scholar] [CrossRef]
- Chin, H.; Yong Yi, M. Exploring the influence of user characteristics on verbal aggression towards social chatbots. Behav. Inf. Technol. 2025, 44, 1576–1594. [Google Scholar]
- Noble, S.M.; Mende, M. The future of artificial intelligence and robotics in the retail and service sector: Sketching the field of consumer-robot-experiences. J. Acad. Mark. Sci. 2023, 51, 747–756. [Google Scholar] [CrossRef]
- Li, T.G.; Zhang, C.B.; Chang, Y.; Zheng, W. The impact of AI identity disclosure on consumer unethical behavior: A social judgment perspective. J. Retail. Consum. Serv. 2024, 76, 103606. [Google Scholar] [CrossRef]
- Black, E.W.; Mezzina, K.; Thompson, L.A. Anonymous social media—Understanding the content and context of Yik Yak. Comput. Hum. Behav. 2016, 57, 17–22. [Google Scholar] [CrossRef]
- Bleize, D.N.; Anschütz, D.J.; Tanis, M.; Buijzen, M. The effects of group centrality and accountability on conformity to cyber aggressive norms: Two messaging app experiments. Comput. Hum. Behav. 2021, 120, 106754. [Google Scholar] [CrossRef]
- Bogdan, A.; Dospinescu, N.; Dospinescu, O. Beyond Credibility: Understanding the Mediators Between Electronic Word-of-Mouth and Purchase Intention. arXiv 2025, arXiv:2504.05359. [Google Scholar] [CrossRef]
- Ganai, A.H.; Hashmy, R.; Khanday, H.A. Finding information diffusion’s seed nodes in online social networks using a special degree centrality. SN Comput. Sci. 2024, 5, 333. [Google Scholar] [CrossRef]



| Category | SimSimi | WildChat | p | ||
|---|---|---|---|---|---|
| Count | Rate | Count | Rate | ||
| Toxicity | 350,015 | 4.02% | 5080 | 0.83% | <0.001 |
| Profanity | 259,677 | 2.98% | 2900 | 0.48% | <0.001 |
| Insult | 47,905 | 0.55% | 356 | 0.06% | <0.001 |
| Iden._Attack | 6934 | 0.08% | 99 | 0.02% | <0.001 |
| Threat | 7537 | 0.09% | 45 | 0.01% | <0.001 |
| S._Toxicity | 146 | 0.00% | 14 | 0.00% | =0.248 |
| Harmful (any) | 672,214 | 7.72% | 8494 | 1.39% | <0.001 |
| Total turns | 8,785,959 | 610,837 | |||
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kwon, O.; Yoon, H.; Chin, H.; Park, J. Temporal Dynamics of Harmful Speech in Chatbot–User Dialogues: A Comparative Study of LLM and Chit-Chat Systems. Appl. Sci. 2025, 15, 13185. https://doi.org/10.3390/app152413185
Kwon O, Yoon H, Chin H, Park J. Temporal Dynamics of Harmful Speech in Chatbot–User Dialogues: A Comparative Study of LLM and Chit-Chat Systems. Applied Sciences. 2025; 15(24):13185. https://doi.org/10.3390/app152413185
Chicago/Turabian StyleKwon, Ohseong, Hyobeen Yoon, Hyojin Chin, and Jisung Park. 2025. "Temporal Dynamics of Harmful Speech in Chatbot–User Dialogues: A Comparative Study of LLM and Chit-Chat Systems" Applied Sciences 15, no. 24: 13185. https://doi.org/10.3390/app152413185
APA StyleKwon, O., Yoon, H., Chin, H., & Park, J. (2025). Temporal Dynamics of Harmful Speech in Chatbot–User Dialogues: A Comparative Study of LLM and Chit-Chat Systems. Applied Sciences, 15(24), 13185. https://doi.org/10.3390/app152413185

