Continual Learning for Saudi-Dialect Offensive-Language Detection Under Temporal Linguistic Drift
Abstract
1. Introduction
2. Related Work
2.1. Offensive-Language Detection in Arabic
2.2. Continual Learning: Approaches and Applications
3. Methodology
3.1. Experimental Framework
3.2. Dataset Construction
3.2.1. Training Data
- Newly emerged offensive terms: Novel expressions without dictionary roots that have developed as offensive terms in everyday discourse and spread through social media, such as “زوط” (zoot) and its morphological variations. These terms arise from natural linguistic evolution, where users organically create and spread new offensive expressions in daily communication.
- Context-shifting expressions: Neutral words that acquire offensive connotations depending on pragmatic or situational context. For example, “العفوية” (spontaneity) functions literally in “العفوية أجمل شي” (spontaneity is the best thing) but becomes sarcastic mockery in “العفوية تتصل بك” (spontaneity is calling you).
3.2.2. Test Scenarios
- Contemporary test set: 500 examples (50% offensive and 50% non-offensive) collected independently from the training dataset (NEW_DS) during the same 2024–2025 period. Offensive examples include newly emerged and context-shifting expressions, while non-offensive examples include both general non-offensive sentences and neutral contextual uses of the same terms, ensuring a balanced distribution.
- Historical test set: Five hundred held-out examples from the original SOD collection period, containing no evolved or context-shifting expressions. This dataset was never used during model development, serving as an unseen benchmark for assessing catastrophic forgetting and knowledge retention.
- Mixed-distribution sets: Two datasets combining contemporary and historical samples in different ratios (20/80 and 40/60 contemporary/historical). These sets evaluate model performance under more realistic conditions, where established offensive patterns from the historical corpus remain predominant, while emerging terms constitute a smaller fraction of encountered content.
- Simulated Realistic test set (SimuReal): Five hundred examples with class distribution approximating real-world conditions: 80% non-offensive; and 20% offensive, comprising 2% newly emerged terms, 3% context-shifting expressions, and 15% general offensive content. This distribution reflects that offensive content is rare in practice, and that drift terms constitute a small fraction of offensive content [36]. Table 2 provides a complete overview of all datasets used in this study, including sizes, collection periods, and composition.
3.3. Experimental Configurations
3.3.1. Baseline Models
- Original: SOD_AraBERT without continual adaptation, serving as the reference for computing Knowledge Retention and Adaptation Gain metrics.
- Naïve fine-tuning: Standard fine-tuning on contemporary data without mechanisms to prevent forgetting, representing the lower bound for knowledge retention.
3.3.2. Core Techniques
- Experience Replay (ER). Reinforces prior knowledge by combining samples from the original SOD training split with samples from the training dataset (NEW_DS). The replay buffer is shuffled during training. Ablation study tested replay ratios ∈ {0.1, 0.2, 0.4, 0.6, 0.8}.
- Elastic Weight Consolidation (EWC). Constrains updates to parameters critical for prior knowledge using the Fisher Information Matrix computed from 1000 historical samples. Ablation study tested λ ∈ {100, 500, 1000, 5000, 10,000}.
- Low-Rank Adaptation (LoRA). Freezes pre-trained weights and introduces trainable low-rank decomposition matrices within transformer attention layers. Ablation study tested ranks r ∈ {8, 16, 32} with both standard (query, key, and value) and extended (including attention output dense) target modules, with no bias adaptation, and dropout rate = 0.1.
3.3.3. Hybrid Configurations
3.3.4. Training Configuration
3.4. Evaluation Metrics
3.4.1. Classification Metrics
3.4.2. Continual-Learning Metrics
3.4.3. Training and Efficiency
4. Results
4.1. Ablation Studies
4.1.1. LoRA Rank and Target Module Selection
4.1.2. EWC Regularization Strength
4.1.3. Experience Replay Buffer Size
4.2. Overall Performance
4.2.1. Diagnostic Boundaries: Historical and Contemporary
4.2.2. Mixed Distributions: Closer to Realistic Temporal Scenarios
4.2.3. SimuReal: Simulated Real-World Conditions
4.3. Stability–Plasticity Trade-Off
4.4. Parameter Efficiency vs. Performance
4.5. Training Efficiency
4.6. Training Convergence
| Method | Final Train Loss | Best Val Loss | Final Val Loss |
|---|---|---|---|
| Naïve FT | 0.002 ± 0.001 | 0.033 ± 0.002 | 0.033 ± 0.002 |
| ER | 0.025 ± 0.005 | 0.044 ± 0.006 | 0.062 ± 0.004 |
| EWC | 0.010 ± 0.001 | 0.039 ± 0.002 | 0.039 ± 0.002 |
| LoRA | 0.389 ± 0.011 | 0.435 ± 0.006 | 0.435 ± 0.006 |
| LoRA+ER | 0.307 ± 0.008 | 0.287 ± 0.002 | 0.287 ± 0.002 |
| Full+ER+EWC | 0.058 ± 0.006 | 0.063 ± 0.007 | 0.081 ± 0.002 |
| LoRA+EWC | 0.448 ± 0.012 | 0.498 ± 0.007 | 0.498 ± 0.007 |
| LoRA+ER+EWC | 0.325 ± 0.008 | 0.332 ± 0.003 | 0.332 ± 0.003 |
5. Discussion
6. Conclusions and Future Work
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Appendix A. Confusion Matrices

References
- Kemp, S. Digital 2025: Saudi Arabia. DataReportal. 3 March 2025. Available online: https://datareportal.com/reports/digital-2025-saudi-arabia (accessed on 12 January 2026).
- Asiri, A.; Saleh, M. SOD: A Corpus for Saudi Offensive Language Detection Classification. Computers 2024, 13, 211. [Google Scholar] [CrossRef]
- Florio, K.; Basile, V.; Polignano, M.; Basile, P.; Patti, V. Time of Your Hate: The Challenge of Time in Hate Speech Detection on Social Media. Appl. Sci. 2020, 10, 4180. [Google Scholar] [CrossRef]
- Arango, A.; Pérez, J.; Poblete, B. Hate Speech Detection is Not as Easy as You May Think: A Closer Look at Model Validation. Inf. Syst. 2020, 101, 101584. [Google Scholar] [CrossRef]
- Jin, M.; Mu, Y.; Maynard, D.; Bontcheva, K. Examining Temporal Bias in Abusive Language Detection. arXiv 2023, arXiv:2309.14146. [Google Scholar] [CrossRef]
- Eisenstein, J.; O’Connor, B.; Smith, N.A.; Xing, E.P. Diffusion of Lexical Change in Social Media. PLoS ONE 2014, 9, e113114. [Google Scholar] [CrossRef]
- Lucy, L.; Bamman, D. Characterizing English Variation across Social Media Communities with BERT. Trans. Assoc. Comput. Linguist. 2021, 9, 538–556. [Google Scholar] [CrossRef]
- Crystal, D. Language and the Internet; Cambridge University Press: Cambridge, UK, 2012; ISBN 9781139164771. [Google Scholar] [CrossRef]
- Dembe, T. The Impact of Social Media on Language Evolution. Eur. J. Linguist. 2024, 3, 1–14. [Google Scholar] [CrossRef]
- Di Marco, N.; Loru, E.; Bonetti, A.; Serra, A.O.G.; Cinelli, M.; Quattrociocchi, W. The Evolution of Language in Social Media Comments. arXiv 2024, arXiv:2406.11450. [Google Scholar] [CrossRef]
- Allison, K.R. Social Norms in Online Communities: Formation, Evolution and Relation to Cyber-Aggression. In Proceedings of the Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems (CHI EA ‘18), Montreal, QC, Canada, 21–26 April 2018; Association for Computing Machinery: New York, NY, USA, 2018; Paper No. DC01. [Google Scholar] [CrossRef]
- Dahou, A.; Dahou, A.H.; Cheragui, M.A.; Abdedaiem, A.; Al-Qaness, M.A.A.; Abd Elaziz, M.; Ewees, A.A.; Zheng, Z. A Survey on Dialect Arabic Processing and Analysis: Recent Advances and Future Trends. ACM Trans. Asian Low-Resour. Lang. Inf. Process. 2024, 24, 84. [Google Scholar] [CrossRef]
- Burnap, P.; Williams, M.L. Cyber Hate Speech on Twitter: An Application of Machine Classification and Statistical Modeling for Policy and Decision Making. Policy Internet 2015, 7, 223–242. [Google Scholar] [CrossRef]
- Mubarak, H.; Darwish, K.; Magdy, W. Abusive Language Detection on Arabic Social Media. In Proceedings of the First Workshop on Abusive Language Online, Vancouver, BC, Canada, 4 August 2017; pp. 52–56. [Google Scholar] [CrossRef]
- Badjatiya, P.; Gupta, S.; Gupta, M.; Varma, V. Deep Learning for Hate Speech Detection in Tweets. In Proceedings of the 26th International Conference on World Wide Web Companion (WWW ‘17 Companion), Perth, Australia, 3–7 April 2017; Association for Computational Linguistics: Stroudsburg, PA, USA, 2017; pp. 759–760. [Google Scholar] [CrossRef]
- Abu Farha, I.; Magdy, W. Multitask Learning for Arabic Offensive Language and Hate-Speech Detection. In Proceedings of the 4th Workshop on Open-Source Arabic Corpora and Processing Tools, with a Shared Task on Offensive Language Detection, Marseille, France, May 2020; European Language Resource Association: Paris, France, 2020; pp. 86–90. Available online: https://aclanthology.org/2020.osact-1.14/ (accessed on 15 July 2025).
- Antoun, W.; Baly, F.; Hajj, H. AraBERT: Transformer-based Model for Arabic Language Understanding. In Proceedings of the 4th Workshop on Open-Source Arabic Corpora and Processing Tools, with a Shared Task on Offensive Language Detection, Marseille, France, May 2020; European Language Resource Association: Paris, France, 2020; pp. 9–15. Available online: https://aclanthology.org/2020.osact-1.2/ (accessed on 15 July 2025).
- Abdul-Mageed, M.; Elmadany, A.; Nagoudi, E.M.B. ARBERT & MARBERT: Deep Bidirectional Transformers for Arabic. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online, August 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 7088–7105. [Google Scholar] [CrossRef]
- Ali, A.R.; Siddiqui, M.A.; Algunaibet, R.; Ali, H.R. A Large and Diverse Arabic Corpus for Language Modeling. Procedia Comput. Sci. 2023, 225, 12–21. [Google Scholar] [CrossRef]
- Mubarak, H.; Darwish, K.; Magdy, W.; Elsayed, T.; Al-Khalifa, H. Overview of OSACT4 Arabic Offensive Language Detection Shared Task. In Proceedings of the 4th Workshop on Open-Source Arabic Corpora and Processing Tools, with a Shared Task on Offensive Language Detection, Marseille, France, May 2020; European Language Resource Association: Paris, France, 2020; pp. 48–52. Available online: https://aclanthology.org/2020.osact-1.7/ (accessed on 15 July 2025).
- Husain, F. OSACT4 Shared Task on Offensive Language Detection: Intensive Preprocessing-Based Approach. In Proceedings of the 4th Workshop on Open-Source Arabic Corpora and Processing Tools, with a Shared Task on Offensive Language Detection, Marseille, France, May 2020; European Language Resource Association: Paris, France, 2020; pp. 53–60. Available online: https://aclanthology.org/2020.osact-1.8/ (accessed on 15 July 2025).
- McCloskey, M.; Cohen, N.J. Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem. Psychol. Learn. Motiv. 1989, 24, 109–165. [Google Scholar] [CrossRef]
- French, R.M. Catastrophic Forgetting in Connectionist Networks. Trends Cogn. Sci. 1999, 3, 128–135. [Google Scholar] [CrossRef] [PubMed]
- Kirkpatrick, J.; Pascanu, R.; Rabinowitz, N.; Veness, J.; Desjardins, G.; Rusu, A.A.; Milan, K.; Quan, J.; Ramalho, T.; Grabska-Barwinska, A.; et al. Overcoming Catastrophic Forgetting in Neural Networks. Proc. Natl. Acad. Sci. USA 2017, 114, 3521–3526. [Google Scholar] [CrossRef]
- Zenke, F.; Poole, B.; Ganguli, S. Continual Learning Through Synaptic Intelligence. In Proceedings of the 34th International Conference on Machine Learning (ICML’17), Sydney, Australia, 6–11 August 2017; PMLR: London, UK, 2017; pp. 3987–3995. Available online: https://proceedings.mlr.press/v70/zenke17a.html (accessed on 15 November 2024).
- Lin, L.-J. Self-Improving Reactive Agents Based on Reinforcement Learning, Planning and Teaching. Mach. Learn. 1992, 8, 293–321. [Google Scholar] [CrossRef]
- Rolnick, D.; Ahuja, A.; Schwarz, J.; Lillicrap, T.P.; Wayne, G. Experience Replay for Continual Learning. In Proceedings of the 33rd International Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019; Curran Associates, Inc.: Red Hook, NY, USA, 2019; pp. 350–360. Available online: https://proceedings.neurips.cc/paper/2019/hash/fa7cdfad1a5aaf8370ebeda47a1ff1c3-Abstract.html (accessed on 15 July 2025).
- Lopez-Paz, D.; Ranzato, M. Gradient Episodic Memory for Continual Learning. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS’17), Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017; pp. 6470–6479. Available online: https://proceedings.neurips.cc/paper/2017/hash/f87522788a2be2d171666752f97ddebb-Abstract.html (accessed on 15 July 2025).
- Chaudhry, A.; Ranzato, M.; Rohrbach, M.; Elhoseiny, M. Efficient Lifelong Learning with A-GEM. arXiv 2019, arXiv:1812.00420. [Google Scholar] [CrossRef]
- Hu, E.J.; Shen, Y.; Wallis, P.; Allen-Zhu, Z.; Li, Y.; Wang, S.; Wang, L.; Chen, W. LoRA: Low-Rank Adaptation of Large Language Models. arXiv 2021, arXiv:2106.09685. [Google Scholar] [CrossRef]
- Houlsby, N.; Giurgiu, A.; Jastrzebski, S.; Morrone, B.; de Laroussilhe, Q.; Gesmundo, A.; Attariyan, M.; Gelly, S. Parameter-Efficient Transfer Learning for NLP. In Proceedings of the 36th International Conference on Machine Learning (ICML 2019), Long Beach, CA, USA, 9–15 June 2019; PMLR: London, UK, 2019; pp. 2790–2799. Available online: https://proceedings.mlr.press/v97/houlsby19a.html (accessed on 15 July 2025).
- Pfeiffer, J.; Kamath, A.; Rücklé, A.; Cho, K.; Gurevych, I. AdapterFusion: Non-Destructive Task Composition for Transfer Learning. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, Online, April 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 487–503. [Google Scholar] [CrossRef]
- Jang, J.; Kim, Y.; Choi, K.; Suh, S. Sequential Targeting: A Continual Learning Approach for Data Imbalance in Text Classification. Expert Syst. Appl. 2021, 179, 115068. [Google Scholar] [CrossRef]
- Agarwal, S.; Chowdary, C.R. Combating Hate Speech Using an Adaptive Ensemble Learning Model with a Case Study on COVID-19. Expert Syst. Appl. 2021, 185, 115632. [Google Scholar] [CrossRef]
- Thompson, B.; Khayrallah, H.; Anastasopoulos, A.; McCarthy, A.D.; Duh, K.; Marvin, R.; McNamee, P.; Gwinnup, J.; Anderson, T.; Koehn, P. Freezing Subnetworks to Analyze Domain Adaptation in Neural Machine Translation. In Proceedings of the Third Conference on Machine Translation: Research Papers, Brussels, Belgium, October 2018; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 124–132. [Google Scholar] [CrossRef]
- Mubarak, H.; Hassan, S.; Chowdhury, S.A. Emojis as Anchors to Detect Arabic Offensive Language and Hate Speech. Nat. Lang. Eng. 2023, 29, 1436–1457. [Google Scholar] [CrossRef]
- Parisi, G.I.; Kemker, R.; Part, J.L.; Kanan, C.; Wermter, S. Continual Lifelong Learning with Neural Networks: A Review. Neural Netw. 2019, 113, 54–71. [Google Scholar] [CrossRef] [PubMed]







| Category | Term | Offensive Context | Non-Offensive Context |
|---|---|---|---|
| Newly Emerged | زوط (zoot) زووط (zooot—elongated) | انت زوط والله (You are a zoot, seriously—enta zoot wallah) | — |
| زووط عليك 😂 (Zoot on you—zoot ʿalayk) | |||
| Context-Shifting | مربع (murabbaʿ—square) | العقل مربع (The mind is square—al-ʿaql murabbaʿ) | مشروع المربع بيكون علامة فارقة SA (The Al-Murabbaʿ project will be a landmark) |
| الفطاير (al-faṭāyir—pastries) | ويش يبغون منك الفطاير؟ (What do they want from you, pastries? —wesh yibghūn minnak alfatayer) | مين يجيب الفطاير؟ (Who’s bringing the pastries?—meen yjeeb alftayer) | |
| رفقًا بهم (rifqan bihim—be gentle) | رفقًا بهم (Be gentle with them—sarcastic) | رفقًا بهم في هذا الجو (Be kind to them in this weather—rifqan bihim fi hādhā al-jaw) | |
| العفوية (al-ʿafawiyah—spontaneity) | العفوية تتصل بك (Spontaneity is calling you—al-ʿafawiyah tittsil bik) | العفوية أجمل شي (Spontaneity is the best thing—al-ʿafawiyah ajmal shay) |
| Dataset | Size | Period | Composition |
|---|---|---|---|
| Training (NEW_DS) | 2000 | 2024–2025 | New offensive terms + context-shifting + general offensive/non-offensive + original-context sentences |
| Historical Test | 500 | 2019–2022 | Held-out from original SOD samples |
| Contemporary Test | 500 | 2024–2025 | 50% offensive, 50% non-offensive |
| Mixed 20–80 | 500 | Mixed | 20% contemporary, 80% historical |
| Mixed 40–60 | 500 | Mixed | 40% contemporary, 60% historical |
| SimuReal Test | 500 | Mixed | 80% non-offensive, 20% offensive (2% new terms, 3% context-shifting, 15% general) |
| Configuration | Parameters | KR | AG |
|---|---|---|---|
| r = 8, standard | 444 K | −0.038 ± 0.002 | +0.129 ± 0.008 |
| r = 8, extended | 591 K | −0.055 ± 0.001 | +0.155 ± 0.001 |
| r = 16, standard | 886 K | −0.077 ± 0.003 | +0.171 ± 0.000 |
| r = 16, extended | 1.2 M | −0.102 ± 0.001 | +0.199 ± 0.006 |
| r = 32, standard | 1.8 M | −0.119 ± 0.002 | +0.194 ± 0.001 |
| r = 32, extended | 2.4 M | −0.130 ± 0.006 | +0.200 ± 0.005 |
| Lambda (λ) | KR | AG |
|---|---|---|
| 100 | −0.064 ± 0.005 | +0.250 ± 0.001 |
| 500 | −0.060 ± 0.002 | +0.253 ± 0.002 |
| 1000 | −0.056 ± 0.002 | +0.253 ± 0.002 |
| 5000 | −0.053 ± 0.002 | +0.255 ± 0.001 |
| 10,000 | −0.062 ± 0.003 | +0.254 ± 0.001 |
| Replay Ratio | KR | AG |
|---|---|---|
| 0.1 | −0.060 ± 0.007 | +0.253 ± 0.002 |
| 0.2 | −0.054 ± 0.005 | +0.258 ± 0.001 |
| 0.4 | −0.044 ± 0.004 | +0.261 ± 0.002 |
| 0.6 | −0.035 ± 0.006 | +0.263 ± 0.003 |
| 0.8 | −0.037 ± 0.010 | +0.261 ± 0.001 |
| Historical | Contemporary | |||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Method | F1 | Acc | Prec | Rec | F1 | Acc | Prec | Rec | ||||
| Original | 0.847 | 0.897 | 0.842 | 0.852 | 0.713 | 0.732 | 0.818 | 0.732 | ||||
| Naïve FT | 0.785 ± 0.007 | 0.833 ± 0.008 | 0.763 ± 0.006 | 0.843 ± 0.005 | 0.965 ± 0.002 | 0.965 ± 0.002 | 0.965 ± 0.002 | 0.965 ± 0.002 | ||||
| ER | 0.812 ± 0.007 | 0.860 ± 0.007 | 0.789 ± 0.007 | 0.854 ± 0.007 | 0.976 ± 0.002 | 0.976 ± 0.002 | 0.976 ± 0.002 | 0.976 ± 0.002 | ||||
| EWC | 0.795 ± 0.002 | 0.838 ± 0.003 | 0.772 ± 0.002 | 0.861 ± 0.004 | 0.968 ± 0.001 | 0.968 ± 0.001 | 0.968 ± 0.001 | 0.968 ± 0.001 | ||||
| LoRA | 0.792 ± 0.001 | 0.838 ± 0.001 | 0.769 ± 0.001 | 0.852 ± 0.001 | 0.868 ± 0.001 | 0.869 ± 0.001 | 0.880 ± 0.001 | 0.869 ± 0.001 | ||||
| LoRA+ER | 0.778 ± 0.003 | 0.822 ± 0.003 | 0.757 ± 0.002 | 0.850 ± 0.002 | 0.914 ± 0.001 | 0.914 ± 0.001 | 0.917 ± 0.001 | 0.914 ± 0.001 | ||||
| Full+ER+EWC | 0.813 ± 0.005 | 0.860 ± 0.005 | 0.790 ± 0.005 | 0.856 ± 0.004 | 0.976 ± 0.001 | 0.976 ± 0.001 | 0.976 ± 0.001 | 0.976 ± 0.001 | ||||
| LoRA+EWC | 0.805 ± 0.002 | 0.850 ± 0.002 | 0.780 ± 0.002 | 0.860 ± 0.001 | 0.870 ± 0.001 | 0.871 ± 0.001 | 0.884 ± 0.001 | 0.871 ± 0.001 | ||||
| LoRA+ER+EWC | 0.784 ± 0.002 | 0.828 ± 0.002 | 0.762 ± 0.002 | 0.852 ± 0.003 | 0.905 ± 0.003 | 0.905 ± 0.003 | 0.909 ± 0.002 | 0.905 ± 0.003 | ||||
| Mixed 20–80 | Mixed 40–60 | SimuReal | ||||||||||
| Method | F1 | Acc | Prec | Rec | F1 | Acc | Prec | Rec | F1 | Acc | Prec | Rec |
| Original | 0.798 | 0.802 | 0.829 | 0.802 | 0.691 | 0.708 | 0.768 | 0.708 | 0.817 | 0.888 | 0.834 | 0.803 |
| Naïve FT | 0.883 ± 0.004 | 0.883 ± 0.004 | 0.885 ± 0.004 | 0.883 ± 0.004 | 0.908 ± 0.003 | 0.909 ± 0.003 | 0.916 ± 0.003 | 0.909 ± 0.003 | 0.825 ± 0.008 | 0.871 ± 0.007 | 0.798 ± 0.008 | 0.880 ± 0.004 |
| ER | 0.909 ± 0.005 | 0.909 ± 0.005 | 0.910 ± 0.005 | 0.909 ± 0.005 | 0.932 ± 0.005 | 0.932 ± 0.005 | 0.936 ± 0.005 | 0.932 ± 0.005 | 0.842 ± 0.005 | 0.889 ± 0.004 | 0.818 ± 0.005 | 0.880 ± 0.006 |
| EWC | 0.894 ± 0.003 | 0.895 ± 0.003 | 0.899 ± 0.004 | 0.895 ± 0.003 | 0.910 ± 0.004 | 0.911 ± 0.004 | 0.919 ± 0.003 | 0.911 ± 0.004 | 0.833 ± 0.002 | 0.876 ± 0.002 | 0.804 ± 0.002 | 0.893 ± 0.004 |
| LoRA | 0.854 ± 0.002 | 0.854 ± 0.002 | 0.854 ± 0.002 | 0.854 ± 0.002 | 0.822 ± 0.002 | 0.822 ± 0.002 | 0.824 ± 0.002 | 0.822 ± 0.002 | 0.802 ± 0.001 | 0.857 ± 0.001 | 0.778 ± 0.001 | 0.847 ± 0.001 |
| LoRA+ER | 0.870 ± 0.001 | 0.870 ± 0.001 | 0.873 ± 0.001 | 0.870 ± 0.001 | 0.852 ± 0.002 | 0.852 ± 0.002 | 0.853 ± 0.002 | 0.852 ± 0.002 | 0.803 ± 0.002 | 0.853 ± 0.002 | 0.776 ± 0.002 | 0.860 ± 0.002 |
| Full+ER+EWC | 0.911 ± 0.003 | 0.911 ± 0.003 | 0.912 ± 0.003 | 0.911 ± 0.003 | 0.933 ± 0.004 | 0.933 ± 0.004 | 0.937 ± 0.003 | 0.933 ± 0.004 | 0.844 ± 0.004 | 0.890 ± 0.004 | 0.820 ± 0.005 | 0.883 ± 0.004 |
| LoRA+EWC | 0.864 ± 0.001 | 0.864 ± 0.001 | 0.864 ± 0.001 | 0.864 ± 0.001 | 0.829 ± 0.001 | 0.829 ± 0.001 | 0.833 ± 0.001 | 0.829 ± 0.001 | 0.813 ± 0.002 | 0.866 ± 0.002 | 0.789 ± 0.003 | 0.853 ± 0.001 |
| LoRA+ER+EWC | 0.863 ± 0.005 | 0.863 ± 0.005 | 0.865 ± 0.006 | 0.863 ± 0.005 | 0.843 ± 0.003 | 0.843 ± 0.003 | 0.843 ± 0.003 | 0.843 ± 0.003 | 0.806 ± 0.001 | 0.857 ± 0.001 | 0.780 ± 0.001 | 0.858 ± 0.001 |
| Method | KR | AG | Historical F1 | Contemporary F1 |
|---|---|---|---|---|
| Naïve FT | −0.062 ± 0.007 | +0.253 ± 0.002 | 0.785 ± 0.007 | 0.965 ± 0.002 |
| ER | −0.035 ± 0.007 | +0.264 ± 0.002 | 0.812 ± 0.007 | 0.976 ± 0.002 |
| EWC | −0.052 ± 0.002 | +0.255 ± 0.001 | 0.795 ± 0.002 | 0.968 ± 0.001 |
| LoRA | −0.055 ± 0.001 | +0.155 ± 0.001 | 0.792 ± 0.001 | 0.868 ± 0.001 |
| LoRA+ER | −0.069 ± 0.003 | +0.202 ± 0.001 | 0.778 ± 0.003 | 0.914 ± 0.001 |
| Full+ER+EWC | −0.034 ± 0.005 | +0.264 ± 0.001 | 0.813 ± 0.005 | 0.976 ± 0.001 |
| LoRA+EWC | −0.043 ± 0.002 | +0.158 ± 0.001 | 0.805 ± 0.002 | 0.870 ± 0.001 |
| LoRA+ER+EWC | −0.063 ± 0.002 | +0.192 ± 0.003 | 0.784 ± 0.002 | 0.905 ± 0.003 |
| Method | F1-Macro | F1-OFF | F1-NOT | Parameters | Training Time |
|---|---|---|---|---|---|
| Full+ER+EWC | 0.844 | 0.760 | 0.929 | 135 M (100%) | 69.0 s |
| ER | 0.842 | 0.757 | 0.928 | 135 M (100%) | 56.4 s |
| EWC | 0.833 | 0.748 | 0.917 | 135 M (100%) | 28.6 s |
| Naïve FT | 0.825 | 0.735 | 0.915 | 135 M (100%) | 23.5 s |
| LoRA+EWC | 0.813 | 0.713 | 0.913 | 591 K (0.44%) | 21.2 s |
| LoRA+ER+EWC | 0.806 | 0.707 | 0.906 | 591 K (0.44%) | 50.0 s |
| LoRA+ER | 0.803 | 0.703 | 0.902 | 591 K (0.44%) | 49.4 s |
| LoRA | 0.802 | 0.699 | 0.906 | 591 K (0.44%) | 20.9 s |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Asiri, A.; Saleh, M. Continual Learning for Saudi-Dialect Offensive-Language Detection Under Temporal Linguistic Drift. Information 2026, 17, 99. https://doi.org/10.3390/info17010099
Asiri A, Saleh M. Continual Learning for Saudi-Dialect Offensive-Language Detection Under Temporal Linguistic Drift. Information. 2026; 17(1):99. https://doi.org/10.3390/info17010099
Chicago/Turabian StyleAsiri, Afefa, and Mostafa Saleh. 2026. "Continual Learning for Saudi-Dialect Offensive-Language Detection Under Temporal Linguistic Drift" Information 17, no. 1: 99. https://doi.org/10.3390/info17010099
APA StyleAsiri, A., & Saleh, M. (2026). Continual Learning for Saudi-Dialect Offensive-Language Detection Under Temporal Linguistic Drift. Information, 17(1), 99. https://doi.org/10.3390/info17010099

