This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Open AccessArticle
Hallucination-Aware Interpretable Sentiment Analysis Model: A Grounded Approach to Reliable Social Media Content Classification
by
Abdul Rahaman Wahab Sait
Abdul Rahaman Wahab Sait 1,*
and
Yazeed Alkhurayyif
Yazeed Alkhurayyif 2,*
1
Department of Documents and Archive, Center of Documents and Administrative Communication, King Faisal University, P.O. Box 400, Hofuf 31982, Saudi Arabia
2
Applied College, Shaqra University, Shaqra 11961, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Electronics 2026, 15(2), 409; https://doi.org/10.3390/electronics15020409 (registering DOI)
Submission received: 18 December 2025
/
Revised: 14 January 2026
/
Accepted: 15 January 2026
/
Published: 16 January 2026
Abstract
Sentiment analysis (SA) has become an essential tool for analyzing social media content in order to monitor public opinion and support digital analytics. Although transformer-based SA models exhibit remarkable performance, they lack mechanisms to mitigate hallucinated sentiment, which refers to the generation of unsupported or overconfident predictions without explicit linguistic evidence. To address this limitation, this study presents a hallucination-aware SA model by incorporating semantic grounding, interpretability-congruent supervision, and neuro-symbolic reasoning within a unified architecture. The proposed model is based on a fine-tuned Open Pre-trained Transformer (OPT) model, using three fundamental mechanisms: a Sentiment Integrity Filter (SIF), a SHapley Additive exPlanations (SHAP)-guided regularization technique, and a confidence-based lexicon-deep fusion module. The experimental analysis was conducted on two multi-class sentiment datasets that contain Twitter (now X) and Reddit posts. In Dataset 1, the suggested model achieved an average accuracy of 97.6% and a hallucination rate of 2.3%, outperforming the current transformer-based and hybrid sentiment models. With Dataset 2, the framework demonstrated strong external generalization with an accuracy of 95.8%, and a hallucination rate of 3.4%, which is significantly lower than state-of-the-art methods. These findings indicate that it is possible to include hallucination mitigation into transformer optimization without any performance degradation, offering a deployable, interpretable, and linguistically complex social media SA framework, which will enhance the reliability of neural systems of language understanding.
Share and Cite
MDPI and ACS Style
Sait, A.R.W.; Alkhurayyif, Y.
Hallucination-Aware Interpretable Sentiment Analysis Model: A Grounded Approach to Reliable Social Media Content Classification. Electronics 2026, 15, 409.
https://doi.org/10.3390/electronics15020409
AMA Style
Sait ARW, Alkhurayyif Y.
Hallucination-Aware Interpretable Sentiment Analysis Model: A Grounded Approach to Reliable Social Media Content Classification. Electronics. 2026; 15(2):409.
https://doi.org/10.3390/electronics15020409
Chicago/Turabian Style
Sait, Abdul Rahaman Wahab, and Yazeed Alkhurayyif.
2026. "Hallucination-Aware Interpretable Sentiment Analysis Model: A Grounded Approach to Reliable Social Media Content Classification" Electronics 15, no. 2: 409.
https://doi.org/10.3390/electronics15020409
APA Style
Sait, A. R. W., & Alkhurayyif, Y.
(2026). Hallucination-Aware Interpretable Sentiment Analysis Model: A Grounded Approach to Reliable Social Media Content Classification. Electronics, 15(2), 409.
https://doi.org/10.3390/electronics15020409
Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details
here.
Article Metrics
Article metric data becomes available approximately 24 hours after publication online.