Detecting Fake News in Urdu Language Using Machine Learning, Deep Learning, and Large Language Model-Based Approaches
Abstract
1. Introduction
- To the best of our knowledge, it is the first time that the performance of the methods used for fake news identification has been compared for the Urdu language in conjunction with deep learning algorithms. In particular, we have performed more than 30 experiments using the existing two Urdu corpora, such as Bend the Truth and Ax-to-Grind. The TFIDF, ACC, ACC2, OR, GINI, and NDM features, with three traditional supervised learning techniques. The traditional techniques include Gaussian Naïve Bayes (GNB), Burnolie Naïve Bayes (BNB), and Support Vector Machine (SVM). Whereas, the deep learning techniques include Long Short-Term Memory (LSTM), and CNN, which uses Embedding, pre-trained Word2vec embedding, and One-Hot Encoding.
- In this study, the fine-tuning of the BERT and GPT2 LLMswas performed. Fine-tuning refers to the process of fine-tuning a model that has been trained on a large dataset and is then adjusted for a specific task. In this research, we used the BERT and GPT2 models and trained them on the Bend the Truth and Ax-to-Grind datasets for fake news identification in Urdu. This process helps the model to capture the language pattern in the Urdu language and various writing styles, and then helps in the process of fake news identification of the Urdu text.
- The performances of supervised learning models and neural network sequential models are compared.
2. Related Work
Ref. | Dataset | Algorithm | Description | Limitations of the Study | Authentic Source | Output |
---|---|---|---|---|---|---|
[28] | Urdu four different Datasets | LLM, and Linear SVM | Linear SVM with TFIDF and LLM | Limited to LLM like GPT and LSVM with TFIDF and does not explore Embeddings | Yes | 87% Significant |
[29] | Urdu Tweets | SVM, LR, DT, RF, NB, RNN, CNN | TFIDF with ML models | Limited to ML models and does not explore other models | yes | 91% Significant |
[15] | Urdu | Stacked Logistic Regression, SVM and Random Forest | BoW with Ngram and Tf-IDF with Stacked LR, Random Forest, SVM | Limited to traditional Machine Learning techniques, did not apply deep learning and transfer learning approaches | Yes | 93% F1 with 5fold validation on Urdu data Significant |
[16] | Urdu | Logistic Regression, SVM, and Multimodal | TF_IDF with LR, N-gram with SVM, and Transfer Learning with Multimodal | Achieved lower performance on Transfer Learning with Multimodal | Yes | 72% F1, 66%, and Accurate result on textual data Significant |
[14] | Urdu News Articles | Multinomial Naïve Bayes, Bernoulli Naïve Bayes, support vector machines | TF-IDF, N-gram Features with Naïve Bayes and Support Vector Machine | Used only traditional Machine Learning Models and does not explore deep learning or transfer learning | Yes | 87% F1 fake and 90% F1 legit Significant |
[5] | Urdu Machine Translated Dataset | Ada-Boost classifier | N-gram Features with Ada-Boost classifier | Limited to traditional classifier and does not explore other models | Yes | 84% F1 fake and 94% ROC-AUC Significant |
[19] | Urdu | Logistic Regression, Random Forest, Naïve Bayes, DNN, CNN, Bi-LSTM, and Transfer Learning BERT Model | N-gram and TFIDF Features with Machine Learning Classifier | Focus mainly on BERT model and does not applied other models like GPT | Yes | 87% F1 achieved by BERT model Significant |
[30] | Spanish Dataset | Random Forest Classifier | BOW, POS tags with Random Forest Classifier | Used only traditional Machine Learning Classifier | Yes | 76% Significant |
[24] | English | Naïve Bayes, SVM and Semantic Analysis Classifier | Hybrid Algorithm of Naïve Bayes and SVM | Limited to only the ML model and does not apply Deep learning | No | Not Significant |
[25] | English Weibo, twitter15 and twitter16 Dataset | Propagation path classification (PPC), CNN and RNN | PPC_RNN, PPC_CNN and PPC_CNN+RNN | Only works for short text from social media and does not generalize to long text from news articles | Yes | 92% Significant |
[26] | Urdu Dataset | XLNet Model | XLNet Pre-trained Model with 6 Layers and 4 Attention Heads | Limited to XLNet model and does not explore other models | Yes | 83% Significant |
[20] | English | CNN and LSTM | Word2Vec Word Embedding with CNN and LSTM | Used only Word2Vec with DL models and does not explore the BERT model | Yes | 97% Significant |
[18] | English | SVM and BiGRU | ELMo Embedding Model with BiGRU | Performance is lower than the recent embedding modes like BERT | Yes | 76% Significant |
[27] | Spanish and English | SVM and Logistic Regression | TF-IDF, N-gram Features with SVM and Logistic Regression | Limited to traditional Machine Learning models and does not explore neural networks or transformers | Yes | 79% Spanish 73% English Significant |
[17] | English | SGD, DT, LR, and KNN | TF-IDF, N-gram Features with SGD, DT, LR, and KNN | Does not explore Neural Networks or transformers | Yes | 92% Significant |
3. Materials and Methods
3.1. Layer 1: Data Acquisition
3.2. Layer 2: Data Pre-Processing
3.3. Layer 3: Feature Extraction and Modeling
3.3.1. Feature Extraction
3.3.2. Term Frequency
3.3.3. Term Frequency-Inverse Document Frequency (Tf-Idf)
3.3.4. Relative Frequency
3.3.5. Raw Frequency
3.3.6. Accuracy (ACC)
3.3.7. Balanced Accuracy (ACC2)
3.3.8. Odds Ratio
3.3.9. Gini Index
3.3.10. Normalized Difference Measure
- An important term should have a high |tpr − fpr| value.
- One of the tpr or fpr values should be closer to zero.
- If two terms have equal |tpr − fpr| values, then the term
- having a lower min (tpr, fpr) value should be assigned a higher rank, where min is the function to find a minimum of the two values.
3.3.11. Word Embedding
3.4. Model Building
3.5. Machine Learning
3.5.1. Support Vector Machine (SVM)
Algorithm 1: Support Vector Machine |
Given: The training dataset with features vectors , Where: • are the feature vectors of -th training sample. • are class labels. Steps Initialize Parameters: • Define a hyperplane in the transformed feature space using a Kernel function . Select a Kernel Function: • To select a kernel function and compute the similarity between samples and . Linear Kernel: Optimize the problem and Compute Decision Boundary: • To find the Lagrange multipliers for each training sample. • Calculate the weight vector and bias in the transformed feature space. Classify Test Samples: • For a given test sample X, predict its class based on sign . Output: Predict the test class for each test sample. |
3.5.2. Naïve Bayes
Algorithm 2: Naïve Bayes |
Given: are class labels. For each class Procedure: • Calculate prior probabilities: • Compute the prior probability as the proportion of the training samples belonging to . • Estimate the likelihood for training feature given : • Calculate the likelihood which represents the probability of observing feature given class For each test sample : • Compute the posterior probability: • Compute the likelihood probability for test class using the prior probability and likelihood. • Predict the output class: Output: Predict the test class for each test sample. |
3.6. Deep Learning
3.6.1. Convolution Neural Network (CNN)
Algorithm 3: Convolution Neural Network |
Given: Input data, such as sequential data. Procedure: Input Layer: • Accept the input data as a multi-dimensional array. Convolutional Layer: • Apply kernel to extract features from the input data by performing convolutional operations. • Use activation function to introduce nonlinearity Pooling and Fully Connected Layer: • Apply pooling layer to down sampling the features. • Flatten the features and pass as fully connected. Output Layer: • Use the activation function to produce class probabilities. Output: Predict the class labels. |
3.6.2. Long Short-Term Memory (LSTM)
3.7. Transformers
BERT and GPT2
Algorithm 4: Transformer |
Given: Input sequence . Procedure: Input Embeddings and Positional Encoding: • Encoded the given input tokens to dense representations. • Positional encoding addition to learn sequential data. Encoder Layer: For encoder layer repeat the following steps: • Apply multi-head self-attention , where Q, K, V are query, key, and a value matrices computed from inputs embeddings. • Apply feed-forward-network , where W is weight of the network and b is bias. • Layer normalization and residual connections have to made after every block. Decoder Layer: • Similar to the encoder, but includes an additional cross-attention layer that attends to encoder outputs. Output Layer: • A softmax activation function is used in order to predict the next token or the next sequence of the output. Output: the last prediction for the sequence. |
3.8. Layer 4: Evaluation Metrics
4. Experiments
5. Results
5.1. Traditional Machine Learning
5.2. Deep Learning Neural Network
5.3. Transformers (BERT and GPT-2)
5.4. Comparative Analysis of Datasets and Model Performance
6. Limitations of the Study
7. Implications of the Study
8. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
NLP | Natural Language Processing |
BERT | Bidirectional Encoder Representations from Transformers |
GPT | Generative Pre-trained Transformer |
CNN | Convolution Neural Network |
LSTM | Long Short-Term Memory |
GRU | Gated Recurrent Unit |
RNN | Recurrent Neural Networks |
MR | Machine Translation |
TF-IDF | Term Frequency-Inverse Document Frequency |
SGD | Stochastic Gradient Descent |
DT | Decision Tree |
LSVM | Linear Support Vector Machine |
LR | Logistic Regression |
SVM | Support Vector Machine |
KNN | K-Nearest Neighbor |
CSV | Comma-Separated Values |
References
- Hunt, E. What Is Fake News. Available online: https://www.theguardian.com/media/2016/dec/18/what-is-fake-news-pizzagate (accessed on 17 December 2016).
- Schlesinger, R. Fake News in Reality. Available online: https://www.usnews.com/opinion/thomas-jefferson-street/articles/2017-04-14/what-is-fake-news-maybe-not-what-you-think (accessed on 14 April 2017).
- Soll, J. The Long and Brutal Story of Fake News. 2016. Available online: https://www.politico.com/magazine/story/2016/12/fake-news-history-long-violent-214535/ (accessed on 20 April 2025).
- Webster, M. The Real Story of Fake News. Available online: https://www.merriam-webster.com/words-at-play/the-real-story-of-fake-news (accessed on 23 March 2017).
- Amjad, M.; Instituto, A.Z.; Adorno, H.G. ‘Bend the truth’: Benchmark dataset for fake news detection in Urdu language and its evaluation. J. Intell. Fuzzy Syst. 2020, 39, 2457–2469. [Google Scholar] [CrossRef]
- Daud, A.; Khan, W.; Che, D. Urdu language processing: A survey. Artif. Intell. Rev. 2017, 47, 279–311. [Google Scholar] [CrossRef]
- Javed, K.; Babri, H.A. Feature selection based on a normalized difference measure for text classification. Inf. Process. Manag. 2017, 53, 473–489. [Google Scholar] [CrossRef]
- Wang, S.; Zhou, W.; Jiang, C. A survey of word embeddings based on deep learning. Computing 2019, 102, 717–740. [Google Scholar] [CrossRef]
- Bondielli, A.; Marcelloni, F. A survey on fake news and rumour detection techniques. Inf. Sci. 2019, 497, 38–55. [Google Scholar] [CrossRef]
- Omran, H.; Elafandy, R.T.; Arsalan, M.; Salama, K.N. Direct Mismatch Characterization of Femtofarad Capacitors. IEEE Trans. Circuits Syst. II Express Briefs 2016, 63, 151–155. [Google Scholar] [CrossRef]
- Sinoara, R.A.; Camacho-Collados, J.; Rossi, R.G.; Navigli, R.; Rezende, S.O. Knowledge-enhanced document embeddings for text classification. Knowl. Based Syst. 2019, 163, 955–971. [Google Scholar] [CrossRef]
- García-Díaz, J.A.; Cánovas-García, M.; Colomo-Palacios, R.; Valencia-García, R. Detecting misogyny in Spanish tweets. An Approach Based on Linguistics Features and Word Embeddings. 2020. Available online: https://www.sciencedirect.com/science/article/abs/pii/S0167739X20301928 (accessed on 10 July 2024).
- Tanishka, G. The Power of BERT: Revolutionizing NLP. Nash Tech. Available online: https://blog.nashtechglobal.com/the-power-of-bert-revolutionizing-nlp/ (accessed on 10 July 2024).
- Amjad, M.; Sidorov, G.; Zhila, A. Data augmentation using machine translation for fake news detection in the Urdu language. In Proceedings of the LREC 2020—12th International Conference on Language Resources and Evaluation, Marseille, France, 13–15 May 2020; pp. 2537–2542. [Google Scholar]
- Farooq, M.S.; Naseem, A.; Rustam, F.; Ashraf, I. Fake news detection in Urdu language using machine learning. PeerJ Comput. Sci. 2023, 9, e1353. [Google Scholar] [CrossRef] [PubMed]
- Salahuddin, R.; Wasim, M. Automatic identification of Urdu fake news using Logistic Regression Model. In Proceedings of the 2022 16th International Conference on Open Source Systems and Technologies (ICOSST), Lahore, Pakistan, 14–15 December 2022. [Google Scholar]
- Ahmed, H.; Traore, I.; Saad, S. Intelligent, Secure, and Dependable Systems in Distributed and Cloud Environments. In Proceedings of the First International Conference on Intelligent, Secure, and Dependable Systems in Distributed and Cloud Environments, Vancouver, BC, Canada, 26–28 October 2017; Volume 10618, pp. 169–181. [Google Scholar] [CrossRef]
- Saikh, T.; De, A.; Ekbal, A.; Bhattacharyya, P. A Deep Learning Approach for Automatic Detection of Fake News. arXiv 2020, arXiv:2005.04938. Available online: http://arxiv.org/abs/2005.04938 (accessed on 11 May 2020).
- Bhawal, S.; Roy, P.K. Fake News Detection in Urdu Language Using BERT. 2021. Available online: http://ceur-ws.org (accessed on 20 April 2025).
- Krešňáková, V.M.; Sarnovský, M.; Butka, P. Deep learning methods for Fake News detection. In Proceedings of the IEEE 19th International Symposium on Computational Intelligence and Informatics and 7th IEEE International Conference on Recent Achievements in Mechatronics, Automation, Computer Sciences and Robotics (CINTI-MACRo), Szeged, Hungary, 14–16 November 2019; pp. 143–148. [Google Scholar]
- Imran, I. CNN-RNN Deep Learning Networks for Pattern Recognition Problems. 2023. Available online: https://ieeexplore.ieee.org/abstract/document/10111363 (accessed on 20 April 2025).
- Tyagi, A.A.A.K. Recurrent Neural Networks: Concepts and Applications. 2022. Available online: https://books.google.com.pk/books?hl=en&lr=&id=5phxEAAAQBAJ&oi=fnd&pg=PR2&dq=Tyagi,+A.+K.,+%26+Abraham,+A.+(Eds.).+(2022).+Recurrent+Neural+Networks:+Concepts+and+Applications.&ots=xPqhl8xZXG&sig=VQUixsNOjMN0kQo-Z8KHBK-yJy8&redir_esc=y#v=onepage&q=Tyagi%252 (accessed on 20 April 2025).
- Kalra, G.C.S.; Kumar, C.H.; Sharma, Y. Multimodal Fake News Detection on Fakeddit Dataset Using Transformer-Based Architectures. In Machine Learning, Image Processing, Network Security and Data Sciences; Springer: Cham, Switzerland, 2022. [Google Scholar] [CrossRef]
- Stahl, K. Fake News Detector in Online Social Media. Int. J. Eng. Adv. Technol. 2019, 9, 58–60. [Google Scholar] [CrossRef]
- Liu, Y.; Wu, Y.B. Early Detection of Fake News on Social Media Through Propagation Path Classification with Recurrent and Convolutional Networks. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 354–361. [Google Scholar]
- Khilji, A.F.U.R.; Laskar, S.R.; Pakray, P.; Bandyopadhyay, S. Urdu fake news detection using generalized autoregressors. CEUR Workshop Proc. 2020, 2826, 452–457. [Google Scholar]
- Vogel, I.; Meghana, M. Fake News Spreader Detection on Twitter Using Character N-Grams Notebook for PAN at CLEF 2020. CEUR Workshop Proc. 2020, 2696, 22–25. [Google Scholar]
- Ali, M.Z.; Wang, Y.; Pfahringer, B.; Smith, T. Detection of Human and Machine-Authored Fake News in Urdu. arXiv 2024, arXiv:2410.19517. Available online: http://arxiv.org/abs/2410.19517 (accessed on 25 October 2024).
- Iqbal, Z.; Khan, F.M.; Khan, I.U.; Khan, I.U. Fake News Identification in Urdu Tweets Using Machine Learning Models. Asian Bull. Big Data Manag. 2024, 4, 52–69. [Google Scholar] [CrossRef]
- Pablo, J.; Gomez-Adorno; Sidorov, G.; Escobar, J.J.M. Detection of fake news in a new corpus for the Spanish language. J. Intell. Fuzzy Syst. 2019, 36, 4868–4876. [Google Scholar] [CrossRef]
- Amjad, M. Fake News Detection in Udru Language. Available online: https://www.urdufake2020.cicling.org/ (accessed on 20 December 2020).
- Harris, S.; Liu, J.; Hadi, H.J.; Cao, Y. Ax-to-Grind Urdu: Benchmark Dataset for Urdu Fake News Detection. Available online: https://arxiv.org/abs/2403.14037 (accessed on 20 March 2024).
- Stanfordnlp Stanfordnlp. Available online: https://stanfordnlp.github.io/stanfordnlp/models.html (accessed on 5 October 2018).
- Sanchhaya Education Private Limited. Support Vector Machine (SVM) Algorithm; GeeksforGeeks: Noida, India, 2025. [Google Scholar]
- Collobert, R.; Weston, J.; Bottou, L.; Karlen, M.; Kavukcuoglu, K.; Kuksa, P. Natural Language Processing (Almost) from Scratch. J. Mach. Learn. Res. 2011, 12, 2493–2537. [Google Scholar]
- Emanuilov, S. The Transformer Revolution; UnfoldAI Building ML System: Roma, Italy, 2024. [Google Scholar]
- Sanchhaya Education Private Limited. BERT Model—NLP; GeeksforGeeks: Noida, India, 2020. [Google Scholar]
- Harrag, F.; Debbah, M.; Darwish, K.; Abdelali, A. BERT Transformer Model for Detecting Arabic GPT2 Auto-Generated Tweets. Available online: https://www.wired.com/story/ai-generated-text-is-the-scariest-deepfake-of-all/ (accessed on 20 April 2025).
Training | Testing | Total | |
---|---|---|---|
Real | 350 | 150 | 500 |
Fake | 288 | 112 | 400 |
Total | 638 | 262 | 900 |
Category | Real | Fake |
---|---|---|
Business | 100 | 50 |
Health | 100 | 100 |
Showbiz | 100 | 100 |
Sports | 100 | 50 |
Technology | 100 | 100 |
Totals | 500 | 400 |
Category | Train | Test | ||
---|---|---|---|---|
Real | Fake | Real | Fake | |
Business | 4640 | 1939 | 2822 | 862 |
Health | 3825 | 3454 | 2283 | 2091 |
Showbiz | 3695 | 3851 | 2919 | 2953 |
Sports | 4948 | 2178 | 3365 | 536 |
Technology | 4494 | 4679 | 2448 | 2458 |
Totals | 13,250 | 10,115 | 8848 | 6610 |
True News | Fake News | Total | |
---|---|---|---|
Training | 3018 | 3031 | 6049 |
Validation | 1006 | 1011 | 2017 |
Testing | 1006 | 1011 | 2017 |
Total | 5030 | 5053 | 10,083 |
Label | Total Documents | Total Words | Unique Words | Avg Sentence Length |
---|---|---|---|---|
True | 5030 | 173,267 | 13,877 | 34.45 |
Fake | 5053 | 585,299 | 27,037 | 115.83 |
Total | 10,083 | 758,566 | 30,448 |
Before Tokenization | After Tokenization | |
---|---|---|
Noun | عمران خان ,سعودی عرب اسلام آباد | ’آباد‘ ,’اسلام‘ ,’عرب‘ ,’سعودی‘ ,’خان‘ ,’عمران‘ |
Words | معاشی اتفاق اعلامیے قر یب | ’معا‘ ,’شی‘ ’اتفا‘ ,’ق‘ ’اعلا‘ ,’میے‘ ’یب‘ ,’قر‘ |
Model | Learning Rate | Batch Size | Optimizer | Loss Function |
---|---|---|---|---|
BERT | 1.00 × 10−5 | 16 | AdamW | CrossEntropyLoss |
GPT2 | 2.00 × 10−5 | 16 | AdamW | CrossEntropyLoss |
Support Vector Machine | ||||||||
---|---|---|---|---|---|---|---|---|
Feature Techniques | Accuracy | Precision | Recall | F1 Score | ||||
Bend the Truth | Ax-to-Grind | Bend the Truth | Ax-to-Grind | Bend the Truth | Ax-to-Grind | Bend the Truth | Ax-to-Grind | |
ACC | 0.54 | 0.63 | 0.56 | 0.60 | 0.91 | 0.77 | 0.69 | 0.67 |
ACC2 | 0.59 | 0.72 | 0.59 | 0.71 | 0.90 | 0.74 | 0.71 | 0.72 |
GINI | 0.60 | 0.67 | 0.60 | 0.63 | 0.90 | 0.84 | 0.72 | 0.72 |
NDM | 0.56 | 0.72 | 0.57 | 0.70 | 0.98 | 0.74 | 0.72 | 0.72 |
OR | 0.57 | 0.72 | 0.57 | 0.71 | 1.0 | 0.74 | 0.72 | 0.72 |
TFIDF | 0.58 | 0.88 | 0.57 | 0.88 | 1.0 | 0.87 | 0.73 | 0.87 |
TFIDF-bigram | 0.58 | 0.89 | 0.57 | 0.90 | 1.0 | 0.87 | 0.73 | 0.89 |
Binary Frequency | 0.51 | 0.88 | 0.54 | 0.90 | 0.86 | 0.86 | 0.67 | 0.88 |
Raw Frequency | 0.59 | 0.88 | 0.59 | 0.87 | 0.98 | 0.88 | 0.73 | 0.88 |
Relative Frequency | 0.57 | 0.60 | 0.57 | 0.60 | 1.0 | 0.60 | 0.72 | 0.60 |
Bernoulli Naïve Bayes | ||||||||
ACC | 0.57 | 0.70 | 0.57 | 0.87 | 1.0 | 0.70 | 0.72 | 0.79 |
ACC2 | 0.57 | 0.72 | 0.57 | 0.71 | 1.0 | 0.74 | 0.72 | 0.72 |
GINI | 0.57 | 0.72 | 0.57 | 0.71 | 1.0 | 0.74 | 0.72 | 0.72 |
NDM | 0.57 | 0.72 | 0.57 | 0.71 | 1.0 | 0.74 | 0.72 | 0.72 |
OR | 0.57 | 0.71 | 0.57 | 0.71 | 1.0 | 0.73 | 0.72 | 0.72 |
TFIDF | 0.61 | 0.67 | 0.80 | 0.61 | 0.44 | 0.61 | 0.56 | 0.74 |
TFIDF-bigram | 0.61 | 0.70 | 0.80 | 0.63 | 0.44 | 0.94 | 0.56 | 0.76 |
Binary Frequency | 0.49 | 0.68 | 0.53 | 0.61 | 0.79 | 0.94 | 0.64 | 0.74 |
Raw Frequency | 0.62 | 0.68 | 0.78 | 0.61 | 0.47 | 0.94 | 0.59 | 0.74 |
Relative Frequency | 0.62 | 0.68 | 0.78 | 0.61 | 0.47 | 0.94 | 0.59 | 0.74 |
Gaussian Naïve Bayes | ||||||||
ACC | 0.45 | 0.57 | 0.63 | 0.54 | 0.12 | 0.97 | 0.21 | 0.69 |
ACC2 | 0.45 | 0.55 | 0.63 | 0.53 | 0.12 | 0.98 | 0.21 | 068 |
GINI | 0.58 | 0.56 | 0.59 | 0.52 | 0.88 | 0.98 | 0.71 | 0.68 |
NDM | 0.58 | 0.59 | 0.59 | 0.55 | 0.88 | 0.96 | 0.71 | 0.70 |
OR | 0.45 | 0.50 | 0.63 | 0.92 | 0.12 | 0.52 | 0.21 | 0.52 |
TFIDF | 0.57 | 0.70 | 0.66 | 0.70 | 0.52 | 0.70 | 0.58 | 0.70 |
TFIDF-bigram | 0.63 | 0.71 | 0.64 | 0.74 | 0.80 | 0.72 | 0.71 | 0.71 |
Binary Frequency | 0.50 | 0.72 | 0.57 | 0.71 | 0.49 | 0.74 | 0.53 | 0.72 |
Raw Frequency | 0.64 | 0.72 | 0.63 | 0.71 | 0.89 | 0.73 | 0.74 | 0.72 |
Relative Frequency | 0.75 | 0.84 | 0.80 | 0.84 | 0.77 | 0.84 | 0.78 | 0.84 |
Convolutional Neural Network (CNN) | ||||
---|---|---|---|---|
Accuracy | Precision | Recall | F1 Score | |
Word2Vec Embedding | 0.68 | 0.68 | 0.78 | 0.72 |
Word Embedding | 0.62 | 0.62 | 0.76 | 0.68 |
Long Short-Term Memory (LSTM) | ||||
Word2Vec Embedding | 0.62 | 0.62 | 0.62 | 0.62 |
Word Embedding | 0.57 | 0.57 | 0.57 | 0.57 |
Convolutional Neural Network (CNN) with Ax-to-Grind Urdu Dataset | ||||
Accuracy | Precision | Recall | F1 Score | |
Word Embedding | 0.86 | 0.87 | 0.86 | 0.86 |
Long Short-Term Memory (LSTM) with Ax-to-Grind Urdu Dataset | ||||
Accuracy | Precision | Recall | F1 Score | |
Word Embedding | 0.91 | 0.91 | 0.88 | 0.89 |
Fine Tuning GPT2 Model | ||||
---|---|---|---|---|
Accuracy | Precision | Recall | F1 Score | |
0.66 | 0.66 | 0.66 | 0.66 | |
Fine Tuning Urdu BERT Model | ||||
Training | 0.97 | 0.97 | 0.97 | 0.97 |
Validation | 0.73 | 0.73 | 0.73 | 0.73 |
Testing | 0.79 | 0.79 | 0.79 | 0.79 |
Fine Tuning GPT2 Model with Ax-to-Grind Urdu Dataset | |||
---|---|---|---|
Accuracy | Precision | Recall | F1 Score |
0.93 | 0.93 | 0.93 | 0.93 |
Fine Tuning Urdu mBERT Model with Ax-to-Grind Urdu Dataset | |||
Accuracy | Precision | Recall | F1 Score |
0.95 | 0.94 | 0.94 | 0.93 |
Research | News Items | Domains | F1-Score | Accuracy | Precision | Recall |
---|---|---|---|---|---|---|
F1-Fake | F1-Real | |||||
Bend the Truth [32] | 900 | 5 | 0.87 | 0.90 | ||
Bend the Truth (BERT) | 900 | 5 | 0.79 | 0.79 | 0.79 | 0.79 |
Ax-to-Grind Urdu [33] | 10,083 | 15 | 0.924 | 0.956 | 0.942 | 0.940 |
Ax-to-Grind Urdu (BERT) | 10,083 | 15 | 0.936 | 0.957 | 0.944 | 0.949 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Farooq, M.S.; Gilani, S.M.A.; Manzoor, M.F.; Shaheen, M. Detecting Fake News in Urdu Language Using Machine Learning, Deep Learning, and Large Language Model-Based Approaches. Information 2025, 16, 595. https://doi.org/10.3390/info16070595
Farooq MS, Gilani SMA, Manzoor MF, Shaheen M. Detecting Fake News in Urdu Language Using Machine Learning, Deep Learning, and Large Language Model-Based Approaches. Information. 2025; 16(7):595. https://doi.org/10.3390/info16070595
Chicago/Turabian StyleFarooq, Muhammad Shoaib, Syed Muhammad Asadullah Gilani, Muhammad Faraz Manzoor, and Momina Shaheen. 2025. "Detecting Fake News in Urdu Language Using Machine Learning, Deep Learning, and Large Language Model-Based Approaches" Information 16, no. 7: 595. https://doi.org/10.3390/info16070595
APA StyleFarooq, M. S., Gilani, S. M. A., Manzoor, M. F., & Shaheen, M. (2025). Detecting Fake News in Urdu Language Using Machine Learning, Deep Learning, and Large Language Model-Based Approaches. Information, 16(7), 595. https://doi.org/10.3390/info16070595