Next Article in Journal
Extracting Flow Characteristics from Single and Multi-Point Time Series Through Correlation Analysis
Previous Article in Journal
Classification of Common Bean Landraces of Three Species Using a Neuroevolutionary Approach with Probabilistic Color Characterization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Framework for Mental Illness Detection Leveraging TOPSIS-ModCHI-Based Feature-Driven Randomized Neural Networks

by
Santosh Kumar Behera
and
Rajashree Dash
*
Department of Computer Science and Engineering, Siksha O Anusandhan University, Bhubaneswar 751030, Odisha, India
*
Author to whom correspondence should be addressed.
Math. Comput. Appl. 2025, 30(4), 67; https://doi.org/10.3390/mca30040067
Submission received: 10 May 2025 / Revised: 28 June 2025 / Accepted: 28 June 2025 / Published: 30 June 2025

Abstract

Mental illness has emerged as a significant global health crisis, inflicting immense suffering and causing a notable decrease in productivity. Identifying mental health disorders at an early stage allows healthcare professionals to implement more targeted and impactful interventions, leading to a significant improvement in the overall well-being of the patient. Recent advances in Artificial Intelligence (AI) have opened new avenues for analyzing medical records and behavioral data of patients to assist mental health professionals in their decision-making processes. In this study performance of four Randomized Neural Networks (RandNNs) such as Board Learning System (BLS), Random Vector Functional Link Network (RVFLN), Kernelized RVFLN (KRVFLN), and Extreme Learning Machine (ELM) are explored for detecting the type of mental illness a user may have by analyzing the random text of the user posted on social media. To improve the performance of the RandNNs during handling the text documents with unbalanced class distributions, a hybrid feature selection (FS) technique named as TOPSIS-ModCHI is suggested in the preprocessing stage of the classification framework. The effectiveness of the suggested FS with all the four randomized networks is assessed over the publicly available Reddit Mental Health Dataset after experimenting on two benchmark multiclass unbalanced datasets. From the experimental results, it is inferred that detecting the mental illness using BLS with TOPSIS-ModCHI produces the highest precision value of 0.92, recall value of 0.66, f-measure value of 0.77, and Hamming loss value of 0.06 as compared to ELM, RVFLN, and KRVFLN with a minimum feature size of 900. Overall, utilizing BLS for mental health analysis can offer a promising avenue toward improved interventions and a better understanding of mental health issues, aiding in decision-making processes.

1. Introduction

The World Health Organization estimates that suicide claims the lives of over 800,000 people each year, making it the fourth most common cause of death for those aged between 15 to 19 [1]. Amidst the COVID-19 pandemic lockdown, there has been a notable increase in mental health challenges and their portrayal on social media platforms. Economic instability and isolation brought on by the pandemic have long-term impacts on people’s mental health and general well-being. This phenomenon has ramifications for clinical psychologists and academic researchers. The reports from August 2021 definitively state that, in England, there were around 1.6 million people on waiting lists for mental health services [2]. Furthermore, around eight million individuals were unable to access specialist help because they did not meet the criteria for being considered sick enough. This situation highlights the necessity of automating the detection of mental health issues from social media data, where individuals freely express their thoughts, beliefs, and emotions [3]. Users are often inclined to share their mental health difficulties or illnesses anonymously on various social media or in online health communities. These online health communities can serve as a platform for sharing empathy and establishing connections with others who are experiencing similar symptoms. On social media, users also frequently look for health information on their symptoms in order to self-diagnose [4]. These massive reservoirs of user-generated data on social media platforms present an emerging alternate for examining behavioral patterns to predict the mental health of the person. The possible effects of untreated mental distress could be lessened by these predictive capabilities, which could enable prompt interventions from healthcare providers and support systems.

1.1. Research Gaps and Motivations

Questionnaires and in-person consultations with psychological consultants are the conventional methods of detecting mental disease. Despite being a very effective approach, very few patients profit from it due to a lack of enough specialists. New opportunities have been opened by recent advancements in Artificial Intelligence (AI) and Machine Learning (ML) to assist mental health professionals to identify mentally sick people by analyzing their behavioral data available in social media [5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20]. Several works utilizing traditional ML algorithms, such as Logistic Regression (LR), Decision Tree (DT), Random Forest (RF), K-Nearest Neighbor (KNN), Naïve Bayes (NB), and Support Vector Machine (SVM) have contributed to identifying mental disorders from data available on social media [5,6,7,8,9]. The majority of these studies concentrate on creating a predictive model for a single, particular illness, treating it as a binary classification problem. However, in the real scenario, differentiating between several mental illnesses, perhaps in a comorbidity scenario, is a more difficult challenge that necessitates for a multiclass classification solution. More recently various Deep Learning (DL) techniques including Convolutional Neural Networks (CNNs) [10,11,12,13,14] Long Short-Term Memory (LSTM) [13,15,16,17], Recurrent Neural Network (RNN) [17,18], and transformers-based DL models [19,20] have been leveraged to figure out a user’s mental condition from their shared information on social media. Despite the immense capability of the deep structures, the majority of the networks suffer from the time-consuming training procedure due to the inclusion of a large number of connecting parameters in filters and layers with several hyper-parameters and complex structures [21,22]. The intricate nature of the deep structure also poses a significant challenge for theoretical analysis, leading researchers to primarily concentrate on tweaking parameters and adding more layers using more and more powerful computing resources in pursuit of improved accuracy [21,22]. In addition, it goes through a complete retraining process if its structure is inefficient to accurately reflect the system.
In contrast to Deep Neural Networks (DNNs), Single Layer Feed Forward Neural Networks (SLFNs) have also drawn interest from researchers for addressing a variety of regression and classification problems due to their simplistic topologies and universal approximation capabilities. The conventional method of training SLFNs using a Back Propagation (BP)-based iterative approach often encounters challenges such as local minima, sensitivity to learning rate, and slow convergence [23,24]. Further, to tackle these challenges, innovative Randomized Neural Networks (RandNNs) like Extreme Learning Machine (ELM) [25] and Random Vector Functional Link Network (RVFLN) [26,27] have been introduced in the literature. These networks assign arbitrary weights and biases for the hidden layer, while the weights for the output layer are computed analytically. Due to the non-iterative training procedure, these networks not only offer fast training speed, but also provide the generalization capability in function approximation [22,26,27]. Based on the idea of RVFLN, another RandNN termed as Broad Learning System (BLS) is proposed in [21,28] as an alternative way of learning deep features. A flat network is used to set up the BLS, where the initial inputs are relocated and kept as mapped features in feature nodes. These are then changed by a nonlinear activation function to create enhancement nodes through which the BLS’s structure is expanded in a wide sense. The most notable aspect of BLS as a SLFN is its use of mapped feature nodes to build enhancement nodes, which introduces a better feature representation capacity. In contrast to the DNN, BLS does not require a gradual search for the optimal parameters of the models using BP [21]. Finding the weights of the linkages between the nodes of hidden and output layers is achieved in a single step during the learning process of BLS by conducting matrix inversion. Although RandNNs are successfully used for binary and multiclass classifications in a number of fields [22,28], they are rarely used for text classification (TC), particularly in the field of mental health analysis.

1.2. Contributions

With an initiative to bridge the research gaps discussed in the previous section, this study offers a novel solution by developing a RandNN-based text classifier that can efficiently identify a wide array of mental illness, including depression, suicide, anxiety, schizophrenia, autism, post-traumatic stress disorder (PTSD), alcoholism, bipolar disorder, and a neutral category. In contrast to several studies that concentrate on detecting the presence or absence of a single mental health condition, this model aims to classify users into a range of potential mental illnesses with the nuanced expressions and linguistic patterns found in user-generated text by utilizing four RandNNs such as BLS [21,28], RVFLN [26,27], KRVFLN [29], and a ELM [25]. To boost the performance of the RandNNs during handling the text documents with unbalanced class distributions, a hybrid FS technique named as TOPSIS-ModCHI is suggested in the preprocessing stage of the classification framework. Unlike traditional chi-square, which merely uses the highest chi-square value to tank features, ModCHI, being an improved version of it, focuses on selecting features with low repetition and high relevance to class [30,31]. It assigns a score to each feature by dividing the amount of relevant documents of each class by the total number of documents in the dataset and then selects k number of highly scored features, resulting in a better handle of multilabeled text documents with unbalanced class distributions. In [30], the authors have analyzed the performance of four ML techniques, such as linear SVM, DT, RF, and KNN with ModCHI for addressing the multiclass unbalanced TC problem. By experimenting with the benchmark Reuters TC dataset, the effectiveness of the techniques is validated by comparing with four other FS techniques. The work is extended by the author in [31] by assessing the performance of RVFLN with ModCHI for TC, including 20 Newsgroups, one more benchmark dataset. In both the previous approaches of application of ModCHI, the ideal feature size is decided by assessing the performance of each classifier with 10 different feature sizes of ModCHI and then taking the minimal feature size producing the average of those 10 assessments as the ideal one. Being motivated by the use of TOPSIS for ideal feature size selection of ReliefF-based FS in [32], this research presents a further extension of the previous two works [31,32] by utilizing TOPSIS to determine the ideal feature size k of ModCHI rather than arbitrarily determining this value. With 10 different k values considered within a range of 10 to 1500, k number of highly scored features is picked by ModCHI for each RandNN-based classifier, and its performance is evaluated using four criteria such as precision, recall, F-measure, and Hamming loss for each k value. Then TOPSIS is used to rank the feature size and to select the optimal feature size for the classifier. The effectiveness of the suggested FS with all the four RandNNs is assessed over the publicly available Reddit Mental Health Dataset after experimenting on two benchmark multiclass unbalanced datasets, such as Reuters and 20 Newsgroups. Based on the experimental findings, it can be concluded that, when compared to RVFLN, KRVFLN, and ELM with a minimum feature size of 900, BLS with TOPSIS-ModCHI yields better results for each assessment criterion when used to detect mental illness.
The objective of this research is to investigate the effectiveness of RandNNs to tackle the challenges in Mental Health Analysis for unbalanced, multiclass classification. This paper contributes to the field by presenting a novel application of RandNN with a hybrid FS technique, with supporting empirical evidence. The paper’s primary contribution can be summarized as follows:
  • Developing a text classifier using RandNN for efficiently categorizing multiclass unbalanced text documents.
  • Selecting an optimal set of features for the classifier by utilizing a hybrid TOPSIS-ModCHI-based FS technique.
  • Assessing the performance of four RandNNs, including BLS, RVFLN, KRVFLN, and ELM, to determine whether a user exhibits linguistic markers indicative of a specific mental illness by analyzing the user’s posts on social media.
  • Addressing mental illness detection as an unbalanced multiclass TC task.
  • Analyzing the impact of hidden layer size and activation functions on the overall performance of each RandNN is evaluated by experimenting with various hidden layer sizes and three different activation functions.
The rest of the paper is presented as follows: Section 2 presents a summary of the literature highlighting the role of ML in the identification of mental health disorders. Section 3 provides a thorough description of the techniques used as well as the schematic layout of the proposed classification model. Section 4 presents the results obtained from the model implementation and their analysis. Finally, Section 5 concludes the work, acknowledging the limitations of the study and outlining the potential avenues for future research directions.

2. Literature Review

In recent years, an ever-growing number of people have turned to popular social media platforms like Twitter, Facebook, Reddit, and Instagram to express their thoughts and connect with others instantly. This trend has led to the generation of extensive social data, which holds valuable insights into individuals’ interests, emotions, and behavior patterns [4]. Social media has had a profound impact on the way individuals self-identify as having a mental health condition and how they connect with others who share similar experiences. People frequently turn to platforms like Reddit and Twitter to ask about treatment options and side effects, feeling less isolated and stigmatized in the process. Analyzing these social media interactions can offer valuable insights into the most pressing concerns of patients, sometimes even surpassing the insight available to their doctors [3,4]. Researchers are moving beyond just looking at how often social media is used and are now using more advanced methods to analyze usage patterns. These methods include delving into the linguistic style, emotional content, and interaction patterns within social networks to predict specific mental health conditions associated with social media posts. In recent years, there has been a surge of interest in various ML techniques within the research community interested in mental health analysis.
The authors of [5] used DT, KNN, SVM, and an ensemble technique to create a binary classification model to conduct a depression analysis on Facebook posts. DT is found to achieve the highest accuracy among the tested models across experiments. Another study has explored the performance of MLP, LR, SVM, and RF for depression analysis on Reddit posts using a Linear Discriminant Analysis (LDA)-based FS technique [6]. Following the selection of key linguistic features, the TF-IDF technique is applied to extract word frequencies. Then, LDA selects the pertinent features, which are fed as input to the binary classifiers. After experimenting with single and combined feature sets, it was found that the MLP classifier using combined feature sets outperforms the other models, achieving an accuracy of 91% and F1-score of 93%. In [7], five ML techniques, including NB, DT, RF, SVM, and KNN, are used to identify three types of mental disorders, such as anxiety, depression, and stress, from the questionnaire prepared from 349 participants. Owing to the imbalanced nature of the dataset, the models are compared based on F1-score, and NB is found to surpass other models. Another comparison of six traditional ML classifiers with three FS techniques and utilizing a SMOTE-based data balancing strategy reveals the superior performance of the AdaBoost classifier with the SelectKbest FS technique for depression detection on a sample dataset prepared by the author in [8]. The performance of NB and SVM is compared in [9] for depression detection from short social media texts using a hybrid FS technique. The proposed FS technique uses the chi-based filter approach in its first stage to rank the features and then uses the whale optimization algorithm in its second phase to retrieve the optimal subset of features. The superiority of the proposed FS technique is validated by experimenting with four benchmark short text datasets.
Recent achievements of deep learning in various NLP tasks have spurred academics to develop models for mental disorder identification utilizing several DNNs, including CNN, LSTM, RNN, and others. In [10], a four-layered CNN model has shown better performance compared to a XGBoost model to classify six types of mental disorders based on the posts of the users submitted in Reddit’s mental health communities. In this study, instead of considering it as a multiclass classification problem, six classifiers are designed for detecting six types of mental illness. Kour & Gupta in [12] have suggested a hybrid CNN-biLSTM model for predicting depressive persons by analyzing their Twitter data. Comparing the hybrid DL model to the CNN and RNN models, the experimental investigations show that the hybrid model obtains the greatest accuracy of 94.28%. Another depression detection model using LSTM is proposed in [13]. The model has employed symptom-based features derived from a one-hot process with Linear Discriminant Analysis (LDA) for feature selection in place of traditional word frequency-based features. Through testing on the Norwegian dataset, the model outperformed three traditional ML models, including LR, DT, and SVM, as well as a CNN model. In [14], a simple four-layer CNN using Word2vec vectorizer is shown to perform better at identifying mental disorders than the traditional XGBoost model. The model was evaluated using user posts available on Twitter and Reddit, suggesting its importance of examining multiple social media platforms to gain a more comprehensive view of public opinion on mental health. Six types of mental illness have been identified by the study, which has addressed it as a multiclass classification problem. When compared to SVM, KNN, LR, RF, AdaBoost, and MLP models, the effective performance of a LSTM model for stress detection is presented in [15] over a self-prepared dataset created using a questionnaire. In [16], authors have explored the use of two traditional ML models such as, KNN, RF, and a LSTM-based DL model for identifying 12 different types of mental disorders. Assessing and comparing the performance of the three models on a mental health disorder dataset collected from Kaggle, the RF model is found to achieve better outcomes compared to the other two models. In [17], another DL model using a transformer and a simple architecture has shown its effectiveness on detecting depression and suicidal ideation from social media content in comparison to traditional ML models. An optimized RNN using a dynamic stabilization technique is proposed in [18] for automatic diagnosis of mental health conditions. The time domain and statistical features extracted using the dual domain feature extraction method are given as input to the RNN model. Experiments on the OSMI dataset from different age groups show that the model classifies mental illness and healthy individuals with 98% and 99.5% accuracy, respectively, when compared to CNN and SVM models. In [19], authors have investigated the automatic classification of nine types of mental disorders from Reddit posts using transformer-based DL models such as BERT, RoBERT, along with XLNet. The NB classifier is employed to choose the most informative features from each type of mental disorder to feed into the DL models. This study has focused on identifying more mental illnesses rather than one or two specific ones, performing general text analysis and classification by posts rather than by individuals or groups. Another hybrid DL model integrating CNN with a pretrained sentence BERT is utilized for detecting depressed Reddit users by utilizing the SMHD data in [20]. By capturing the semantic meaning of Reddit posts using SBERT and then identifying their behavioral patterns using CNN, the model is able to achieve an accuracy of 86% by outperforming other ML models such as CNN, LSTM, BERT, RoBERT, XLNET, LSVM, LR, and XGBoost. Table 1 presents a summary of ML and DL models used in identifying different types of mental illness via social media posts.
Even with DL models’ tremendous upside, their primary drawback is the complexity of their deep structure, which leads to a laborious training process and requires researchers to concentrate mostly on adjusting their parameters and building more layers with even more potent computational resources [21,22]. As an alternative to these DL models, researchers are turning to RandNNs for a variety of binary and multiclass classification tasks because of their non-iterative training process and comparatively simple structure [26,27,28,29]. Chauhan & Tiwari (2022) have explored four RandNNs, such as RVFLN, KRVFLN, BLS, and fuzzy BLS, in [29] to address the multilabel classification problem. The model exhibits competitive performance as measured by metrics such a HL and average precision with less training time when tested on the Bibtex dataset with 159 classes, Emotions, and Scene dataset each including six classes. Another application of RVFLN for unbalanced multiclass TC is suggested by Behera & Dash (2022) in [31]. For enhancing the performance of RVFLN, a ModCHI-based FS technique is suggested in their work. When tested on the Reuters dataset containing 90 classes and 20 Newsgroups dataset with 20 classes, the model has shown superior performance in terms of precision, recall, and HL value compared to four traditional ML techniques. An ELM supplemented with a regularization factor termed as RELM is applied for TC in [33]. Using singular value decomposition on the TF-IDF matrix, the relevant features are chosen to feed to RELM. Experimenting on two benchmark TC datasets, such as the Reuters-21578 dataset with 10 classes and the WebKB dataset with four classes, the RELM-based TC model has shown better performance compared to ELM, BP, and SVM-based models. In [34], the performance of ELM along with three traditional Ml models such as SVM, KNN, and NB are compared for TC. The authors of this paper have suggested a modified frequency-based term weighting strategy to enhance the TC task by capturing the important terms present in the imbalanced text datasets. Experiments on three benchmark TC datasets show that the suggested FS technique with SVM classifier performs better than the other three models in terms of F1-score, recall, and precision. Assessing the performance of an ensemble ELM with an information entropy-based FS approach in [35] has shown better TC performance compared to weighted ELM, RELM, ELM, and three traditional ML techniques when experimented on Reuters-52 dataset with 52 classes, 20 Newsgroup with 20 classes, and WebKB with 4 classes. In [36], it is demonstrated how using max-min document frequency-based FS has improved the performance of an ELM-based TC system in categorizing the ninety different classes of the Reuters-21578 dataset. The effectiveness of combining a transformer model like BERT with BLS is highlighted in [37] to achieve both high accuracy and computationally efficiency in text emotion classification. The model has used BERT for feature extraction and BLS for efficient and fast learning, leading to improved emotion classification. As a convincing substitute for DL models, Du et al. (2020) have demonstrated the benefits of BLS for accurate and computationally efficient TC in [38]. Both word significance and sequence information is learnt by the suggested BLS for TC. When evaluated on 13 diverse datasets covering sentiment analysis and TC, the proposed model showed a considerable improvement in accuracy and training efficiency compared to the LSTM model. A node slicing BLS-based TC framework is suggested in [39]. An enhancement layer reflecting global information is developed based on the feature layer, once it is created by including rich words through one to one correspondence between the words and feature node groups. After that, a few nodes from feature layer groups are turned on and combined with the enhancement layer to produce outputs using ridge regression. The suggested BLS has shown better performance compared to CNN-BLS, LSTM-BLS, and ELM models when tested for sentiment analysis over Reddit sentiment and Twitter sentiment, each one containing three classes. A review on different types of RandNN used for TC, the type of FS technique used in those models, and the type of dataset specifying the number of classes used for validation of those models is shown in Table 2.
Based on the foregoing discussion, it is concluded that the bulk of ML models suggested for mental health analysis have focused primarily on a specific illness, addressing it as a binary classification task. This study attempted to get over this constraint by approaching mental illness detection as an unbalanced multiclass TC task. Despite the advantages of RandNNs over DL models as delivered by several researchers, the review has also revealed their infrequent application in mental health analysis. Again, less work is done related to the efficient FS of RandNNs for TC. Therefore, we proposed a TC framework using RandNN for efficiently identifying different categories of mental illness and a hybrid TOPSIS-ModCHI-based FS technique to select the optimal set of features for the classifier.

3. Materials and Methods

The fundamental concept of RandNN, with the detailed working principle of the suggested TOPSIS-ModCHI-based RandNN classification model, is covered in this section.

3.1. Randomized Neural Networks

RandNNs were put forward as a counterpart to gradient descent-based NNs to overcome their shortcomings of a highly time consumable learning process, local minima, sensitivity to learning rate, and so on. These networks have a naturally random character since they use a non-iterative process to determine the output layer’s weight and maintain all other weights constant during training [23]. Because of this randomness, RandNNs can learn faster with fewer adjustable settings and without requiring high end hardware [23,24]. ELM, RVFLN, and BLS are some popular RandNNs that have been successfully used for solving various classification and prediction problems [21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39]. Figure 1 presents the basic structural differences among ELM, RVFLN, and BLS.

3.1.1. Extreme Learning Machine (ELM)

The Extreme Learning Machine (ELM) algorithm, developed by Huang [25], is a powerful SLFN that is widely utilized for tasks such as classification and regression. It uses single step training with randomly initialized weights and biases between the input and hidden layers. Only the output weights are analytically computed using the Moore-Penrose generalized inverse of the hidden layer’s output matrix. This approach distinguishes ELM from conventional NNs, which frequently use an iterative training process, by allowing it to learn from training data in a single step. ELM has been used in various text classification problems [33,34,35,36] due to its rapid training, minimal need for parameter tuning intervention, and its ability to support a variety of activation functions.

3.1.2. Random Vector Functional Link Neural Network (RVFLN)

Like ELM, RVFLN is another RandNN with three layers representing the input, output, and hidden layer. After being received by the input layer, the input data are processed by the neurons of the hidden layer using randomly assigned weights. Finally, the output layer produces the result of the computation. The only difference in the architecture of ELM and RVFLN is that the output layer receives the original features as input through direct linkages. The direct links help in improving the generalization ability of the NN. The network also has the advantage of not using the iterative training steps [22,26]. The details of RVFLN-based text classification are detailed in [31].

3.1.3. Board Learning System

As an extension to RVFLN, another RandNN, termed as BLS, is proposed in [21] as a substitute for learning deep features. In contrast to RVFLN, the BLS maps the input layer feature matrix to feature layer F, and the feature layer F is mapped to the enhancement layer E by nonlinear activation functions. The output layer is connected to both the feature layer and the enhancement layer. BLS aims to provide an alternating approach to deep learning. Unlike DNN, BLS does not require a gradual search for the optimal parameters of the model using BP. In order to determine the weights of the linkages between the hidden and output layer, the BLS learning process only involves the one step matrix inversion.
The BLS network possesses a unique horizontal expansion and vertical fix, setting it apart from conventional deep neural networks. With a significantly reduced number of layers, its distinctive structure is illustrated in Figure 2. If X is the input feature matrix represented as [X1, X2, X3, …. Xn]. The feature node Fi in feature layer maps the feature matrix from the input layer using the following equation:
F i = θ i X W i + B i ,   i = 1 ,   2 ,   . n
where X act as feature matrix, and Wi and Bi act as weight and bias from the input layer to the feature layer. Here, θ i acts as nonlinear activation function. Fi acts as feature node in feature layer F. All feature nodes can be represented as Fn = {F1, F2, F3 …. Fn}. The feature nodes act as input to the enhancement layer E, producing the enhancement nodes using the following equation:
E j = α j F n W j + B j ,   j = 1 ,   2 ,   . m
where Fn act as feature node from the feature layer. Wj and Bj act as weight and bias used between the feature layer and enhancement layer. Here, α j acts as a nonlinear activation function. All enhancement nodes can be represented as Em = [E1, E2, E3, … Em]. The output node in output layer is obtained using the following:
Y =   ( F 1 ,   F 2 ,   F 3 ,   F n )     ( E 1 ,   E 2 ,   E 3 ,   E m )   ]   W n + m
Here, W n + m represents the total weights used in between feature layer to output layer and in between enhancement layer to output layer.
If ( F 1 , F 2 , F 3 , F n ) ( E 1 , E 2 , E 3 , E m ) ] is represented as [ H 1 , H 2 , H 3 H m ] , then the output node Y can be further calculated as follows:
Y =   F n E m   ]   W n + m
If A =   F n E m   ] , then
Y = A W n + m
In BLS, the parameters such as Wi, Wj, Bi, and Bj are randomly assigned. As the values of matrix A and Y are already known in the learning process, so the weights represented by W n + m are analytically computed using the Moore–Penrose generalized inverse as follows:
W n + m = A + Y
Here, A+ represents the pseudo inverse of matrix A.

3.2. Proposed RandNN-Based Text Classification Model Utilizing TOPSIS-ModCHI-Based FS

This section provides a comprehensive guide on the steps involved in utilizing RandNNs for text classification with the TOPSIS-ModCHI-based FS method. Figure 3 depicts the proposed RandNN-based text classification model.
As the main objective of this study is to analyze the performance of RandNN for detecting mental disorder of a person from his/her social media posts, a publicly available dataset named “Reddit Mental Health Dataset” is used for the validation of the model. As this problem refers to a multiclass unbalanced text classification problem, so to enhance the performance of the classifier, TOPSIS-ModCHI based, a hybrid FS technique is used in the classification framework.
Data preprocessing is a crucial step in text classification that involves transforming raw data into a format that can be used for text classification. In this process, each document in the corpus goes through the steps of converting all the text to lower case or upper case, eliminating unnecessary characters, abbreviations, emojis, and punctuation. Additionally, stop words and duplicate entries are removed, and spelling errors are corrected. Lemmatization and tokenization further refine the data by converting words to their base forms and splitting text into individual units. Stemming is a linguistic process that involves breaking down a word into its root form by removing any prefixes, infixes, or suffixes. This technique is commonly used to identify the core meaning of a word and to group together words with similar roots. To carry out this process, the most widely recognized and commonly used stemmer, known as the Porter stemmer, has been employed in this study. The Porter stemmer algorithm is a well-established technique that effectively removes suffixes from words while preserving the core meaning. Lemmatization is a linguistic process that is used to determine the base form or root word of a given word. It performs a thorough analysis of the morphology of a word and provides the root form of all its inflectional forms. This approach is more accurate than stemming, which may not always produce the correct root word.
Once the data pre-processing is complete, the raw text data are transformed into feature vectors through a process of feature extraction. This vital step involves the removal of any redundant features, which helps to decrease the dimensionality of the attributes. The process of converting raw text into a feature vector involves the use of the Term Frequency-Inverse Document Frequency (TF-IDF) method. This method not only helps to identify the frequency of individual terms in a document but also the importance of these terms about the entire corpus. By assigning higher weights to words that are rare in the corpus but frequent in the document, the TF-IDF method helps to create a feature vector that captures the most relevant information and can be used for further analysis.
Following the feature extraction, a hybrid FS technique is utilized in the next step to include the pertinent set of features by narrowing the feature space. The TF-IDF matrix calculated in the previous step and the total count of required k features are given as input to this FS technique. For each class of each attribute, the chi-square value is computed using the following equation:
C h i s q u a r e t i , c j = N P S R Q 2 P + R Q + S P + Q R + S .
where N presents the number of documents in a corpus, P specifies the documents in class cj containing terms ti, Q is the documents in other class with term ti, R is the documents of class cj not containing terms ti, and S is the documents in the other class without term ti. Then, for each feature in the feature set, each feature’s class number with highest chi value and the maximum chi value for each feature are found. Following it, a triplet containing feature list, highest chi value, and class number is formed for each class. Based on their chi value, the triplets are arranged in descending order. The number of document belonging to same class is multiplied by the number of features, and then the result is divided by the total number of document in the corpus. This yields a score based on each class’s feature and the quantity of documents in that class. After that, the feature list of those class numbers is chosen if the value shown in the triplet corresponds to the score that was determined. This approach will increase the number of documents with fewer features in the training set while giving significant attention to relevant classes. Another challenge involved with this technique is identifying its optimal k value. Instead of randomly setting this value, further, the ideal feature size of ModCHI is decided by TOPSIS, a Multi Criteria Decision Making (MCDM) approach, as specified in [32]. The performance of the classifier is observed with different k value of ModCHI with respect to different criteria to form the decision matrix, which is then passed as input to TOPSIS. The highest ranked k value obtained by applying the steps of TOPSIS is considered as the ideal one. The flow chart of TOPSIS-based ranking is presented in Figure 4.
Then the structure of four RandNNs, such as ELM, RVFLN, KRVFLN, and BLS, is established by specifying the number of nodes used in their different layers and the activation functions used for those nodes. KRVFLN is an extension of RVFLN that uses the kernel function for mapping input features to the hidden layers, so as to increase the stability of standard RVFLN [29]. Kernel functions are a powerful tool for measuring the similarity between two vectors, providing a greater value for similar vectors compared to dissimilar ones. When using the polynomial kernel, two vectors will yield the maximum value if they are parallel, even if they are not similar in length. In this study, we have used the polynomial kernel as follows:
K X , Y = X T   Y + C d
where X and Y represent the input vectors and C is the constant term.
For each RandNN, according to the feature size k, that number of nodes is fixed in the input layer, and the output layer size is fixed to the total classes. For ELM, RVFLN, and KRVFLN, hidden layer size is set to the summation of one-third of the feature size with the total number of output nodes [31]. With k number of selected features and y number of output classes, the hidden layer size m is calculated as follows:
m = k 3 + y   .    
The same size is also taken for the feature and enhancement layer in BLS. During the training of ELM, RVFLN, and KRVFLN, the weights between the input and hidden layers are initialized randomly, and the hidden layer output matrix is calculated using a suitable activation function. Then, the output layer weights are calculated in a single step by finding the pseudo inverse of the hidden layer output matrix. Similarly in case of BLS, the weights used in between input and feature layer and in between feature and enhancement layer are initialized randomly, and then the output layer weights are calculated in a single step using Equation (6). Finally the performance of the model is accessed using Hamming loss and the micro averaged value of precision, recall, and F-measure as calculated as follows:
H L = 1 N L i = 1 N j = 1 C X O R t a r g e t i j , p r e d i c t e d i j .
where N is the number of samples, C is the number of classes, targetij is the true class label for instance i and class j, and predictedij is the predicted label for instance i and class j.
P m i c r o = i = 1 C t p i i = 1 C t p i + f p i .
where tp is the true positive, and fp is the false positive.
R m i c r o = i = 1 C t p i i = 1 C t p i + f n i
F m i c r o = 2 p m i c r o r m i c r o p m i c r o + r m i c r o

4. Results

This section presents a comprehensive comparison of four RandNNs leveraging ModCHI-based FS for classifying mental health conditions from social media text by experimenting with the publicly available “Reddit Mental Health Dataset”. Reddit is a social media platform comprising user-generated content organized into themed communities called “subreddits” [40]. Users can post, share, and discuss content ranging from news and entertainment to niche interests. The original dataset contains posts from 28 subreddits from the year 2018 to 2020. This paper has used a small piece of data including 75,558 posts for analysis from the original dataset that comprises of 8 different classes of mental illnesses, which are depression, suicide, anxiety, schizophrenia, autism, PTSD, alcoholism, and bipolar disorder, and a set of neutral data class of the year 2019 only. Before experimenting over the mental health dataset, the performance of the proposed model is analyzed over two benchmark datasets used for text classification, such as Reuters-21578 and 20 Newsgroups. The trials are run on a Core i3 computer with 4GB of RAM running Ubuntu 16.04.6 LTS. Python 3.8.18 is used to put the algorithms and techniques into implementation.
Table 3 shows the number of classes, along with the total number of documents of each dataset. The distribution of documents in the Reddit Mental Health Dataset based on their respective classes is shown in Table 4, and its graphical representation using a pie chart is shown in Figure 5. A sample of a Reddit posts with their corresponding class/subreddit is shown in Table 5. Referring to the hold-out validation approach used in several works, particularly in the context of mental illness detection [6,8,10,12], the dataset is split into a ratio of 80:20 in this study. The class-wise distribution of documents related to the top and bottom three classes of Reuters and 20 Newsgroup is presented in Table 6. Preliminary analysis reveals that all the three datasets include uneven distribution of samples among different classes.
After applying the steps of preprocessing, the feature vectors for each dataset are created using the TF-IDF vectorizer. Then, the performance of the four RandNNs is observed with the ModCHI-based FS technique by systematically varying the value of selected features between 10 to 1500. Table 7, Table 8 and Table 9 present the results for Pmicro, Rmicro, Fmicro, and HL values produced by four RandNNs utilizing the ModCHI with various feature sizes for the Reuters, 20 Newsgroups, and Reddit datasets, respectively. The best value observed related to each criterion for each RandNN is highlighted as bold in the tables.
By analyzing the results depicted in Table 7, it is observed that for ELM the highest Pmicro value of 0.95 is achieved with a feature size of 700, the highest Rmicro value of 0.73 is achieved with a feature size of 1500, and the highest Fmicro value of 0.82 and the smallest HL value of 0.004 are achieved with feature sizes of 1000 and 1500 both. Similarly analyzing the outcomes of RVFLN with different k values of ModCHI, it is observed that the highest Pmicro value of 0.95 is achieved with a feature size of 700, and the highest Rmicro value of 0.81 and the highest Fmicro value of 0.87 are achieved with feature sizes of 1000 and 1500 both. In contrast, the smallest HL value of 0.003 is achieved with feature sizes ranging from 700 to 1500. For KRVFLN, the highest Pmicro value of 0.97 is achieved with feature sizes of 900 and 1500, and the highest Rmicro value of 0.81 and the highest Fmicro value of 0.88 are achieved with feature sizes of 900 and 1500 both. In contrast, the smallest HL value of 0.003 is achieved with feature sizes ranging from 900 to 1500. Similarly analyzing the outcomes of BLS with different feature sizes of ModCHI in the Reuters dataset, it is observed that the highest Pmicro value of 0.98 and the highest Fmicro value of 0.90 are achieved with feature sizes of 700, 900, and 1500, and the highest Rmicro value of 0.83 is achieved with feature sizes ranging from 700 to 1500. In contrast, the smallest HL value of 0.002 is achieved with the same feature sizes of 500, 700, 1000, and 1500.
As depicted in Table 8, for the 20 Newsgroup dataset, the highest Pmicro value of 0.92 is achieved with a feature size of 900, the highest Rmicro value of 0.65 and the highest Fmicro value of 0.76 are achieved with a feature size of 1500, and the smallest HL value of 0.03 is achieved with feature sizes of 900, 1000, and 1500 for the ELM network. Similarly analyzing the outcomes of RVFLN with different k values of ModCHI for the 20 Newsgroups dataset, it is observed that the highest Pmicro value of 0.92 is achieved with the feature sizes of 100, 300, and 500, and the highest Rmicro value of 0.68 and the highest Fmicro value of 0.77 are achieved with a feature size of 1500. In contrast, the smallest HL value of 0.02 is achieved with a feature size ranging from 300 to 1500. For KRVFLN, the highest Pmicro value of 0.94 is achieved with feature sizes of 1000 and 1500, and the highest Rmicro value of 0.67 and the highest Fmicro value of 0.78 are achieved with the feature sizes of 900, 1000 and 1500. In contrast, the smallest HL value of 0.02 is achieved with the feature size ranging from 100 to 1500. Similarly analyzing the outcomes of BLS over the 20 Newsgroups dataset, it is observed that the highest Pmicro value of 0.96 is achieved with feature sizes ranging from 700 to 1500, the highest Fmicro value of 0.70 is achieved with the feature size of 1500, the highest Rmicro value of 0.82 is achieved with a feature size of 900, and the smallest HL value of 0.02 is achieved with the feature sizes ranging from 300 to 1500.
Analyzing the outcomes of four RandNNs for Reddit Mental Health Dataset, depicted in Table 9, it is observed that, the highest Pmicro value of 0.87 is achieved with feature sizes of 900, 1000, and 1500. In addition, the highest Rmicro value of 0.59 and the highest Fmicro value of 0.70 are achieved when the feature size is 1500, and the smallest HL value of 0.06 is achieved with the feature sizes ranging from 700 to 1500 for the ELM network. Similarly analyzing the outcomes of RVFLN for Reddit dataset, it is observed that the highest Pmicro value of 0.89, the highest Rmicro value of 0.63, and the highest Fmicro value of 0.74 are achieved with feature sizes of 1000 and 1500 both, whereas the smallest HL value of 0.06 is achieved with the feature sizes ranging from 900 to 1500. For KRVFLN, the highest Pmicro value of 0.92 is achieved with the feature sizes of 1000 and 1500, and the highest Rmicro value of 0.65 and the highest Fmicro value of 0.76 are achieved when the feature size is 1000. In contrast, the smallest HL value of 0.06 is achieved with the feature size ranging from 500 to 1500. Similarly analyzing the outcomes of BLS for Reddit dataset, it is observed that the highest Pmicro value of 0.92 is achieved with feature sizes ranging from 900 to 1500, and the highest Fmicro value of 0.77 and the highest Rmicro value of 0.66 are achieved with the feature sizes of 900 and 1500. In contrast, the smallest HL value of 0.06 is achieved with the feature sizes ranging from 700 to 1500.
After conducting a similar type of analysis on Table 7, Table 8 and Table 9 for each RandNN with 10 different k values of ModCHI across four evaluation criteria, it is found that a particular feature size does not give the best result for all performance measures. Thus, it points to a sensible situation of forming a more reliable decision considering multiple criteria. Further, selecting the optimal feature size for each RandNN is addressed as a Multi Criteria Decision Making problem, and TOPSIS is implemented to rank 10 alternates of each RandNN with respect to four criteria. Table 10 presents the TOPSIS ranking of all the feature sizes for each RandNN for all the three datasets. The optimal feature size selected by TOPSIS for each RandNN related to each dataset is further specified in Table 11. Analyzing the results of this table, it is clearly observed that the minimal feature size is selected for BLS with respect to each dataset.
To assess the impact of activation functions on the overall performance of the RandNNs, the performance of each NN is next evaluated with three activation functions, such as sigmoid, hyperbolic tangent (Tanh), and rectified linear unit (Relu), considering the selected features obtained by TOPSIS-ModCHI. Table 12 shows the outcome of each RandNN with the optimal feature set with respect to the three activation functions. Analyzing the results for all the three datasets, it is observed that in all the three cases BLS with sigmoid activation function is producing better values for each criterion.
Table 13 shows the outcome of BLS utilizing the sigmoid activation function and the optimal feature set selected by the hybrid FS technique with five different hidden layer sizes to assess the impact of hidden layer size on its performance. It has been observed from Table 12 that, for most of the criteria, the better values are observed by considering the hidden layer size 323 for Reuters, 320 for 20 Newsgroups, and 309 for the Reddit dataset that was set using the approach specified in Equation (9).
Further analyzing the comparative performance of the four RandNNs with respect to the four criteria represented in Figure 6, Figure 7 and Figure 8, it is also clear that BLS with the least number of features selected by TOPSIS-ModCHI performs better as compared to the other three RandNNs for all three datasets. The true positive document selection of TOPSIS-ModCHI-based FS for all four RandNNs depicted in Figure 9, Figure 10 and Figure 11 for the three datasets also clearly shows that BLS is able to choose a high proportion of true positive documents with fewer attributes compared to the remaining three NNs.
Further, the statistical significance of the BLS model over KRVFLN, RVFLN, and ELM classifiers is evaluated using a resampled paired t-test [41]. In this technique, the splitting procedure used in the hold out method for the dataset is repeated for n times. In each iteration, the two models A and B used for comparison are trained on the training set and evaluated on the testing set. The performance difference of the two models is calculated for each iteration, resulting in n performance differences. Subsequently, the t statistic is computed using the following equation:
t = p ¯ n i = 1 n ( P ( i ) p ¯ ) 2 / ( n 1 )
with n − 1 degrees of freedom, under the null hypothesis that models A and B perform equally. Here, P(i) computes the difference between the model performance in the ith iteration, and P ( i ) = P A ( i ) P B ( i ) and p ¯ represents the average difference between the classifier performance.
p ¯ = 1 n i = 1 n p ( i )   .
After calculating the t statistic, the next step is to compute the p value and compare it to the chosen significance level α. If the p value is less than α, we reject the null hypothesis and conclude that there is a significant difference between the two models. In this study, the paired t-test is conducted using Fmicro as the evaluation criterion with the significance level of 0.05.
Table 14 presents the t statistics and the corresponding p-value obtained from the resampled paired t-test with the n value set to 10. The results shown in Table 14 indicate that, the p-value is less than the predetermined significance level of α = 0.05 for all the three datasets. In each case, the p-value falls below the significance threshold, allowing us to reject the null hypothesis. This suggests a significant difference in the performance of the BLS model compared to the KRVFLN, RVFLN, and ELM models.

5. Discussion

In light of the fact that social media data might offer valuable insights into individual well-being, this study has investigated the use of four RandNNs to forecast mental illness. In order to enhance the performance of the RandNNs when processing text documents with uneven class distributions, a hybrid TOPSIS-ModCHI-based FS technique is recommended in the classification framework. As an improved version of chi-square, ModCHI emphasizes the selection of features with high class significance and low repetition. The ideal feature size k of ModCHI for each RandNN is not determined at random; instead, TOPSIS is used to figure out this value. Using two benchmark multiclass text datasets and the publicly available Reddit Mental Health Dataset, the efficacy of the proposed model is assessed through experimentation in various circumstances.
Analyzing the Pmicro, Rmicro, Fmicro, and HL values depicted in Table 7, Table 8 and Table 9 related to the four RandNNs using ModCHI by systematically varying its k value between 10 and 1500, it is observed that a single k value of ModCHI does not yield the best result for all performance measures. Further approaching the selection of the optimal k value of ModCHI for each RandNN as a MCDM problem, TOPSIS is applied to rank the 11 feature sizes with respect to four criteria. As evident from the values depicted in Table 10, it is found that in the case of the Reuters dataset, the ideal feature sizes of ModCHI for ELM, RVFLN, KRVFLN, and BLS are 1500, 1000, 900, and 700, respectively. For the 20 Newsgroups dataset, the ideal feature sizes of ModCHI for ELM, RVFLN, KRVFLN, and BLS are 1500, 1500, 1000, and 900, respectively. In case of the Reddit dataset, the ideal feature sizes of ModCHI for ELM, RVFLN, KRVFLN, and BLS are 1500, 1000, 1000, and 900, respectively. Upon further examination of the ideal feature size of ModCHI for each RandNN, as shown in Table 11, it comes to light that the proposed hybrid FS approach selects the fewest features for BLS with respect to each dataset.
The comparative performance of the four RandNNs with least number of features selected by TOPSIS-ModCHI with three different activation functions, as shown in Table 12, reveals that in the Reuters dataset, the BLS model using sigmoid function produces an improvement of 4.25%, 4.25%, and 1.03% in Pmicro value; 13.69%, 2.46%, and 2.46% in Rmicro value; 14.6%, 3.44%, and 2.27% in Fmicro value; and 50%, 33.33%, and 33.33% in HL value as compared to ELM, RVFLN, and KRVFLN, respectively. For the 20 Newsgroups dataset, the BLS model with sigmoid outperforms the ELM, RVFLN, and KRVFLN by 5.49%, 7.86%, and 2.12% in the Pmicro value; 9.23%, 4.41%, and 5.97% in the Rmicro value; 7.89%, 6.49%, and 5.12% in Fmicro; and 33.33%, 0%, and 0% in the HL value. In the Reddit Mental Health Dataset, with a sigmoid function, the BLS produces same Pmicro value as that of KRVFLN with less selected features and an improved Pmicro value by 5.74% and 3.37% as compared to ELM and RVFLN, respectively. It improves the Rmicro value by 11.86%, 4.76%, and 1.53% as well as the Fmicro value by 10%, 4.05%, and 1.31% as compared to ELM, RVFLN, and KRVFLN, respectively. Moreover, it yields the same HL value as RVFLN and KRVLN with less selected features, and it produces an improvement of 14.28% in the HL value when compared to the ELM model. It is also evident from the overall comparison of the four RandNNs based on four criteria shown in Figure 6, Figure 7 and Figure 8 that the BLS with the fewest features chosen by TOPSIS-ModCHI outperforms the other three RandNNs for each of the three datasets.
From Table 13, it can be depicted that the BLS model gives the best performance with a feature size of 700 and a hidden layer node size of 323 in the Reuters dataset with a Pmicro value of 0.98, Rmicro value of 0.83, Fmicro value of 0.90, and HL value of 0.003. For the 20 Newsgroups dataset, the BLS model gives the best performance with a feature size of 900 and hidden layer node size of 320 with a Pmicro value of 0.96, Rmicro value of 0.71, the Fmicro value of 0.82, and HL value of 0.02. In case of Reddit Mental Health Dataset, the BLS model gives the best performance with a feature size of 900 and hidden layer node size of 309 with a Pmicro value of 0.92, Rmicro value of 0.66, Fmicro value of 0.77, and HL value of 0.02.
As evidenced by the classification results in Table 12 and Table 13, the BLS consistently outperforms the other three RandNNs across multiple performance metrics, particularly in handling the complexities of multiclass and imbalanced mental health datasets. Further, the statistical analysis presented in Table 14 underscores BLS’s capability in handling multiclass unbalanced mental health datasets with minimal features, offering a promising approach for a scalable and efficient mental health analysis. Unlike ELM and RVFLN, which use a single-pass training process, the combination of broad feature mapping with enhancement nodes helps to improve the fitting ability of BLS compared to other RandNNs. The layered structure allows BLS to capture the nonlinear relationships in complex and sparse textual data in a more efficient way. In summary, the comparative analysis reinforces that BLS delivers a promising trade-off among performance, computational cost, and generalization, positioning it as a preferred choice for efficient mental health detection.

6. Conclusions

In conclusion, this study highlights the potential of leveraging ML in mental health analysis, particularly through the application of RandNNs. The exploration of the four RandNNs, such as BLS, RVFLN, KRVFLN, and ELM, coupled with the innovative TOPSIS-ModCHI-based FS technique, brings new insights in handling multiclass text data with unbalanced class distributions, which are common in mental health datasets. The integration of ModCHI for FS by focusing on the distribution of relevant documents across classes with the TOPSIS for optimal feature size selection showcases the potential of combining statistical and decision making methodologies in enhancing model robustness by ensuring better FS and improved classification accuracy. The experimental results and discussion presented related to two benchmark multiclass text datasets and the publicly available Reddit Mental Health Dataset clearly reveal that, among the four RandNNs, BLS emerged as the most effective model providing the highest Pmicro, Rmicro, and Fmicro values and the lowest HL value surpassing the other models. This underscores BLS’s capability in handling multiclass unbalanced mental health datasets with minimal features, offering a promising approach for scalable and efficient mental health analysis. The insights gained from this research could help mental health professionals leverage social media data effectively to understand and predict mental health trends.
However, there are some additional aspects that remain unexplored in our research. First, we performed all the experiments with a single mental health dataset, considering eight categories of mental disorders. In future work, the model could be trained and tested using a broader range of social media datasets and additional mental health categories to improve its generalizability and robustness. Although we have tested our model with a multiclass unbalanced dataset, it can be extended by exploring the performance of RandNNs for multilabeled text classification. This would be valuable in scenarios where a single social media post may indicate multiple co-occurring mental disorders. The performance of BLS could also be enhanced by integrating suitable optimization techniques in its weight updating phase as well as the FS process. RandNN-based ensemble models could be developed and explored for mental health analysis. Again, the current work focuses on achieving high predictive performance; it does not yet address the interpretability of the models, which is critical for adoption in clinical settings. Future work will focus on developing a deployable solution in the form of a web-based dashboard or API for use by the clinicians. Additionally, involving domain experts in model evaluation and interpretation will be a key step toward ensuring that predictions align with clinical knowledge and practice.

Author Contributions

Conceptualization, S.K.B. and R.D.; methodology, S.K.B. and R.D.; software, S.K.B.; validation, S.K.B.; formal analysis, S.K.B. and R.D.; investigation, S.K.B.; resources, S.K.B.; data curation, S.K.B.; writing—original draft preparation, S.K.B. and R.D.; writing—review and editing, R.D.; visualization, S.K.B. and R.D.; supervision, R.D.; project administration, R.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The dataset used in the study is publicly available from https://zenodo.org/records/3941387#.Y5L6O_fMKUl (accessed on 10 May 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. WHO. 2021. Available online: https://www-who-int.ezproxy.uio.no/news-room/fact-sheets/detail/suicide (accessed on 10 May 2024).
  2. Available online: https://www.theguardian.com/society/2021/aug/29/strain-on-mental-health-care-leaves-8m-people-without-help-say-nhs-leaders (accessed on 10 May 2024).
  3. Giuntini, F.T.; Cazzolato, M.T.; Reis, M.d.J.D.d.; Campbell, A.T.; Traina, A.J.M.; Ueyama, J. A review on recognizing depression in social networks: Challenges and opportunities. J. Ambient. Intell. Humaniz. Comput. 2020, 11, 4713–4729. [Google Scholar] [CrossRef]
  4. Garg, M. Mental Health Analysis in Social Media Posts: A Survey. Arch. Comput. Methods Eng. 2023, 30, 1819–1842. [Google Scholar] [CrossRef] [PubMed]
  5. Islam, R.; Kabir, M.A.; Ahmed, A.; Kamal, A.R.M.; Wang, H.; Ulhaq, A. Depression detection from social network data using machine learning techniques. Health Inf. Sci. Syst. 2018, 6, 8. [Google Scholar] [CrossRef] [PubMed]
  6. Tadesse, M.M.; Lin, H.; Xu, B.; Yang, L. Detection of depression-related posts in reddit social media forum. IEEE Access 2019, 7, 44883–44893. [Google Scholar] [CrossRef]
  7. Priya, A.; Garg, S.; Tigga, N.P. Predicting anxiety, depression and stress in modern life using machine learning algorithms. Procedia Comput. Sci. 2020, 167, 1258–1267. [Google Scholar] [CrossRef]
  8. Zulfiker, S.; Kabir, N.; Biswas, A.A.; Nazneen, T.; Uddin, M.S. An in-depth analysis of machine learning approaches to predict depression. Curr. Res. Behav. Sci. 2021, 2, 100044. [Google Scholar] [CrossRef]
  9. Priya, S.K.; Karthika, K.P. An embedded feature selection approach for depression classification using short text sequences. Appl. Soft Comput. 2023, 147. [Google Scholar]
  10. Kim, J.; Lee, J.; Park, E.; Han, J. A deep learning model for detecting mental illness from user content on social media. Sci. Rep. 2020, 10, 11846. [Google Scholar] [CrossRef]
  11. Li, Z.; Li, W.; Wei, Y.; Gui, G.; Zhang, R.; Liu, H.; Chen, Y.; Jiang, Y. Deep learning based automatic diagnosis of first-episode psychosis, bipolar disorder and healthy controls. Comput. Med Imaging Graph. 2021, 89, 101882. [Google Scholar] [CrossRef]
  12. Kour, H.; Gupta, M.K. An hybrid deep learning approach for depression prediction from user tweets using feature-rich CNN and bi-directional LSTM. Multimed. Tools Appl. 2022, 81, 23649–23685. [Google Scholar] [CrossRef]
  13. Uddin, M.Z.; Dysthe, K.K.; Følstad, A.; Brandtzaeg, P.B. Deep learning for prediction of depressive symp-toms in a large textual dataset. Neural Comput. Appl. 2022, 34, 721–744. [Google Scholar] [CrossRef]
  14. Ang, C.S.; Venkatachala, R. Generalizability of Machine Learning to Categorize Various Mental Illness Using Social Media Activity Patterns. Societies 2023, 13, 117. [Google Scholar] [CrossRef]
  15. Bhavani, B.H.; Naveen, N.C. An Approach to Determine and Categorize Mental Health Condition using Machine Learning and Deep Learning Models. Eng. Technol. Appl. Sci. Res. 2024, 14, 13780–13786. [Google Scholar] [CrossRef]
  16. Alkahtani, H.; Aldhyani, T.H.H.; Alqarni, A.A. Artificial Intelligence Models to Predict Disability for Mental Health Disorders. J. Disabil. Res. 2024, 3, 20240022. [Google Scholar] [CrossRef]
  17. Ezerceli, Ö.; Dehkharghani, R. Mental disorder and suicidal ideation detection from social media using deep neural networks. J. Comput. Soc. Sci. 2024, 7, 2277–2307. [Google Scholar] [CrossRef]
  18. Revathy, J.S.; Maheswari, N.U.; Sasikala, S.; Venkatesh, R. Automatic diagnosis of mental illness using optimized dynamically stabilized recurrent neural network. Biomed. Signal Process. Control 2024, 95, 106321. [Google Scholar] [CrossRef]
  19. Dinu, A.; Moldovan, A.C. Automatic detection and classification of mental illnesses from general social media texts. In Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP 2021), Online, 1–3 September 2021; pp. 358–366. [Google Scholar]
  20. Chen, Z.; Yang, R.; Fu, S.; Zong, N.; Liu, H.; Huang, M. Detecting Reddit Users with Depression Using a Hybrid Neural Network SBERT-CNN. In Proceedings of the 2023 IEEE 11th International Conference on Healthcare Informatics (ICHI), Houston, TX, USA, 26–29 June 2023; pp. 193–199. [Google Scholar]
  21. Chen, C.L.P.; Liu, Z. Broad learning system: An effective and efficient incremental learning system without the need for deep architecture. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 10–24. [Google Scholar] [CrossRef]
  22. Malik, A.; Gao, R.; Ganaie, M.; Tanveer, M.; Suganthan, P.N. Random vector functional link network: Recent developments, applications, and future directions. Appl. Soft Comput. 2023, 143, 110377. [Google Scholar] [CrossRef]
  23. Zhang, L.; Suganthan, P.N. A survey of randomized algorithms for training neural networks. Inf. Sci. 2016, 364–365, 146–155. [Google Scholar] [CrossRef]
  24. Suganthan, P.N.; Katuwal, R. On the origins of randomization-based feedforward neural networks. Appl. Soft Comput. 2021, 105, 107239. [Google Scholar] [CrossRef]
  25. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  26. Pao, Y.H.; Park, G.H.; Sobajic, D.J. Learning and generalization characteristics of the random vector functional-link net. Neurocomputing 1994, 6, 163–180. [Google Scholar] [CrossRef]
  27. Zhang, L.; Suganthan, P.N. A comprehensive evaluation of random vector functional link networks. Inf. Sci. 2016, 367, 1094–1105. [Google Scholar] [CrossRef]
  28. Liu, Z.; Huang, S.; Jin, W.; Mu, Y. Broad learning system for semi-supervised learning. Neurocomputing 2021, 444, 38–47. [Google Scholar] [CrossRef]
  29. Chauhan, V.; Tiwari, A. Randomized neural networks for multilabel classification. Appl. Soft Comput. 2022, 115, 108184. [Google Scholar] [CrossRef]
  30. Behera, S.K.; Dash, R. A novel feature selection technique for enhancing performance of unbalanced text classification problem. Intell. Decis. Technol. 2022, 16, 51–69. [Google Scholar] [CrossRef]
  31. Behera, S.K.; Dash, R. Performance Enhancement of the Unbalanced Text Classification Problem Through a Modified Chi Square-Based Feature Selection Technique: A Mod-Chi based FS technique. Int. J. Intell. Inf. Technol. (IJIIT) 2022, 18, 1–23. [Google Scholar] [CrossRef]
  32. Dash, R.; Dash, R.; Rautray, R. Enhancing performance of a LENN based CCFD model with a hybrid TOPSIS-ReliefF based feature selection technique. In Proceedings of the 2022 2nd Odisha International Conference on Electrical Power Engineering, Communication and Computing Technology (ODICON), Bhubaneswar, India, 11–12 November 2022; pp. 1–6. [Google Scholar]
  33. Zheng, W.; Qian, Y.; Lu, H. Text categorization based on regularization extreme learning machine. Neural Comput. Appl. 2012, 22, 447–456. [Google Scholar] [CrossRef]
  34. Sabbah, T.; Selamat, A.; Selamat, M.H.; Al-Anzi, F.S.; Viedma, E.H.; Krejcar, O.; Fujita, H. Modified fre-quency-based term weighting schemes for text classification. Appl. Soft Comput. 2017, 58, 193–206. [Google Scholar] [CrossRef]
  35. Li, M.; Xiao, P.; Zhang, J. Text classification based on ensemble extreme learning machine. arXiv 2018, arXiv:1805.06525. [Google Scholar]
  36. Behera, S.K.; Dash, R. Performance of ELM using max-min document frequency-based feature selection in multilabeled text classification. In Intelligent and Cloud Computing: Proceedings of ICICC 2019; Springer: Singapore, 2021; Volume 1, pp. 425–433. [Google Scholar]
  37. Peng, S.; Zeng, R.; Liu, H.; Chen, G.; Wu, R.; Yang, A.; Yu, S. Emotion classification of text based on BERT and broad learning system. In Web and Big Data, Proceedings of the 5th International Joint Conference, APWeb-WAIM 2021, Guangzhou, China, 23–25 August 2021, Proceedings, Part I 5; Springer International Publishing: Berlin/Heidelberg, Germany, 2021; pp. 382–396. [Google Scholar]
  38. Du, J.; Vong, C.-M.; Chen, C.L.P. Novel Efficient RNN and LSTM-Like Architectures: Recurrent and Gated Broad Learning Systems and Their Applications for Text Classification. IEEE Trans. Cybern. 2020, 51, 1586–1597. [Google Scholar] [CrossRef] [PubMed]
  39. Liu, F.; Wu, X.; Li, C. Node slicing broad learning system for text classification. In Proceedings of the ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Virtual, 7–13 May 2022; pp. 4788–4792. [Google Scholar]
  40. Low, D.M.; Rumker, L.; Talkar, T.; Torous, J.; Cecchi, G.; Ghosh, S.S. Natural Language Processing Reveals Vulnerable Mental Health Support Groups and Heightened Health Anxiety on Reddit During COVID-19: Observational Study. J. Med Internet Res. 2020, 22, e22635. [Google Scholar] [CrossRef] [PubMed]
  41. Dietterich, T.G. Approximate Statistical Tests for Comparing Supervised Classification Learning Algorithms. Neural Comput 1998, 10, 1895–1923. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Basic structural differences among ELM, RVFLN, and BLS.
Figure 1. Basic structural differences among ELM, RVFLN, and BLS.
Mca 30 00067 g001
Figure 2. Structure of BLS.
Figure 2. Structure of BLS.
Mca 30 00067 g002
Figure 3. Proposed RandNN-based text classifier for mental illness detection.
Figure 3. Proposed RandNN-based text classifier for mental illness detection.
Mca 30 00067 g003
Figure 4. Flow chart of TOPSIS-based ranking.
Figure 4. Flow chart of TOPSIS-based ranking.
Mca 30 00067 g004
Figure 5. Class-wise documents distribution percentage of the Reddit Mental Health Dataset.
Figure 5. Class-wise documents distribution percentage of the Reddit Mental Health Dataset.
Mca 30 00067 g005
Figure 6. Performance comparison of RandNNs utilizing TOPSIS-ModCHI-based FS for the Reuters dataset.
Figure 6. Performance comparison of RandNNs utilizing TOPSIS-ModCHI-based FS for the Reuters dataset.
Mca 30 00067 g006
Figure 7. Performance comparison of RandNNs utilizing TOPSIS-ModCHI-based FS for the 20 Newsgroups dataset.
Figure 7. Performance comparison of RandNNs utilizing TOPSIS-ModCHI-based FS for the 20 Newsgroups dataset.
Mca 30 00067 g007
Figure 8. Performance comparison of RandNNs utilizing TOPSIS-ModCHI-based FS for the Reddit Mental Health Dataset.
Figure 8. Performance comparison of RandNNs utilizing TOPSIS-ModCHI-based FS for the Reddit Mental Health Dataset.
Mca 30 00067 g008
Figure 9. True positive rates obtained the four RandNNs utilizing TOPSIS-ModCHI-based FS for the Reuters dataset.
Figure 9. True positive rates obtained the four RandNNs utilizing TOPSIS-ModCHI-based FS for the Reuters dataset.
Mca 30 00067 g009
Figure 10. True positive rates obtained the four RandNNs utilizing TOPSIS-ModCHI-based FS for the 20 Newsgroups dataset.
Figure 10. True positive rates obtained the four RandNNs utilizing TOPSIS-ModCHI-based FS for the 20 Newsgroups dataset.
Mca 30 00067 g010
Figure 11. True positive rates obtained the four RandNNs utilizing TOPSIS-ModCHI-based FS for the Reddit Mental Health Dataset.
Figure 11. True positive rates obtained the four RandNNs utilizing TOPSIS-ModCHI-based FS for the Reddit Mental Health Dataset.
Mca 30 00067 g011
Table 1. A survey on ML and DL models used for identifying mental illness.
Table 1. A survey on ML and DL models used for identifying mental illness.
ReferenceAuthor
(Year)
ML ModelFS TechniqueDatasetType of Mental Illness AddressedEvaluation Metrics
[5]Islam et al. (2018)DT, KNN, SVM, ensemble---Facebook dataDepressionAccuracy
[6]Tadesse et al., (2019)MLP,
LR, SVM, RF
LDARedditDepressionAccuracy, F1-score
[7]Priya et al., (2020)NB, DT, RF, SVM, KNN---DASS-21Anxiety, depression, stressAccuracy, precision, recall, specificity, F1-score
[8]Zulfiker et al., (2021)KNN, AdaBoost, Gradient boosting (GB), Extreme GB, Bagging, Weighted votingSelectKBest, mRMR, BorutaSelf-prepared datasetDepressionSensitivity, specificity, precision, F1-score, AUC
[9]Priya & Karthika, (2023)NB, SVMHybrid FS using Chi and Whale OptimizationFour bench mark short text datasets such as Stack Overflow, Short Messaging Service (SMS), Sentiment Labeled Sentences (SLS), and Sentiment 140 datasetDepressionAccuracy, F1-scores, sensitivity
[10]Kim et al., (2020)CNN,
XGBoost
---RedditDepression, anxiety, bipolar, borderline personality disorder, schizophrenia, autismAccuracy, precision, recall, F1-score
[12]Kour & Gupta (2022)CNN-biLSTM,
CNN,
RNN
Embedding layer of CNNTwitter dataDepressionPrecision, recall, F1-score, accuracy, specificity, AUC
[13]Uddin et al., (2022)LSTM,
LR, DT, SVM,
CNN
LDAYoung user’s text data obtained from ung.no
Norwegian
DepressionPrecision, recall, F1-score, support
[14]Ang & Venkatachala,
(2023)
CNN,
XGBoost
---Reddit,
Twitter
Depression, anxiety, bipolar, BPD, schizophrenia, autismPrecision, recall, F1-score, accuracy
[15]Bhavani & Naveen, (2024)LSTM,
KNN, RF, SVM, LR, AdaBoost, MLP
---Self-prepared dataset through questionariesStressAccuracy, precision, recall, F1-score
[16]Alkahtani (2024)LSTM,
KNN, RF
---Mental disorder dataset collected from KaggleADHD, ASD, loneliness, MDD, OCD, PDD, PTSD, anxiety, BD, eating disorder, psychotic depression, and sleeping disorderAccuracy, precision, recall, F1-score
[17]Ezerceli & Dehkharghani, (2024)RNN, CNN, LSTM, BERT, SVM,
NB, RF, LR, DT
---RedditDepression/suicidal ideationAUC, precision, recall F1-score
[18]Revathy et al., (2024)Dynamically Stabilized RNN, CNN, SVMSiberian tiger optimization algorithmOSMI datasetMentally ill for getting treatment or notF1-score, accuracy
[19]Dinu & Moldovan, (2021)BERT, XLNet, RoBERTNBRedditSchizophrenia,
bipolar, depression, anxiety, obsessive-compulsive disorders, eating disorders, autism,
post-traumatic stress disorder,
attention-deficit/hyperactivity disorder
Precision, Recall, F1-score
[20]Chen et al., (2023)SBERT-CNN, CNN, LSTM, BERT, RoBERT, XLNET, LSVM, LR, XGBoost---SMHD dataset including Reddit postsDepressionAccuracy, Precision, Recall, F1-score
Our approachBLS, ELM, RVFLN, K-RVFLNTOPSIS-ModCHIRedditDepression, suicide, anxiety, schizophrenia, autism, PSTD, alcoholism, bipolar, neutralPmicro, Rmicro, Fmicro, Haming loss (HL), true positive rate
Table 2. A review on RandNNs used for text classification.
Table 2. A review on RandNNs used for text classification.
ReferenceAuthor
(Year)
Type of RandNNFS TechniqueDataset
(Number of Classes)
[29]Chauhan & Tiwari (2022)RVFLN, KRVFLN, BLS, Fuzzy BLS---Bibtex (159), Emotions (6), Scene (6)
[31]Behera & Dash (2022)RVFLN,
LSVM, RF, DT, MLKNN
ModCHI,
Chi,
MI,
Tf-idf
Reuters-21578 (90)
20 Newsgroups (20)
[33]Zheng et al. (2013)RELM,
ELM,
BP,SVM
Singular value decomposition applied on TFIDF matrixReuters-21578
(10), WebKB (4)
[34]Sabbah et al. (2017)SVM, KNN,ELM, NBModified frequency-based term weighting schemeReuters-21578 (8)
20 Newsgroups (20),
WebKB (7)
[35]Li et al., (2018)AE1WELM,
Weighted ELM,
RELM, ELM, SVM, KNN, NB
Information entropyReuters-52 (52)
20 Newsgroups (20),
WebKB (4)
[36]Behera & Dash (2021)ELMMax-min document frequencyReuters-21578
(90)
[38]Du et al., (2020)Recurrent BLS, Gated BLS, LSTM---New Years resolutions (10), Political-media (9), Twitter sentiment (6), Objective-sentence (5), Apple-Twitter sentiment (4), Corporate messaging (4), Progressive tweet (4), Weather sentiment (4), Claritin-October twitter (3), Airline-sentiment (3), Tweet-global warming (3), Electronic sentiment (2), Books sentiment (2)
[39]Liu et al., (2022)NSBLS,
ELM,
CNN-BLS,
LSTM-BLS
---Reddit sentiment (3),
Twitter sentiment (3)
Our approachBLS, RVFLN, KRVFLN, ELMTOPSIS-ModCHIReuters-21578 (90)
20 Newsgroups (20),
Reddit Mental Health (9)
Table 3. Description of the dataset.
Table 3. Description of the dataset.
Dataset NameTotal Number of DocumentsTotal Number of ClassTotal Number of Training DocumentsTotal Number of Testing Documents
Reddit Mental Health75,559960,44615,112
Reuters-2157810,7889077693019
20 Newsgroups52262041801046
Table 4. Class-wise distribution of documents in the Reddit Mental Health Dataset.
Table 4. Class-wise distribution of documents in the Reddit Mental Health Dataset.
Class NameNumber of DocumentsNumber of Training DocumentsNumber of Testing Documents
Depression33,55426,8386716
Suicide14,59211,6792913
Anxiety13,21310,5902623
Schizophrenia15501224326
Autism14011111290
PSTD13971127270
Alcoholism942755187
Bipolar910702208
Neutral799964201579
Total75,55860,44615,112
Table 5. Sample user’s post related to each class from the Reddit Mental Health Dataset.
Table 5. Sample user’s post related to each class from the Reddit Mental Health Dataset.
Sl.noPostSubreddit
1HELP If life was a dream, what would we have to do to wake up from it.Depression
2How should I end my life Girls get a ton of attention so they either ignore me or treat me like shit. The human race sucks. Im going to be alone foreverSuicide
3I got a job interview on Monday at 11. Not going as I’m too anxious Thats it. First interview after studying and getting my diploma. I’m too nervous that ill screw it over. Advice neededAnxiety
4I put out a lit cigarette on my leg. My thoughts are uncontrollable. The violent urges and thoughts are uncontrollable now, they just keep coming.Schizophrenia
5Is autism genetic? I’m a teenager with autism spectrum disorder. My dad also has it. If I eventually have kids, what are the chances that they will also be autistic?Autism
6I am afraid to attribute to my trauma every bad thing that I feel. Hey guys, Just wanted to know if anyone can relate to. It has been roughly a year since my traumatic episode took place and ever since my life has became a mess.PSTD
7Been an alcoholic for past 3 years Like my title says I’ve been struggling with drinking for past few years. My drink of choice is vodka. I occasionally drink beer as well.Alcoholism
8Enjoy life with this one weird trick. Yes, you are bipolar--or maybe you are the SO of a bipolar person. Here is one trick that will make 2019 better than any other year has been.Bipolar
9I’m meeting up with one of my besties tonight! Can’t wait!!Neutral
Table 6. Class-wise distribution of documents of benchmark datasets.
Table 6. Class-wise distribution of documents of benchmark datasets.
Reuters dataset
Top 3 class name with training and testing documentsClass nameTraining: TestingBottom 3 class name with training and testing documentsClass nameTraining: Testing
Earn2877:1087Sun-meal1:1
Acq1650:719Lin-oil1:1
Money-fx538:179Coconut-oil1:2
20 Newsgroups dataset
Top 3 class name with training and testing documentsClass nameTraining: TestingBottom 3 class name with training and testing documentsClass nameTraining: Testing
rec.autos485:109soc.religion.christian58:12
rec.sport.hockey476:124alt.atheism45:8
rec.motorcycles475:123talk.politics.misc43:9
Table 7. Performance of RandNNs with ModCHI for the Reuters dataset.
Table 7. Performance of RandNNs with ModCHI for the Reuters dataset.
Performance with ELM
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.90.910.920.930.940.940.950.940.940.94
Rmicro0.330.430.560.680.690.680.70.710.720.73
Fmicro0.480.580.700.790.800.790.810.810.820.82
HL0.0090.0080.0070.0070.0060.0060.0050.0050.0040.004
Performance with RVFLN
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.940.940.950.940.950.930.950.940.940.94
Rmicro0.380.470.670.710.770.780.790.790.810.81
Fmicro0.540.630.790.810.850.850.860.860.870.87
HL0.0080.0070.0060.0040.0040.0040.0030.0030.0030.003
Performance with KRVFLN
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.940.940.950.960.960.970.960.970.960.97
Rmicro0.40.470.680.720.780.790.80.810.80.81
Fmicro0.560.630.790.820.860.870.870.880.870.88
HL0.0080.0070.0060.0060.0040.0040.0040.0030.0030.003
Performance with BLS
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.940.940.950.950.970.970.980.980.970.98
Rmicro0.380.500.670.720.780.790.830.830.830.83
Fmicro0.540.650.790.820.860.870.900.900.890.90
HL0.0080.0070.0060.0040.0040.0020.0020.0030.0020.002
Table 8. Performance of RandNNs with ModCHI for the 20 Newsgroups dataset.
Table 8. Performance of RandNNs with ModCHI for the 20 Newsgroups dataset.
Performance with ELM
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.790.810.840.870.890.910.910.920.910.91
Rmicro0.190.280.320.360.480.550.600.610.620.65
Fmicro0.310.420.460.510.620.690.720.730.740.76
HL0.050.040.040.040.040.040.040.030.030.03
Performance with RVFLN
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.830.850.890.920.920.920.910.900.900.89
Rmicro0.230.330.340.390.540.590.620.620.620.68
Fmicro0.360.480.490.550.680.720.740.730.730.77
HL0.040.030.030.030.020.020.020.020.020.02
Performance with KRVFLN
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.820.860.890.920.920.930.930.930.940.94
Rmicro0.250.370.380.400.550.600.640.670.670.67
Fmicro0.380.520.530.560.690.730.760.780.780.78
HL0.030.030.030.020.020.020.020.020.020.02
Performance with BLS
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.840.870.90.930.940.950.960.960.960.96
Rmicro0.240.360.390.410.570.650.680.710.70.70
Fmicro0.370.510.540.570.710.770.800.820.810.81
HL0.040.040.030.030.020.020.020.020.020.02
Table 9. Performance of RandNNs with ModCHI for the Reddit Mental Health Dataset.
Table 9. Performance of RandNNs with ModCHI for the Reddit Mental Health Dataset.
Performance with ELM
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.740.780.790.810.830.840.860.870.870.87
Rmicro0.340.390.410.450.470.490.520.550.570.59
Fmicro0.470.520.540.580.600.620.650.670.690.70
HL0.080.080.080.070.070.070.060.060.060.06
Performance with RVFLN
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.770.790.790.810.830.850.860.880.890.89
Rmicro0.350.390.440.480.530.560.600.610.630.63
Fmicro0.480.520.560.600.630.650.680.690.740.74
HL0.080.080.070.070.070.070.070.060.060.06
Performance with KRVFLN
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.780.790.810.830.840.850.870.890.920.92
Rmicro0.350.380.450.490.540.570.610.630.650.63
Fmicro0.480.510.580.620.660.680.720.740.760.75
HL0.080.080.070.070.070.060.060.060.060.06
Performance with BLS
Evaluation MetricsFeature Size (k)
10204010030050070090010001500
Pmicro0.790.800.820.840.860.880.900.920.920.92
Rmicro0.360.390.450.500.550.570.620.660.640.66
Fmicro0.490.520.580.630.670.690.730.770.750.77
HL0.080.080.080.070.070.070.060.060.060.06
Table 10. TOPSIS ranking of different k values of ModCHI for each RandNN.
Table 10. TOPSIS ranking of different k values of ModCHI for each RandNN.
DatasetType of RandNNFeature Size (k)
10204010030050070090010001500
ReutersELM10987564321
RVFLN10987563412
KRVFLN10987645132
BLS10987641532
20 NewsgroupELM10987654321
RVFLN10987652341
KRVFLN10987654312
BLS10987654123
RedditELM10987654321
RVFLN10987654312
KRVFLN10987654312
BLS10987654132
Table 11. Optimal feature size selected by TOPSIS for each RandNN.
Table 11. Optimal feature size selected by TOPSIS for each RandNN.
ModelsReuters 20 Newsgroups Reddit
Feature SizeFeature SizeFeature Size
ELM150015001500
RVFLN100015001000
KRVFLN90010001000
BLS700900900
Table 12. Assessing the impact of activation functions for each RandNN with optimal feature set selected using the TOPSIS-ModCHI-based FS technique.
Table 12. Assessing the impact of activation functions for each RandNN with optimal feature set selected using the TOPSIS-ModCHI-based FS technique.
Dataset: Reuters
ModelsActivation functions
SigmoidTanhRelu
PmicroRmicroFmicroHLPmicroRmicroFmicroHLPmicroRmicroFmicroHL
ELM0.940.730.820.0040.930.710.800.0040.930.720.810.004
RVFLN0.940.810.870.0030.920.780.840.0030.920.780.840.003
KRVFLN0.970.810.880.0030.950.770.850.0040.940.670.780.005
BLS0.980.830.900.0020.970.780.860.0030.970.770.850.003
Dataset: 20 Newsgroups
ModelsActivation functions
SigmoidTanhRelu
PmicroRmicroFmicroHLPmicroRmicroFmicroHLPmicroRmicroFmicroHL
ELM0.910.650.760.030.900.620.730.030.890.640.740.03
RVFLN0.890.680.770.020.870.660.750.030.870.650.740.02
KRVFLN0.940.670.780.020.920.650.760.020.910.630.740.03
BLS0.960.710.820.020.950.670.780.030.920.650.760.03
Dataset: Reddit
ModelsActivation functions
SigmoidTanhRelu
PmicroRmicroFmicroHLPmicroRmicroFmicroHLPmicroRmicroFmicroHL
ELM0.870.590.700.060.860.570.680.060.840.550.660.07
RVFLN0.890.630.740.060.870.610.710.060.870.600.710.06
KRVFLN0.920.650.760.060.900.630.740.060.900.620.730.06
BLS0.920.660.770.060.910.660.760.070.900.650.750.06
Table 13. Assessing the impact of hidden layer size for BLS.
Table 13. Assessing the impact of hidden layer size for BLS.
Dataset: ReutersDataset: 20 NewsgroupsDataset: Reddit
Evaluation MetricsHidden Layer SizeHidden Layer SizeHidden Layer Size
100200323400500100200320400500100200309400500
Pmicro0.950.960.980.980.980.910.940.960.950.970.890.900.920.920.93
Rmicro0.790.820.830.830.820.660.680.710.720.710.620.640.660.650.64
Fmicro0.860.880.900.900.890.760.780.820.810.810.730.740.770.760.75
HL0.0030.0030.0030.0030.0030.030.020.020.020.020.050.050.060.060.06
Table 14. Paired t-test statistics.
Table 14. Paired t-test statistics.
DatasetBLS/KRVFLNBLS/RVFLNBLS/ELM
t-Statisticspt-Statisticspt-Statisticsp
Reuters6.197.99 × 10−514.238.9 × 10−833.414.74 × 10−11
20 Newsgroup20.124.29 × 10−914.457.77 × 10−828.941.70 × 10−10
Reddit8.516.72 × 10−615.464.33 × 10−833.464.67 × 10−11
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Behera, S.K.; Dash, R. A Novel Framework for Mental Illness Detection Leveraging TOPSIS-ModCHI-Based Feature-Driven Randomized Neural Networks. Math. Comput. Appl. 2025, 30, 67. https://doi.org/10.3390/mca30040067

AMA Style

Behera SK, Dash R. A Novel Framework for Mental Illness Detection Leveraging TOPSIS-ModCHI-Based Feature-Driven Randomized Neural Networks. Mathematical and Computational Applications. 2025; 30(4):67. https://doi.org/10.3390/mca30040067

Chicago/Turabian Style

Behera, Santosh Kumar, and Rajashree Dash. 2025. "A Novel Framework for Mental Illness Detection Leveraging TOPSIS-ModCHI-Based Feature-Driven Randomized Neural Networks" Mathematical and Computational Applications 30, no. 4: 67. https://doi.org/10.3390/mca30040067

APA Style

Behera, S. K., & Dash, R. (2025). A Novel Framework for Mental Illness Detection Leveraging TOPSIS-ModCHI-Based Feature-Driven Randomized Neural Networks. Mathematical and Computational Applications, 30(4), 67. https://doi.org/10.3390/mca30040067

Article Metrics

Back to TopTop