Next Article in Journal
Dynamic Behavior of the Stenting & Shielding Hernia System Fosters Neomyogenesis in Experimental Porcine Model
Previous Article in Journal
Development of a Computationally Efficient CFD Method for Blood Flow Analysis Following Flow Diverter Stent Deployment and Its Application to Treatment Planning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Supervised Learning and Large Language Model Benchmarks on Mental Health Datasets: Cognitive Distortions and Suicidal Risks in Chinese Social Media

1
College of Computer Science, Beijing University of Technology, Beijing 100124, China
2
Institut du Cerveau—Paris Brain Institute-ICM, Sorbonne Université, CNRS, Inria, Inserm, AP-HP, Hôpital de la Pitié-Salpêtrière, 75013 Paris, France
3
School of Nursing, Wuhan University, Wuhan 430071, China
*
Author to whom correspondence should be addressed.
Bioengineering 2025, 12(8), 882; https://doi.org/10.3390/bioengineering12080882
Submission received: 7 July 2025 / Revised: 11 August 2025 / Accepted: 14 August 2025 / Published: 19 August 2025
(This article belongs to the Section Biosignal Processing)

Abstract

On social media, users often express their personal feelings, which may exhibit cognitive distortions or even suicidal tendencies on certain specific topics. Early recognition of these signs is critical for effective psychological intervention. In this paper, we introduce two novel datasets from Chinese social media: SOS-HL-1K for suicidal risk classification, which contains 1249 posts, and SocialCD-3K, a multi-label classification dataset for cognitive distortion detection that contains 3407 posts. We conduct a comprehensive evaluation using two supervised learning methods and eight large language models (LLMs) on the proposed datasets. From the prompt engineering perspective, we experiment with two types of prompt strategies, including four zero-shot and five few-shot strategies. We also evaluate the performance of the LLMs after fine-tuning on the proposed tasks. Experimental results show a significant performance gap between prompted LLMs and supervised learning. Our best supervised model achieves strong results, with an F1-score of 82.76% for the high-risk class in the suicide task and a micro-averaged F1-score of 76.10% for the cognitive distortion task. Without fine-tuning, the best-performing LLM lags by 6.95 percentage points in the suicide task and a more pronounced 31.53 points in the cognitive distortion task. Fine-tuning substantially narrows this performance gap to 4.31% and 3.14% for the respective tasks. While this research highlights the potential of LLMs in psychological contexts, it also shows that supervised learning remains necessary for more challenging tasks.

1. Introduction

The prevalence of mental illness, particularly depression, continues to present significant challenges worldwide. According to a 2023 report from the World Health Organization (WHO), an estimated 3.8% of the global population experiences depression [1]. Specifically in China, mental health challenges are equally severe. The results of a landmark nationwide study, the China Mental Health Survey (CMHS), published by Huang et al. (2019) [2], revealed that the weighted lifetime prevalence of depressive disorders among Chinese adults is 6.8%, highlighting the pervasiveness of these mental health issues in the country. Such severe depression can often precipitate suicidal behaviors [3]. Platforms such as Twitter and Sina Weibo have become channels for emotional outbursts as social media has become more widespread [4]. Within these platforms, a certain subset of topics recur, with users frequently expressing deep-seated negative emotions and even showing suicidal tendencies [5,6].
Artificial intelligence (AI), especially the branches of deep learning and natural language processing, is a promising solution to this challenge [7]. Over recent years, AI research has resulted in the formulation of several algorithms tailored for emotion recognition within textual data [8]. However, traditional sentiment analysis, which primarily categorizes text as positive, negative, or neutral, is often insufficient. It fails to capture the underlying maladaptive reasoning patterns that define cognitive distortions, which often involve complex patterns of irrational thinking beyond surface sentiment and are more indicative of mental health risks. Moreover, these advancements are not without obstacles [9]. Constructing a potent deep learning model often demands considerable time and financial resources. The cost of data labeling, the need to engage domain experts, and the variability of model performance when moved to different application domains present significant challenges [10]. In specialized domains such as mental health, annotation must be performed by trained clinical professionals rather than crowd workers. For example, in a study on cognitive reframing, Sharma et al. [11] recruited 15 practicing mental health professionals and clinical psychology graduate students to label data, paying them CNY 37.50 per hour. This significantly exceeds the typical cost of annotation on crowdsourcing platforms, making it financially burdensome to scale high-quality datasets. This highlights a critical need for more flexible and adaptable algorithmic solutions, especially in the mental health domain [12]. In this context, the emergence of large language models (LLMs) is particularly noteworthy, and they have already attracted significant attention in the mental health domain [13].
Characterized by their large parameter and training datasets, LLMs stand as the state of the art in the framework of computational linguistics [14,15]. Their potential lies in their ability to comprehend and emulate human-like text nuances. Despite their promising potential, several studies have attempted to validate their practical performance. For instance, Xu et al. [16] examined four public datasets related to online social media sentiment detection. However, their study focused only on English data, and the classification task was relatively broad. This reveals a critical gap in Chinese-language mental health research, where cultural and linguistic nuances, such as indirect expressions, euphemisms, and idiomatic usage, affect how psychological distress is conveyed. These differences underscore the necessity of Chinese-specific datasets and highlight the novelty of our work on Chinese social media. A nuanced understanding of emotional expression is particularly important for the early detection of mental health issues, enabling timely and accurate intervention. Such capabilities further emphasize the importance of fine-grained emotion recognition in psychological assessment and therapy. Nevertheless, due to the lack of large-scale evaluations and real-world testing, concerns remain about the reliability of LLMs in high-stakes domains such as medicine and healthcare [15].
However, so far, there is a lack of large-scale datasets for psychology-related tasks, especially cognitive distortions. Sharma et al. [17] proposed an English dataset consisting of 600 entries with 15 labels, 14 of which pertain to cognitive distortions, making it a multi-label classification dataset. Despite the diversity of cognitive distortion labels, the dataset’s sample size is relatively small. The dataset, named Therapist Q&A, is available at https://www.kaggle.com/arnmaud/therapist-qa (accessed on 22 May 2025). It includes 2507 entries with 11 labels, 10 of which are cognitive distortion labels, also categorized as a multi-label classification dataset. In contrast, the landscape for Chinese datasets has been particularly constrained, not only in scale but also in annotation complexity. To the best of our knowledge, one of the first public resources is the C2D2 dataset [18]. However, to save costs, its authors abandoned the multi-label annotation approach in favor of single-labeling. This trend of simplification is further evidenced in the latest Chinese dataset by Lin et al. [19], which circumvents the challenge of classification altogether by focusing only on the binary detection of whether a distortion exists. In practice, it is common for a single text to reflect multiple cognitive distortions simultaneously. Our work is constructed with a label scheme methodologically distinct from existing works. We directly adopt the classic framework from Burns [20], a foundational standard in Cognitive-Behavioral Therapy (CBT), whereas other datasets often utilize inconsistent labeling systems. Consequently, our work represents the first systematic application of this authoritative clinical framework to the task of Chinese multi-label classification, ensuring our dataset’s professional quality and practical utility.
Despite increasing attention to suicide risk detection, data accessibility remains a major challenge. For example, the CLPsych 2024 shared task utilized the dataset to identify evidence of suicidality in Reddit posts, but the dataset was not publicly available due to privacy concerns [21]. The CLPsych 2025 task similarly adopted a password-protected dataset requiring formal agreements [22]. These restrictions underscore the broader difficulty of accessing high-quality annotated resources, even in English. In response, some efforts have aimed to improve openness. The RSD-15K dataset [23], for instance, provides around 15,000 Reddit posts with four-level suicide risk annotations and is publicly available upon request. However, such resources remain largely English-centric. In contrast, the lack of publicly available datasets in Chinese is even more severe. Cao et al. [24] created a binary classification dataset based on Sina Weibo, though not all posts reflect explicit suicidal intent. Huang et al. [25] collected data from 130 individuals who died by suicide, but the dataset is no longer accessible. These limitations highlight the urgent need for high-quality Chinese-language resources to advance suicide detection research in non-English contexts.
To provide more contribution to the research community, we present two datasets: SOS-HL-1K for high and low suicide risk classification, and SocialCD-3K for a cognitive distortion task. The SOS-HL-1K dataset contains 1249 social media posts reflecting high and low suicide risk, and the SocialCD-3K dataset contains 3407 social media posts annotated in a multi-label way. Both datasets were annotated by multiple domain experts. LLMs are prompt-driven and characterized by their flexibility and ease of use across different tasks. However, their specific performance in psychological tasks, especially in the Chinese domain, has not been fully explored. To address this, we evaluate the performance of eight LLMs and test the performance of four zero-shot and five few-shot prompt strategies on the two proposed datasets. We also test the performance of different types of LLMs fine-tuned for these tasks. The experimental results show a significant performance gap between prompted LLMs and supervised learning. This gap is 6.95 percentage points based on the F1-score for the high-risk class in the suicide task, widening to a more pronounced 31.53 points on the micro-averaged F1-score for the cognitive distortion task. Fine-tuning substantially narrows these gaps to 4.31% and 3.14%, respectively, though supervised learning retains the lead. To the best of our knowledge, our proposed dataset is the first Chinese cognitive distortion multi-label classification dataset. We evaluate the performance of LLMs from different angles, providing a valuable reference for researchers using LLMs in psychological tasks.
In summary, this work makes the following primary contributions:
1.
We introduce SocialCD-3K, the first Chinese multi-label cognitive distortion dataset, constructed from Burns’ classic Cognitive-Behavioral Therapy (CBT) framework, unlike prior simplified or inconsistent schemes.
2.
We present SOS-HL-1K, a publicly available Chinese social media dataset for suicide risk classification, addressing the scarcity of non-English resources.
3.
We provide a benchmark evaluating two supervised methods and eight LLMs across zero-shot, few-shot, and fine-tuning settings, offering a reference for future research.
4.
We analyze performance gaps between supervised methods and LLMs on simple vs. complex psychological tasks, revealing the current limitations of LLMs in specialized domains.

2. Related Work

2.1. Text Sentiment Analysis

Social media platforms have become central channels for the expression of emotions worldwide. Accurately and quickly identifying the emotions embedded in this data presents a big challenge to computational algorithms [8]. Fu et al. [26] developed a distant supervision approach for classifying high and low suicide risk on Chinese social media platforms, using annotations by experts at different levels, significantly reducing the dependency on extensive manual effort by professional experts. Employing this approach, coupled with the integration of key psychological features extracted from user blogs, they achieved an F1-score of 77.98%. Singh et al. [27] focused on COVID-19-related sentiment analysis on social media platforms, training a BERT-based model on these two datasets, achieving an accuracy of 94%. Wan [28] introduced a method for sentiment analysis of Sina Weibo comments, leveraging deep neural networks. The data undergo feature extraction through multi-level pooling and convolution layers, and key features are subsequently extracted from the feature matrix using a CNN. For the final classification and sentiment analysis, the softmax logistic regression method is employed. Zhang et al. [29] introduced a factor graph-based emotion recognition model that integrates correlations among emotion labels, social interactions, and temporal patterns into a unified framework, adeptly identifying multiple emotions through a multi-label learning approach.
Although deep learning algorithms demonstrate impressive performance, the requirement for significant volume of labeled data makes them difficult to apply. The utilization of the distant supervision approach proposed by Fu et al. [26] reduces the need for labeling but still involves expert groups. Given these limitations, there is a growing demand for efficient methods for emotion detection on social media. Recent advancements in LLMs offer potential solutions, attracting increased attention in the mental health domain.

2.2. Large Language Model and Its Applications in Medical Domain

The advent of Large Language Models (LLMs) like OpenAI’s GPT-4 has revolutionized natural language processing [30], significantly outperforming smaller models [31]. These models have diverse applications, including content generation [32], medical report assistance [33], coding assistance [34], education [35], and answering medical-related questions [36]. The large size of the model enables them to generate complex, contextually relevant content, making them flexible for use in downstream tasks.
LLMs have garnered significant attention in the medical domain [15]. Jiang et al. [37] developed the clinical LLM NYUTron to assist physicians, achieving AUC scores between 78.7 and 94.9%. It has been deployed in a prospective trial, indicating its potential for real-world application. In psychology, LLMs have also been applied to analyze suicidality, for instance by identifying linguistic markers [38], detecting risk cues [39], or extracting evidence from text [21,40]. However, a significant limitation of these pioneering works is their predominant focus on English-language data from platforms like Reddit, leaving the nuances of mental health expression in other cultural and linguistic contexts, such as Chinese social media, largely unexplored.
Other studies have explored LLMs as simulators for psychiatrist–patient interactions [41] or as support systems for non-professional counselors [42]. While insightful, these applications often rely on general-purpose models like ChatGPT, whose performance on specific, high-stakes psychological tasks requires deeper validation. For instance, a recent evaluation by Xu et al. [16] benchmarked several LLMs on English mental health datasets. Crucially, their study was confined to English and did not include complex multi-label classification, which is essential for phenomena like cognitive distortions, where multiple issues co-occur. Yang et al. [43] similarly found that while ChatGPT surpasses traditional methods, it still lags behind advanced, task-specific approaches, highlighting the need for more rigorous, domain-specific evaluations.
In summary, the existing literature, while promising, reveals two critical gaps that this paper aims to fill: (1) a scarcity of research focused on non-English languages like Chinese, and (2) a lack of systematic benchmarks on complex, clinically grounded tasks such as multi-label cognitive distortion detection. Our work directly confronts these limitations by introducing two novel Chinese datasets and providing a comprehensive benchmark analysis.

3. Methods

To comprehensively evaluate the capabilities of LLMs in psychologically relevant classification tasks, we construct a multi-stage comparison framework that encompasses traditional supervised learning, zero-shot and few-shot prompt-based inference, and fine-tuning of LLMs. Specifically, we focus on two tasks derived from Chinese social media: binary suicide risk classification and multi-label cognitive distortion detection. Our framework contrasts two supervised baselines with eight LLMs, including both general-purpose and Chinese-adaptive models, under multiple prompting strategies. We design a series of zero-shot and few-shot prompts tailored to psychological scenarios, and further explore fine-tuning techniques to enhance LLM performance. The overall goal is to understand the strengths and limitations of different approaches in this domain and to provide empirical guidance for practical deployment. The structure of our framework is illustrated in Figure 1.

3.1. Baseline Supervised Learning Models

  • LSAN [44]: The LSAN model uses label semantics to identify relationships between labels and documents, creating a label-specific document representation. It employs a self-attention mechanism to focus on this representation and uses an adaptive fusion strategy for multi-label text classification, proving effective in predicting low-frequency labels.
  • BERT [45]: BERT uses a bidirectional approach, facilitated by the Transformer architecture [46], to understand the full context of words. It is pre-trained with a masked language model objective, predicting masked words. BERT excels in various NLP tasks, such as question answering and sentiment analysis, especially when fine-tuned on specific data.

3.2. Large Language Models

For the LLM comparison experiments, we evaluated eight popular models, including two general LLMs (GPT-3.5 and GPT-4 [47]) and six Chinese-adaptive LLMs (ChatGLM2-6B, GLM-130B [48], GLM-4 [49], Chinese-Alpaca-2-7B [50], Chinese-LLaMA-2-7B [50], and LLaMA2-Chinese-7b-Chat [51]).

3.2.1. General LLMs: GPTs

  • GPT-3.5: GPT-3.5 is an advanced iteration of the GPT-3 [52] language model, offering improvements in conversational capabilities. It is designed to provide more coherent and context-aware responses in conversational applications, reflecting ongoing developments in language modeling techniques.
  • GPT-4 [47]: GPT-4 is a groundbreaking multimodal model that processes both images and text to generate text outputs. It performs at a human level on various benchmarks, including scoring in the top 10% on a simulated bar exam. Built on the Transformer architecture [46], GPT-4 is trained to predict tokens and undergoes post-training alignment for improved accuracy. Despite its capabilities, GPT-4 has limitations like occasional content hallucinations and a constrained context window.

3.2.2. Chinese LLMs

  • ChatGLM2-6B: ChatGLM2-6B is an open-source bilingual model with 6.2 billion parameters, optimized for Chinese question-answering and dialogue. Trained on about 1 TB of Chinese and English text, it can be fine-tuned through various techniques like supervised learning and human feedback, allowing for diverse language processing applications.
  • GLM-130B [48]: GLM-130B is a bilingual model with 130 billion parameters, optimized for English and Chinese. It aims to provide an open-source alternative comparable to GPT-3, outperforming GPT-3 175B on multiple English benchmarks and surpassing ERNIE TITAN 3.0 260B [53] on Chinese benchmarks.
  • GLM-4 [49]: GLM-4 uses a bidirectional Transformer [46] structure with the Masked and Prefix Language Model (MPCT) pre-training strategy, combining MLM and PLM strengths. It employs Rotary Position Embedding (RoPE) for long text sequences and adaptive masking for varied tasks, and supports multi-task learning.
  • Chinese-LLaMA-2-7B [50]: Chinese-LLaMa-2-7B, based on LLaMa-2 [54], uses a new vocabulary of 55,296 entries to enhance Chinese text coverage. It employs Low-Rank Adaptation (LoRA) [55] for fine-tuning, supporting multi-task learning, and demonstrating superior performance in Chinese tasks.
  • Chinese-Alpaca-2-7B [50]: Chinese-Alpaca-2-7B focuses on conversational and instruction-following capabilities. Using LoRA [55] for efficient training, it excels in multi-turn dialogues and complex command parsing after supervised fine-tuning.
  • LLaMA2-Chinese-7b-Chat [51]: Built on LLaMA-2 [54], LLaMA2-Chinese-7b-Chat is optimized for Chinese through continuous pre-training on a large-scale dataset. It leverages LoRA [55] for training, enhancing conversational capabilities with Chinese-specific commands and data.

3.3. LLM Prompt Strategies

LLMs use prompts as guides and are valued for their flexibility in performing downstream tasks. However, with LLMs it is crucial to design the prompts precisely, as this directly affects its understanding of the given task and the output of the results [56]. To this end, we design the following prompt strategies, including zero-shot and few-shot approaches. Zero-shot prompts denote direct interaction with the LLM without providing any task examples, while few-shot prompts involve interaction with the LLM by giving some examples within tasks, similar to the training data used in the supervised learning.

3.3.1. Zero-Shot Prompting

We begin our experiment with a prompt design tailored to tasks within a zero-shot paradigm, and it includes several strategies: direct task requests (represented as basic), role definition, scene definition, and a hybrid strategy. For illustrative purposes, the cognitive distortion classification task is used as an example. Note that, since our experimental data are in Chinese, we also use Chinese prompts. We provide translations of the English prompts in the following explanation. To avoid repetitions in prompts, we select the key part of each strategy and represent it as Strategy [Prompt]. When the prompt is a combination of several strategies from 1 to n, it is referred to as Strategy 1 + Strategy 2 + + Strategy n .
1.
Basic: A direct task statement without specific contextual emphasis. Prompt English translation: “ Basic [Please perform a multi-label classification task to determine whether it reflects any of the specified 12 cognitive distortions: (list of cognitive distortions categories)].”
2.
Role-definition prompting: This strategy delineates the role of the LLM (in this case, a psychologist) and emphasizes the need for psychological expertise. Prompt English translation: “ Role [Assuming the role of a psychologist with professional psychological experiences] + Basic
3.
Scene-definition prompting: Introduces the context of a social media environment, highlighting user identifiers to eliminate ambiguity. Prompt English translation: “ Scene [Given the user ID u and the associated posts on social media, based on the post content], + Basic
4.
Hybrid prompting: A combination of both role and scene definitions that provides an integrated prompt, and it can be represented as Scene + Role + Basic .

3.3.2. Few-Shot Prompting

Few-shot prompting is construed as a method to integrate prior knowledge or a batch of n training samples into LLMs, thereby enabling them to learn this information and adeptly execute the task. We experiment with the following setup.
1.
Background knowledge: This strategy is provided with psychological definitions, supplemented by representative cases, followed by one of the four prompting strategies devised from zero-shot prompting. Prompts that integrate background knowledge and employ the hybrid strategy from zero-shot prompting are detailed as follows: “ Background [There are the definitions of 12 cognitive distortions (list of cognitive distortions definitions), and there are the representative cases (list of cognitive distortions cases). Consider these cognitive distortions definitions and cases], + Scene + Role + Basic .”
2.
Prompting with n reference samples per category: In this strategy, reference data ref n are randomly selected from the training set for each category to construct prompts for the LLM, followed by one of the four prompting strategies designed from zero-shot prompting. Prompts that incorporate the reference data and employ the hybrid strategy from zero-shot prompting are detailed as follows: “ Reference [There are some examples of the target task with the ground truth label (list of reference samples). Consider these cognitive distortions examples], + Scene + Role + Basic .”
3.
Background knowledge and prompting with n reference samples per category: This approach investigates whether enhancing explanations in few-shot prompting can improve the LLM in terms of its understanding of psychological tasks. It incorporates psychological definitions, and provides n reference samples per category for LLM prompt conduction. The following example integrates background knowledge and reference instances, and employ the hybrid strategy from zero-shot prompting. It can be represented as Background + Reference + Scene + Role + Basic .

3.4. LLM Fine-Tuning for Downstream Task

Fine-tuning LLM is the process of adapting a pre-trained language model to a specific task or domain by continuing its training on a smaller, task-specific dataset [57]. Fine-tuning typically involves updating the model’s weights using supervised learning with labeled data, enhancing its performance in the targeted downstream task. Following the fine-tuning, our evaluation paradigm retained the role, scene, and hybrid definitions from the zero-shot prompting for consistency and comparative assessment as introduced in Section 3.3.1.

4. Experiments and Results

4.1. Datasets and Evaluation Metrics

We proposed two psychology-related classification datasets for high/low suicide risk classification and cognitive distortion multi-label classification, named “SOS-HL-1K” and “SocialCD-3K”. The suicide risk task primarily differentiates between high and low suicide risks, while the cognitive distortion task focuses on the classifications defined by Burns [20]. We sourced our data by crawling comments from the “Zoufan” blog within the Weibo social platform, a known space for users to express psychological distress, covering the period from 1 January 2017, to 31 December 2019. To ensure data relevance and quality, we first identified comments using keywords such as “depression”, “suicide”, and “despair”. Following this, all candidate posts underwent a manual review and cleaning process. We specifically excluded (1) positive and encouraging expressions, and (2) commercial advertisements, spam, and purely symbolic emoji comments. This multi-step preprocessing ensured our final datasets are composed of high-quality, genuine expressions of personal struggle suitable for psychological analysis.
To ensure the quality and reliability of our annotations, we employed three annotators with backgrounds in clinical psychology or counseling. Prior to annotation, all annotators underwent unified training using a detailed guide based on Burns’ cognitive distortion framework to ensure a shared understanding of each category’s definition. Each post in our dataset was independently annotated. To measure inter-rater reliability (IRR), we randomly sampled 20% of the data from both datasets for double-blind annotation. For the binary suicide risk classification task (SOS-HL-1K), we calculated Fleiss’ κ , achieving a score of 0.82, which indicates strong agreement. For the more complex multi-label cognitive distortion task (SocialCD-3K), we computed Krippendorff’s α , resulting in a value of 0.78, reflecting substantial agreement. Disagreements during the annotation process were resolved using a discussion-based strategy. If the initial annotators could not reach a consensus, a senior annotator, a licensed psychologist with over five years of clinical experience, served as the final arbitrator. After establishing this robust annotation standard, the remaining 80% of the data was annotated by the team. To ensure transparency and support future research, the detailed annotation guide, which includes definitions and representative examples for each category, is publicly available at https://github.com/HongzhiQ/SupervisedVsLLM-EfficacyEval (accessed on 28 May 2024).
The suicide detection dataset is a binary classification dataset consisting of 1249 samples. Among these, there are 648 posts with low suicide risk and 601 posts with high suicide risk. The cognitive distortion dataset contains a total of 3407 posts. This is a multi-label classification dataset, with each sample corresponding to one or more cognitive distortion categories. The classification labels used for this data follow the definitions provided by Burns [20]. For both sets of data, the training set and test set are divided according to the ratio of 4:1. The data distribution and specific categories of these two datasets are listed in Table 1. We utilize three primary evaluation metrics to measure the performance of different algorithms for our two tasks: precision, recall, and F1-score. Given the distinct nature of the two classification tasks, for the binary suicide risk classification task, we report micro-averaged and macro-averaged metrics as well as the specific precision, recall, and F1-score for the high-risk class. We emphasize the F1-score of the high-risk class in our main findings, as this positive class is the primary target for clinical intervention and early detection. For the multi-label cognitive distortion task, we report both micro-averaged and macro-averaged F1-scores to provide a comprehensive view of model performance. In line with best practices for imbalanced datasets, we primarily highlight the micro-averaged F1-score in our analysis. This method aggregates the contributions of all classes to compute an overall score, which is particularly useful in scenarios with a skewed distribution of class instances. The macro-averaged F1-score is presented as a complementary metric, offering insights into the model’s unweighted average performance across all categories, treating each class equally regardless of its frequency.

4.2. Experiment Design

Our experiments focus on validating the two proposed tasks, particularly testing different strategies for using LLMs. This includes four zero-shot and three few-shot prompt construction strategies, as well as the performance of fine-tuned LLMs. In total, we compare two types of supervised learning methods and eight LLMs. Using cognitive distortions as an example to show our points, our evaluations spanned several dimensions:
  • Prompt design: Initially, we assessed four prompting strategies within the zero-shot learning conduction. Subsequently, based on their performance, the top two strategies were selected for further evaluation in the few-shot learning setting across various LLMs.
  • LLM ability: In our experiments, we found that the performance of the ChatGLM2-6B and GLM-130B models performed poorly. Therefore, we decided to use the latest version of the GLM series, the GLM-4 model, to evaluate performance on the newly expanded dataset. For GPT-3.5, its token limitation prevented us from entering five samples for each category during few-shot prompting. Consequently, we reserved the ref 5 approach exclusively for GPT-4.
  • LLM fine-tuning: OpenAI recently introduced a fine-tuning feature for GPT-3.5, and official reports suggest that, under certain conditions, fine-tuned GPT-3.5 may outperform GPT-4 [58]. Therefore, we experimented with fine-tuning GPT-3.5. Since the current version of GPT-4 lacks fine-tuning capabilities, we were unable to assess its potential. Additionally, we explored the fine-tuning performance of three other open-source Chinese large models in the cognitive distortion recognition task.
Note that, since the experimental performance of the three Chinese LLMs (Chinese-Alpaca-2-7B, Chinese-LLaMA-2-7B, and LLaMA2-Chinese-7b-Chat) on cognitive distortion was similar to the performance of GPT-3.5 after fine-tuning, we did not conduct experiments with these three LLMs in the suicide risk classification task. Also, ChatGLM2-6B’s performance in the cognitive distortion classification task was too poor to even output in the correct format, so we did not include its experimental results. By the time we experimented with the cognitive distortion task, GLM-130B had been replaced by a newer version, GLM-4, so we evaluated GLM-4 in this task instead.

4.3. Implementation Details

Prior to model input, all text underwent a standardized data preprocessing pipeline. First, to ensure user privacy and data cleanliness, personally identifiable information such as usernames was masked, and all URLs were removed. Subsequently, all emojis were removed from the text. To preserve the full contextual integrity required by modern language models, we deliberately retained stopwords. Finally, the preprocessed text was fed into each respective model, where its native tokenizer automatically performed the necessary steps, including tokenization, conversion to numerical IDs, and the addition of special tokens (e.g., [CLS] and [SEP]) to convert the text into the required input format.
  • LSAN: We used Word2vec [59] to train 300-dimensional embeddings for both the document and randomly initialized label texts. The function of the attention mechanism is to compute word contributions to labels and create label-specific document representations. Dot products between these document and label vectors refined these relationships further. These two types of document representations were then fused using weighted combinations. For predictions, we employed a fully connected layer, followed by ReLU and a sigmoid function. We used cross-entropy as the loss function.
  • BERT: We employed a Chinese pre-trained BERT model to extract 768-dimensional vectors from the sentences. To avoid overfitting, a dropout function [60] was applied to these sentence vectors. Subsequently, a fully connected layer was introduced for classification. The sigmoid function served as the activation function for the output layer.
  • LLM-zero shot: Both GPT-3.5 and GPT-4 [47] are closed-source and available through OpenAI’s API. We used Gpt-3.5-turbo because it is one of the most capable and cost-effective models in the GPT-3.5 family, and GPT-4 for its advanced capabilities. For GLM models, we deployed the open-source ChatGLM2-6B [48] on our server, and tested the larger GLM-130B [48] via its official website due to deployment challenges. We averaged the performance over five rounds of experiments for all models to minimize randomness. For GPT-3.5, GPT-4, ChatGLM2-6B, and GLM-4, we set the temperature to 0.1, 0.3, 0.5, 0.7, and 0.9, and averaged the results over five rounds of experiments. For GLM-130B, we could not adjust the temperature due to its limitations.
  • LLM-few shot: We conducted the experiments using the top two performing prompt strategies from the zero-shot performance as determined by their F1-scores. Given the different input token constraints of each model, we selected varying amounts of reference data ref n according to the requirements of the corresponding models.
  • LLM-fine-tuning: For the closed-source model GPT-3.5, we used the API provided by OpenAI to fine-tune the GPT-3.5 Turbo model for suicide risk and cognitive distortions tasks. In this experiment, the training epoch was set to 3. For other hyperparameters, we did not explicitly specify them; instead, OpenAI selected the default values based on the dataset size. We also explored the performance of three open-source Chinese LLMs (Chinese-LLaMA-2-7B, Chinese-Alpaca-2-7B, and LLaMA2-Chinese-7b-Chat) on this fine-tuning task. These models were fine-tuned locally on an NVIDIA A100 GPU. The training process was configured with a batch size of 8, a learning rate of 6 × 10 5 using a cosine scheduler, and a weight decay of 0.01, and was run for 5 epochs, with model checkpoints saved every 300 steps. We utilized Low-Rank Adaptation (LoRA) for parameter-efficient fine-tuning. The LoRA-specific configuration was set to a rank of 8, an alpha of 32, and a lora_dropout rate of 0.05. The LoRA modules were applied to the attention layers’ query and value projection matrices (q_proj and v_proj). To ensure full reproducibility, the complete configuration file for these experiments is available in our GitHub repository.
To ensure the robustness and reproducibility of our findings, all experiments were conducted multiple times. For the supervised learning models, we performed five independent runs using different random seeds (42, 52, 62, 72, and 82) for data shuffling and model initialization, and reported the average and standard deviation of the results. For all LLM-based evaluations (zero-shot, few-shot, and fine-tuning), we conducted experiments under five different temperature settings (0.1, 0.3, 0.5, 0.7, and 0.9). The final reported LLM scores were reported as the mean and standard deviation across these five runs. This approach allows us to assess the stability of the supervised models against random initialization and the sensitivity of the LLMs to the temperature hyperparameter. To assess the statistical significance of performance differences between experimental conditions, we conducted paired-samples t-tests with a significance level of α = 0.05 .
To ensure transparency and reproducibility, we detail the computational environment used for our experiments. All local experiments were conducted on a workstation equipped with a single NVIDIA A100 GPU (40 GB VRAM) and 256 GB of RAM. For supervised learning models (e.g., BERT and LSAN), the average fine-tuning time was approximately 1 h. Fine-tuning LLMs with LoRA (e.g., Chinese-Alpaca-2-7B and LLaMA2-Chinese-7b-Chat) required between 3 and 5 h depending on the model size. For GPT-3.5, fine-tuning was performed via OpenAI’s API service. Due to API rate limits and network latency, each run took approximately 4 to 5 h to complete. Although not GPU-bound locally, GPT-3.5 fine-tuning was the most time consuming among all models. All zero-shot and few-shot evaluations were completed within seconds per instance and conducted using the same A100 GPU, except for GPT-3.5 and GPT-4, which were accessed via the OpenAI API.

5. Results

5.1. Effect of Prompting Strategies on Task Performance

In our study, we primarily aimed to answer two questions: how can LLMs be most simply used to accomplish a task, and what is the performance potential of LLMs on both tasks? The order of using LLMs ranges from simple to complex, starting with zero-shot and few-shot prompt engineering to the fine-tuning. This analysis is detailed in the following sections, and the results of suicide classification and cognitive distortions classification can be seen in Table 2 and Table 3, respectively.
In the zero-shot prompting strategy, model performance exhibited a clear dependence on task complexity. For the relatively simple suicide risk classification task, all LLMs, except for the smaller ChatGLM2-6B, achieved F1-scores close to 70% on the high-risk class. As shown in Table 2, the performance gap among the top-performing models was small. GPT-4, using the scene-definition strategy, achieved an average F1-score of 73.39% for high-risk detection, slightly outperforming the second-best GLM-130B (72.20%, hybrid strategy). Compared to other models, GPT-4 significantly outperformed GPT-3.5 (68.59%, t = 7.18 , p < 0.001 ) and ChatGLM2-6B (53.74%, t = 4.59 , p < 0.01 ). Although GPT-4 achieved the highest mean F1-score among all models, its difference from GLM-130B was not statistically significant, indicating that GLM-130B is a competitive alternative in the zero-shot setting. Overall, the best-performing GPT-4 and GLM-130B models showed micro-F1 and macro-F1-scores that were close to each other but generally lower than their high-risk class F1 (by approximately 1–8 percentage points), suggesting that while overall and macro-averaged performance was stable, high-risk sample detection remained the most discriminative indicator.
The importance of prompt engineering was also validated in this experiment. Different models benefited from different prompt construction strategies; for example, the hybrid strategy worked best for GLM-130B, whereas the role-definition strategy was optimal for GPT-3.5. Notably, for GPT-4, its best prompt strategy (scene-definition) outperformed its weakest strategy (basic) by approximately 1.67 percentage points in average F1, but the difference did not reach statistical significance ( p = 0.094 ). This suggests that prompt design can sometimes improve performance but may yield limited and unstable gains. However, zero-shot prompting proved inadequate for the more complex cognitive distortion multi-label classification task. The best-performing zero-shot model, GLM-4 (role-definition strategy), achieved a micro-F1 of only 28.59%, with an even lower macro-F1 of 22.11%. Compared to the supervised BERT model (micro-F1: 76.10%, macro-F1: 70.05%), this represents substantial performance gaps of 47.51 and 47.94 percentage points, respectively, both highly significant (micro-F1: t = 57.40 , p < 0.001 ; macro-F1: t = 44.68 , p < 0.001 ). Furthermore, the significantly lower macro-F1 compared to micro-F1 (e.g., GPT-4: t = 5.10 , p < 0.01 ) can be attributed to the downward pull from long-tail categories in the evaluation phase. Some categories had very few test samples, making their F1-scores highly sensitive to single prediction successes or failures, often dropping close to zero with minor misclassifications. Consequently, while zero-shot prompting may suffice for well-defined binary classification tasks, it falls far short for complex multi-label reasoning tasks like cognitive distortion detection.
The few-shot prompting strategy involves incorporating a small number of labeled examples into the prompt as references to guide the model in completing the task. In the suicide risk classification task, this approach significantly improved upon the zero-shot baseline. For example, GPT-4’s high-risk class mean F1 increased from 73.39% to 75.81% ( t = 2.63 , p < 0.05 ), with both micro-F1 and macro-F1 increasing by 8.88 and 12.68 percentage points, respectively, each reaching statistical significance (micro-F1: t = 6.37 , p < 0.01 ; macro-F1: t = 5.72 , p < 0.01 ). For ChatGLM2-6B, this strategy yielded an 11.04 percentage point improvement in high-risk F1, with the best result under the background + ref12 + hybrid setting. However, increasing the example count to ref 30 did not further improve performance and instead caused a decline (F1 from 64.78% to 51.88%), though this drop was not statistically significant ( p > 0.05 ), suggesting that overly long contexts may introduce distracting information and hinder reasoning. In the cognitive distortion multi-label classification task, overall performance under few-shot settings remained low (best F1 only 43.78%) but still improved significantly compared to zero-shot. For example, GLM-4’s micro-F1 increased by 15.19 percentage points ( p < 0.001 ), and GPT-4’s micro-F1 improved by 13.78% ( t = 26.05 , p < 0.001 ). In summary, few-shot prompting can deliver statistically significant gains across tasks, but the improvement magnitude strongly depends on the initial performance baseline: for tasks already performing well (e.g., suicide risk classification), gains are modest but consistent; for low-baseline, complex reasoning tasks (e.g., cognitive distortion detection), improvements are much larger.
Fine-tuned LLMs achieved better performance in both tasks. In the suicide classification task, the fine-tuned GPT-3.5 model significantly outperformed the zero-shot model, improving high-risk F1 by 9.86 percentage points ( t = 63.02 , p < 0.001 ). In the cognitive distortion task, the performance gain was even more pronounced: compared with the best zero-shot prompt strategy, the fine-tuned GPT-3.5’s micro-average F1 improved dramatically (mean: 72.21% vs. 12.06%), with the difference being highly significant ( t = 76.01 , p < 0.001 ). However, as the performance gaps among fine-tuned LLMs in the cognitive distortion task were small, no further fine-tuned LLM comparisons were conducted for the suicide risk classification task. Despite the substantial improvements from fine-tuning, supervised learning maintained a performance advantage in both tasks. In the suicide classification task, BERT achieved the highest high-risk F1-score of 82.76%, surpassing the best fine-tuned GPT-3.5 model (mean: 78.45%), with the difference confirmed as statistically significant ( t = 53.52 , p < 0.001 ). Similarly, in the cognitive distortion detection task, BERT outperformed all LLM-based methods, achieving a micro-F1 of 76.10% compared to 72.96% for the best LLM (Chinese-Alpaca-2-7B), a difference that was also statistically significant ( t = 26.71 , p < 0.001 ). More importantly, we observed that all models achieved much higher micro-F1-scores than macro-F1-scores in this task (e.g., BERT: 76.10% vs. 70.05%; Chinese-Alpaca-2-7B: 72.96% vs. 66.89%). This significant gap indicates that due to the imbalanced distribution of cognitive distortion categories in the dataset, models performed well on frequent categories but poorly on rare ones. Micro-F1, being dominated by large classes, was thus inflated, whereas macro-F1, giving equal weight to all classes, exposed the models’ shortcomings in handling long-tail categories. Furthermore, while task-specific fine-tuning can enhance LLM performance, it still cannot surpass supervised learning. Fine-tuning also increases LLM usage complexity and cost, reducing flexibility and versatility. LLMs typically have far more parameters than supervised models, making them more expensive to train and deploy.
Overall, LLMs are effective for performing relatively simple tasks using only zero-shot prompt, but they need to be properly designed with prompt such as scene settings. For more difficult tasks, LLMs cannot perform adequately through prompt-only solution. The few-shot strategy, which introduces training data into the prompt, is effective but does not result in a significant performance improvement. Fine-tuning LLMs for specific tasks can achieve substantial improvements, but their performance still falls short of that achieved by supervised learning. Supervised learning remains the best choice for low-cost and high-performance in domain-specific tasks.

5.2. Temperature Sensitivity Analysis

To examine the effect of decoding temperature on model performance, we analyzed the results across different temperature values ( T 0.1 , 0.3 , 0.5 , 0.7 , 0.9 ) for both the zero-shot and few-shot settings of the best-performing LLM in each task as illustrated in Figure 2. For the suicide risk classification task, GPT-4 under the zero-shot scene-definition setting achieved its highest high-risk F1-score at T = 0.3 (75.16%), with stable performance for T 0.5 and a gradual decline thereafter ( T = 0.9 , 71.52%). In the few-shot background+ref30+hybrid setting, performance peaked at T = 0.1 (82.89%), dropped markedly at T = 0.5 (75.02%), and remained moderate for higher temperatures. Notably, at T = 0.9 , the score decreased substantially to 69.57%, indicating that few-shot prompting becomes less stable when the temperature is set too high, likely due to increased randomness in token sampling.
For the cognitive distortion multi-label classification task, GLM-4 in the zero-shot role-definition setting showed its best micro-average F1 at T = 0.3 (31.24%), with a consistent downward trend as temperature increased. In the few-shot background + ref2 + hybrid setting, performance was highest at T = 0.1 (45.12%) and decreased steadily to 42.54% at T = 0.9 . These results indicate that lower temperatures (0.1–0.3) generally yield better and more stable classification outcomes for both tasks, while higher temperatures introduce greater output variability and reduced accuracy, consistent with the precision requirements of psychological risk detection.

6. Expert Evaluation and Feedback

As mentioned above, using LLMs with prompts alone, without fine-tuning, is the best case scenario for their role. However, LLMs do not perform well under zero-shot and few-shot prompts. For this reason, we selected the output of GPT-4, which performs relatively well under the “ ref 2 + basic” strategy, for further analysis. The examples can be seen in Figure 3.
Overall, the model’s judgments tend to be overly predictive and overly sensitive. When dealing with simple texts and scenarios, such as in Example A, the model can often make accurate predictions. However, simplicity does not necessarily imply clarity; in fact, seemingly simple statements may contain ambiguous or implicit meanings that are more difficult to evaluate. For instance, in Example B, the model identifies negative expectations regarding exam outcomes and associated emotions, classifying the text under “the fortune teller error” and “should statements”. However, it fails to capture the core cognitive pattern: the user’s generalized conclusion based on past failures. A more appropriate categorization would be “overgeneralization”, as the writer extrapolates that “the exams will never be passed” based on a limited number of unsuccessful attempts.
These examples demonstrate that while the model can provide generally plausible classifications, it often misses the central semantic feature of the text, leading to mislabeling or incorrect predictions. When processing longer and more complex texts, as in Examples C and D, the model often succeeds in identifying one or more correct categories but exhibits a consistent tendency toward over-prediction. In Example C, the model correctly labels the sentence with “mental filter” but also incorrectly assigns three additional categories: “disqualifying the positive”, “the fortune teller error”, and “blaming oneself”. These false positives are not substantiated by the content—there is no explicit denial of positive experiences, no exaggerated future prediction, nor any self-blaming element. Similarly, in Example D, the model correctly identifies “mental filter” and “magnification”, reflecting the user’s focus on negative details and exaggeration of the problem’s severity. However, it further predicts three unsupported distortions—“emotional reasoning”, “blaming oneself”, and “blaming others”—none of which are explicitly expressed in the original text.
This over-prediction behavior may be influenced by the prompting strategy. The hybrid prompt, which frames the model as a psychologist, may lead to hyper-vigilant responses. In an effort to fulfill this role comprehensively, the model becomes overly sensitive to subtle cues, resulting in the identification of both well-supported distortions and additional categories based on weak or inferred signals. This is particularly evident in Example D, where the model appears to over-interpret the user’s emotional intensity, resulting in multiple false positives.
Overall, these examples highlight that while the model often aligns with the general direction of the user’s intent, it tends to over-interpret nuanced language, thereby reducing prediction precision. Future work should focus on enhancing the model’s ability to prioritize central semantic features over peripheral cues, which may help reduce over-prediction and improve overall classification accuracy.

7. Conclusions

In this paper, we present two psychologically relevant datasets based on social media data: SocialCD-3K for cognitive distortion multi-label classification, and SOS-HL-1K for high/low suicide risk classification. We validate the performance of two supervised learning models and eight LLMs, including both Chinese and English models, on these tasks. Our experiments evaluate the performance of LLMs under different prompt designs. The results show that prompt-only LLMs can perform reasonably well on the suicide risk classification task but fall short in more complex scenarios such as cognitive distortion recognition. While fine-tuning improves performance, it still underperforms compared to supervised learning methods. Additionally, fine-tuning may reduce the flexibility and ease of use of LLMs in deployment. Nevertheless, our study provides new research resources to the community and offers detailed experimental insights for applying LLMs in the mental health domain. The SocialCD-3K dataset could also support early-warning systems for suicide prevention hotlines by offering a rich source of annotated linguistic patterns associated with different levels of psychological distress. As future work, we will expand the datasets and conduct cross-platform and out-of-time evaluations, while exploring confidence calibration, selective prediction, and prompt-robustness strategies to reduce over-prediction and improve real-world generalization.

Author Contributions

Conceptualization, H.Q. and G.F.; Data curation, D.L., S.L. and Y.Y.; Formal analysis, G.F.; Funding acquisition, Q.Z.; Investigation, J.L. and B.Y.; Methodology, H.Q., C.S. and W.Z.; Project administration, D.L.; Resources, B.Y. and Q.Z.; Supervision, J.L., D.L. and Q.Z.; Validation, H.Q., G.F., C.S., W.Z., S.L., Y.Y.; Writing—original draft, H.Q.; Writing—review and editing, J.L., B.Y., Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Beijing Natural Science Foundation (Grant No. 7254302); the National Natural Science Foundation of China (Grant Nos. 72174152 and 72474166); the Fundamental Research Funds for the Central Universities (Grant Nos. 2042022kf1218 and 2042022kf1037); and the Young Top-notch Talent Cultivation Program of Hubei Province. The APC was funded by Qing Zhao.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the use of publicly available social media data (Weibo), which were collected in compliance with the platform’s terms of service and subsequently anonymized to ensure that no personally identifiable information was retained.

Informed Consent Statement

Patient consent was waived due to the use of publicly available and anonymized social media data (Weibo), which do not contain any personally identifiable information.

Data Availability Statement

The dataset generated and analyzed in this study is publicly available at https://github.com/HongzhiQ/SupervisedVsLLM-EfficacyEval (accessed on 28 May 2024). It includes annotated Weibo posts and evaluation code used for comparing supervised methods and large language models in efficacy evaluation tasks.

Acknowledgments

We would like to express our gratitude to the anonymous reviewers for their insightful comments and constructive feedback, which greatly helped to improve the quality of this manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. World Health Organization. Depressive Disorder (Depression); World Health Organization: Geneva, Switzerland, 2023. [Google Scholar]
  2. Huang, Y.; Wang, Y.; Wang, H.; Liu, Z.; Yu, X.; Yan, J.; Yu, Y.; Kou, C.; Xu, X.; Lu, J.; et al. Prevalence of mental disorders in China: A cross-sectional epidemiological study. Lancet Psychiatry 2019, 6, 211–224. [Google Scholar] [CrossRef]
  3. World Health Organization. Suicide; World Health Organization: Geneva, Switzerland, 2023. [Google Scholar]
  4. Keles, B.; McCrae, N.; Grealish, A. A systematic review: The influence of social media on depression, anxiety and psychological distress in adolescents. Int. J. Adolesc. Youth 2020, 25, 79–93. [Google Scholar] [CrossRef]
  5. Robinson, J.; Cox, G.; Bailey, E.; Hetrick, S.; Rodrigues, M.; Fisher, S.; Herrman, H. Social media and suicide prevention: A systematic review. Early Interv. Psychiatry 2016, 10, 103–121. [Google Scholar] [CrossRef]
  6. Luxton, D.D.; June, J.D.; Fairall, J.M. Social media and suicide: A public health perspective. Am. J. Public Health 2012, 102, S195–S200. [Google Scholar] [CrossRef]
  7. Coppersmith, G.; Leary, R.; Crutchley, P.; Fine, A. Natural language processing of social media as screening for suicide risk. Biomed. Inform. Insights 2018, 10, 1178222618792860. [Google Scholar] [CrossRef] [PubMed]
  8. Nandwani, P.; Verma, R. A review on sentiment analysis and emotion detection from text. Soc. Netw. Anal. Min. 2021, 11, 81. [Google Scholar] [CrossRef] [PubMed]
  9. Acheampong, F.A.; Wenyu, C.; Nunoo-Mensah, H. Text-based emotion detection: Advances, challenges, and opportunities. Eng. Rep. 2020, 2, e12189. [Google Scholar] [CrossRef]
  10. Saunders, D. Domain adaptation and multi-domain adaptation for neural machine translation: A survey. J. Artif. Intell. Res. 2022, 75, 351–424. [Google Scholar] [CrossRef]
  11. Sharma, A.; Rushton, K.; Lin, I.W.; Wadden, D.; Lucas, K.G.; Miner, A.S.; Nguyen, T.; Althoff, T. Cognitive reframing of negative thoughts through human-language model interaction. arXiv 2023, arXiv:2305.02466. [Google Scholar] [CrossRef]
  12. Laparra, E.; Bethard, S.; Miller, T.A. Rethinking domain adaptation for machine learning over clinical language. JAMIA Open 2020, 3, 146–150. [Google Scholar] [CrossRef]
  13. He, T.; Fu, G.; Yu, Y.; Wang, F.; Li, J.; Zhao, Q.; Song, C.; Qi, H.; Luo, D.; Zou, H.; et al. Towards a psychological generalist AI: A survey of current applications of large language models and future prospects. arXiv 2023, arXiv:2312.04578. [Google Scholar] [CrossRef]
  14. Zhao, W.X.; Zhou, K.; Li, J.; Tang, T.; Wang, X.; Hou, Y.; Min, Y.; Zhang, B.; Zhang, J.; Dong, Z.; et al. A survey of large language models. arXiv 2023, arXiv:2303.18223. [Google Scholar]
  15. Thirunavukarasu, A.J.; Ting, D.S.J.; Elangovan, K.; Gutierrez, L.; Tan, T.F.; Ting, D.S.W. Large language models in medicine. Nat. Med. 2023, 29, 1930–1940. [Google Scholar] [CrossRef]
  16. Xu, X.; Yao, B.; Dong, Y.; Yu, H.; Hendler, J.; Dey, A.K.; Wang, D. Leveraging Large Language Models for Mental Health Prediction via Online Text Data. arXiv 2023, arXiv:2307.14385. [Google Scholar]
  17. Sharma, A.; Rushton, K.; Lin, I.; Wadden, D.; Lucas, K.; Miner, A.; Nguyen, T.; Althoff, T. Cognitive Reframing of Negative Thoughts through Human-Language Model Interaction. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, Canada, 9–14 July 2023; Rogers, A., Boyd-Graber, J., Okazaki, N., Eds.; pp. 9977–10000. [Google Scholar] [CrossRef]
  18. Wang, B.; Deng, P.; Zhao, Y.; Qin, B. C2D2 Dataset: A Resource for the Cognitive Distortion Analysis and Its Impact on Mental Health. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Singapore, 6–10 December 2023. [Google Scholar]
  19. Lin, S.; Wang, Y.; Dong, J.; Ni, S. Detection and Positive Reconstruction of Cognitive Distortion Sentences: Mandarin Dataset and Evaluation. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2024, Bangkok, Thailand, 11–16 August 2024; Ku, L.W., Martins, A., Srikumar, V., Eds.; pp. 6686–6701. [Google Scholar] [CrossRef]
  20. Burns, D.D. Feeling Good; Signet Book: New York, NY, USA, 1981; Volume 4. [Google Scholar]
  21. Zhu, J.; Xu, A.; Tan, M.; Yang, M. XinHai@CLPsych 2024 Shared Task: Prompting Healthcare-oriented LLMs for Evidence Highlighting in Posts with Suicide Risk. In Proceedings of the 9th Workshop on Computational Linguistics and Clinical Psychology (CLPsych 2024), St. Julians, Malta, 21 March 2024; Yates, A., Desmet, B., Prud’hommeaux, E., Zirikly, A., Bedrick, S., MacAvaney, S., Bar, K., Ireland, M., Ophir, Y., Eds.; pp. 238–246. [Google Scholar]
  22. Kim, L. A Baseline for Self-state Identification and Classification in Mental Health Data: CLPsych 2025 Task. In Proceedings of the 10th Workshop on Computational Linguistics and Clinical Psychology (CLPsych), St. Julians, Malta, 3 May 2025; pp. 218–224. [Google Scholar]
  23. Zheng, S.; Tao, Y.; Zhou, T. RSD-15K: A Large-Scale User-Level Annotated Dataset for Suicide Risk Detection on Social Media. arXiv 2025, arXiv:2507.11559. [Google Scholar] [CrossRef]
  24. Cao, L.; Zhang, H.; Feng, L.; Wei, Z.; Wang, X.; Li, N.; He, X. Latent Suicide Risk Detection on Microblog via Suicide-Oriented Word Embeddings and Layered Attention. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; Inui, K., Jiang, J., Ng, V., Wan, X., Eds.; pp. 1718–1728. [Google Scholar] [CrossRef]
  25. Huang, X.; Xing, L.; Brubaker, J.R.; Paul, M.J. Exploring timelines of confirmed suicide incidents through social media. In Proceedings of the 2017 IEEE International Conference on Healthcare Informatics (ICHI), Park City, UT, USA, 23–26 August 2017; pp. 470–477. [Google Scholar]
  26. Fu, G.; Song, C.; Li, J.; Ma, Y.; Chen, P.; Wang, R.; Yang, B.X.; Huang, Z. Distant supervision for mental health management in social media: Suicide risk classification system development study. J. Med. Internet Res. 2021, 23, e26119. [Google Scholar] [CrossRef]
  27. Singh, M.; Jakhar, A.K.; Pandey, S. Sentiment analysis on the impact of coronavirus in social life using the BERT model. Soc. Netw. Anal. Min. 2021, 11, 33. [Google Scholar] [CrossRef]
  28. Wan, F. Sentiment analysis of Weibo comments based on deep neural network. In Proceedings of the 2019 International Conference on Communications, Information System and Computer Engineering (CISCE), Haikou, China, 5–7 July 2019; pp. 626–630. [Google Scholar]
  29. Zhang, X.; Li, W.; Ying, H.; Li, F.; Tang, S.; Lu, S. Emotion detection in online social networks: A multilabel learning approach. IEEE Internet Things J. 2020, 7, 8133–8143. [Google Scholar] [CrossRef]
  30. Kaddour, J.; Harris, J.; Mozes, M.; Bradley, H.; Raileanu, R.; McHardy, R. Challenges and Applications of Large Language Models. arXiv 2023, arXiv:2307.10169. [Google Scholar] [CrossRef]
  31. Wei, J.; Tay, Y.; Bommasani, R.; Raffel, C.; Zoph, B.; Borgeaud, S.; Yogatama, D.; Bosma, M.; Zhou, D.; Metzler, D.; et al. Emergent abilities of large language models. arXiv 2022, arXiv:2206.07682. [Google Scholar] [CrossRef]
  32. Liebrenz, M.; Schleifer, R.; Buadze, A.; Bhugra, D.; Smith, A. Generating scholarly content with ChatGPT: Ethical challenges for medical publishing. Lancet Digit. Health 2023, 5, e105–e106. [Google Scholar] [CrossRef]
  33. Jeblick, K.; Schachtner, B.; Dexl, J.; Mittermeier, A.; Stüber, A.T.; Topalis, J.; Weber, T.; Wesp, P.; Sabel, B.; Ricke, J.; et al. ChatGPT makes medicine easy to swallow: An exploratory case study on simplified radiology reports. arXiv 2022, arXiv:2212.14882. [Google Scholar] [CrossRef] [PubMed]
  34. Surameery, N.M.S.; Shakor, M.Y. Use ChatGPT to solve programming bugs. Int. J. Inf. Technol. Comput. Eng. (IJITC) 2023, 3, 17–22. [Google Scholar]
  35. Kasneci, E.; Seßler, K.; Küchemann, S.; Bannert, M.; Dementieva, D.; Fischer, F.; Gasser, U.; Groh, G.; Günnemann, S.; Hüllermeier, E.; et al. ChatGPT for good? On opportunities and challenges of large language models for education. Learn. Individ. Differ. 2023, 103, 102274. [Google Scholar] [CrossRef]
  36. Yeo, Y.H.; Samaan, J.S.; Ng, W.H.; Ting, P.S.; Trivedi, H.; Vipani, A.; Ayoub, W.; Yang, J.D.; Liran, O.; Spiegel, B.; et al. Assessing the performance of ChatGPT in answering questions regarding cirrhosis and hepatocellular carcinoma. Clin. Mol. Hepatol. 2023, 29, 721–732. [Google Scholar] [CrossRef]
  37. Jiang, L.Y.; Liu, X.C.; Nejatian, N.P.; Nasir-Moin, M.; Wang, D.; Abidin, A.; Eaton, K.; Riina, H.A.; Laufer, I.; Punjabi, P.; et al. Health system-scale language models are all-purpose prediction engines. Nature 2023, 619, 357–362. [Google Scholar] [CrossRef]
  38. Bauer, B.; Norel, R.; Leow, A.; Rached, Z.A.; Wen, B.; Cecchi, G. Using Large Language Models to Understand Suicidality in a Social Media–Based Taxonomy of Mental Health Disorders: Linguistic Analysis of Reddit Posts. JMIR Ment. Health 2024, 11, e57234. [Google Scholar] [CrossRef]
  39. Koushik, L.; Vishruth, M.; Anand Kumar, M. Detecting suicide risk patterns using hierarchical attention networks with large language models. In Proceedings of the 9th Workshop on Computational Linguistics and Clinical Psychology (CLPsych 2024), St. Julians, Malta, 21 March 2024; pp. 227–231. [Google Scholar]
  40. Alhamed, F.; Ive, J.; Specia, L. Using large language models (llms) to extract evidence from pre-annotated social media data. In Proceedings of the 9th Workshop on Computational Linguistics and Clinical Psychology (CLPsych 2024), St. Julians, Malta, 21 March 2024; pp. 232–237. [Google Scholar]
  41. Chen, S.; Wu, M.; Zhu, K.Q.; Lan, K.; Zhang, Z.; Cui, L. LLM-empowered Chatbots for Psychiatrist and Patient Simulation: Application and Evaluation. arXiv 2023, arXiv:2305.13614. [Google Scholar] [CrossRef]
  42. Fu, G.; Zhao, Q.; Li, J.; Luo, D.; Song, C.; Zhai, W.; Liu, S.; Wang, F.; Wang, Y.; Cheng, L.; et al. Enhancing Psychological Counseling with Large Language Model: A Multifaceted Decision-Support System for Non-Professionals. arXiv 2023, arXiv:2308.15192. [Google Scholar] [CrossRef]
  43. Yang, K.; Ji, S.; Zhang, T.; Xie, Q.; Kuang, Z.; Ananiadou, S. Towards Interpretable Mental Health Analysis with ChatGPT. arXiv 2023, arXiv:2304.03347. [Google Scholar] [CrossRef]
  44. Xiao, L.; Huang, X.; Chen, B.; Jing, L. Label-specific document representation for multi-label text classification. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 466–475. [Google Scholar]
  45. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  46. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998. [Google Scholar]
  47. OpenAI. GPT-4 Technical Report. arXiv 2023. [Google Scholar] [CrossRef]
  48. Zeng, A.; Liu, X.; Du, Z.; Wang, Z.; Lai, H.; Ding, M.; Yang, Z.; Xu, Y.; Zheng, W.; Xia, X.; et al. GLM-130b: An open bilingual pre-trained model. arXiv 2022, arXiv:2210.02414. [Google Scholar]
  49. ZHIPU AI DevDay GLM-4. 2024. Available online: https://zhipuai.cn/en/devday (accessed on 23 April 2024).
  50. Cui, Y.; Yang, Z.; Yao, X. Efficient and Effective Text Encoding for Chinese LLaMA and Alpaca. arXiv 2023, arXiv:2304.08177. [Google Scholar]
  51. Llama Chinese Community and AtomEcho. Llama2-Chinese-7b-Chat. Model Continuously Updated with New Parameters, Training Process Available at Llama.family. Deployment, Training, and Fine-Tuning Methods Detailed at GitHub Repository: Llama-Chinese. 2024. Available online: https://huggingface.co/FlagAlpha/Llama2-Chinese-7b-Chat (accessed on 10 April 2024).
  52. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  53. Wang, S.; Sun, Y.; Xiang, Y.; Wu, Z.; Ding, S.; Gong, W.; Feng, S.; Shang, J.; Zhao, Y.; Pang, C.; et al. Ernie 3.0 titan: Exploring larger-scale knowledge enhanced pre-training for language understanding and generation. arXiv 2021, arXiv:2112.12731. [Google Scholar]
  54. Touvron, H.; Martin, L.; Stone, K.; Albert, P.; Almahairi, A.; Babaei, Y.; Bashlykov, N.; Batra, S.; Bhargava, P.; Bhosale, S.; et al. Llama 2: Open foundation and fine-tuned chat models. arXiv 2023, arXiv:2307.09288. [Google Scholar] [CrossRef]
  55. Hu, E.J.; Shen, Y.; Wallis, P.; Allen-Zhu, Z.; Li, Y.; Wang, S.; Wang, L.; Chen, W. LoRA: Low-rank adaptation of large language models. arXiv 2021, arXiv:2106.09685. [Google Scholar]
  56. Lin, Z. How to write effective prompts for large language models. Nat. Hum. Behav. 2024, 8, 611–615. [Google Scholar] [CrossRef]
  57. Ding, N.; Qin, Y.; Yang, G.; Wei, F.; Yang, Z.; Su, Y.; Hu, S.; Chen, Y.; Chan, C.M.; Chen, W.; et al. Parameter-efficient fine-tuning of large-scale pre-trained language models. Nat. Mach. Intell. 2023, 5, 220–235. [Google Scholar] [CrossRef]
  58. OpenAI. GPT-3.5 Turbo Fine-Tuning and API Updates. 2023. Available online: https://openai.com/index/gpt-3-5-turbo-fine-tuning-and-api-updates (accessed on 10 April 2024).
  59. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar] [CrossRef]
  60. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
Figure 1. Workflow of our framework. In the figure, the thick, light-orange arrows represent the main data flow, while the dark green arrows indicate decision steps or inference loops in our experimental design.
Figure 1. Workflow of our framework. In the figure, the thick, light-orange arrows represent the main data flow, while the dark green arrows indicate decision steps or inference loops in our experimental design.
Bioengineering 12 00882 g001
Figure 2. Effect of decoding temperature on model performance for the best LLM settings in each task.
Figure 2. Effect of decoding temperature on model performance for the best LLM settings in each task.
Bioengineering 12 00882 g002
Figure 3. Examples of expert annotated labels and GPT-4 predictions in cognitive distortion classification task.
Figure 3. Examples of expert annotated labels and GPT-4 predictions in cognitive distortion classification task.
Bioengineering 12 00882 g003
Table 1. Data distribution of the proposed datasets: SOS-HL-1K and SocialCD-3K.
Table 1. Data distribution of the proposed datasets: SOS-HL-1K and SocialCD-3K.
DatasetCategories N c
SOS-HL-1KHigh-risk601
Low-risk648
N t r a i n = 999, N t e s t = 250, L = 1, L ¯ = 1, C ¯ = 47.79
SocialCD-3KAll-or-nothing thinking77
Over-generalization141
Mental filter378
Disqualifying the positive27
Mind reading121
The fortune teller error652
Magnification321
Emotional reasoning16
Should statements84
Labeling and mislabeling1961
Blaming oneself188
Blaming others27
N t r a i n = 2725, N t e s t = 682, L = 12, L ¯ = 1.71, C ¯ = 42.56
Note. N c : Number of posts in each category. N t r a i n / N t e s t : Size of training/test set. L: Total number of label types. L ¯ : Average number of labels per post. C ¯ : Average number of words per post.
Table 2. Results for the suicide binary classification task. All these methods were evaluated on the same test set containing 250 posts. The terms “zero-shot” and “few-shot” refer to different prompt construction without training. “Train from scratch” and “fine-tuning” indicate that the corresponding method was trained or fine-tuned using the training set. Reported metrics include micro-averaged, macro-averaged, and high-risk class precision, recall, and F1-score, presented as mean (±standard deviation) over five runs.
Table 2. Results for the suicide binary classification task. All these methods were evaluated on the same test set containing 250 posts. The terms “zero-shot” and “few-shot” refer to different prompt construction without training. “Train from scratch” and “fine-tuning” indicate that the corresponding method was trained or fine-tuned using the training set. Reported metrics include micro-averaged, macro-averaged, and high-risk class precision, recall, and F1-score, presented as mean (±standard deviation) over five runs.
Model CategoryModelTypeSub-TypeTrain DataMicro-AverageMacro-AverageHigh-Risk
F1-Score (%) Precision (%) Recall (%) F1-Score (%) Precision (%) Recall (%) F1-Score (%)
Supervised learningLSANTrain from scratch-99983.20 (±1.15)82.75 (±1.05)83.75 (±0.90)82.81 (±0.98)74.59 (±1.21)87.50 (±0.95)80.53 (±0.87)
BERTFine-tuning-99987.20 (±0.61)87.41 (±0.65)85.54 (±0.78)86.23 (±0.68)88.42 (±0.75)77.78 (±1.02)82.76 (±0.56)
LLMChatGLM2-6BZero-shotbasic060.80 (±4.91)63.41 (±5.32)60.61 (±4.93)58.26 (±5.99)69.07 (±7.02)37.10 (±8.55)48.07 (±8.99)
role058.56 (±4.20)60.94 (±5.81)58.38 (±4.18)56.19 (±3.97)65.77 (±9.00)35.81 (±4.25)46.15 (±4.48)
scene060.56 (±3.22)61.51 (±3.18)60.44 (±3.24)59.49 (±3.60)64.52 (±3.57)45.16 (±6.57)53.01 (±5.33)
hybrid061.20 (±6.74)62.39 (±7.34)61.08 (±6.75)60.04 (±7.12)65.68 (±9.61)46.13 (±10.90)53.74 (±9.47)
Few-shotbackground + scene057.44 (±6.28)57.98 (±6.48)57.36 (±6.34)56.47 (±6.59)58.56 (±6.23)47.26 (±16.11)51.45 (±10.71)
background + hybrid062.80 (±16.62)64.29 (±18.11)62.86 (±16.65)62.23 (±16.60)60.37 (±13.17)70.64 (±23.99)64.41 (±17.14)
ref 12 + scene2465.68 (±6.41)65.88 (±6.33)65.63 (±6.43)65.48 (±6.54)67.19 (±6.20)59.52 (±9.43)63.04 (±7.88)
ref 12 + hybrid2461.04 (±8.50)61.75 (±9.08)60.95 (±8.50)60.38 (±8.58)64.29 (±11.40)49.20 (±9.18)55.56 (±9.56)
background + ref 12 + scene2450.00 (±9.17)50.09 (±9.80)50.05 (±9.16)49.53 (±9.26)49.74 (±7.41)56.61 (±12.12)52.70 (±8.70)
background + ref 12 + hybrid2460.64 (±9.29)61.57 (±9.99)60.74 (±9.26)59.51 (±10.15)58.91 (±7.92)73.23 (±11.86)64.78 (±7.59)
ref 30 + scene6055.28 (±8.32)55.83 (±11.10)55.05 (±8.36)50.78 (±10.80)57.71 (±16.53)26.77 (±14.14)36.14 (±16.74)
ref 30 + hybrid6052.16 (±5.99)51.27 (±8.31)51.94 (±6.04)47.57 (±8.85)50.60 (±12.52)24.84 (±13.85)32.60 (±15.48)
background + ref 30 + scene6060.64 (±12.21)61.33 (±12.79)60.58 (±12.20)60.23 (±12.07)62.97 (±14.85)52.90 (±13.93)57.02 (±12.67)
background + ref 30 + hybrid6058.56 (±5.52)59.11 (±5.37)58.47 (±5.59)57.41 (±6.39)60.14 (±4.83)47.10 (±16.21)51.88 (±11.30)
GLM-130BZero-shotbasic058.32 (±2.37)69.53 (±3.65)58.62 (±2.35)51.69 (±3.88)54.58 (±1.49)95.81 (±2.57)69.52 (±1.36)
role059.68 (±2.79)69.34 (±3.52)59.96 (±2.77)54.16 (±4.13)55.51 (±1.77)94.84 (±2.18)70.02 (±1.66)
scene058.96 (±1.43)68.04 (±4.36)59.24 (±1.44)53.53 (±1.07)55.05 (±0.75)93.87 (±3.20)69.40 (±1.46)
hybrid062.80 (±0.49)74.79 (±2.02)63.07 (±0.49)57.97 (±0.85)57.37 (±0.35)97.42 (±1.75)72.20 (±0.47)
Few-shotbackground + role060.80 (±1.85)66.90 (±3.41)61.03 (±1.86)57.25 (±1.72)56.55 (±1.10)90.32 (±3.28)69.55 (±1.77)
background + hybrid061.52 (±2.29)69.00 (±3.86)61.77 (±2.29)57.62 (±2.73)56.91 (±1.47)92.42 (±3.45)70.43 (±1.82)
ref 12 + role2455.36 (±1.28)58.17 (±2.47)55.58 (±1.30)51.78 (±0.88)53.18 (±0.74)83.23 (±3.29)64.89 (±1.53)
ref 12 + hybrid2458.80 (±0.57)63.91 (±1.71)59.03 (±0.58)55.02 (±0.73)55.30 (±0.32)88.39 (±3.05)68.02 (±0.97)
background + ref 12 + role2461.60 (±1.10)64.44 (±1.05)61.77 (±1.09)59.80 (±1.32)57.84 (±0.87)83.38 (±0.44)68.30 (±0.67)
background + ref 12 + hybrid2466.24 (±1.08)71.07 (±0.71)66.38 (±1.06)64.29 (±1.50)60.88 (±1.07)90.00 (±1.47)72.61 (±0.48)
GPT-3.5Zero-shotbasic053.76 (±4.28)57.82 (±7.62)54.03 (±4.28)47.76 (±4.81)52.00 (±2.25)88.23 (±1.98)65.42 (±1.81)
role056.32 (±0.95)67.70 (±2.39)56.64 (±0.95)48.44 (±1.59)53.31 (±2.05)96.13 (±1.55)68.59 (±1.67)
scene054.08 (±2.57)58.84 (±4.69)54.36 (±2.58)47.89 (±2.93)52.16 (±2.11)89.03 (±2.01)65.76 (±1.94)
hybrid054.80 (±2.12)61.06 (±3.34)55.10 (±2.10)47.69 (±3.36)52.55 (±1.98)92.26 (±1.82)66.95 (±1.78)
Few-shotbackground + role058.24 (±1.69)63.49 (±2.95)58.48 (±1.69)54.17 (±2.04)54.90 (±1.07)88.55 (±3.84)67.76 (±1.63)
background + hybrid058.80 (±1.67)64.06 (±2.06)59.04 (±1.67)54.85 (±2.11)55.27 (±1.14)89.03 (±1.22)68.19 (±1.09)
ref 12 + role2459.76 (±1.37)62.98 (±1.69)59.95 (±1.37)57.50 (±1.52)56.34 (±0.96)83.39 (±1.51)67.22 (±1.08)
ref 12 + hybrid2460.96 (±1.37)64.29 (±1.69)61.15 (±1.37)58.77 (±1.52)57.19 (±0.96)84.68 (±1.51)68.27 (±1.08)
background + ref 12 + role2463.20 (±1.57)65.53 (±2.62)63.35 (±1.59)61.98 (±1.31)59.37 (±0.83)81.61 (±4.50)68.71 (±2.04)
background + ref 12 + hybrid2462.08 (±2.25)64.86 (±3.31)62.25 (±2.26)60.46 (±2.14)58.26 (±1.44)82.90 (±4.13)68.41 (±2.26)
Fine-tuningrole99983.12 (±0.55)82.25 (±0.49)81.44 (±0.72)81.80 (±0.58)84.76 (±0.51)71.77 (±0.98)77.73 (±0.42)
scene99983.36 (±0.51)82.81 (±0.56)82.02 (±0.69)82.40 (±0.54)84.11 (±0.63)72.58 (±0.85)77.92 (±0.49)
hybrid99983.60 (±0.48)83.83 (±0.55)82.05 (±0.61)82.70 (±0.45)84.26 (±0.45)73.39 (±0.81)78.45 (±0.38)
GPT-4Zero-shotbasic062.64 (±2.05)72.38 (±3.40)62.90 (±2.05)58.33 (±2.57)57.43 (±1.31)95.48 (±2.10)71.72 (±1.42)
role062.64 (±2.74)74.26 (±3.90)62.91 (±2.73)57.77 (±3.50)57.29 (±1.73)97.26 (±2.03)72.10 (±1.82)
scene064.88 (±2.11)76.03 (±2.59)65.14 (±2.09)60.86 (±2.68)58.81 (±1.46)97.58 (±1.51)73.39 (±1.37)
hybrid062.96 (±2.60)74.67 (±4.57)63.23 (±2.60)58.22 (±3.02)57.47 (±1.57)97.42 (±2.38)72.30 (±1.89)
Few-shot ref 12 + scene2466.88 (±1.21)74.28 (±1.59)67.10 (±1.21)64.32 (±1.52)60.70 (±0.93)94.35 (±1.51)73.87 (±0.83)
ref 12 + hybrid2464.00 (±3.36)66.96 (±4.09)64.16 (±3.35)62.52 (±3.60)59.77 (±2.60)84.19 (±4.51)69.87 (±2.75)
background + ref 12 + scene2469.28 (±5.60)70.58 (±5.11)69.38 (±5.56)68.74 (±6.05)65.44 (±5.37)81.77 (±2.77)72.63 (±4.07)
background + ref 12 + hybrid2468.96 (±1.99)69.85 (±1.78)69.04 (±1.97)68.65 (±2.14)65.65 (±2.48)78.87 (±2.70)71.60 (±1.36)
ref 30 + scene6067.04 (±1.56)73.13 (±1.97)67.24 (±1.55)64.87 (±1.98)61.11 (±1.28)92.42 (±2.32)73.56 (±1.09)
background + ref 30 + hybrid6073.76 (±5.63)74.59 (±5.41)73.83 (±5.61)73.54 (±5.77)70.16 (±5.76)82.58 (±4.21)75.81 (±4.74)
Table 3. Results for the cognitive distortion multi-label classification task. All these methods were evaluated on the same test set containing 682 posts. The micro and macro metrics are presented as mean (±standard deviation) over five runs.
Table 3. Results for the cognitive distortion multi-label classification task. All these methods were evaluated on the same test set containing 682 posts. The micro and macro metrics are presented as mean (±standard deviation) over five runs.
Model CategoryModelTypeSub-TypeTrain DataMicro-AverageMacro-Average
Precision (%) Recall (%) F1-Score (%) Precision (%) Recall (%) F1-Score (%)
Supervised learningLSANTrain from scratch-272575.53 (±0.85)71.48 (±0.91)73.45 (±0.88)70.10 (±1.15)67.00 (±1.21)67.28 (±1.18)
BERTFine-tuning-272585.43 (±0.68)68.62 (±0.85)76.10 (±0.45)75.00 (±0.98)68.30 (±1.15)70.05 (±0.85)
LLMGLM-4Zero-shotbasic017.39 (±2.41)46.95 (±2.95)25.38 (±2.68)16.52 (±3.11)45.15 (±3.65)21.65 (±3.28)
role020.56 (±2.25)46.95 (±2.81)28.59 (±2.51)19.88 (±2.95)45.82 (±3.51)22.11 (±3.21)
scene017.19 (±2.55)45.33 (±3.08)24.93 (±2.85)16.81 (±3.25)44.92 (±3.78)21.05 (±3.55)
hybrid019.54 (±2.31)47.32 (±2.88)27.66 (±2.59)18.95 (±3.01)43.93 (±3.58)21.67 (±3.29)
Few-shotbackground + hybrid030.63 (±2.01)48.82 (±2.15)37.64 (±2.08)25.50 (±2.61)41.21 (±2.75)25.30 (±2.68)
background + role028.76 (±2.11)47.20 (±2.28)35.74 (±2.18)27.01 (±2.71)43.43 (±2.88)25.65 (±2.78)
ref 2 + hybrid2432.25 (±1.85)43.21 (±2.05)36.94 (±1.95)23.96 (±2.45)37.05 (±2.65)23.05 (±2.55)
ref 2 + role2434.82 (±1.72)44.58 (±1.91)39.10 (±1.81)23.64 (±2.32)30.17 (±2.51)22.40 (±2.41)
background + ref 2 + hybrid2441.74 (±1.55)47.82 (±1.48)44.57 (±1.51)27.04 (±2.05)40.96 (±1.98)25.59 (±2.01)
background + ref 2 + role2441.91 (±1.51)45.83 (±1.62)43.78 (±1.58)28.75 (±2.01)34.27 (±2.22)27.03 (±2.08)
ref 5 + hybrid6026.32 (±2.58)34.74 (±2.71)29.95 (±2.65)23.11 (±3.28)29.98 (±3.41)18.80 (±3.35)
ref 5 + role6029.67 (±2.45)36.61 (±2.65)32.78 (±2.52)23.72 (±3.15)32.71 (±3.35)20.74 (±3.22)
GPT-3.5Zero-shotbasic010.63 (±2.65)13.95 (±3.11)12.06 (±2.88)9.85 (±3.35)12.16 (±3.81)8.76 (±3.58)
role012.17 (±2.81)10.21 (±2.98)11.10 (±2.91)11.05 (±3.51)9.86 (±3.68)7.21 (±3.61)
scene010.33 (±2.77)11.83 (±3.25)11.03 (±3.01)9.79 (±3.47)10.02 (±3.95)7.55 (±3.71)
hybrid010.59 (±2.71)11.83 (±3.11)11.18 (±2.85)9.96 (±3.41)10.62 (±3.81)6.48 (±3.55)
Few-shotbackground + hybrid020.16 (±2.51)12.20 (±2.75)15.21 (±2.62)18.14 (±3.21)11.96 (±3.45)8.36 (±3.32)
background + basic025.84 (±2.35)11.46 (±2.61)15.88 (±2.48)23.31 (±3.05)10.48 (±3.31)9.60 (±3.18)
ref 2 + hybrid2416.76 (±2.68)19.68 (±2.99)18.10 (±2.81)14.37 (±3.38)18.01 (±3.69)10.14 (±3.51)
ref 2 + basic2417.61 (±2.61)14.69 (±2.88)16.02 (±2.75)15.14 (±3.31)8.87 (±3.58)6.09 (±3.45)
Fine-tuningscene272572.03 (±0.78)71.86 (±0.85)71.95 (±0.65)63.81 (±0.95)68.55 (±1.05)64.98 (±0.85)
role272569.88 (±0.95)70.49 (±1.01)70.18 (±0.98)61.05 (±1.15)66.92 (±1.21)62.73 (±1.18)
hybrid272571.94 (±0.75)72.48 (±0.81)72.21 (±0.68)64.22 (±0.95)69.01 (±1.01)65.53 (±0.88)
Chinese-Alpaca-2-7BFine-tuningscene272572.61 (±0.61)71.07 (±0.88)71.83 (±0.55)63.10 (±0.81)68.03 (±1.08)64.51 (±0.75)
role272573.60 (±0.45)72.32 (±0.71)72.96 (±0.38)65.48 (±0.65)70.02 (±0.91)66.89 (±0.58)
hybrid272572.02 (±0.58)69.95 (±0.95)70.97 (±0.49)62.83 (±0.78)67.59 (±1.15)63.81 (±0.69)
Chinese-LLaMA-2-7BFine-tuningscene272573.56 (±0.51)71.45 (±0.81)72.49 (±0.43)64.91 (±0.71)69.46 (±1.01)66.04 (±0.63)
role272572.82 (±0.65)68.83 (±1.11)70.77 (±0.58)61.33 (±0.85)68.50 (±1.31)63.75 (±0.78)
hybrid272573.46 (±0.48)70.07 (±0.99)71.73 (±0.51)63.08 (±0.68)69.72 (±1.19)65.29 (±0.71)
LLaMA2-Chinese-
7b-Chat
Fine-tuningscene272566.22 (±0.88)61.85 (±1.21)63.96 (±1.05)56.19 (±1.08)59.84 (±1.41)57.23 (±1.25)
role272568.59 (±0.75)65.09 (±1.01)66.80 (±0.81)58.75 (±0.95)64.01 (±1.21)60.18 (±1.01)
hybrid272569.44 (±0.69)64.59 (±1.05)66.93 (±0.79)59.52 (±0.89)63.88 (±1.25)60.95 (±0.99)
GPT-4Zero-shotbasic017.13 (±2.02)59.65 (±2.25)26.61 (±2.11)16.43 (±2.52)58.79 (±2.75)22.95 (±2.61)
role017.30 (±1.98)57.53 (±2.15)26.61 (±2.05)16.62 (±2.48)56.51 (±2.65)22.45 (±2.55)
scene018.29 (±2.18)42.71 (±2.58)25.62 (±2.35)17.25 (±2.68)41.69 (±3.08)21.73 (±2.85)
hybrid017.41 (±2.01)56.41 (±2.21)26.61 (±2.09)16.93 (±2.51)55.72 (±2.71)20.74 (±2.59)
Few-shotbackground + hybrid030.68 (±1.15)53.67 (±1.25)39.04 (±1.18)28.97 (±1.55)50.16 (±1.65)27.35 (±1.58)
background + basic031.38 (±1.11)56.66 (±1.19)40.39 (±1.13)28.10 (±1.51)55.14 (±1.59)28.58 (±1.53)
ref 2 + hybrid2434.62 (±1.05)40.35 (±1.25)37.26 (±1.15)23.37 (±1.45)37.45 (±1.65)22.78 (±1.55)
ref 2 + basic2437.56 (±0.98)40.97 (±1.21)39.19 (±1.09)25.59 (±1.38)43.23 (±1.61)27.14 (±1.49)
background + ref 2 + hybrid2426.25 (±1.55)43.84 (±1.81)32.84 (±1.68)20.60 (±2.05)42.87 (±2.31)22.04 (±2.18)
background + ref 2 + basic2437.56 (±1.01)40.97 (±1.25)39.19 (±1.11)27.93 (±1.41)36.12 (±1.65)23.65 (±1.51)
ref 5 + hybrid6026.37 (±1.48)47.82 (±1.65)34.00 (±1.55)20.86 (±1.98)48.03 (±2.15)23.70 (±2.05)
ref 5 + basic6029.77 (±1.35)53.80 (±1.51)38.33 (±1.41)22.42 (±1.85)53.05 (±2.01)26.04 (±1.91)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qi, H.; Fu, G.; Li, J.; Song, C.; Zhai, W.; Luo, D.; Liu, S.; Yu, Y.; Yang, B.; Zhao, Q. Supervised Learning and Large Language Model Benchmarks on Mental Health Datasets: Cognitive Distortions and Suicidal Risks in Chinese Social Media. Bioengineering 2025, 12, 882. https://doi.org/10.3390/bioengineering12080882

AMA Style

Qi H, Fu G, Li J, Song C, Zhai W, Luo D, Liu S, Yu Y, Yang B, Zhao Q. Supervised Learning and Large Language Model Benchmarks on Mental Health Datasets: Cognitive Distortions and Suicidal Risks in Chinese Social Media. Bioengineering. 2025; 12(8):882. https://doi.org/10.3390/bioengineering12080882

Chicago/Turabian Style

Qi, Hongzhi, Guanghui Fu, Jianqiang Li, Changwei Song, Wei Zhai, Dan Luo, Shuo Liu, Yijing Yu, Bingxiang Yang, and Qing Zhao. 2025. "Supervised Learning and Large Language Model Benchmarks on Mental Health Datasets: Cognitive Distortions and Suicidal Risks in Chinese Social Media" Bioengineering 12, no. 8: 882. https://doi.org/10.3390/bioengineering12080882

APA Style

Qi, H., Fu, G., Li, J., Song, C., Zhai, W., Luo, D., Liu, S., Yu, Y., Yang, B., & Zhao, Q. (2025). Supervised Learning and Large Language Model Benchmarks on Mental Health Datasets: Cognitive Distortions and Suicidal Risks in Chinese Social Media. Bioengineering, 12(8), 882. https://doi.org/10.3390/bioengineering12080882

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop