Next Article in Journal
Correction: Visani et al. miR-196B-5P and miR-200B-3P Are Differentially Expressed in Medulloblastomas of Adults and Children. Diagnostics 2020, 10, 265
Next Article in Special Issue
AI-Enabled Algorithm for Automatic Classification of Sleep Disorders Based on Single-Lead Electrocardiogram
Previous Article in Journal
Intracoronary Thrombogenicity in Patients with Vasospastic Angina: An Observation Using Coronary Angioscopy
Previous Article in Special Issue
Predicting Depression in Community Dwellers Using a Machine Learning Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Machine Learning-Based Definition of Symptom Clusters and Selection of Antidepressants for Depressive Syndrome

1
Department of Psychiatry, Hanyang University Guri Hospital, Guri 11923, Korea
2
Graduate School of Medical Science and Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon 34141, Korea
3
Department of Psychiatry, Hanyang University College of Medicine, Seoul 04763, Korea
*
Author to whom correspondence should be addressed.
Diagnostics 2021, 11(9), 1631; https://doi.org/10.3390/diagnostics11091631
Submission received: 30 June 2021 / Revised: 3 September 2021 / Accepted: 3 September 2021 / Published: 7 September 2021
(This article belongs to the Special Issue Machine Learning and Big Data in Psychiatric and Sleep Disorders)

Abstract

:
The current polythetic and operational criteria for major depression inevitably contribute to the heterogeneity of depressive syndromes. The heterogeneity of depressive syndrome has been criticized using the concept of language game in Wittgensteinian philosophy. Moreover, “a symptom- or endophenotype-based approach, rather than a diagnosis-based approach, has been proposed” as the “next-generation treatment for mental disorders” by Thomas Insel. Understanding the heterogeneity renders promise for personalized medicine to treat cases of depressive syndrome, in terms of both defining symptom clusters and selecting antidepressants. Machine learning algorithms have emerged as a tool for personalized medicine by handling clinical big data that can be used as predictors for subtype classification and treatment outcome prediction. The large clinical cohort data from the Sequenced Treatment Alternatives to Relieve Depression (STAR*D), Combining Medications to Enhance Depression Outcome (CO-MED), and the German Research Network on Depression (GRND) have recently began to be acknowledged as useful sources for machine learning-based depression research with regard to cost effectiveness and generalizability. In addition, noninvasive biological tools such as functional and resting state magnetic resonance imaging techniques are widely combined with machine learning methods to detect intrinsic endophenotypes of depression. This review highlights recent studies that have used clinical cohort or brain imaging data and have addressed machine learning-based approaches to defining symptom clusters and selecting antidepressants. Potentially applicable suggestions to realize machine learning-based personalized medicine for depressive syndrome are also provided herein.

1. Introduction

Depression is one of the most burdensome disorders worldwide, with a lifetime prevalence of approximately 20% of the global population [1]. Depression remission after the first antidepressant trial is only 30% [2,3]. This low remission rate is partly because diagnosing depression does not guarantee heterogeneous symptom subtypes [4]. Inevitably, the concept that depression is characterized by symptomatic heterogeneity, such as atypical [5], melancholic [6], and anxious [7] subtypes, has gained considerable attention. In addition, it has been reported that the heterogeneity of depressive syndrome can theoretically result from the polythetic and operational criteria of major depression [8,9,10,11,12]. According to the Diagnostic and Statistical Manual of Mental Disorders, fifth edition (DSM-5) [13], a confirmed diagnosis of major depressive disorder requires both the presence of five or more symptoms among, nine symptoms, including depressed mood, diminished interest or pleasure, weight loss or gain, insomnia or hypersomnia, psychomotor retardation or agitation, fatigue or loss of energy, feelings of worthlessness or excessive guilt, diminished thinking ability or indecisiveness, recurrent thoughts of death or recurrent suicidal ideation, and the presence of either depressed mood or diminished interest or pleasure. Herein, the subset of k draws from n distinguishable objects without replacement and without regard to order that (nCk) can calculate from the theoretical number of different combinations meeting the polythetic and operational criteria of major depressive disorder in DSM-5. Thus, 227 different diagnostic symptom combinations were calculated that can fulfill the DSM-5 diagnostic criteria for major depressive disorder [14,15,16,17]. In terms of psychiatric taxonomy, the heterogeneity of depressive syndrome has been criticized by the concept of a language game in Wittgensteinian philosophy [18]. Wittgenstein suggested the analogy as follows [19]:
Consider for example the proceedings that we call games. I mean board-games, card-games, ball-games, Olympic games, and so on. What is common to them all?—Don’t say: “There must be something common, or they would not be called games”—but look and see whether there is anything common to all.—For if you look at them you will not see something that is common to all, but similarities, relationships, and a whole series of them at that. To repeat: don’t think, but look!—the concept game is a concept with blurred edges.—“But is a blurred concept a concept at all?”—Is an indistinct photograph a picture of a person at all? Is it even always an advantage to replace an indistinct picture by a sharp one? Isn’t the indistinct one often exactly what we need? (Wittgenstein, 2001).
It is also proposed that cases of depressive syndrome are conceptually related by the “family resemblance” rather than the “essence.” Thus, it is concluded that the heterogeneity of depressive syndrome is consistent with Wittgensteinian’s analogy [18]. Thus, the nomenclature of depressive syndrome can be consistent not with the categorical approach, but the dimensional approach, in the context of the heterogeneity of major depressive disorder [14]. Furthermore, based on the theoretical construct change from chemical imbalance to dysfunctional circuitry, the symptom-based approach, but not the diagnosis-based approach, has been emphasized by Thomas Insel in his work on the next generation of treatments for mental disorders [20]. Along with the heterogeneity concept, the therapeutic approach also shifts toward selecting antidepressants according to specific symptom clusters [21]. Each cluster of depression symptoms may be thought to react to specific antidepressants, thus potentially improving the current low remission rates. The theorem supporting depression heterogeneity has not generated notable clinical utility in that theory-driven classification of symptom clusters and subsequent antidepressant selection have only produced low accuracies in treatment outcome predictions [22]. However, the clinical utility of the depression heterogeneity concept in diagnostics and therapeutics is increasingly acknowledged with the use of data-driven machine learning approaches.
Machine learning approaches can be more beneficial in the study of depression compared with traditional methods. Factor analysis, for instance, may generate complicated combinations of heterogeneous symptoms within specific dimensions [23]. These analytic approaches also can be vulnerable to experimenter bias in that a researcher has to choose the number of components or clusters in data, as such in k means clustering method [24]. Hierarchical clustering, a type of machine learning method, is an easy-to-implement, deterministic approach in which each of the symptoms is assigned to a single cluster even without predetermining the desired number of clusters.
Clinical expertise in diagnostics and therapeutics currently progresses in accordance with the advent of machine learning algorithms for handling clinical big data [25]. The machine learning approach allows clinicians to consider that clinical data can be a useful source of predictors for the classification of depressive symptom clusters and selection of antidepressants. Thus, data-driven approaches, rather than clinician-based, diagnostics, and therapeutic approaches, are now incorporated into depression research to disentangle the questions as to what subtypes of depression exist, and how these subtypes react to different antidepressants. However, from a cautious perspective, clinical expertise can still be a keystone in machine learning-based diagnostics and therapeutics in depression, as clinical knowledge and experience are inevitable not only for choosing clinical data input to algorithms but also for interpreting the algorithm prediction outcomes. The next generation of machine learning-based depression practice might be dependent on the degree to which clinical expertise is fully incorporated into unmanned learning approaches to future precision medicine [26]. Thus, clinicians are required to not only be updated of the latest findings from the machine learning-based depression research but also yield clinical expertise-based perspectives for the learning algorithms in terms of better translation of the machine learning findings to clinical applications.
This review is not intended to be systematic or comprehensive for all relevant studies, but rather narrative for emphasizing predominant papers which seem of practical interest for readership. In particular, this review focuses on the studies in which machine learning-based prediction models for treatment outcome in depression were built in large clinical cohorts with several hundreds of subjects, including the Sequenced Treatment Alternatives to Relieve Depression (STAR*D) [27], Combining Medications to Enhance Depression Outcome (CO-MED) [3], and German Research Network on Depression (GRND) [28]. These cohorts encompass accessible clinical data that offer benefits such as cost-effectiveness, compared to the genetics and imaging data, and interchangeable usability of variable characteristics between cohorts for the algorithm outcome validations, in part guaranteeing generalizability. This review also provides suggestions for advancing machine learning-based depression research for better symptom classification and antidepressant selection.

2. Data-Driven Classification of Symptom Clusters in Depression

Depression is viewed as a heterogeneous mental construct [8]. Machine learning-based analysis underlines the benefits of addressing symptom heterogeneity by subject stratification and the application of data-driven therapeutic responses instead of summed scores of clinical scales. Although previous studies have suggested a conventional clinician experience-based (theory-driven) classification of depression subtypes, the efforts resulted in poor predictive values for therapeutic outcomes for different antidepressants. For instance, Uher et al. [21] suggested three types of depression, including melancholic, atypical, and anxious features, for which therapeutic outcomes were evaluated with escitalopram and nortriptyline. However, this clinician-driven approach to grouping depression subtypes resulted in low to modest accuracy in the prediction of therapeutic outcomes. Likewise, a similar study using clinician-driven classification of depression subtypes also showed only low prediction accuracies, thus limiting clinical utility [22]. After these trials, the science of big data was incorporated into depression research, supporting the advantages of data-driven phenotypes. Indeed, a recent review suggested evidence for the considerable prognostic value of data-driven depression subtype classification [29]. These research trends are rooted in discovering clinical signatures of predictions for response of specific symptoms to specific antidepressants, and thus require established clinical data with a large number of depressive subjects. Indeed, with the increased accessibility of large databases, multivariate models exploiting clinical data have been introduced to psychiatric research in recent years [25]. The cohort data consisted of sociodemographic (sex and age), diagnostic (scale scores), and therapeutic variables (antidepressant classes), which can be assumed to be useful candidate predictors for the classification of depression subtypes. In particular, the STAR*D database in the US and the GRND and Group for Studies of Resistant Depression (GSRD) databases in Europe have facilitated progress in predicting therapeutic outcomes at each patient level [30,31,32]. Even though there are approaches of integrating more comprehensive data, including pharmacogenomics [33], in elucidating the machine learning-based study of depression, this may be beyond the scope of the review. Herein, we highlight some findings from representative studies that used clinical cohort data and machine learning-based approaches to classify depression subtypes.

3. Machine Learning-Based Symptom Clustering of Depression

Machine learning algorithms based on clinical data suggest clinical patterns for three to four subtypes of depression according to different studies. For instance, Kautzky et al. [34] examined 1079 patients with acute depression from a longitudinal multicenter study, conducted by the GRND. The researchers attempted to demonstrate the interplay between clinical and sociodemographic variables and their predictive influence on therapeutic outcomes. Hierarchical symptom clustering resulted in three subtypes of depression: emotional, anxious, and sleep- and appetite-related symptoms. Similarly, Chekroud et al. [35] used patient-reported data from the STAR*D trial (n = 4039) and applied hierarchical clustering to the Hamilton Depression Rating Scale (HAM-D) [36] and Quick Inventory of Depression Symptomatology Self-Report (QIDS-SR) items [37], thereby suggesting three clusters including “core emotional,” “atypical,” and “sleep” subtypes. Interestingly, the “emotional type” resembles the traditional melancholic subtype of depression, supporting the converging of data-driven classification of subtypes with theory-driven symptom grouping [38]. Furthermore, the symptom clustering findings from these studies are consistent with symptom-specific therapeutic responses. Whereas response rates for the clusters, including the “emotional” and “vegetative function” types, were comparable to the total HAM-D response rate, the rate for the anxiety cluster was significantly lower at 48.5% [34], indicating poor therapeutic response for anxious depression [39]. The poor response might be associated with the treatment side effects that manifest within the anxious cluster and can negatively impact any therapeutic effects on other symptoms. This mechanism was is in line with recent analyses of SSRI response [40]. Taken together, the recent clinical cohort studies resulted in similar symptom clusters that were driven by machine learning algorithms and agreed with clinical experience. Machine learning-based symptom clustering is presented in Table 1.
Symptom classification using machine learning algorithms should be conducted from the perspective of clinical experience and conventional theories [38]. Although symptom clustering is performed using learning algorithms without researcher intervention, the ratification of the resulting items is entirely based on researcher experience and knowledge. For instance, Chekroud et al. and Kautzky et al. similarly applied hierarchical clustering methods and named several resulting items “emotional clusters.” Both “core emotional clusters” include symptoms related to mood, energy, concentration, interest, and self-worth [41,42,43]. However, the emotional cluster by Kautzky et al. also included suicidality, whereas that by Chekroud et al. did not. In clinical practice, suicidality can occasionally be interpreted as an atypical feature of depression, rather than as a core emotional component, and may differ between atypical and other types of depression, according to previous analyses [44]. Other studies, however, have suggested that suicidality risk can be enhanced by the presence of anhedonia, which is regarded as a core emotional component [45,46]. The characteristics of suicidality congruent with emotional components are also supported by factorial analysis using HAM-D [47]. As the conventional concept of atypical depression involves hyperphagia and hypersomnia, further ratification of atypical clusters is required to adopt clinical scales such as QIDS-SR, rather than HAM-D, which does not assess the excess of vegetative symptoms. Meanwhile, contrary to the analysis results of Chekroud et al., anxiety symptoms were not included in the core emotional cluster but rather suggested to produce a separate cluster, along with psychomotor agitation, loss of sleep, and appetite. Independent appetite loss was also suggested in a previous factorial analysis [47]. The anxious cluster may be corroborated by studies that used GRND and GSRD subjects that recently established an anxious subtype of depression, in which anxiety symptoms were associated with somatic symptoms and generally poor therapeutic outcomes [39]. Taken together, data-driven symptom clustering can be further interpreted based on clinical experiences and ratified using the clinical consensus.

4. Exploring Prediction Model for Treatment Outcomes

In pharmacotherapy, personalized medicine holds promise in the treatment of depression, a heterogeneous disorder [48] for which no single antidepressant is identically effective, and for which numerous patients are given several treatments before a correct regimen is identified. At a population level, large-scale clinical trials, such as STAR*D and CO-MED, have shown that approximately 30% of patients achieve symptomatic remission for a given treatment [2,3]. However, personalized medicine shifts emphasis away from remission rates and treatment efficacy at the population level and attempts to identify the specific drug that is the best candidate for 30% of patients. For example, although general remission rates for all drugs were similar (48–52%), the analysis by Chekroud et al., which aimed to prospectively predict therapeutic outcomes, was more diverse (51–65%) [35]. Evidently, the development of generally effective antidepressants is crucial for public health. Until then, the application of pioneering statistical methods to choose the best drug candidate for each patient may offer an interim solution [49,50,51]. These findings are the first step in the era of personalized medicine for psychiatry. However, the performances are apparently inadequate in comparison with other areas of medicine.
Some representative studies have investigated methods to build prediction models for estimating antidepressant responses. Kautzky et al. [52] demonstrated that a random forest model for therapeutic response accurately identified 25% of melancholia responders by using three SNPs and clinical variables. Patel et al. showed that an alternating decision tree model estimated therapeutic response with an accuracy of 89% using age, structural imaging, and mini-mental status examination scores [53]. Chekroud et al. [35] demonstrated that a machine learning model forecasted remission with 59% accuracy using 25 clinical variables. Iniesta et al. [54] also showed that regularized regression models using demographic variables predict therapeutic response with clinically meaningful accuracy. Maciukiewicz et al. [55] recently reported that a support vector machine model predicted therapeutic response with 52% accuracy using SNPs. More precise prediction models, with accuracies of approximately 70%, have also been suggested for antidepressant therapeutic outcomes from large-scale clinical databases, including the GSRD and GRND in Europe, and the STAR*D cohorts in the US [30,31,32,56,57]. Table 2 describes some representative studies that have addressed the prediction of treatment response in depression.
The success of these models depends on their ability to be generalized. In general, there are two important precautions. First, all variable selection and model building should occur in the direction of ensuring validity. A vital challenge for prediction is determining the variables to be used. For example, Kautzky et al. [52] used a random forest procedure to identify 88 predictors that distinguish patients with unfavorable therapeutic outcomes from those with favorable outcomes, considering the number of interaction effects between predictive variables. Chekroud et al. adopted a penalized logistic regression with advantages: coefficients of correlated predictors are minimized toward each other, and uninformative features are uninvolved from the model [58,59]. They used the method to identify 25 predictive variables. Second, the way the model is trained on a large antidepressant cohort would be performed in other clinical trial cohorts with other treatment protocols, with different recruitment criteria and distributions of symptoms. An external validation analysis, for instance, showed that a citalopram model, trained in the STAR*D cohort, accurately predicted outcomes for the escitalopram treatment group of CO-MED. The model also showed significant prediction accuracy in the escitalopram-bupropion group, but not in the combination group of venlafaxine-mirtazapine [35]. This result shows that the model may generalize to an independent cohort sample and represent some caveats of therapeutic specificity. The finding that the model poorly predicts response to the venlafaxine-mirtazapine group indicates that the model does not predict a broad therapeutic response, nor does it predict equivalently for all treatment subgroups of CO-MED. The use of wholly independent validation cohorts also showed that although some predictors could still yield comparable model performance in the STAR*D escitalopram cohort, the model does not generalize to the escitalopram group of an independent clinical cohort, emphasizing the importance of external validation.
Further, statistical (and biomarker) approaches showed a performance above chance to be clinically usable. In STAR*D analyses, an accuracy of 53.13% would have surpassed the chance accuracy of 51.3% regarding the sample size, by the conventional statistics of p value < 0.05. The model by Chekroud et al. achieved an accuracy of 64.6%, outperforming this benchmark significantly. However, clinical experience-based predictions of who would respond to which therapeutics are generally poor [60]. Similarly, in a pilot sample of psychiatrists, the average accuracy of the 23 clinicians in predicting therapeutic outcomes for 26 STAR*D patients was 49.4%, given that the chance prediction was 53.9% [57]. These findings suggest that machine learning approaches can be a useful tool for predicting therapeutic outcomes with clinically meaningful accuracy.

5. Antidepressant Selection Specific to Symptom Clusters

Predicting who will respond to which treatment is a major challenge in personalized medicine. Machine learning algorithms are available tools for predicting differential responses to various antidepressants. The current research attempts to measure the extent of the differential predictive power of the algorithm approach for distinct medications. These efforts reflect the distinctive neural mechanisms probed in large-scale clinical cohort studies and heterogeneity in the underlying neurobiology of depression among the patients who entered the studies. Here, we address a recent representative study that explored antidepressant treatment outcomes according to each symptom cluster.
Chekroud et al. [35] used a citalopram-treated STAR*D cohort (n = 1962) with the aim of identifying differential response trajectories from three symptom clusters based on QIDS-SR scores. The unsupervised clustering approach yielded core emotional, atypical, and sleep/insomnia clusters among the citalopram STAR*D cohort. The CO-MED (n = 640) cohort was independently adopted as a test sample for the mixed-effects regression analysis. The CO-MED cohort included three subgroups of escitalopram (n = 151), escitalopram-bupropion (n = 134), and venlafaxine-mirtazapine (n = 140). Core emotional symptoms were predicted with significantly above-chance performance in the escitalopram and venlafaxine-mirtazapine groups. Sleep symptoms were predicted above chance for the escitalopram-bupropion group. These findings may translate into clinical applications. Escitalopram alone or a venlafaxine-mirtazapine combination may be beneficial for the core emotional symptoms related to energy/fatigability, concentration/decision-making, interest loss, sad mood, and feelings of worthlessness. Escitalopram-bupropion combination may have a beneficial impact on symptoms related to sleep onset insomnia, midnocturnal insomnia, and termination insomnia. In other words, no antidepressant was equally effective for all three symptom clusters, and for each symptom cluster, there were significant differences in treatment efficacy between drugs. In general, antidepressants worked best in treating the core emotional and sleep symptoms and showed less effectiveness in treating atypical symptoms. The extent of these differences suggests that selection of the best antidepressant for a given symptom allows for more benefit than that gained by the application of an active compound versus a placebo. Taken together, therapeutic outcomes at the symptom cluster level remain predictable using the machine learning algorithm of self-reported patient data.
These findings may help guide future research on personalized medicines for antidepressant selection. The finding by Chekroud et al.—better trajectories for core emotional symptoms with citalopram—is consistent with the findings of the genome-based therapeutic drugs for depression (GENDEP) study where symptom dimensions of cognition and mood were significantly better with escitalopram than with nortriptyline [23]. Whereas large-scale comparative studies of combined severity demonstrated modest differences between antidepressant classes [61,62], the results by Chekroud et al., at the symptom cluster level, indicate considerable differences between medications both within and across antidepressant classes.
Future clinical research should determine whether these clusters can be generalized to other clinical cohorts and reflect good candidates for a true depressive symptom structure [63,64]. The cluster structures by Chekroud et al. resemble those of other scales in other large samples of patients [61,65,66,67], although a recent review insisted that the argument is still ongoing [63]. These findings are largely consistent with those of Chekroud et al. in terms of independent sleep symptoms and core emotional symptoms, including sad mood, anhedonia, and self-worthlessness. However, the use of many different rating scales for depression impedes direct comparisons between cohorts [64].

6. Brain Imaging Techniques and Machine Learning in Depression

Machine learning algorithms are wildly applicable to diverse data of patients for elucidation of the complex nature of depression. In particular, in addition to the aforementioned clinical cohort data, there has been growing attention toward using brain imaging methods to detect endophenotypes of depression that can be clinically significant and feasible for translation to diagnosis [53,68]. Brain MRIs are one of the most widely used techniques that help identify the potential biological markers of depression. Importantly, MRI techniques with machine learning algorithms can produce a classification with brain networks and prediction of treatment response in depression. First, some representative studies adopted graph theory approaches [69,70,71,72] to detect defective functional and structural brain networks of depressed patients. Gong et al. [73] enumerated diverse brain network features, such as alterations in regional and connectivity patterns of different MRI modalities for depression, which include the regional betweenness and degree centrality in structural MRI, region-of-interest-based analysis in functional MRI, and white matter structural connectivity in the diffusion tensor image. Zeng et al. [74] examined the whole-brain functional connectivity at resting state to distinguish the depressed patients from the controls, which yielded 100% sensitivity. The most discriminant functional connectivity was found within or across the affective network, default mode network, and visual cortical areas, which seem to play a critical role in the neuromechanism of depression. Second, other representative studies sought to find alterations in brain network activity at resting state as potential endophenotypes for prediction of therapeutic outcomes. Drysdale et al. [75] suggested that four different patterns of fronto-striatal and limbic functional connectivity be defined as depression biomarkers from functional MRI analyses. The biomarkers were also related to distinctive profiles of clinical symptoms. For instance, biomarker 1 was associated with severe fatigue and anhedonia, and showed best response to repetitive transcranial magnetic stimulation. Redlich et al. [76] examined whether changes in gray matter volume predict response to electroconvulsive therapy. Support vector regression was accompanied with univariate analysis of the Hamilton Depression Rating Scale score, which successfully predicted the response to electroconvulsive therapy and reduction in HDRS. Jiang et al. [77] predicted remission after electroconvulsive therapy using the gray matter of depressed patients, in which six different gray matter networks were suggested as predictors of response to electroconvulsive therapy. Thus, the connectome-based endophenotypes may yield novel opportunities to define the diagnosis of depression and improve therapeutic response.

7. Future Research into Treatment Selection Models

We address the major steps involved in building antidepressant selection models from a clinical database that involves values, for each patient, on variables that represent clinical and demographic characteristics, therapeutics applied to the patient, and observed outcomes from the therapeutics. Understanding the sequential steps is crucial for interpreting and evaluating the utility of findings from the antidepressant selection studies.
The first step was to establish candidate predictor variables. Appropriate candidate predictor variables are those that are acquired prior to the treatment assignment and that credibility could be related to outcome, either generally or differentially between treatments. If a previous study has suggested that a variable can predict an outcome, then it should be involved as a potential predictor variable. However, as the literature on predictors of psychiatric disorders is still relatively scarce, considering other putative variables is recommended.
Variables should be free of significant missingness, and systematic missingness should be examined to ensure the appropriateness of imputation [78]. Variables should also show considerable variability. For instance, it does not make sense to involve sex if 90% of the sample is male. Selecting variables used for prediction is reliant on situations in which predictors exhibit high collinearity. Therefore, it is plausible to test the covariance structure of the putative predictors and take measures to reduce the high collinearity [78]. Other suggestions for identifying putative predictors include addressing outliers, making categorical variables binary, and converting variables for hypothetical reasons or handling highly skewed distributions.
Once putative predictor variables were selected, the next step was to construct the prediction model. This is typically a two-step procedure that includes variable selection and model specifications. Many different variable selection approaches have been suggested for treatment selection, all of which attempt to identify which variables, among the putative predictors, contribute significantly to the prediction outcome. Gillan and Whelan [79] presented an outstanding discussion of data-driven versus theory-driven approaches to model specifications. Typical approaches depend on parametric regression models [80] that select only variables with statistically significant contributions to the outcome. Another approach includes penalties with the goal of limiting the number of selected variables [81]. Others utilize bootstrapping processes that help maximize the generalizability of the models [58,82,83]. Progress in statistical modeling has led to feature selection methods, which are largely based on machine learning algorithms that can compliantly model and identify predictors, even with higher-order interactions [84]. Gillan and Whelan [79] provided an in-depth discussion of the merits of machine learning in the field of psychiatric disorders.

8. Conclusions

Research into depression is complicated by patient heterogeneity, which has motivated the search for homogeneous symptom clusters via data-driven computational approaches to identify predictor signatures in a cohort database. Along with the availability of clinical cohort data that are cost-effective and generalizable, machine learning algorithms enable symptom grouping and antidepressant selection. In particular, antidepressant selection in depression aims to help each patient receive the treatment, among the available options, which are most likely to yield a beneficial outcome for them. Symptom clusters putatively include core emotional, anxious, sleep/appetite, and atypical types, which seem congruent with the theory-driven grouping of depression, underlining that data-driven classification can agree with theory-driven approaches. Treatment outcomes differ in distinct antidepressant classes according to symptom clusters, although this pioneering finding should be replicated in other cohorts in the future. In addition, machine learning is a flexible platform to which brain imaging data can also be subjected to yield favorable performance in classification and therapeutics prediction in depression. Machine learning-based symptom clustering and antidepressant selection may be the first step in facilitating personalized medicine for better diagnostics and therapeutics for depression.

Author Contributions

I.B.K. and S.-C.P. designed the study and wrote the manuscript. Both authors have read and agreed to the published version of the manuscript.

Funding

This work has no funding sources.

Institutional Review Board Statement

Approval from institutional review board is wavied for the nature of review.

Informed Consent Statement

Informed consent is not required for the study.

Data Availability Statement

No further or supplementary data is stated in the study.

Conflicts of Interest

The authors have no conflict of interest relevant to this article.

References

  1. Vos, T.; Allen, C.; Arora, M.; Barber, R.M.; Bhutta, Z.A.; Brown, A.; Carter, A.; Casey, D.C.; Charlson, F.J.; Chen, A.Z. Global, regional, and national incidence, prevalence, and years lived with disability for 310 diseases and injuries, 1990–2015: A systematic analysis for the Global Burden of Disease Study 2015. Lancet 2016, 388, 1545–1602. [Google Scholar] [CrossRef] [Green Version]
  2. Rush, A.J.; Wisniewski, S.R.; Warden, D.; Luther, J.F.; Davis, L.L.; Fava, M.; Nierenberg, A.A.; Trivedi, M.H. Selecting among second-step antidepressant medication monotherapies: Predictive value of clinical, demographic, or first-step treatment features. Arch. Gen. Psychiatry 2008, 65, 870–880. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Rush, A.J.; Trivedi, M.H.; Stewart, J.W.; Nierenberg, A.A.; Fava, M.; Kurian, B.T.; Warden, D.; Morris, D.W.; Luther, J.F.; Husain, M.M.; et al. Combining Medications to Enhance Depression Outcomes (CO-MED): Acute and Long-Term Outcomes of a Single-Blind Randomized Study. Am. J. Psychiatry 2011, 168, 689–701. [Google Scholar] [CrossRef] [PubMed]
  4. Fried, E. Moving forward: How depression heterogeneity hinders progress in treatment and research. Expert Rev. Neurother. 2017, 17, 423–425. [Google Scholar] [CrossRef] [Green Version]
  5. Łojko, D.; Rybakowski, J.K. Atypical depression: Current perspectives. Neuropsychiatr. Dis. Treat. 2017, 13, 2447. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Day, C.V.; Williams, L.M. Finding a biosignature for melancholic depression. Expert Rev. Neurother. 2012, 12, 835–847. [Google Scholar] [CrossRef] [PubMed]
  7. Ionescu, D.F.; Niciu, M.J.; Mathews, D.C.; Richards, E.M.; Zarate, C.A., Jr. Neurobiology of anxious depression: A review. Depress. Anxiety 2013, 30, 374–385. [Google Scholar] [CrossRef] [Green Version]
  8. Park, S.-C.; Kim, J.-M.; Jun, T.-Y.; Lee, M.-S.; Kim, J.-B.; Yim, H.-W.; Park, Y.C. How many different symptom combinations fulfil the diagnostic criteria for major depressive disorder? Results from the CRESCEND study. Nord. J. Psychiatry 2017, 71, 217–222. [Google Scholar] [CrossRef]
  9. Park, S.-C.; Kim, D. The centrality of depression and anxiety symptoms in major depressive disorder determined using a network analysis. J. Affect. Disord. 2020, 271, 19–26. [Google Scholar] [CrossRef]
  10. Park, S.C.; Jang, E.Y.; Xiang, Y.T.; Kanba, S.; Kato, T.A.; Chong, M.Y.; Lin, S.K.; Yang, S.Y.; Avasthi, A.; Grover, S. Network analysis of the depressive symptom profiles in Asian patients with depressive disorders: Findings from the Research on Asian Psychotropic Prescription Patterns for Antidepressants (REAP-AD). Psychiatry Clin. Neurosci. 2020, 74, 344–353. [Google Scholar] [CrossRef] [Green Version]
  11. Kim, Y.-K.; Park, S.-C. An alternative approach to future diagnostic standards for major depressive disorder. Prog. Neuro-Psychopharmacol. Biol. Psychiatry 2020, 105, 110133. [Google Scholar] [CrossRef]
  12. Park, S.-C.; Kim, Y.-K. Challenges and Strategies for Current Classifications of Depressive Disorders: Proposal for Future Diagnostic Standards. Major Depress. Disord. Rethink. Underst. Recent Discov. 2021, 1305, 103. [Google Scholar]
  13. American Psychiatric Association. Diagnostic and Statistical Manual of Mental Disorders, 5th ed.; American Psychiatric Publishing: Washington, DC, USA, 2014. [Google Scholar]
  14. Østergaard, S.D.; Jensen, S.; Bech, P. The heterogeneity of the depressive syndrome: When numbers get serious. Acta Psychiatr. Scand. 2011, 124, 495–496. [Google Scholar] [CrossRef]
  15. Zimmerman, M.; Ellison, W.; Young, D.; Chelminski, I.; Dalrymple, K. How many different ways do patients meet the diagnostic criteria for major depressive disorder? Compr. Psychiatry 2015, 56, 29–34. [Google Scholar] [CrossRef] [PubMed]
  16. Park, S.-C.; Kim, Y.-K. Contemporary issues in depressive disorders. Psychiatry Investig. 2019, 16, 633. [Google Scholar] [CrossRef]
  17. Park, S.-C.; Kim, Y.-K. Diagnostic issues of depressive disorders from Kraepelinian dualism to the Diagnostic and Statistical Manual of Mental Disorders. Psychiatry Investig. 2019, 16, 636. [Google Scholar] [CrossRef]
  18. Rosenman, S.; Nasti, J. Psychiatric diagnoses are not mental processes: Wittgenstein on conceptual confusion. Aust. N. Z. J. Psychiatry 2012, 46, 1046–1052. [Google Scholar] [CrossRef]
  19. Wittgenstein, L. Philosophical Investigations, the German Text, with a Revised English Translation; Anscombe, G.E.M., Translator; Blackwell: Malden, MA, USA, 2001. [Google Scholar]
  20. Insel, T.R. Next-generation treatments for mental disorders. Sci. Transl. Med. 2012, 4, ps119–ps155. [Google Scholar] [CrossRef]
  21. Uher, R.; Dernovsek, M.Z.; Mors, O.; Hauser, J.; Souery, D.; Zobel, A.; Maier, W.; Henigsberg, N.; Kalember, P.; Rietschel, M. Melancholic, atypical and anxious depression subtypes and outcome of treatment with escitalopram and nortriptyline. J. Affect. Disord. 2011, 132, 112–120. [Google Scholar] [CrossRef]
  22. Arnow, B.A.; Blasey, C.; Williams, L.M.; Palmer, D.M.; Rekshan, W.; Schatzberg, A.F.; Etkin, A.; Kulkarni, J.; Luther, J.F.; Rush, A.J. Depression subtypes in predicting antidepressant response: A report from the iSPOT-D trial. Am. J. Psychiatry 2015, 172, 743–750. [Google Scholar] [CrossRef] [PubMed]
  23. Uher, R.; Maier, W.; Hauser, J.; Marušič, A.; Schmael, C.; Mors, O.; Henigsberg, N.; Souery, D.; Placentino, A.; Rietschel, M. Differential efficacy of escitalopram and nortriptyline on dimensional measures of depression. Br. J. Psychiatry 2009, 194, 252–259. [Google Scholar] [CrossRef] [PubMed]
  24. Hartigan, J.A.; Wong, M.A. Algorithm AS 136: A k-means clustering algorithm. J. R. Stat. Soc. Ser. C (Appl. Stat.) 1979, 28, 100–108. [Google Scholar] [CrossRef]
  25. Passos, I.C.; Mwangi, B.; Kapczinski, F. Personalized Psychiatry: Big Data Analytics in Mental Health; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  26. Tonekaboni, S.; Joshi, S.; McCradden, M.D.; Goldenberg, A. What clinicians want: Contextualizing explainable machine learning for clinical end use. In Proceedings of the 4th Machine Learning for Healthcare Conference, Ann Arbor, MI, USA, 8 August 2019; pp. 359–380. [Google Scholar]
  27. Sinyor, M.; Schaffer, A.; Levitt, A. The sequenced treatment alternatives to relieve depression (STAR* D) trial: A review. Can. J. Psychiatry 2010, 55, 126–135. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Pfeiffer-Gerschel, T.; Schlee, A.; Hegerl, U. The German Research Network on depression and suicidality. Arch. Suicide Res. 2006, 101, 505–508. [Google Scholar]
  29. Van Loo, H.M.; De Jonge, P.; Romeijn, J.-W.; Kessler, R.C.; Schoevers, R.A. Data-driven subtypes of major depressive disorder: A systematic review. BMC Med. 2012, 10, 156. [Google Scholar] [CrossRef] [Green Version]
  30. Kautzky, A.; Baldinger-Melich, P.; Kranz, G.S.; Vanicek, T.; Souery, D.; Montgomery, S.; Mendlewicz, J.; Zohar, J.; Serretti, A.; Lanzenberger, R. A new prediction model for evaluating treatment-resistant depression. J. Clin. Psychiatry 2017, 78, 215–222. [Google Scholar] [CrossRef]
  31. Riedel, M.; Möller, H.-J.; Obermeier, M.; Adli, M.; Bauer, M.; Kronmüller, K.; Brieger, P.; Laux, G.; Bender, W.; Heuser, I. Clinical predictors of response and remission in inpatients with depressive syndromes. J. Affect. Disord. 2011, 133, 137–149. [Google Scholar] [CrossRef] [PubMed]
  32. Perlis, R.H. A clinical risk stratification tool for predicting treatment resistance in major depressive disorder. Biol. Psychiatry 2013, 74, 7–14. [Google Scholar] [CrossRef] [Green Version]
  33. Athreya, A.P.; Neavin, D.; Carrillo-Roa, T.; Skime, M.; Biernacka, J.; Frye, M.A.; Rush, A.J.; Wang, L.; Binder, E.B.; Iyer, R.K. Pharmacogenomics-driven prediction of antidepressant treatment outcomes: A machine-learning approach with multi-trial replication. Clin. Pharmacol. Ther. 2019, 106, 855–865. [Google Scholar] [CrossRef] [Green Version]
  34. Kautzky, A.; Möller, H.J.; Dold, M.; Bartova, L.; Seemüller, F.; Laux, G.; Riedel, M.; Gaebel, W.; Kasper, S. Combining machine learning algorithms for prediction of antidepressant treatment response. Acta Psychiatr. Scand. 2021, 143, 36–49. [Google Scholar] [CrossRef] [PubMed]
  35. Chekroud, A.M.; Gueorguieva, R.; Krumholz, H.M.; Trivedi, M.H.; Krystal, J.H.; McCarthy, G. Reevaluating the efficacy and predictability of antidepressant treatments: A symptom clustering approach. JAMA Psychiatry 2017, 74, 370–378. [Google Scholar] [CrossRef] [PubMed]
  36. Sharp, R. The Hamilton rating scale for depression. Occup. Med. 2015, 65, 340. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Reilly, T.J.; MacGillivray, S.A.; Reid, I.C.; Cameron, I.M. Psychometric properties of the 16-item Quick Inventory of Depressive Symptomatology: A systematic review and meta-analysis. J. Psychiatr. Res. 2015, 60, 132–140. [Google Scholar] [CrossRef] [PubMed]
  38. Musil, R.; Seemüller, F.; Meyer, S.; Spellmann, I.; Adli, M.; Bauer, M.; Kronmüller, K.T.; Brieger, P.; Laux, G.; Bender, W. Subtypes of depression and their overlap in a naturalistic inpatient sample of major depressive disorder. Int. J. Methods Psychiatr. Res. 2018, 27, e1569. [Google Scholar] [CrossRef]
  39. Gaspersz, R.; Nawijn, L.; Lamers, F.; Penninx, B.W. Patients with anxious depression: Overview of prevalence, pathophysiology and impact on course and treatment outcome. Curr. Opin. Psychiatry 2018, 31, 17–25. [Google Scholar] [CrossRef]
  40. Hieronymus, F.; Lisinski, A.; Nilsson, S.; Eriksson, E. Influence of baseline severity on the effects of SSRIs in depression: An item-based, patient-level post-hoc analysis. Lancet Psychiatry 2019, 6, 745–752. [Google Scholar] [CrossRef]
  41. Bech, P. Rating scales for affective disorders: Their validity and consistency. Acta Psychiatr. Scand. 1981, 64, 11–101. [Google Scholar]
  42. Maier, W.; Philipp, M.; Heuser, I.; Schlegel, S.; Buller, R.; Wetzel, H. Improving depression severity assessment—I. Reliability, internal validity and sensitivity to change of three observer depression scales. J. Psychiatr. Res. 1988, 22, 3–12. [Google Scholar] [CrossRef]
  43. McIntyre, R.; Kennedy, S.; Bagby, R.M.; Bakish, D. Assessing full remission. J. Psychiatry Neurosci. 2002, 27, 235. [Google Scholar]
  44. Seemüller, F.; Riedel, M.; Wickelmaier, F.; Adli, M.; Mundt, C.; Marneros, A.; Laux, G.; Bender, W.; Heuser, I.; Zeiler, J. Atypical symptoms in hospitalised patients with major depressive episode: Frequency, clinical characteristics, and internal validity. J. Affect. Disord. 2008, 108, 271–278. [Google Scholar] [CrossRef]
  45. Loas, G.; Lefebvre, G.; Rotsaert, M.; Englert, Y. Relationships between anhedonia, suicidal ideation and suicide attempts in a large sample of physicians. PLoS ONE 2018, 13, e0193619. [Google Scholar] [CrossRef]
  46. Winer, E.S.; Drapeau, C.W.; Veilleux, J.C.; Nadorff, M.R. The association between anhedonia, suicidal ideation, and suicide attempts in a large student sample. Arch. Suicide Res. 2016, 20, 265–272. [Google Scholar] [CrossRef]
  47. Trivedi, M.H.; Morris, D.W.; Grannemann, B.D.; Mahadi, S. Symptom clusters as predictors of late response to antidepressant treatment. J. Clin. Psychiatry 2005, 66, 1064–1070. [Google Scholar] [CrossRef]
  48. Fried, E.I.; Nesse, R.M. Depression is not a consistent syndrome: An investigation of unique symptom patterns in the STAR*D study. J. Affect. Disord. 2015, 172, 96–102. [Google Scholar] [CrossRef] [Green Version]
  49. Insel, T.R.; Cuthbert, B.N. Brain disorders? precisely. Science 2015, 348, 499–500. [Google Scholar] [CrossRef] [PubMed]
  50. Chekroud, A.M.; Krystal, J.H. Personalised pharmacotherapy: An interim solution for antidepressant treatment? BMJ Br. Med. J. (Online) 2015, 350. [Google Scholar] [CrossRef] [PubMed]
  51. Paulus, M.P. Pragmatism instead of mechanism: A call for impactful biological psychiatry. JAMA Psychiatry 2015, 72, 631–632. [Google Scholar] [CrossRef] [PubMed]
  52. Kautzky, A.; Baldinger, P.; Souery, D.; Montgomery, S.; Mendlewicz, J.; Zohar, J.; Serretti, A.; Lanzenberger, R.; Kasper, S. The combined effect of genetic polymorphisms and clinical parameters on treatment outcome in treatment-resistant depression. Eur. Neuropsychopharmacol. 2015, 25, 441–453. [Google Scholar] [CrossRef] [PubMed]
  53. Patel, M.J.; Andreescu, C.; Price, J.C.; Edelman, K.L.; Reynolds, C.F., III; Aizenstein, H.J. Machine learning approaches for integrating clinical and imaging features in late-life depression classification and response prediction. Int. J. Geriatr. Psychiatry 2015, 30, 1056–1067. [Google Scholar] [CrossRef] [PubMed]
  54. Iniesta, R.; Malki, K.; Maier, W.; Rietschel, M.; Mors, O.; Hauser, J.; Henigsberg, N.; Dernovsek, M.Z.; Souery, D.; Stahl, D. Combining clinical variables to optimize prediction of antidepressant treatment outcomes. J. Psychiatr. Res. 2016, 78, 94–102. [Google Scholar] [CrossRef]
  55. Maciukiewicz, M.; Marshe, V.S.; Hauschild, A.-C.; Foster, J.A.; Rotzinger, S.; Kennedy, J.L.; Kennedy, S.H.; Müller, D.J.; Geraci, J. GWAS-based machine learning approach to predict duloxetine response in major depressive disorder. J. Psychiatr. Res. 2018, 99, 62–68. [Google Scholar] [CrossRef] [PubMed]
  56. Cohen, Z.D.; DeRubeis, R.J. Treatment selection in depression. Annu. Rev. Clin. Psychol. 2018, 14, 209–236. [Google Scholar] [CrossRef] [PubMed]
  57. Chekroud, A.M.; Zotti, R.J.; Shehzad, Z.; Gueorguieva, R.; Johnson, M.K.; Trivedi, M.H.; Cannon, T.D.; Krystal, J.H.; Corlett, P.R. Cross-trial prediction of treatment outcome in depression: A machine learning approach. Lancet Psychiatry 2016, 3, 243–250. [Google Scholar] [CrossRef]
  58. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 2005, 67, 301–320. [Google Scholar] [CrossRef] [Green Version]
  59. Friedman, J.; Hastie, T.; Tibshirani, R. Regularization paths for generalized linear models via coordinate descent. J. Stat. Softw. 2010, 33, 1. [Google Scholar] [CrossRef] [Green Version]
  60. Leuchter, A.F.; Cook, I.A.; Marangell, L.B.; Gilmer, W.S.; Burgoyne, K.S.; Howland, R.H.; Trivedi, M.H.; Zisook, S.; Jain, R.; McCracken, J.T. Comparative effectiveness of biomarkers and clinical indicators for predicting outcomes of SSRI treatment in major depressive disorder: Results of the BRITE-MD study. Psychiatry Res. 2009, 169, 124–131. [Google Scholar] [CrossRef] [PubMed]
  61. Cipriani, A.; Furukawa, T.A.; Salanti, G.; Geddes, J.R.; Higgins, J.P.; Churchill, R.; Watanabe, N.; Nakagawa, A.; Omori, I.M.; McGuire, H. Comparative efficacy and acceptability of 12 new-generation antidepressants: A multiple-treatments meta-analysis. Lancet 2009, 373, 746–758. [Google Scholar] [CrossRef]
  62. Gartlehner, G.; Hansen, R.A.; Morgan, L.C.; Thaler, K.; Lux, L.; Van Noord, M.; Mager, U.; Thieda, P.; Gaynes, B.N.; Wilkins, T. Comparative benefits and harms of second-generation antidepressants for treating major depressive disorder: An updated meta-analysis. Ann. Intern. Med. 2011, 155, 772–785. [Google Scholar] [CrossRef] [Green Version]
  63. Fried, E.I.; van Borkulo, C.D.; Epskamp, S.; Schoevers, R.A.; Tuerlinckx, F.; Borsboom, D. Measuring depression over time… Or not? Lack of unidimensionality and longitudinal measurement invariance in four common rating scales of depression. Psychol. Assess. 2016, 28, 1354. [Google Scholar] [CrossRef] [PubMed]
  64. Fried, E.I. The 52 symptoms of major depression: Lack of content overlap among seven common depression scales. J. Affect. Disord. 2017, 208, 191–197. [Google Scholar] [CrossRef] [PubMed]
  65. Shafer, A.B. Meta-analysis of the factor structures of four depression questionnaires: Beck, CES-D, Hamilton, and Zung. J. Clin. Psychol. 2006, 62, 123–146. [Google Scholar] [CrossRef] [PubMed]
  66. Li, Y.; Aggen, S.; Shi, S.; Gao, J.; Tao, M.; Zhang, K.; Wang, X.; Gao, C.; Yang, L.; Liu, Y. The structure of the symptoms of major depression: Exploratory and confirmatory factor analysis in depressed Han Chinese women. Psychol. Med. 2014, 44, 1391–1401. [Google Scholar] [CrossRef] [Green Version]
  67. Romera, I.; Delgado-Cohen, H.; Perez, T.; Caballero, L.; Gilaberte, I. Factor analysis of the Zung self-rating depression scale in a large sample of patients with major depressive disorder in primary care. BMC Psychiatry 2008, 8, 1–8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  68. Haslam, N.; Beck, A.T. Categorization of major depression in an outpatient sample. J. Nerv. Ment. Dis. 1993, 181, 725–731. [Google Scholar] [CrossRef] [PubMed]
  69. Yoshida, K.; Shimizu, Y.; Yoshimoto, J.; Takamura, M.; Okada, G.; Okamoto, Y.; Yamawaki, S.; Doya, K. Prediction of clinical depression scores and detection of changes in whole-brain using resting-state functional MRI data with partial least squares regression. PLoS ONE 2017, 12, e0179638. [Google Scholar] [CrossRef] [PubMed]
  70. Zhong, X.; Shi, H.; Ming, Q.; Dong, D.; Zhang, X.; Zeng, L.-L.; Yao, S. Whole-brain resting-state functional connectivity identified major depressive disorder: A multivariate pattern analysis in two independent samples. J. Affect. Disord. 2017, 218, 346–352. [Google Scholar] [CrossRef]
  71. Wang, X.; Ren, Y.; Zhang, W. Depression disorder classification of fMRI data using sparse low-rank functional brain network and graph-based features. Comput. Math. Methods Med. 2017, 2017, 3609821. [Google Scholar] [CrossRef] [Green Version]
  72. Fang, P.; Zeng, L.-L.; Shen, H.; Wang, L.; Li, B.; Liu, L.; Hu, D. Increased cortical-limbic anatomical network connectivity in major depression revealed by diffusion tensor imaging. PLoS ONE 2012, 7, e45972. [Google Scholar] [CrossRef] [Green Version]
  73. Gong, Q.; He, Y. Depression, neuroimaging and connectomics: A selective overview. Biol. Psychiatry 2015, 77, 223–235. [Google Scholar] [CrossRef]
  74. Zeng, L.-L.; Shen, H.; Liu, L.; Wang, L.; Li, B.; Fang, P.; Zhou, Z.; Li, Y.; Hu, D. Identifying major depression using whole-brain functional connectivity: A multivariate pattern analysis. Brain 2012, 135, 1498–1507. [Google Scholar] [CrossRef] [Green Version]
  75. Drysdale, A.T.; Grosenick, L.; Downar, J.; Dunlop, K.; Mansouri, F.; Meng, Y.; Fetcho, R.N.; Zebley, B.; Oathes, D.J.; Etkin, A. Resting-state connectivity biomarkers define neurophysiological subtypes of depression. Nat. Med. 2017, 23, 28–38. [Google Scholar] [CrossRef] [Green Version]
  76. Redlich, R.; Opel, N.; Grotegerd, D.; Dohm, K.; Zaremba, D.; Bürger, C.; Münker, S.; Mühlmann, L.; Wahl, P.; Heindel, W. Prediction of individual response to electroconvulsive therapy via machine learning on structural magnetic resonance imaging data. JAMA Psychiatry 2016, 73, 557–564. [Google Scholar] [CrossRef]
  77. Jiang, R.; Abbott, C.C.; Jiang, T.; Du, Y.; Espinoza, R.; Narr, K.L.; Wade, B.; Yu, Q.; Song, M.; Lin, D. SMRI biomarkers predict electroconvulsive treatment outcomes: Accuracy with independent data sets. Neuropsychopharmacology 2018, 43, 1078–1087. [Google Scholar] [CrossRef] [Green Version]
  78. Jamshidian, M.; Jalal, S. Tests of homoscedasticity, normality, and missing completely at random for incomplete multivariate data. Psychometrika 2010, 75, 649–674. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  79. Gillan, C.M.; Whelan, R. What big data can do for treatment in psychiatry. Curr. Opin. Behav. Sci. 2017, 18, 34–42. [Google Scholar] [CrossRef]
  80. Fournier, J.C.; DeRubeis, R.J.; Shelton, R.C.; Hollon, S.D.; Amsterdam, J.D.; Gallop, R. Prediction of response to medication and cognitive therapy in the treatment of moderate to severe depression. J. Consult. Clin. Psychol. 2009, 77, 775. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  81. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B (Methodol.) 1996, 58, 267–288. [Google Scholar] [CrossRef]
  82. Austin, P.C.; Tu, J.V. Bootstrap methods for developing predictive models. Am. Stat. 2004, 58, 131–137. [Google Scholar] [CrossRef]
  83. Garge, N.R.; Bobashev, G.; Eggleston, B. Random forest methodology for model-based recursive partitioning: The mobForest package for R. BMC Bioinform. 2013, 14, 125. [Google Scholar] [CrossRef] [Green Version]
  84. Bleich, J.; Kapelner, A.; George, E.I.; Jensen, S.T. Variable selection for BART: An application to gene regulation. Ann. Appl. Stat. 2014, 8, 1750–1781. [Google Scholar] [CrossRef] [Green Version]
Table 1. Data-driven symptom clustering of depression from clinical cohorts with over 500 depressed patients.
Table 1. Data-driven symptom clustering of depression from clinical cohorts with over 500 depressed patients.
ClusterSubtyping Based onRelevant PhenotypeStatistical MethodCohortStudy
Core emotionalQIDS-SREnergy/fatigability STAR*D (n = 4017)Chekroud et al. [35]
HAM-DConcentration/decision makingHierarchical clusteringCO-MED (n = 640)Kautzky et al. [34]
Loss of interest GRND (n = 1079)
Mood (sadness) GSRD (n = 1568)
Worthlessness
SleepQIDS-SRSleep-onset insomniaHierarchical clusteringSTARD*D (n = 4017)Chekroud et al. [35]
HAM-DMid-nocturnal insomnia CO-MED (n = 640)Kautzky et al. [34]
Early-morning insomnia GRND (n = 1079)-
GSRD (n = 1568)
AtypicalQIDS-SRPsychomotor agitationHierarchical clusteringSTARD*D (n = 4017)Chekroud et al. [35]
HAM-DPsychomotor retardation CO-MED (n = 640)
Suicidal ideation
Hypersomnia
Hypochondriasis
Somatic/anxiousHAM-DPsychomotor agitationHierarchical clusteringGRND (n = 1079)Kautzky et al. [34]
Anxiety somatic GSRD (n = 1568)
Anxiety psychic
Somatic symptoms
Genital symptoms
Hypochondriasis
Appetite/weightHAM-DAppetiteHierarchical clusteringGRND (n = 1079)Kautzky et al. [34]
Weight changes GSRD (n = 1568)
Table 2. Prediction models for treatment response in depression.
Table 2. Prediction models for treatment response in depression.
StudyFeature Sample SizeStatistical MethodPrediction Accuracy on Treatment Response
Kautzky et al. [52]three SNPs and clinical variables225 depressed patients from GSRDRandom forest model25% of melancholia
Patel et al. [53]age, structual imaging, and mini-mental status examination scores33 late-life depression patients and 35 non-depressed individualsLinear and non-linear decision tree model89%
Iniesta et al. [54]clinical and demographic variables793 depressed patientsRegularized regression model72%
Chekroud et al. [35]25 clinical variables4039 depressed patients from STAR*DMixed-effects regression analysis59%
Maciukiewicz et al. [55]SNPs186 MDD patientsSupporter vector machine model52%
Kautzky et al. [30]48 clinical, sociodemographic, and psychosocial predictors400 training samples and 80 test samplesRandom forest model85%
Michael et al. [31] HAMD-21 baseline score, episode length < 24 months and fewer previous hospitalizations1014 naturalistically treated inpatientsLogistic regression and third-CART analyses72%
Perlis et al. [32]Clinical or sociodemographic variables 4039 depressed patients from STAR*DLogistic regression model71%
Kautzky et al. [34]88 clinical predictors1079 GRND samplesRandom forest model85%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, I.B.; Park, S.-C. Machine Learning-Based Definition of Symptom Clusters and Selection of Antidepressants for Depressive Syndrome. Diagnostics 2021, 11, 1631. https://doi.org/10.3390/diagnostics11091631

AMA Style

Kim IB, Park S-C. Machine Learning-Based Definition of Symptom Clusters and Selection of Antidepressants for Depressive Syndrome. Diagnostics. 2021; 11(9):1631. https://doi.org/10.3390/diagnostics11091631

Chicago/Turabian Style

Kim, Il Bin, and Seon-Cheol Park. 2021. "Machine Learning-Based Definition of Symptom Clusters and Selection of Antidepressants for Depressive Syndrome" Diagnostics 11, no. 9: 1631. https://doi.org/10.3390/diagnostics11091631

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop