Next Article in Journal
Doctors’ Perceptions of E-Prescribing upon Its Mandatory Adoption in Poland, Using the Unified Theory of Acceptance and Use of Technology Method
Previous Article in Journal
Sex Differences in the Influence of Sleep on Body Mass Index and Risk of Metabolic Syndrome in Middle-Aged Adults
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Identifying High-Risk Factors of Depression in Middle-Aged Persons with a Novel Sons and Spouses Bayesian Network Model

1
SKK Business School, Sungkyunkwan University, Seoul 03063, Korea
2
Airports Council International (ACI) World, Montreal, QC H4Z 1G8, Canada
3
Department of Health Sciences & Technology, Samsung Advanced Institute for Health Sciences & Technology (SAIHST), Sungkyunkwan University, Seoul 06355, Korea
*
Author to whom correspondence should be addressed.
Healthcare 2020, 8(4), 562; https://doi.org/10.3390/healthcare8040562
Submission received: 6 November 2020 / Revised: 6 December 2020 / Accepted: 11 December 2020 / Published: 15 December 2020
(This article belongs to the Section Health Informatics and Big Data)

Abstract

:
It has been reported repeatedly that depression in middle-aged people may cause serious ramifications in public health. However, previous studies on this important research topic have focused on utilizing either traditional statistical methods (i.e., logistic regressions) or black-or-gray artificial intelligence (AI) methods (i.e., neural network, Support Vector Machine (SVM), ensemble). Previous studies lack suggesting more decision-maker-friendly methods, which need to produce clear interpretable results with information on cause and effect. For the sake of improving the quality of decisions of healthcare decision-makers, public health issues require identification of cause and effect information for any type of strategic healthcare initiative. In this sense, this paper proposes a novel approach to identify the main causes of depression in middle-aged people in Korea. The proposed method is the Sons and Spouses Bayesian network model, which is an extended version of conventional TAN (Tree-Augmented Naive Bayesian Network). The target dataset is a longitudinal dataset employed from the Korea National Health and Nutrition Examination Survey (KNHANES) database with a sample size of 8580. After developing the proposed Sons and Spouses Bayesian network model, we found thirteen main causes leading to depression. Then, genetic optimization was executed to reveal the most probable cause of depression in middle-aged people that would provide practical implications to field practitioners. Therefore, our proposed method can help healthcare decision-makers comprehend changes in depression status by employing what-if queries towards a target individual.

1. Introduction

Depression is a common illness that affects a vast amount of the population [1]. Some estimates in western-based samples have depression levels as (potentially) high as half of the population, although these figures have been scrutinized [2]. Until now, it has been widely accepted that middle-aged adults are among the least depressed age group in society. Based on the maturity hypothesis, it is thought that middle-aged adults have greater self-esteem and positive self-image [3]. Thus, prior research has continuously shown a U-shaped distribution in the data with pre-30s and post-70s shown to be the most venerable [3]. Unfortunately, this does not mean that middle-aged adults are not at risk of depression. In fact, research suggests that depression in middle-aged adults can start around their 40s and continue in an upward trend [3].
As depression is not attributable to a small number of factors, trying to understand all of the potential causes can be a complex problem. For example, it has been widely shown that people who suffer from depression are more likely to engage in excessive smoking and drinking, creating a doubled-edged sword of harm towards one’s mental and physical health [4]. Further, trends in depression have been strongly linked to differences within gender [5], as well as socioeconomic and health-related reasons, such as obesity [6]. Moreover, depression has been shown to intrude on fundamental biological processes that are used for regulating certain bodily functions, such as sleep and appetite [7].
One common solution proposed for dealing with these issues is antidepressant medication. These drugs have been shown to be effective as quick fixes for people suffering from depression; however, they are yet to be proven as effective, long-lasting solutions [8]. Based on numbers provided by the American Psychological Association (APA) in 2017, 12.7% of the American population were actively taking antidepressants, with 16.6% of the population falling between the ages of 40 and 59 [9]. In South Korea, for example, short-/long-term use of antidepressants for the same age group was shown to be 21.25% [10]. Based on these figures, it could be suggested that depression is potentially more prevalent in middle ages compared to what was previously found [3].
As the use of antidepressant drugs in the middle age population is seen to be relatively high, in this paper we wanted to explore and propose the use of machine learning and artificial intelligence (AI) to identify depression risk factors. These days, deep-learning and logistical reasoning have dominated the conceptual frameworks within the field of artificial intelligence (AI). This has meant that less attention has been focused on the probabilistic theoretical paradigms. Due to real-world demands, probabilistic models are being chosen more often for tackling complex problems—this is especially true when compared to black box models, such as deep learning [11].
Given the complexity of understanding depression, there are valid grounds for a research agenda that implements health informatic techniques. Therefore, the goal of this research paper is to provide an alternative solution for understanding depression and exploring potential early warning signs that can guide preventative measures. Specifically, we explore depression-related factors using Bayesian statistics. This approach involves three stages, including feature selection, a supervised Sons and Spouses Bayesian Network (SS) [12], and genetic optimization through a genetic algorithm for understanding what solutions show the greatest risk to inducing depression. Although not common within the medical field, this approach has been successfully implemented in exploring other medical fields, including adolescent obesity [13], and thus we believe this methodology is well adapted for the current issue presented in this study.
Henceforth, we look to combine the use of probabilistic AI in the pursuit of analyzing the most probable risk factors on depression; therefore, this research is guided by the following research questions:
RQ1: based on a machine-learned probabilistic Bayesian Network, what are the most probable antecedents of depression within middle-aged people?
RQ2: through an AI-based genetic algorithm, what solutions show the greatest risk to inducing depression within middle-aged people?

2. Materials and Methods

2.1. Data Used for Experimentation

This study is an analysis using data from the seventh Korea National Health and Nutrition Examination Survey (KNHANES) (2016–2018), conducted by the Korea Centers for Disease Control and Prevention. Among the total samples of 24,269 in the seventh KNHANES database, this study was conducted on 10,120 people that fell into a middle-aged category, including people in their 30s, 40s, and 50s. After excluding 1540 people whose data had too many missing values, the final sample included 8580 people. This selection process was influenced by prior work that dealt with the topic of depression from the KNHANES public dataset [14].
Independent variables were divided into eight variables based on demographic characteristics and 18 variables based on health-related traits. First, gender, age, income (quartile), marital status, job status, drinking, smoking, and body mass index (BMI) were selected for the demographic characteristic variables. Variables related to health include subjective health conditions, doctor’s diagnosis for 13 diseases (hypertension, stroke, myocardial infarction, angina, arthritis, osteoporosis, pulmonary tuberculosis, asthma, diabetes, thyroid disease, breast cancer, cervical cancer, and thyroid cancer), oral examination, and values of glycated hemoglobin, hemoglobin, and high-sensitivity of the C-reactive protein.
The dependent variable was selected from the Euro quality of life questionnaire 5-dimensional classification (EQ-5D) developed by the EuroQol. In this study, the samples that showed little or high anxiety/depression were merged into one, representing depression is present. Detailed classifications are shown in Table A1.

2.2. Machine Learning Approach Based on Bayesian Statistics

In order to explore our first research question, we implemented a Bayesian Network (BN). A BN is a directed acyclic graph in which nodes represent domain variables, and arcs between nodes represent probabilistic dependencies. A BN is used to model data by computing the conditional probability of a given node based upon values assigned to other nodes within the data space [15]. In the distinct case, and provided that the probability of B is not equal to zero, Bayes’ theorem relates the conditional and marginal probabilities of events A and B through the following calculation [15]:
P r [ A | B ] = P ( A ) * P r [ B | A ] P r [ B ]
The original proposition of Bayes’ theorem stated that the P(A) of the equation is the prior probability or the “unconditional” probability of A. In other words, this means that it does not take into consideration any of the information of B. Further, B does not have to emerge after the event A. Next, for a Bayes’ model to produce evidence of a probable outcome, it must calculate the joint probability distribution of a network. As a Bayes’ network can produce exponential amounts of probabilities depending on the data size, Bayesian networks use a localized conditional distribution that acts as a compressor in terms of calculations. This joint probability distribution then must be calculated in full. This is where the chain rule is used to analyze all of the joint probabilities of the network created. This is represented with the following equation [15]:
P ( x i , , x n ) =   i P ( x i | p a i )  
In order to learn the Bayes’ network, there are two competing methods: mutual information and the Kullback–Leibler (KL) Divergence metric. In this research, the KL Divergence was employed as this approach allows for the comparison of two probability distributions [16]. If we represent two probability distributions as A and B, then we can calculate the strength of the direct relationship between two of the nodes (variables):
D K L ( A ( X ) B ( X ) ) =   X A ( X ) log 2 A ( X ) B ( X )
BNs have an advantage over other various predictive models, such as K-Nearest Neighbor (KNN) or Decision Trees (DT). This is because the Bayesian network structure represents the interrelations amongst attributes in the dataset explicitly [12]. This provides the benefit of being able to compute inference “omnidirectionally.” [17]. In this research, three Bayesian networks were considered: Naïve Bayes [15], Tree Augmented Naïve Bayes (TAN) [12], and Sons and Spouses (SS) [18]. First, Naïve Bayes was not used. This is because it is unable to filter out non-relevant nodes due to the fact it works in a Naïve fashion connecting all nodes to the parent node. Next, due to performance reasons, we selected to use the SS compared to the TAN model. SS is a variation of TAN; however, it allows for the possibility of relationships between the child nodes (sons), including two parent nodes (spouses; see Figure 1).

2.3. Artificial Intelligence Approach Based on a Genetic Algorithm

The purpose of this section is so that we can pursue research question two. Through the implementation of the previous approach using a BN, one can identify the relationships within data through machine learning. Despite this, it is hard to infer whether these relationships are causal. For this reason, an alternative or next step approach is necessary to further build upon the BN model that can be used as a predictor rather than observer. Based on the target optimization, we implemented the evolutionary search heuristic based on a Genetic algorithm (GA). This approach allows for a search of a “good explanation” based on a precise and concise exploration. In this setting, precise refers to removing the surprise value that is seen for explaining the relationships [19].
GA is a type of evolutionary algorithm designed to replicate real life. The search space is represented as a collection of individuals represented by the character strings, also known as chromosomes. Measured on an objective function, GAs attempt to find the best “genetic material” based on the given search space, otherwise known as the population [20]. GAs work in the following way: a population is chosen, and the quality of all data points are examined. Parents are selected from the population that then produce children. Newly spawned individuals in the population have a probability near zero so that they “mutate” into new potential individuals. Next, individuals are removed from the population based on a selection criterion until the population is the same size as before, known as one iteration or generation. In BN, children are selected based on two operators that help with the mutation, namely the crossover and mutation operators. Within the GA, mutation is responsible for exploring new states and avoiding local optima, whereas crossover looks to increase the average quality of the population. Coupled with a reduction factor, a GA can quickly find an optimum solution in a short amount of iterations [20].
For the purposes of our research, we use the generalized Bayes factor (GBF) as the measurement for selecting the best solutions provided by the GA. The GBF measure allows for the comparison of competing hypothesis that are presented to it, i.e., the best solutions are provided by the GA based on mutation and crossover of the BN. The GBF of a given explanation x out of all possible explanations x ¯ given the evidence e can be defined in the following way [19]:
G B F ( x ; e ) P ( e | x ) P ( e | x ¯ )
The GBF has a unique property when it measures P ( e | x ¯ ) . Specifically, when it measures two competing explanations x, it has a symmetry in explanation. In other words, as one increases, the other decreases in a symmetric fashion. Given this fact, GBF essentially can be seen as the measure of change in the odds ratio of an explanation. However, as previously seen, BNs are based on multiple pieces of evidence, and thus implements the use of the chain rule. The same applies to GBF whereby a new explanation of y must be considered. Thus, the calculation is updated to [19]:
G B F ( y ; e | x ) P ( e | y , x ) P ( e | y ¯ , x )
and applied to the chain rule whereby e 1 represents the evidence to be true based on multiple pieces of evidence:
G B F ( x ; e 1 , e 2 , , e n ) = G B F ( x ; e 1 ) i = 2 n G B F ( x ; e i | e 1 , e 2 , , e i 1 )

2.4. Benchmarked Classifiers

In order to gauge the performance of the selected method, benchmark classifiers were selected. This allowed for an easy comparison of the superior nature of Bayesian networks for the given target problem. Overall, nine commonly known algorithms that fall under various types of learning were selected, these included classical machine learning approaches, Bayesian methods, ensemble methods, and neural networks. The included algorithms were Tree Augmented Naïve Bayes (TAN) [12], Logistic Regression (LG) [15], Decision Tree (DT) [21], Neural Network (NN) [15], Support Vector Machine (SVM) [22], AdaBoost (ADA) [23], Bagging (BA) [24], Random Forest (RF) [25], and Random Subspace (RSS) [26].

3. Results

3.1. Feature Selection through Sons and Spouses Network Formulation

The first task was to find the variables within the dataset that had the most information upon the target variable depression. This process not only helped to find the most significant variables as antecedents of depression, but it also helped to remove the redundant features from the network. The SS Bayes network was able to find 13 variables that had a direct or indirect relationship with the target variable (i.e., all the children and the parents of these children (spouses); see Figure 2). Furthermore, Table 1 shows all of the direct variables (children) that are having a direct relationship with the target node (see Table 1). According to this result, we can see that the top three variables influencing depression are found to be subjective health conditions, income quartile, and marital status. Based on this supervised machine-learning network, we next tested its predictive performance against the other classification models.

3.2. Benchmarking the Performance of the Obtained Sons and Spouses Model

In order to validate the use of Bayes statistics, we first compared the performance of the SS algorithm against well-known machine learning and ensemble learning techniques. Further, for all of the models, we implemented 10-fold cross-validation in order to validate the predictive results. As depression is a mental disease, it is not as common within the population and hence presents an imbalanced dataset. Thus, following prior research, we analyzed the performance based on accuracy, precision, and the F-Measure metric [13]. However, in medical classification tasks, the F-Measure is usually considered the best predictor. This is because the F-Measure considers both the recall and precision through a weighted average, meaning it punishes the classifier for false positives, an important feature when the dataset is sometimes about life and death situations [22].
Table 2 presents the results of the classification. As can be seen, the logistic classifier outperforms all of the other classifiers in terms of accuracy, with 93.32% accuracy. Compared with SS (93.04%), LOG was only a small margin greater in performance. For the AUC and F-Measure, the SS model outperforms all the other classifiers with 76.97% and 91.81%, respectively. Compared with LOG (AUC: 75.90%; F-Measure: 90.70%) and TAN (AUC: 75.20%; F-Measure: 92.59%), SS showed good performance, and thus justified our selection criteria. It is interesting to note that many of the classifiers, including popular models, such as SVM and NN, struggled to deal with the unbalanced data, and thus returned no F-Measure metric. This means that their prediction was unusable for comparison.

3.3. Identifying the Main Causes of Inducing and Reducing Depression with Genetic Optimization

Attempting to find the underlying causes and suggesting solutions can be a laborious task that is fraught with human error. For this reason, we next turned to AI. Based on a GA, we looked to analyze the probabilistic causes that are strong predictors of depression. As illustrated in Table 3, five models have been built showing a generalized Bayes factor of nearly 15 for each hypothesis: this means that the models provide strong evidence of having depression. Compared to other potential hypotheses found by the GA, these models have a greater probable cause of depression when all factors are considered in the BN. Specifically it presents three males and two females. All are in their 50s, except for one of the females. All are sufferers of arthritis and do not have current economic employment. This also means that they have low income. Diabetes was present in all males, yet none had a diagnosis of hypertension. Interestingly, all five solutions had varying conditions in their marital status and smoking experience. Interestingly, the data show that there is a potential link between being unmarried and smoking; however, this is not consistent. Lastly, three solutions showed very bad subjective health conditions, with one male having good, and another showing not bad. Lastly, as none of the authors are medical professionals, it is hard to analyze the remaining variables, however, the link between glycated hemoglobin and diabetes is well known, and thus higher levels, as seen in two males, may be a reflection of a twofold problem.
As shown, the five solutions represent a hypothetical person. We most note however, the guiding principle of a best solution in this context is not for understanding the most probable cause of depression with 100% accuracy, it is more to show what factors need to be present at one time to show a greatly increased likelihood of depression. Although the GA has modelled five humans with 100% probability of having depression, reality tells us that these models are unattainable, yet useful in their application of identifying depression with just several of the factors. Hence, we show the benefits of this implemented system using a what-if analysis. For demonstration purposes, we explore the first three variables seen in Table 1 and the demographic factors of gender and age. As seen in Figure 3, compared with the original score of 5.48% depression amongst the whole sample, the genetic algorithm was able to find hypothetical patient conditions that would be high-risk for depression. As seen in outcomes 1, 2, and 5, the probability rate of having depression is above the threshold of 65%. In the case of outcomes 3 and 4, a rise in potential depression was seen to reach over 25% and 10%, respectively.

4. Discussion

Until now, research into middle-age depression has been limited to trend analysis [3]. Although the maturity hypothesis has been put forward to explain why this age group is at low risk, numbers indicating the use of depressive drugs within this age group suggest a different reality [9,10]. If we consider the pace of the modern economy and how quickly one’s life can change, for good and bad, analysis of the risk factors for depression in this age group has valid grounds for further research. Additionally, up until now, no research has focused on attempting to provide an early warning risk assessment model, in which practitioners can use to interpret depression in this age group. Thus, this research set out to tackle this problem using Bayesian statistics and AI. As AI techniques have developed in recent times, many options exist that can show strong predictive power. Unfortunately, in the case of methods such as deep learning, one cannot understand the decision-making process within the algorithm. Thus, it is hard to make grounded implications for medical purposes when multiple causes are at play [27].
To help overcome this problem, we employed the use of Bayesian statistics through a supervised network analysis. Specifically, we employed the Sons and Spouses (SS) network model in order to identify the antecedents that were having the greatest effect on depression. As Bayesian networks use the power of machine learning, there is no subjective insights that have led to its conclusions. As seen in Figure 1, the SS model provided seven direct sons and six spouses that were influencing depression. Furthermore, Table 1 showed the most important factors based on their relative binary mutual information, including: subjective health condition, income quartile, marital status, economic status, diagnosed arthritis, gender, and diagnosed diabetes. Next, due to the power of Bayesian networks and their interpretability, we employed a genetic algorithm (GA) in order to suggest the most probable causes of depression within our sample. By doing this, we were able to find variables that were suggested as high-risk causes of depression.
Compared with the 5.48% depression rate seen in the original data, based on the proposed solutions in this study, and the manipulation of five key variables’ values presented in the GA results, we found vastly different results. For example, in the first identified solution, the probability of depression within that hypothetical male example went up to 75.30% probability of being depressed. A similar effect was seen within two of the female solutions, whereby the proposed solutions showed an increase in the probability of depression to 68.95% and 65.40%, respectively. In addition, two other hypothetical male scenarios showed strong effects in terms of their probability of having depression rising to 28.84% and 13.36%, respectively.
The implications of this study can be seen through both an academic and practical viewpoint. First, academically, this research implemented a machine-learning Bayesian network approach, which was also seen in prior work within the medical field and public health [13]. However, prior work using Bayesian networks relies on subjective reasoning; thus, this is usually only accessible for medical practitioners. Therefore, to deal with this problem, we employed an AI-based genetic algorithm to help explore the target problem that did not require complex subjective knowledge. Based on this approach, an early warning risk application can be built that is induced with data-driven results opposed to one’s subjective approach. Practically, this has some key advantages. First, it makes this approach flexible. This is essential for depression research as cultural variations and local differences can be addressed in a quick and effective manner. Further, as this model is based on Bayesian statistics that use conditional probabilities, an evidence-based analysis can be made based on the features that a person has when they are being diagnosed. In other words, practitioners could implement this system with regionally localized data and find a number of solutions that predict the probability of depression. With this, they can then diagnose a patient’s probability of depression through updating the network with simple questions, such as age, gender, marital status, and subjective health conditions, to name a few.

5. Conclusions

This research paper looked to analyze depression within an understudied population: the middle-aged. By doing so, this research’s main goal was to create an approach based on a non-subjective machine learning and AI implemented system. The results showed that this methodology was effective in identifying highly probable causes of depression within middle-aged people. In addition, we were able to isolate the most probably causes of depression. Our results showed that this approach was able to effectively show probabilistic causes without the need for subjective knowledge. In implementing such a system, any practitioner could use this as an early warning risk application, by implementing a simple what-if analysis based on the probabilistic models found by a genetic optimized solution. This is because this method allows for an evidence-based analysis constructed on the features that a person has when they are being diagnosed. Thus, this could potentially help identify depression early on, potentially decreasing the long-term risk of severe depression in a patient.
This research has some flaws that need to be addressed. Firstly, the selection criteria for the data meant large amounts of people were removed due to missing data and variables that were too messy. Thus, there may have been some key samples removed in the preprocessing stage that could have had an effect on the overall outcome. However, as the conclusion in this research is more about the approach over the data implications, per se, this issue can be carefully explored with a more reliable survey. Moreover, questions over the use of the given models can be made. However, as there are endless numbers of models just within the Bayesian family, a decision had to be made on limiting this. Future research can explore the various models that exist in the Bayesian family, and compare and contrast their performance in this given problem.

Author Contributions

Methodology, K.C.L. and F.J.C.; conceptualization, F.J.C. and K.C.L.; software, C.M.K. and C.K.; validation, C.K. and C.M.K.; investigation, F.J.C.; resources, K.C.L., C.K., and C.M.K.; data curation, C.M.K., C.K.; writing—original draft preparation, F.J.C.; writing—review and editing, F.J.C., C.K., and K.C.L.; visualization, F.J.C. and C.K.; supervision, K.C.L.; project administration, K.C.L.; funding acquisition, K.C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by BK21 FOUR research fund of the National Research Foundation of Korea.

Acknowledgments

The authors would like to give thanks to Dong Eun Lee for his valuable contribution in helping to obtain the dataset as well as his invaluable insights into obtaining reliable results.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Variables used for analysis extracted from the Korea National Health and Nutrition Examination Survey (KNHANES) dataset based in South Korea.
Table A1. Variables used for analysis extracted from the Korea National Health and Nutrition Examination Survey (KNHANES) dataset based in South Korea.
IDNameClassification
Sex2Gender1. Male
2. Female
age2Age29~38 (30 s)
39~48 (40 s)
40~58 (50 s)
incmIncome Quartile (individual)1. Low
2. Middle low
3. Middle high
4. High
marri_2Marital status1. Married, live together
2. Married, live separated
3. Separated by death
4. Divorced
5. Unmarried
EC1_1Current economic status1. Yes (Employed)
2. No (Unemployed, economically inactive)
BD1_11Drinking frequency in one year0. N/A
1. I haven’t drunk at all in the last one year.
2. Less than once a month
3. About once a month
4. 2–4 times a month
5. Two or three times a week.
6. More than four times a week
BS1_1Smoking experience1. Less than 5 packs (100 pieces)
2. Five packs (100 pieces) or more
3. Non-smoker
HE_BMIBody mass index1. Under Weigh (less than 18.5)
2. Normal Weight (18.5~22.9)
3. Over-Weight (23~29.9)
4. Mild Obesity (30~34.9)
5. Obese (35~39.9)
6. Severe Obesity (more than 39.9)
D_1_1Subjective health condition1.Very good
2.Good
3.Normal
4.Bad
5.Very bad
DI1_dgDoctor’s diagnosis status for hypertension0. No
1. Yes
DI3_dgDoctor’s diagnosis status for stroke
DI5_dgDoctor’s diagnosis status for myocardial infarction
DI6_dgDoctor’s diagnosis status for angina
DM1_dgDoctor’s diagnosis status for arthritis
DM4_dgDoctor’s diagnosis status for osteoporosis
DJ2_dgDoctor’s diagnosis status for pulmonary tuberculosis
DJ4_dgDoctor’s diagnosis status for asthma
DE1_dgDoctor’s diagnosis status for diabetes
DE2_dgDoctor’s diagnosis status for thyroid
DC4_dgDoctor’s diagnosis status for breast cancer
DC5_dgDoctor’s diagnosis status for cervical cancer
DC7_dgDoctor’s diagnosis status for thyroid cancer
HE_HbA1cGlycated hemoglobin__%
HE_HBhemoglobin__g/dL
HE_hsCRPHigh-sensitivity of C-reactive protein__mg/L
OR1_2Oral examination in the last year0. No
1. Yes
LQ_5EQLEuroQoL: Anxiety/Depression1. I’m not anxious/depressed
2. I’m anxious/depressed

References

  1. Chang, Y.-S.; Fan, C.-T.; Lo, W.-T.; Hung, W.-C.; Yuan, S.-M. Mobile cloud-based depression diagnosis using an ontology and a Bayesian network. Future Gener. Comput. Syst. 2015, 43–44, 87–98. [Google Scholar] [CrossRef]
  2. Andrews, G.; Poulton, R.; Skoog, I. Lifetime risk of depression: Restricted to a minority or waiting for most? Br. J. Psychiatry 2005, 187, 495–496. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Mirowsky, J.; Ross, C.E. Age and Depression. J. Health Soc. Behav. 1992, 33, 187. [Google Scholar] [CrossRef] [PubMed]
  4. Leem, A.Y.; Han, C.H.; Ahn, C.M.; Lee, S.H.; Kim, J.Y.; Chun, E.M.; Yoo, K.H.; Jung, J.Y.; Shim, J.J.; Jung, T.H.; et al. Factors associated with stage of change in smoker in relation to smoking cessation based on the Korean National Health and Nutrition Examination Survey II-V. PLoS ONE 2017, 12, 1–16. [Google Scholar] [CrossRef] [PubMed]
  5. Mirowsky, J. Age and the Gender Gap in Depression. J. Health Soc. Behav. 1996, 37, 362. [Google Scholar] [CrossRef]
  6. Miech, R.A.; Shanahan, M.J. Socioeconomic Status and Depression over the Life Course. J. Health Soc. Behav. 2000, 41, 162. [Google Scholar] [CrossRef]
  7. Gold, P.W.; Machado-Vieira, R.; Pavlatou, M.G. Clinical and biochemical manifestations of depression: Relation to the neurobiology of stress. Neural Plast. 2015, 2015, 7–9. [Google Scholar] [CrossRef]
  8. Cipriani, A.; Geddes, J.R.; Furukawa, T.A.; Barbui, C. Metareview on short-term effectiveness and safety of antidepressants for depression: An evidence-based approach to inform clinical practice. Can. J. Psychiatry 2007, 52, 553–562. [Google Scholar] [CrossRef]
  9. Pratt, L.A.; Brody, D.J.; Gu, Q. Antidepressant Use among Persons Aged 12 and Over: United States, 2011–2014. NCHS Data Brief 2017, 283, 1–8. [Google Scholar]
  10. Kim, M.J.; Kim, N.; Shin, D.; Rhee, S.J.; Hyung Keun Park, C.; Kim, H.; Cho, S.J.; Lee, J.W.; Kim, E.Y.; Yang, B.; et al. The epidemiology of antidepressant use in South Korea: Does short-term antidepressant use affect the relapse and recurrence of depressive episodes? PLoS ONE 2019, 14, 1–15. [Google Scholar] [CrossRef]
  11. How, M.L.; Cheah, S.M.; Chan, Y.J.; Khor, A.C.; Say, E.M.P. Artificial intelligence-enhanced decision support for informing global sustainable development: A human-centric Ai-Thinking approach. Information 2020, 11, 39. [Google Scholar] [CrossRef] [Green Version]
  12. Madden, M.G. On the classification performance of TAN and general Bayesian networks. Knowl. Based Syst. 2009, 22, 489–495. [Google Scholar] [CrossRef] [Green Version]
  13. Kim, C.; Costello, F.J.; Lee, K.C.; Li, Y.; Li, C. Predicting factors affecting adolescent obesity using general bayesian network and what-if analysis. Int. J. Environ. Res. Public Health 2019, 16, 4684. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Oh, J.; Yun, K.; Maoz, U.; Kim, T.S.; Chae, J.H. Identifying depression in the National Health and Nutrition Examination Survey data using a deep learning algorithm. J. Affect. Disord. 2019, 257, 623–631. [Google Scholar] [CrossRef]
  15. Witten, I.H.; Frank, E.; Hall, M.A. Data Mining: Practical Machine Learning Tools and Techniques, 3rd ed.; Morgan Kaufmann Publishers Inc.: Burlington, MA, USA, 2011. [Google Scholar]
  16. Cliff, O.M.; Prokopenko, M.; Fitch, R. Minimising the kullback-leibler divergence for model selection in distributed nonlinear systems. Entropy 2018, 20, 51. [Google Scholar] [CrossRef] [Green Version]
  17. Barber, D. Bayesian Reasoning and Machine Learning. Bayesian Reason. Mach. Learn. 2011. [Google Scholar] [CrossRef] [Green Version]
  18. Prabhakaran, R.; Krishnaprasad, R.; Nanda, M.; Jayanthi, J. System Safety Analysis for Critical System Applications Using Bayesian Networks. Procedia Comput. Sci. 2016, 93, 782–790. [Google Scholar] [CrossRef] [Green Version]
  19. Yuan, C.; Lim, H.; Lu, T.C. Most relevant explanation in bayesian networks. J. Artif. Intell. Res. 2011, 42, 309–352. [Google Scholar] [CrossRef]
  20. Larranaga, P. Structure learning of bayesian networks by genetic algorithms: A performance analysis of control parameters. IEEE Trans. Pattern Anal. Mach. Intell. 1996, 18, 912–926. [Google Scholar] [CrossRef] [Green Version]
  21. Olson, D.L.; Delen, D.; Meng, Y. Comparative analysis of data mining methods for bankruptcy prediction. Decis. Support Syst. 2012, 52, 464–473. [Google Scholar] [CrossRef]
  22. Huang, M.W.; Chen, C.W.; Lin, W.C.; Ke, S.W.; Tsai, C.F. SVM and SVM ensembles in breast cancer prediction. PLoS ONE 2017, 12, 1–14. [Google Scholar] [CrossRef] [PubMed]
  23. Freund, Y.; Schapire, R.E. A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef] [Green Version]
  24. Ballings, M.; van den Poel, D.; Hespeels, N.; Gryp, R. Evaluating multiple classifiers for stock price direction prediction. Expert Syst. Appl. 2015, 42, 7046–7056. [Google Scholar] [CrossRef]
  25. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  26. Ho, T.K. The random subspace method for constructing decision forests. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 832–844. [Google Scholar] [CrossRef] [Green Version]
  27. Wang, L.; Wang, X.; Chen, A.; Jin, X.; Che, H. Prediction of Type 2 Diabetes Risk and Its Effect Evaluation Based on the XGBoost Model. Healthcare 2020, 8, 247. [Google Scholar] [CrossRef]
Figure 1. The main difference between the Tree Augmented Naïve Bayes model and the Sons and Spouses model. As can be seen, the Sons and Spouses model allows for relationships that are not directly linked to the target (father) node.
Figure 1. The main difference between the Tree Augmented Naïve Bayes model and the Sons and Spouses model. As can be seen, the Sons and Spouses model allows for relationships that are not directly linked to the target (father) node.
Healthcare 08 00562 g001
Figure 2. Variables selected with the Sons and Spouses machine learning algorithm and the connected nodes and arcs. As seen, marital status, economic status, income quartile, subjective health condition, gender, diagnosed with arthritis, and diagnosed with diabetes are direct child nodes of the target node depression.
Figure 2. Variables selected with the Sons and Spouses machine learning algorithm and the connected nodes and arcs. As seen, marital status, economic status, income quartile, subjective health condition, gender, diagnosed with arthritis, and diagnosed with diabetes are direct child nodes of the target node depression.
Healthcare 08 00562 g002
Figure 3. Based on the most probable causes found by the genetic optimization seen in Table 3, we explore the first three variables in Table 1, as well as gender and age, as these are non-subjective variables. The results show the difference from the original status of the population sampled (5.48%) and if a person was to have the values seen in these five variables suggested by the genetic algorithm.
Figure 3. Based on the most probable causes found by the genetic optimization seen in Table 3, we explore the first three variables in Table 1, as well as gender and age, as these are non-subjective variables. The results show the difference from the original status of the population sampled (5.48%) and if a person was to have the values seen in these five variables suggested by the genetic algorithm.
Healthcare 08 00562 g003
Table 1. Descriptive statistics from the Sons and Spouses network (target variable: depression).
Table 1. Descriptive statistics from the Sons and Spouses network (target variable: depression).
NodeRelative Binary Mutual InformationPosterior Mean ValueMax Bayes Factor for No DepressionMin Bayes Factor for No DepressionMax Bayes Factor for DepressionMin Bayes Factor for Depression
Subjective Health Condition9.74%1.59Very Good
(1.0647)
Very Bad
(0.6807)
Very Bad
(0.6807)
Very Good
(1.0647)
Income Quartile2.12%1.50High
(1.0266)
Low
(0.9530)
Low
(0.9530)
High
(1.0266)
Marital Status1.84%2.15Married Live Together (1.0127)Divorced
(0.9183)
Divorced
(0.9183)
Married Live Together
(1.0127)
Economic Status1.78%0.77Yes
(1.0150)
No
(0.9537)
No
(0.9537)
Yes
(1.0150)
Diagnosed Arthritis0.56%0.05No
(1.0035)
Yes
(0.9308)
Yes
(0.9308)
No
(1.0035)
Gender0.39%0.46Male
(1.0128)
Female
(0.9893)
Female
(0.9893)
Male
(1.0128)
Diagnosed Diabetes0.190.04No
(1.0018)
Yes
(0.9579)
Yes
(0.9579)
No
(1.0018)
Table 2. Performance of the Sons and Spouses algorithm against common benchmark models.
Table 2. Performance of the Sons and Spouses algorithm against common benchmark models.
SSTANLOGSVMNNDTADABARSSRF
F-Measure
Total91.8191.0090.70N/A90.00N/AN/A90.50N/A90.20
Accuracy
Total93.0492.5993.3993.3291.3193.3293.3293.2693.3292.97
AUC
Total76.9775.2075.9050.0065.4049.8073.3074.2073.9071.30
Precision
Total93.1490.0191.10N/A89.00N/AN/A90.30N/A88.70
Recall
Total90.6092.6093.4093.3091.3093.3093.3090.5093.3093.30
SS (Sons and Spouses), TAN (Tree Aumented Naïve Bayes), LOG (Logarithm), NN (Neural Network), DT (Decision Tree), ADA (Adaboost), BA (Bagging), RSS (Random Subspace), RF (Random Forest).
Table 3. Variables selected based on the genetically induced most probable cause of depression.
Table 3. Variables selected based on the genetically induced most probable cause of depression.
Most Probable Cause of Depression
AgeDiagnosed ArthritisDiagnosed DiabetesDiagnosed HypertensionEconomic StatusGenderGlycated HemoglobinHemoglobin
150sYesYesNoNoMale≤8(3/4)≤10(1/7)
240sYesNoYesNoFemale≤8(3/4)≤16(6/7)
350sYesYesNoNoMale>8(4/4)≤10(1/7)
450sYesYesNoNoMale>8(4/4)≤10(1/7)
550sYesYesNoNoFemale≤6(2/4)≤15(5/7)
Income QuartileMarital StatusOral ExaminationSmoking ExperienceSubjective Health ConditionGeneralized Bayes Factor
1LowUnmarriedYesLess than 5 packsVery Bad14.9738
2LowDivorcedYesNon-smokerVery Bad14.9738
3LowMarried Live SeparatedYesLess than 5 packsNot Bad14.9738
4LowMarried Live SeparatedYesNon-smokerGood14.9738
5LowUnmarriedNoLess than 5 packsVery Bad14.9738
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Costello, F.J.; Kim, C.; Kang, C.M.; Lee, K.C. Identifying High-Risk Factors of Depression in Middle-Aged Persons with a Novel Sons and Spouses Bayesian Network Model. Healthcare 2020, 8, 562. https://doi.org/10.3390/healthcare8040562

AMA Style

Costello FJ, Kim C, Kang CM, Lee KC. Identifying High-Risk Factors of Depression in Middle-Aged Persons with a Novel Sons and Spouses Bayesian Network Model. Healthcare. 2020; 8(4):562. https://doi.org/10.3390/healthcare8040562

Chicago/Turabian Style

Costello, Francis Joseph, Cheong Kim, Chang Min Kang, and Kun Chang Lee. 2020. "Identifying High-Risk Factors of Depression in Middle-Aged Persons with a Novel Sons and Spouses Bayesian Network Model" Healthcare 8, no. 4: 562. https://doi.org/10.3390/healthcare8040562

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop