Next Article in Journal
Steel Surface Defect Diagnostics Using Deep Convolutional Neural Network and Class Activation Map
Next Article in Special Issue
Transfer Learning with Convolutional Neural Networks for Diabetic Retinopathy Image Classification. A Review
Previous Article in Journal
A Multiagent System Prototype of a Tacit Knowledge Management Model to Reduce Labor Incident Resolution Times
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Child’s Target Height Prediction Evolution

by
João Rala Cordeiro
1,*,
Octavian Postolache
1 and
João C. Ferreira
2,3,4,5
1
Instituto de Telecomunicações, IT-IUL, Instituto Universitário de Lisboa, ISCTE-IUL, 1649-026 Lisbon, Portugal
2
INOV INESC Inovação—Instituto de Novas Tecnologias, 1000-029 Lisbon, Portugal
3
Instituto Superior Técnico, 1049-001 Lisbon, Portugal
4
ISTAR-IUL, Instituto Universitário de Lisboa (ISCTE-IUL), 1649-026 Lisbon, Portugal
5
Centro ALGORITMI, University of Minho, 4800-058 Guimarães, Portugal
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(24), 5447; https://doi.org/10.3390/app9245447
Submission received: 22 October 2019 / Revised: 5 December 2019 / Accepted: 8 December 2019 / Published: 12 December 2019
(This article belongs to the Special Issue Medical Artificial Intelligence)

Abstract

:
This study is a contribution for the improvement of healthcare in children and in society generally. This study aims to predict children’s height when they become adults, also known as “target height”, to allow for a better growth assessment and more personalized healthcare. The existing literature describes some existing prediction methods, based on longitudinal population studies and statistical techniques, which with few information resources, are able to produce acceptable results. The challenge of this study is in using a new approach based on machine learning to forecast the target height for children and (eventually) improve the existing height prediction accuracy. The goals of the study were achieved. The extreme gradient boosting regression (XGB) and light gradient boosting machine regression (LightGBM) algorithms achieved considerably better results on the height prediction. The developed model can be usefully applied by pediatricians and other clinical professionals in growth assessment.

1. Introduction

Children’s growth assessments are a very important aspect of health status monitoring, for example, in order to identify development deviations, and diseases or interventions/treatments impacts (i.e., obesity care or stunting). Monitoring the children’s growth ensures children develop in the best possible way [1,2,3,4].
But what exactly is growth and growth study? Growth is a complex interaction between heredity and environment. Growth refers mainly to changes in magnitude, increments in the size of organs, increases in the tissue thickness, or changes in the size of individuals as a whole [5].
The study of growth is focused on monitoring individuals that have not reached the “mature” stage, where full growth potential should be [6].
Knowing the importance of growth assessment several growth standards have been developed in the past decades, for instance (the most popular) NCHS/WHO, CDC 2000, Tanner standards, or Harvard standards [1,7].
In 1978, the NCHS—National Center for Health Statistics growth reference (the most-used reference until then) was replaced after a recommendation of The World Health Assembly in 1994, which considered the reference did not represent early childhood growth [8].
The NCHS reference was based on a longitudinal study of children from a single community in the USA, which were measured every three months [9,10].
The WHO-World Health Organization Standards were developed using data collected on the WHO Multicentre Growth Reference Study (combines a longitudinal study with a cross-sectional study) between 1997 and 2003. Data include about 8500 children from South America (Brazil), Africa (Ghana), Asia (India, Oman), Europe (Norway), and North America (USA) [10,11,12].
A global survey was conducted in 178 countries to understand the use and the recognized importance of the new standards. The importance of this work was recognized after huge scrutiny and a wide adoption for more than 125 countries in just five years after the work was released (April 2006). Child growth charts are currently one of the most used tools for assessing children’s wellbeing, and allow the physiological needs for growth to be followed and establishing whether the expected development of the child would be met [1,11].
Still, the main reason for non-WHO Standards adoption was the preference for local (or national) references, considered more appropriate [1].
Studies compared WHO growth standards with local/national growth references, considering different geographic areas with genetic and environment diversity. These studies reveal that the local references are more faithfully than the WHO standards [3,13].
The WHO standards do not consider ethnicity, socio-economic status, feeding mode, or other factors. They consider a healthy and “common” population and those not appropriated to individuals that belong to a risk population segment, such as the poor, belonging to an ethnic minority, parents, the unemployed, or those with some particular health issue with a high influence on the child [1,4].
As mentioned above the current standards present some limitations. These limitations can be overcome using personalized medicine, a concept that it’s becoming more popular every day.
Personalized medicine has achieved better results than a “standard” medicine, including in children, and with the evolution of technology, it is expected to bring true revolution on patient care on the near future. This new approach to health care can bring considerable advances in disease prediction, preventive medicine, diagnostics, and treatments [14,15].
But what is personalized medicine? The US National Institutes of Health, for instance, describes it as, “the science of individualized prevention and therapy”. It can simply be described as a way of synthesizing an individual’s health history, including information family history, clinical data, genetics, or the environment in a way to benefit individuals’ health and wellbeing [15,16].
The current study aims to contribute to improving growth assessments on children, creating a new tool that combines the main concepts above, namely expected growth and personalized medicine on children. Having a more accurate target height will allow pediatricians, and other health professionals, to apply growth standards in a more accurate way and provide a better assessment on children’s development.
Until now, the growth standards have been applied by professionals by considering the actual “state” (growth measure) for the child and the existing methods described in the state-of-the-art, which offer a less accurate prediction. We have created a new tool that offers better outcomes and can be adapted to specific populations, by considering their specificities.
Artificial Intelligence and machine learning are contributing, through many studies and projects all over the world, to more personalized medicine, thereby delivering better outcomes. We believe our study is a step towards the medicine of the future.
For achieving the proposed goals, we have decided to research and analyze the existing methods, through state-of-the-art research, followed by the application of the CRISP-DM, an open standard methodology for data mining. This methodology includes a process that goes from the business understanding to the solution deployment, which allows our work to deliver value (a solution) to society.
Finally, we have documented our results and conclusions, where we also describe potential future work.

2. State-of-the-Art

As discussed in the previous chapter, the height assessment on a child assumes critical importance when improving child health and wellbeing.
Prediction for adult height can be used both for diagnosis or following treatments and provides a valuable tool in pediatrics. For instance, it can help detect growth hormone deficiency with hGH and follow treatment performance. In diagnosis, the height of the child can be compared with the so-called “target height”, by considering the parent’s height and child’s gender [5].
There are two known methods for assessing the “target height”.
Horace Grey has defined the following model to forecast the height (when adult) for a child [17]:
F e m a l e   h e i g h t   ( i n c h e s ) = + f a t h e r   h e i g h t   × 12 13 + m o t h e r   h e i g h t   2
M a l e   h e i g h t   ( i n c h e s ) = + f a t h e r   h e i g h t   + m o t h e r   h e i g h t × 12 13   2
Other common used model is the Mid-Parental Method (described by Tanner) [18,19]:
F e m a l e   h e i g h t   ( i n c h e s ) = + f a t h e r   h e i g h t   + m o t h e r   h e i g h t   2 2.5
M a l e   h e i g h t   ( i n c h e s ) = + f a t h e r   h e i g h t   + m o t h e r   h e i g h t   2 + 2.5
There is a lot of information regarding the applicability of the methods, but information about the studies’ bases and characteristics are scarce, probably because they have been performed some decades ago. Still, it was possible to comprehend that both methods are based on statistics approaches where a simple regression method is applied to the population and the “optimal” formula is defined. In a generic way, these approaches try to find the mathematical formula that best fits the general characteristics of the population.
Horace grey approach reports to a study published in 1948 and is based on 53 collected cases. This method introduces a different weight on the mother’s or father’s height when calculating the output prediction. It emphasizes that a mother and father’s influence is different for different child’s genders [17].
The method described by Tanner was introduced in 1970 and has been a standard procedure for assessing children’s growth since. It was not possible to find a description of the dataset used in the study, but it was still possible to understand that this method approaches the height of parents on the same proportion, given the emphasis on children’s gender, which is critical to defining height [20].
As we observed, the methods did not introduce personalization to the individual level, besides the influence of gender, which is one of the advantages that machine learning can introduce in these areas of study.
There are other studies besides parent’s height can use bone age, actual age, or child’s current height. All these studies were based on longitudinal population projects [21,22].
These projects, also called cohort studies, involve observing and measuring population variables (the same individuals–cohort) for long periods, which can involve years or decades.
In the current study, we will only focus on information gathered at childbirth, and as far as we know, the two methods described before are the only existing “tools” for predicting a child’s height when adults. They are quite popular and it is easy to find applications of these methods in several child development areas.
We have also found research on applications of machine learning methods that address “target” height prediction. We did not find studies addressing this subject.
Knowing that this subject was never approached on this perspective represents, in our view, an opportunity, and was one of the main reasons for proceeding with this work. This study develops a new method in forecasting children’s target height using a completely different approach based on machine learning tools.

3. Methodology

To pursuit the goal defined for this study we have considered as possible methodologies CRISP-DM—CRoss Industry Standard Process for Data Mining, DMME—Data Mining Methodology for Engineering Applications or SEMMA—Sample, Explore, Modify, Model, and Assess, the most common approaches on data mining projects [23,24].
CRISP-DM it’s an open standard, developed by IBM in cooperation with other industrial companies. It has the particularly of not specifying the data acquisition phase as it assumes the data has been already collected. It’s a popular methodology that has proven to increase success on data mining challenges, including in the medical area [25,26,27].
DMME it’s an extension of the CRISP-DM methodology that includes the process of data acquisition [23].
SEMMA was developed by the SAS Institute and is more commonly used by SAS tools users as it is integrated into SAS tools such as Enterprise Miner [23].
On our study the data to use was already been collected and those we will not need to perform this task. Considering this particularity, we have decided to use CRISP-DM methodology that is clearly the most popular approach among the community for the current type of challenge that we are addressing.
CRISP-DM model, represented on Figure 1, is a hierarchical and cyclic process that breaks data mining process into the following phases (resume description) [28,29]:
  • Business Understanding: Determination of the goal.
  • Data Understanding: Become familiar with data and identify data quality issues.
  • Data Preparation: Also known as pre-processing and will have as output the final dataset.
  • Modeling: where a model that represents the existing knowledge it’s built through the use of several techniques. The model can be calibrated in a way to achieve a more accurate prediction of the target value defined as the goal on the 1st phase.
  • Evaluation: Evaluate the model performance and utility. This moment defines if a new CRISP-DM iteration will be needed or if to continue to the final phase.
  • Deployment: Deployment of the solution.

4. Work Methodology Proposed

4.1. Business Understanding

The main goal for this study is to contribute to a better growth assessment creating a new method to forecast a child’s height when they are an adult, also known as the target height. The model output will be personalized to the child (person-centric) and is expected to deliver a more accurate prediction when compared with the existing methods, already described in the previous chapters.
We began search for a “better” tool that provided a way to compare, for instance, a metric. We extensively researched this subject and found several methods for measure/comparing models’ performance, such as: Accuracy, confusion matrix, and several associate measures such as sensitivity, specificity, precision, recall or f-measure (very useful when facing classification problems with imbalanced classes), MSE—mean square error, ROC—Receiver Operating Characteristic Curve, AUC—area under the roc curve, RMSE—root mean squared error, MAE—mean absolute error, R²—R Squared, Log-loss, Cohen’s kappa, Sum squared error, Silhouette coefficient, Euclidean distance, or Manhattan distance [30,31,32,33].
The several methods/measures present different approaches for different problems (regression, classification, and clustering) with advantages and drawbacks. In this paper we will not describe or proceed with an extensive analysis on this subject, since other works, for instance, the ones mentioned as references, have already provided excellent analysis. Accuracy is a simple method (easy to compute and low complexity) able to discriminate and select the best (optimal) solution. It has limitations when dealing with classes and can result in sub-optimal solutions when classes are not balanced.
Our problem is not a classification problem, but a regression problem, and our goal is to provide better performance on our predictions. Considering our challenge characteristics, we believe the measure that best suits our goal is accuracy.
The performance results for the state-of-the-art height prediction when adult models were measured are presented in Table 1. The best accuracy is obtained by the Mid-Parental Method which presents an accuracy of 59.53%.

4.2. Data Understanding

For the proposed study, was decided to use a dataset based on the famous study 1885 of Francis Galton (an English statistician that founded concepts such as correlation, quartile, percentile and regression) that was used to analyse the relations between child’s heights when adult to their parent’s heights.
The dataset is based on the observation of 928 children and the corresponding parents (205 pairs), where several variables have been recorded, such as father and mother height or child gender and adult height [34].
Galton’s study made an important contribution in understanding multi-variable relationships and introduced the concept of regression analysis. It has proved that parents’ height were two Gaussian distributions (also known as normal), one for the mother and one for the father, that, when joined together, could be described as bivariate normal. It also has proved that adult children’s height could be described by a Gaussian distribution [35].
The dataset used to this experiment is composed of 898 cases and the measuring unit is the inch. The dataset file includes the following information:
  • Family: Family identification, one for each family.
  • Father: Father’s height (in inches).
  • Mother: Mother’s height (in inches)
  • Sex: Child’s gender: F or M.
  • Height: the child’s height as an adult (in inches).
  • Nkids: the number of adult children in the family, or, at least, the number whose heights Galton recorded.
Figure 2 plots the distribution of heights on mother, fathers, child’s (regardless of gender), female child’s and male child’s. The horizontal axis represents the height, in inches, and on the vertical axis we can find the distribution density or another way the probability of a given value on the horizontal axis.
Analyzing the graphic provides the following conclusions:
  • Childs gender it’s a feature with high impact on height;
  • All “segments” heights (mothers, fathers, child’s, female child’s and male child’s) correspond to Gaussian distributions;
  • Female child’s and mother’s height distributions are similar. The same happens with the male child’s and parent’s heights.
The following charts allow visualizing the relation between father’s heights, represented on the horizontal axis, and child’s heights on the vertical axis. On the first chart (Figure 3) we used all dataset which made it possible to understand that the relation had two nuclei. This means that there are two “focus” on the relationship between father’s and child’s heights.
To understand these two different densities, we decide to split the dataset on male and female child’s what can be observed in Figure 4 and Figure 5, respectively.
Figure 4 and Figure 5 clearly identity a relation of father’s height with child’s height that is different when considering child’s gender. For the same father height a child will be taller when male and shorter when female.
The same exercise was made using the mother’s height (rather than father’s) and we obtain similar shapes but show different influence dimensions. This analysis outputs a high degree of influence caused by the child’s gender as well as different influences from mother and father.
Understanding the influence on child height were different on father and mother we have performed a correlation analysis using person correlation measure.
Several approaches were implemented, such as analyzing children’s height by gender or parents, according to height distribution quarters. The results can be found in Table 2.
Based on the obtained results we considered the following conclusions:
  • Parent’s height correlation with the child’s height differs between different child’s genders. This is evident for example when considering (all) father influence on female (0.459) and male (0.391) children. We observe a bigger influence on females.
  • Father height and mother height have different weights/influences on the child’s height. This can be observed for example on females’ child’s height correlated with father height (0.495) and mother height (0.314).
  • The parent’s height influence on child’s height it’s not constant, it varies among the height range.
Based on the study dataset, a linear regression analysis was performed. We have taken several approaches, such as considering on the horizontal axis the sum of parent heights (Figure 6), the mother height or father height individually and considering for vertical axis child’s height (regardless of gender), or considering the child’s gender (Figure 7 for female gender).
We found a multilinear regression, where the child’s height dependended on the individual’s variable mother and father height. We also observed a high variance on the relations as it could easily be observed on the plots.
This phase also includes verifying the dataset for incomplete or missing data, errors, duplicates or other data issues. In Table 3, it is possible to observe one of the tools used in our verification for checking information, such as the number of rows, that every row has values to every feature, or number of unique values for each feature.
There was no need for data cleaning or other tasks on the dataset. This was expected, since our dataset was previously used for other scientific purposes.

4.3. Data Preparation

The data preparation phase includes several tasks to achieve the final dataset to process, including feature selection, which is crucial to improving the models’ outcomes.
Feature selection can be performed using several techniques. These techniques can be segmented, a first step, supervised, unsupervised, and semi-supervised approaches. Inside the supervised segment, the category applied to our case, we can find different categories namely filter methods, wrapper methods, embedded methods, among others. Example of feature selection methods that fit these categories, include information gain, Relief, Fisher score, gain ratio, and Gini ind47ex [36,37,38].
According to previous data analysis, the “family” feature defines the family to which the child belongs. The machine learning models can use the “family” feature to extract the brothers and sister heights when adults use this information to forecast their child’s height. In our scenario we did not want to use brothers’ and sisters’ heights, since in real cases, they are unlikely to be adults yet, and consequently we will not have this information. Therefore, we decided to exclude the family column from the dataset and eliminate family relations.
We have also excluded the “nkids” feature which refers to the number of brothers of an individual, since machine learning can also use this information. This information is not valid since, for instance, at the first brother born there is no information about how many brothers he will have in the future.
Considering feature selection, we also analysed the remaining features. Through the analysis performed in the previous chapter, we can verify that all of them seem to present a degree of influence on our prediction goal, the target height. Also, the algorithms that will be used on this study wrap feature selections techniques and those can address feature selection automatically. Considering these facts, we have decided to include all the remaining features in the modelling phase.
Finally, the gender feature was converted into a binary feature defining whether the child “is male?” so it can be used on the generality of machine learning models.
For applying machine learning supervised methods is necessary to divide the entire dataset on training and test sets. Usually a ratio of 80/20 or 70/30 is applied. We decided to use the first option. Once divided it’s important to assure that both (sub) datasets maintain the original dataset shape, which was assured as it can be observed in Table 4 and Table 5.

4.4. Modelling

To resolve the challenge, there was the need to choose which machine learning “tools” to use. For this study the criteria adopted was to choose the most popular machine learning algorithms, namely: Linear regression, SVR—Support Vector Regression (several kernels), Lasso, Elastic Net, Ridge Regression, K-Neighbors Regression, Bayesian Regression, AdaBoost Regression, SGD—Stochastic Gradient Descent Regression, Decision Tree Regression, XGB—Extreme Gradient Boosting Regression and LGBM—Light Gradient Boosting Machine Regression [39,40].
To pursue our goals, we start by applying all the previously described algorithms to our dataset, according to the partition described in the previous chapter.
For this study was used, as a development environment, Jupyter notebooks with Python 3.6 and the algorithms libraries Sklearn, XGBoost and Lightgbm.
Regarding the algorithm’s “configuration”, we have decided to start our modelling phase using the default (hyper) parameters which are defined by the libraries used on the study, mention above. The results obtained can be visualized in Table 6.
Using this approach, we have achieved the goals outlined for the study, and several algorithms (such as SVR Linear, Ridge, Bayesian or Linear regression) were able to overcome the accuracy in the state-of-the-art models. Most of these algorithms achieved a small improvement, but XGB and LightGBM achieved a considerable improvement.
Based on the present approach, SVR Poly and SGD were difficult to converge and were not able to produce a valid output.
Although, the goals of the study, forecasting target height, and where possible, overcome the existing models, Horace Grey and Mid Parental method accuracy were achieved. It would be important to achieve the best accuracy possible, as the output is a valuable contribution to the growth assessment.
When using machine learning algorithms, it is possible to improve results using techniques like scaling and hyper-parameters tuning.
Regarding the scaling technique, our study has tried several approaches, namely techniques, such as StandardScaler, MinMaxScaler, MaxAbsScaler, RobustScaler and Scale. The best results were obtained with the MaxAbsScaler, which rescales data in a way that all feature values are in the range [0, 1].
Table 7 highlights the technique impact obtain on accuracy (column impact) when using MaxAbsScaler. The value “No change” includes very small challenges, mainly justified by the algorithms common variations.
Although, this technique improved the accuracy of some algorithms, an improvement was not made on the height prediction parameter.
Note, in this case, SVR Poly and SGD were able to converge and present a valid output, as the new scale overcomes the previous verified convergence issue with these two algorithms.
For hyperparameters tuning we considered two possible strategies: Grid search and Random Search. As Grid search requires a lot of resources and time to compute all possible algorithms and parameter dimensional space and those that adopted the random search strategy performed several iterations.
The applicability of this technique to the study dataset can be observed in Table 8.
Hyper-parameter tuning has improved the accuracy of many algorithms, and most importantly, has achieved progress on the study goal, increasing children’s “target” height prediction accuracy.
Besides the importance of the performance output, it is also important to compare the complexities and costs regarding the algorithm’s applicability. We have decided to perform an analysis of the computation resources consumption required for each of the methods used above.
Since CPU consumption can present some variability, we decided to analyse CPU time mean and standard deviation by considering 1000 executions (Table 9).
The computer supporting this study has the following characteristics:
  • Processor: Intel Core i7-3770 3.40GHz.
  • RAM: 8 GB.
  • Operating System: Windows 10, 64 bits, x64-based processor.

4.5. Evaluation

Through the output of the data mining process, we have obtainead several models that can forecast the child “target” height. Ten of those models have overcome the results obtained on the state-of-the-art reference model, Mid-Parental Method, and two of the models produced, XGB and LightGBM, with significant improvements. These two models have achieved the second goal defined in the business understating phase, improving the existing prediction accuracy.
Considering these results, it was possible to proceed to the next phase.
Regarding the computation resources consumption between different methods, we can verify that memory consumption presents a low variance (15%) with a minimum of 140 Mb and a maximum of 163 Mb. In our option, there was not a significant difference, even when dealing with large files, considering the availability of resources nowadays.
We can divide algorithms into four classes based on CPU time. The algorithms on the milliseconds range (i.e., Linear, SGD or KNeighbors), those on the tens of milliseconds range (i.e., AdaBoost or 26.6 ms), the class for the hundreds of milliseconds (i.e., SVR RBF or LightGBM), and finally the SVR Poly on the long-time processing class.
The CPU consumption is directly proportional to the algorithm’s complexity and it was expected to have a low processing consumption for a “simple” linear regression and bigger processing needs for support vector regressions. Remembering that linear regression method builds a linear equation (i.e., y = β0 + β1 × 1) that fits the best linear relationship between a set of input data to predict the output value, while SVR algorithms try to fit data using a hyper-plane where the support vectors threshold can assume different shapes (including nonlinear). These can vary according to the kernel defined as parameter [41,42].
Focusing on the algorithms that present the best improvements, XGB and LightGBM, we can observe a huge difference on the CPU consumption, where the last requests quadruple the time consumed by the first one.
Both algorithms are gradient boosted tree-based ensemble methods. XGB grows trees horizontally, as most of the tree base algorithms, while LightGBM grows tree vertically. That means that LightGBM grows from the leaf with greater loss (leaf-wise), while the other considers the level (level wise), what can reduce loss and result in better accuracy [43,44,45].
LightGBM was developed after XGB and uses a novel sampling method called Gradient-based One-Side Sampling. It was developed with the purposes of achieving faster training, lower memory needs and better accuracy. Our results can confirm the accuracy overcome but we observe the inverse for the resources consumption. That can easily be explained as LightGBM was developed for large datasets, which was not the case in our study.

4.6. Deployment

After achieving the goals defined in this study, it is important to deliver an easy way for clinical professionals, particularly undertaking growth assessments, to benefit from this evolution. For that we have developed a basic interface (Figure 8), connected by a web service to the model, that allows us to quickly introduce the inputs and generate the output prediction. This interface, model and service can be easily deployed to an existing IT infrastructure on hospitals, clinics, or other care facilities.

5. Results

In this study, we used several machine learning algorithms as a way to forecast a child’s height prediction when they turn into adults and we have also been successful in improving accuracy when compared with the currently existing models.
The models that presented the best accuracy were XGB and mainly LightGBM with considerable improvements when compared with the best-known existing model, the Mid-Parental method.
It is important to understand how LightGBM accuracy has been able to surpass the accuracy of the Mid-Parental method (the best model on state-of-the-art). Also, we believe that understanding the fundamentals of the prediction will make the model outputs more trustworthy for its users.
In analyzing the model outputs, we first decide to compare the outputs (vertical axis) against the sum of parents’ height (horizontal axis) using Figure 9 where we can find LightGBM (red) and Mid-Parental methods (green) results. In Figure 10 we also add the real values for the child height (blue).
In the previous figures (Figure 9 and Figure 10), was possible to verify that the Mid-Parental predicts the same child height, for the same gender, when considering the same parent’s height sum, while LightGBM model presents variations. When comparing both models LightGBM and Mid-Parental outputs with the real values it’s possible to understand that the first model is able to achieve predictions that are closer to the real values.
Focusing on gender influence on predictions, we verify that both models (LightGBM and Mid-Parental) were able to influence of gender the child height prediction.
The Mid-Parental considers a constant difference of 5 inches between male and female children. In Figure 9, we can easily observe two parallel green lines formed by several green points, one for each gender, with a 5 inches difference.
Figure 11 shows the influence of gender on LightGBM in the dataset. The cadet blue colour presents male child’s (variable is_male = true) and dark orchid colour corresponds to female child’s (variable is_male = false). The influence of gender it’s not a constant of 5 inches, but otherwise its centre around 5 inches presents some small variability. The variability is about 0.5 inches on female (dark orchid area) and 0.7 on male (cadet blue area). The variability is not related to any trend on child height, but instead with the parents’ individual heights influence.
Previously, in the data exploration and visualization chapter, we verified that the influence of mother and father is different for different genders and for different mother and father heights. In Figure 9, Observing figure 9 we can verify that our model (LightGBM) it’s able to capture and embed that variations that are not linear and where same parents’ heights produce different outputs.
We have verified average fathers’ and mothers’ heights’ influence our model, according to children’s genders, as presented in Figure 12. Importantly, fathers’ and mothers’ heights influence is different for different children’s genders. This fact confirms the results obtained on data exploration and visualization phase where we found, for instance, that the correlation of father height and female children was 0.459 and 0.391 for male children.
In Figure 12, it is possible to observe the average influence of parent’s height. In Figure 13, we observe the detail influence of mother and father height in the dataset records for a female child.
The graphic below represents all female children’s order from tallest to lowest. The red colour represents a positive influence and the blue colour a negative influence on the child’s height.
The most important fact to extract from the previous figure is that parents’ height influences their children’s height differently. The tallest female child, above the mean, mainly had a positive influence of fathers height, while the same is not true for mothers’ height, where we can see a lot of fluctuations.
By observing the oscillations, it is possible to see that the relationship is not linear, as defined by the Mid-Parental method, and that presents variance, mainly on the mother case. This variance was already observed in the data analysis phase, where we observed that the correlation between mother and female children have fluctuations.
Following the figure above (Figure 13) we can verify that our model is able to incorporate the different influences of the parent’s heights on different situations. This leads to a model with better accuracy.
For the evaluation described above, we have used some resources from SHAP library, including some graphics generated by the tool. SHAP (SHapley Additive exPlanations) is a framework that combines features contributions and game theory to explain the machine learning predictions [46,47].

6. Conclusions

In our study, we assessed the ability to predict children’s height when they became adults, and we challenges ourselves to overcome the existing predictive models using machine learning techniques. Based on the developed work, our research group was able to find better approaches to forecast children’s height when they become adults. For most of the approaches, the improvements were minor but with the XGB and LightGBM algorithms were possible for achieving a considerable improvement on accuracy.
By analyzing the LightGBM model, we understood that our model was able to incorporate the different degree of influence of fathers’ and mothers’ heights in different stages, and on different children’s genders, what has contributed to improving the results obtained when compared with the previously existing models (Horace Grey and Mid-Parental method).
At the end of the study, we found that a step forward toward more personalized medicine, creating a new “tool” where growth assessments can be made using as reference the “real” potential growth for analysis in children.
The final model still has some variation when compared to real values. We believe it can be reduced by adding other signification variables to the model, such as grandparent’s height, existing pathologies, economical level and alimentation access/quality. This will be probably approached in future work.
In future work, we would like to develop three new studies:
  • Use the procedures defined in this article for a current population, using a bigger dataset. Contact with Portuguese public institutions/services is being made to explore the current Portuguese height records.
  • Develop a new approach where the height of a child is forecast from birth to adulthood, monitoring the full development by year, and where data from previous years will probably take prominence.
  • Use the methodology defined for this experience using a dataset with more features. We would like to add to the existing features variables, such as grandparent’s height, existing pathologies, economical level and alimentation access/quality. We are already developing contacts with Geração XXI project (Portugal) that has created a cohort study with a considerable number of child’s (about 8000), that can help to deepen this study.
We believe that other features can reduce the variability on the model, and consequently increase accuracy.

Author Contributions

J.R.C. is a PhD student that performed all development work. O.P. is thesis supervisor that performs final revision. J.C.F organized work in the computer science subject. All authors have read and agree to the published version of the manuscript

Funding

This work has been partially supported by Fundação para a Ciência e Tecnologia Project UID/EEA/50008/2019 and Instituto de Telecomunicações.

Acknowledgments

João C Ferreira receives support from Portuguese National funds through FITEC - Programa Interface, with reference CIT “INOV - INESC INOVAÇÃO - Financiamento Base.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. De Onis, M.; Onyango, A.; Borghi, E.; Siyam, A.; Blössner, M.; Lutter, C. Worldwide implementation of the WHO Child Growth Standards. Public Health Nutr. 2012, 15, 1603–1610. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Victora, C.G.; Adair, L.; Fall, C.; Hallal, P.C.; Martorell, R.; Richter, L. Maternal and child undernutrition. Consequences for adult health and human capital. Lancet 2008, 371, 340–357. [Google Scholar] [CrossRef] [Green Version]
  3. Ziegler, E.E.; Nelson, S.E. The WHO growth standards: Strengths and limitations. Curr. Opin. Clin. Nutr. Metab. Care 2012, 15, 298–302. [Google Scholar] [CrossRef] [PubMed]
  4. Tanner, J.M. Growth as a Mirror of the Condition of Society: Secular Trends and Class Distinctions. Pediatr. Int. 1987, 29, 96–103. [Google Scholar] [CrossRef] [PubMed]
  5. Tanner, J.M. Normal growth and techniques of growth assessment. Clin. Endocrinol. Metab. 1986, 15, 411–451. [Google Scholar] [CrossRef]
  6. Garn, S.M. Physical growth and development. Am. J. Phys. Anthropol. 1952, 10, 169–192. [Google Scholar] [CrossRef] [PubMed]
  7. Bilukha, O.; Talley, L.; Howard, C. Impact of New WHO Growth Standards on the Prevalence of Acute Malnutrition and Operations of Feeding Programs-Darfur, Sudan. JAMA J. Am. Med Assoc. 2009, 302, 484–485. [Google Scholar]
  8. Seal, A.; Kerac, M. Operational implications of using 2006 World Health Organization growth standards in nutrition programmes: Secondary data analysis. BMJ 2007, 334, 733. [Google Scholar] [CrossRef] [Green Version]
  9. World Health Organization. WHO Child Growth Standards: Length/Height-for-Age, Weight-for-Age, Weight-for-Length, Weight-for-Height and Body Mass Index-for-Age: Methods and Development; WHO: Geneva, Switzerland, 2006. [Google Scholar]
  10. Wang, Y.; Moreno, L.A.; Caballero, B.; Cole, T.J. Limitations of the Current World Health Organization Growth References for Children and Adolescents. Food Nutr. Bull. 2007, 27, 175–188. [Google Scholar] [CrossRef] [Green Version]
  11. Borghi, E.; de Onis, M.; Garza, C.; Van den Broeck, J.; Frongillo, E.A.; Grummer-Strawn, L. Construction of the World Health Organization child growth standards: Selection of methods for attained growth curves. Stat. Med. 2006, 25, 247–265. [Google Scholar] [CrossRef]
  12. World Health Organization. The WHO Multicentre Growth Reference Study (MGRS); WHO: Geneva, Switzerland, 2003. [Google Scholar]
  13. Cameron, N.; Hawley, N.L. Should the UK use WHO growth charts? Paediatr. Child Health 2010, 20, 151–156. [Google Scholar] [CrossRef]
  14. Pijnenburg, M.W.; Szefler, S. Personalized medicine in children with asthma. Paediatr. Respir. Rev. 2015, 16, 101–107. [Google Scholar] [CrossRef] [PubMed]
  15. Cornetta, K.; Brown, C.G. Balancing personalized medicine and personalized care. Acad. Med. 2013, 88, 309–313. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Chan, I.S.; Ginsburg, G.S. Personalized Medicine: Progress and Promise. Annu. Rev. Genom. Hum. Genet. 2011, 12, 217–244. [Google Scholar] [CrossRef]
  17. Gray, H. Prediction of Adult Stature. Child Dev. 1948, 19, 167–175. [Google Scholar] [CrossRef]
  18. Tanner, J.M.; Goldstein, H.; Whitehouse, R.H. Standards for children’s height at ages 2–9 years allowing for heights of parents. Arch. Dis. Child. 1970, 45, 755–762. [Google Scholar] [CrossRef] [Green Version]
  19. Wright, C.M.; Cheetham, T.D. The strengths and limitations of parental heights as a predictor of attained height. Arch. Dis. Child. 1999, 81, 257–260. [Google Scholar] [CrossRef] [Green Version]
  20. Cole, T.J.; Wright, C.M. A chart to predict adult height from a child’s current height. Ann. Hum. Biol. 2011, 38, 662–668. [Google Scholar] [CrossRef]
  21. Tanner, J.M.; Whitehouse, R.H.; Marshall, W.A.; Carter, B.S. Prediction of adult height from height, bone age, and occurrence of menarche, at ages 4 to 16 with allowance for midparent height. Arch. Dis. Child. 1975, 50, 14–26. [Google Scholar] [CrossRef] [Green Version]
  22. Tanner, J.M.; Landt, K.W.; Cameron, N.; Carter, B.S.; Patel, J. Prediction of adult height from height and bone age in childhood: A new system of equations (TW Mark II) based on a sample including very tall and very short children. Arch. Dis. Child. 1983, 58, 767–776. [Google Scholar] [CrossRef] [Green Version]
  23. Huber, S.; Wiemer, H.; Schneider, D.; Ihlenfeldt, S. DMME: Data Mining Methodology for Engineering Applications-A Holistic Extension to the CRISP-DM Model. Procedia CIRP 2018, 79, 403–408. [Google Scholar] [CrossRef]
  24. Mariscal, G.; Marban, O.; Fernandez, C. A survey of data mining and knowledge discovery process models and methodologies. Knowl. Eng. Rev. 2010, 25, 137–166. [Google Scholar] [CrossRef]
  25. Moro, S.; Laureano, R.; Cortez, P. Using Data Mining for Bank Direct Marketing: An Application of the CRISP-DM Methodology. In Proceedings of the European Simulation and Modelling Conference, Guimaraes, Portugal, 24–26 October 2011. [Google Scholar]
  26. Chapman, P.; Clinton, J.; Kerber, R.; Khabaza, T.; Reinartz, T.; Shearer, C.; Wirth, R. CRISP-DM 1.0: Step-by-Step Data Mining Guide; SPSS: Copenhagen, Denmark, 2000. [Google Scholar]
  27. Caetano, N.; Cortez, P.; Laureano, R.M. Using Data Mining for Prediction of Hospital Length of Stay: An Application of the CRISP-DM Methodology; Lecture Notes in Business Information Processing; Springer: Cham, Switzerland, 2015; pp. 149–166. [Google Scholar]
  28. CRISP-DM by Smart Vision Europe-DM Methodology. Available online: http://crisp-dm.eu/home/crisp-dm-methodology/ (accessed on 10 September 2019).
  29. Pllana, S.; Janciak, I.; Brezany, P.; Wohrer, A. A Survey of the State of the Art in Data Mining and Integration Query Languages. In Proceedings of the 14th International Conference on Network-Based Information Systems, Tirana, Albania, 7–9 September 2011; pp. 341–348. [Google Scholar]
  30. Hossin, M.; Sulaiman, M.N. A Review on Evaluation Metrics for Data Classification Evaluations. Int. J. Data Min. Knowl. Manag. Process 2015, 5, 1–11. [Google Scholar]
  31. Flach, P. The Geometry of ROC Space: Understanding Machine Learning Metrics through ROC Isometrics. In Proceedings of the ICML03, Washington, DC, USA, 21–24 August 2003; pp. 194–201. [Google Scholar]
  32. Sokolova, M.; Japkowicz, N.; Szpakowicz, S. Beyond Accuracy, F-Score and ROC: A Family of Discriminant Measures for Performance Evaluation. In AI 2006: Advances in Artificial Intelligence; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2006; pp. 1015–1021. [Google Scholar]
  33. Namlı, E.; Erdal, H.; Erdal, H.I. Artificial Intelligence-Based Prediction Models for Energy Performance of Residential Buildings. In Recycling and Reuse Approaches for Better Sustainability; Springer International Publishing: Cham, Switzerland, 2019; pp. 141–149. [Google Scholar]
  34. Wachsmuth, A.; Wilkinson, L.; Dallal, G.E. Galton’s Bend: A Previously Undiscovered Nonlinearity in Galton’s Family Stature Regression Data. Am. Stat. 2003, 57, 190–192. [Google Scholar] [CrossRef]
  35. Frees, E.W.; Valdez, E.A. Understanding Relationships Using Copulas. N. Am. Actuar. J. 1998, 2, 1–25. [Google Scholar] [CrossRef]
  36. Chadha, A.N.; Zaveri, M.A.; Sarvaiya, J.N. Optimal feature extraction and selection techniques for speech processing: A review. In Proceedings of the International Conference on Communication and Signal Processing (ICCSP), Melmaruvathur, India, 6–8 April 2016; pp. 1669–1673. [Google Scholar]
  37. Tang, J.; Alelyani, S.; Liu, H. Feature Selection for Classification: A Review. In Data Classification: Algorithms and Applications; CRC Press: New York, NY, USA, 2014; pp. 37–64. [Google Scholar]
  38. Saeys, Y.; Inza, I.; Larrañaga, P. A review of feature selection techniques in bioinformatics. Bioinformatics 2007, 23, 2507–2517. [Google Scholar] [CrossRef] [Green Version]
  39. Wu, X.; Kumar, V.; Quinlan, J.R.; Ghosh, J.; Yang, Q.; Motoda, H. Top 10 Algorithms in Data Mining. Knowl. Inf. Syst. 2007, 14, 1–37. [Google Scholar] [CrossRef] [Green Version]
  40. Orzechowski, P.; La Cava, W.; Moore, J.H. Where Are We Now?: A Large Benchmark Study of Recent Symbolic Regression Methods. In Proceedings of the Genetic and Evolutionary Computation Conference, Kyoto, Japan, 15–19 July 2018; pp. 1183–1190. [Google Scholar]
  41. Brereton, R.G.; Lloyd, G.R. Support Vector Machines for classification and regression. Analyst 2010, 135, 230–267. [Google Scholar] [CrossRef]
  42. Drucker, H.; Burges, C.J.; Kaufman, L.; Smola, A.J.; Vapnik, V. Support Vector Regression Machines. Adv. Neural Inf. Process. Syst. 2003, 9, 155–161. [Google Scholar]
  43. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 3149–3157. [Google Scholar]
  44. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  45. Daoud, E.A. Comparison between XGBoost, LightGBM and CatBoost Using a Home Credit Dataset. Int. J. Comput. Inf. Eng. 2019, 13, 6–10. [Google Scholar]
  46. Lundberg, S.M.; Lee, S.I. A Unified Approach to Interpreting Model Predictions. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4765–4774. [Google Scholar]
  47. Lundberg, S.M.; Erion, G.G.; Lee, S.I. Consistent Individualized Feature Attribution for Tree Ensembles. arXiv 2018, arXiv:1802.03888. [Google Scholar]
Figure 1. Phases of the Industry Standard Process for Data Mining (CRISP-DM) reference model (adapted from the original diagram) [28].
Figure 1. Phases of the Industry Standard Process for Data Mining (CRISP-DM) reference model (adapted from the original diagram) [28].
Applsci 09 05447 g001
Figure 2. Heights distributions, namely mother, father, both parents, child’s, female child’s and male child’s.
Figure 2. Heights distributions, namely mother, father, both parents, child’s, female child’s and male child’s.
Applsci 09 05447 g002
Figure 3. Relation between father height and child height (regardless of gender).
Figure 3. Relation between father height and child height (regardless of gender).
Applsci 09 05447 g003
Figure 4. Relation between father height and male child height.
Figure 4. Relation between father height and male child height.
Applsci 09 05447 g004
Figure 5. Relation between father height and female child height.
Figure 5. Relation between father height and female child height.
Applsci 09 05447 g005
Figure 6. Regression analysis considering (both) parents and female child height.
Figure 6. Regression analysis considering (both) parents and female child height.
Applsci 09 05447 g006
Figure 7. Regression analysis considering father and female child height.
Figure 7. Regression analysis considering father and female child height.
Applsci 09 05447 g007
Figure 8. Interface for Child’s Target Height Prediction.
Figure 8. Interface for Child’s Target Height Prediction.
Applsci 09 05447 g008
Figure 9. LightGBM (red) and Mid-Parental (green) prediction outputs.
Figure 9. LightGBM (red) and Mid-Parental (green) prediction outputs.
Applsci 09 05447 g009
Figure 10. LightGBM (red), Mid-Parental (green) prediction and child’s real height (blue).
Figure 10. LightGBM (red), Mid-Parental (green) prediction and child’s real height (blue).
Applsci 09 05447 g010
Figure 11. Influence caused by gender on child’s height prediction (dark orchid for the female gender and cadet blue for the male gender).
Figure 11. Influence caused by gender on child’s height prediction (dark orchid for the female gender and cadet blue for the male gender).
Applsci 09 05447 g011
Figure 12. Parent’s influence on child height, considering child gender.
Figure 12. Parent’s influence on child height, considering child gender.
Applsci 09 05447 g012
Figure 13. Influence of parents height on each female child height (red for positive influence and blue for negative influence).
Figure 13. Influence of parents height on each female child height (red for positive influence and blue for negative influence).
Applsci 09 05447 g013
Table 1. Models identified on our existing literature research (our reference models) outputs.
Table 1. Models identified on our existing literature research (our reference models) outputs.
Mean Squared ErrorAccuracy
Horace Grey20.698358.55%
Mid Parental5.283659.53%
Table 2. Height correlation considering different perspectives, namely child’s regardless of gender and considering gender and father and mother height quarters.
Table 2. Height correlation considering different perspectives, namely child’s regardless of gender and considering gender and father and mother height quarters.
Child’sMale Female
Father All0.2750.3910.459
quarter 10.1150.0550.130
≤mean0.2150.2630.216
>mean0.1440.2550.356
quarter 40.0440.1560.187
Mother All0.2020.3340.314
quarter 10.1800.2750.236
≤mean0.1340.1970.291
>mean0.2170.2490.351
quarter 40.1590.3110.247
Table 3. Dataset resume.
Table 3. Dataset resume.
FamilyFatherMotherSexHeightNkids
count898898.000000898.000000898898.000000898.000000
unique197--2--
top185--M--
freq15--465--
mean-69.23285164.084410-66.7606906.135857
std-2.4702562.307025-3.5829182.685156
min-62.00000058.000000-56.0000001.000000
25%-68.00000063.000000-64.0000004.000000
50%-69.00000064.000000-66.5000006.000000
75%-71.00000065.500000-69.7000008.000000
max-78.50000070.500000-79.00000015.00000
Table 4. Train dataset provide to our model’s training.
Table 4. Train dataset provide to our model’s training.
FatherMotherSexHeight
count718.000000718.000000718718.000000
unique--2-
top--M-
freq--370-
mean69.35849664.108774-66.839972
std2.5182362.290169-3.570904
min62.00000058.000000-56.000000
25%68.00000063.000000-64.000000
50%69.50000064.000000-67.000000
75%71.00000065.500000-70.000000
max78.50000070.500000-79.000000
Table 5. Test dataset to evaluate our model’s performance.
Table 5. Test dataset to evaluate our model’s performance.
FatherMotherSexHeight
count180.000000180.000000180180.000000
unique--2-
top--M-
freq--95-
mean68.73166763.987222-66.444444
std2.2049082.377103-3.623191
min62.50000058.000000-57.000000
25%67.37500063.000000-63.500000
50%69.00000064.000000-66.250000
75%70.00000065.500000-69.050000
max75.00000070.500000-76.500000
Table 6. Machine learning algorithms performance with default parameter.
Table 6. Machine learning algorithms performance with default parameter.
Mean Squared ErrorAccuracy
Linear5.178160.33%
SVR RBF6.126953.08%
SVR Linear5.216760.04%
SVR PolyInvalid output----
SVR Sigmoid13.2621.58%
Lasso9.860024.47%
Elastic Net9.989623.48%
Ridge5.182960.30%
KNeighbors6.378551.14%
Bayesian5.180560.32%
AdaBoost5.394158.68%
SGDInvalid output----
Decision Tree5.501757.86%
XGB4.829063.01%
LightGBM4.793863.28%
Table 7. Machine learning algorithms performance with MaxAbsScaler.
Table 7. Machine learning algorithms performance with MaxAbsScaler.
Mean 2 ErrorAccuracyImpact
Linear5.178160.33%No change
SVR RBF5.313159.30%Improved
SVR Linear5.239859.86%Deteriorate
SVR Poly5.794355.61%Valid output
SVR Sigmoid5.471558,09%Improved
Lasso10.679818.19%Deteriorate
Elastic Net10.846116.92%Deteriorate
Ridge5.191660.23%Deteriorate
KNeighbors5.477958.03%Improved
Bayesian5.179860.32%No change
AdaBoost5.260759.70%Improved
SGD17.354632.93%Valid output
Decision Tree5.529957.64%No change
XGB4.829063.01%No change
LightGBM4.793863.28%No change
Table 8. Machine learning algorithms performance after hyper-parameters tuning.
Table 8. Machine learning algorithms performance after hyper-parameters tuning.
Mean 2 ErrorAccuracyImpact
Linear5.178160.33%No change
SVR RBF5.313159.30%No change
SVR Linear5.202160.15%Improved
SVR Poly5.296559.42%Improved
SVR Sigmoid5.304459.36%Improved
Lasso5.169760.40%Improved
Elastic Net5.169760.39%Improved
Ridge5.178260.33%Improved
KNeighbors5.294559.44%Improved
Bayesian5.179860.32%No change
AdaBoost5.260759.70%No change
SGD5.187160.27%Improved
Decision Tree5.533957.60%No change
XGB4.712563.90%Improved
LightGBM4.615964.64%Improved
Table 9. Machine learning algorithms resources comptions.
Table 9. Machine learning algorithms resources comptions.
CPU Time Mean CPU Time StdMemory Consumption (MB)
Linear1.87 ms20 µs138.44
SVR RBF197.00 ms3.33 ms141.62
SVR Linear101.00 ms1.66 ms140.76
SVR Poly>10.00 m ***
SVR Sigmoid11.60 ms211 µs156.61
Lasso1.50 ms26.2 µs156.87
Elastic Net1.65 ms58.1 µs154.52
Ridge1.18 ms39.8 µs154.91
KNeighbors868 µs34.8 µs156.00
Bayesian1.44 ms33.7 µs156.02
AdaBoost15.00 ms260.0 µs156.19
SGD5.85 ms68.4 µs156.32
Decision Tree1.36 ms28.8 µs156.57
XGB26.6 ms1.01 ms157.84
LightGBM107.00 ms1.26 ms163.16
* the purpose of this analysis is to compare execution consumptions which usually stand on the millisecond scale. Support Vector Regression (SVR) Poly algorithm consumes more than 10 min processing and those we have decided to classify it simply has “long time processing” (>10 m).

Share and Cite

MDPI and ACS Style

Cordeiro, J.R.; Postolache, O.; Ferreira, J.C. Child’s Target Height Prediction Evolution. Appl. Sci. 2019, 9, 5447. https://doi.org/10.3390/app9245447

AMA Style

Cordeiro JR, Postolache O, Ferreira JC. Child’s Target Height Prediction Evolution. Applied Sciences. 2019; 9(24):5447. https://doi.org/10.3390/app9245447

Chicago/Turabian Style

Cordeiro, João Rala, Octavian Postolache, and João C. Ferreira. 2019. "Child’s Target Height Prediction Evolution" Applied Sciences 9, no. 24: 5447. https://doi.org/10.3390/app9245447

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop