Next Article in Journal
Testing the Impact of Familiarity with Health Benefits Information on Dietary Supplement Choice in Pregnancy: An Online Choice Experiment
Previous Article in Journal
Factors Associated with (Exclusive) Breastfeeding Duration—Results of the SUKIE-Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Prospects and Pitfalls of Machine Learning in Nutritional Epidemiology

1
EcoVision Lab, Photogrammetry and Remote Sensing Group, ETH Zürich, 8092 Zurich, Switzerland
2
Department of Human Sciences and Quality of Life Promotion, San Raffaele University, 00166 Rome, Italy
3
Unit of Clinical and Molecular Epidemiology, IRCCS San Raffaele Roma, 00163 Rome, Italy
*
Author to whom correspondence should be addressed.
Nutrients 2022, 14(9), 1705; https://doi.org/10.3390/nu14091705
Submission received: 18 March 2022 / Revised: 13 April 2022 / Accepted: 14 April 2022 / Published: 20 April 2022
(This article belongs to the Section Nutritional Epidemiology)

Abstract

:
Nutritional epidemiology employs observational data to discover associations between diet and disease risk. However, existing analytic methods of dietary data are often sub-optimal, with limited incorporation and analysis of the correlations between the studied variables and nonlinear behaviours in the data. Machine learning (ML) is an area of artificial intelligence that has the potential to improve modelling of nonlinear associations and confounding which are found in nutritional data. These opportunities notwithstanding, the applications of ML in nutritional epidemiology must be approached cautiously to safeguard the scientific quality of the results and provide accurate interpretations. Given the complex scenario around ML, judicious application of such tools is necessary to offer nutritional epidemiology a novel analytical resource for dietary measurement and assessment and a tool to model the complexity of dietary intake and its relation to health. This work describes the applications of ML in nutritional epidemiology and provides guidelines to avoid common pitfalls encountered in applying predictive statistical models to nutritional data. Furthermore, it helps unfamiliar readers better assess the significance of their results and provides new possible future directions in the field of ML in nutritional epidemiology.

1. Introduction

The ever increasing hardware and Internet accessibility and the novel big data sources are just some of the key elements that are now enabling technologies to unlock the power of artificial intelligence (AI) and machine learning (ML). ML is an area of AI that works by learning from data. Here, analytical model building is automated to provide predictions without being explicitly programmed to do so. Therefore, it does not come as a surprise that ML methods are becoming progressively pervasive in our society, where most fields of science have undergone a big data revolution.
Nutritional epidemiology uses dietary analysis to study the complex link between nutritional intake and health. Given the ongoing technology revolution resulting in an increased amount of available electronic data, nutritional epidemiology research has recently been seeing a rapid expansion of ML applications, especially in the field of data fusion, modelling of nonlinear associations and feature reduction. Nevertheless, reconciling between the well established scientific principles of nutritional epidemiology and this novel field is proving to be difficult.
Given the still young state of ML in nutritional epidemiology, the goal of this work is to provide a critical overview of the applications of ML in nutritional epidemiology research, its limitations and future perspectives. Clearly, this demands an ever deeper understanding of such approaches at hand. Many researchers have addressed this problem by teaming up with or contracting mathematically qualified scientists, but the shortage of suitably trained professionals means that research groups often rely on untrained or inexperienced scientists to perform complex data analysis. Although previous commentaries have offered recommendations and guidelines to avoid common pitfalls associated with these analysis tasks, in practice, following these guidelines has proven more difficult than expected. Furthermore, these frameworks have never been provided for the field of nutritional epidemiology. By bringing in a novel perspective on some of the key statistical issues, we hope that the exposition below will help readers better apply such techniques and interpret results from studies. In the next sections, we will first review current areas of research and limitations of nutritional epidemiology. Following that, we provide the introductory concepts and a literature review on ML and big data. We then describe the current approaches of ML in nutritional epidemiology and focus on its benefits, limitations and common pitfalls to avoid when applying data-driven models to the field of nutritional epidemiology. We conclude by proposing future directions for research on the applications of ML in nutritional epidemiology.

2. Nutritional Epidemiology

Nutritional epidemiology is a discipline that studies the impact of diet and/or nutritional elements on disease occurrence in populations. The findings from nutritional epidemiology are generally applied for creating dietary policies, food fortification interventions, restriction of substances, or recommendations for prevention of cancer, chronic diseases, obesity and congenital malformations (Satija et al. [1]).
To do so, nutritional epidemiology consists of three main research areas (Illner et al. [2]): (1) exposure and outcome measurement (e.g., nutritional intake and disease occurrence) through data collection on large groups of participants; (2) choice of the study design; and (3) analytical and statistical techniques to assess the magnitude of the association between these two measures. Nutritional epidemiology is characterised by uncertainty and an incredible number of factors that can play a role in the measured association. Therefore, the process of causal inference is complex, and it is generally rare for a causal relationship to be considered unequivocal. The major challenges in the process of causal inference are described below.

2.1. Errors in Measurement Methods

Sources of error in dietary assessments can be divided in two categories (Thornton and Villamor [3]): random errors are generally unintentional mistakes such as marking the wrong frequency column, or copying a wrong number when switching from a paper to an electronic form. Such errors are considered noise in the sense that they cover the real signal, i.e., the association between cause and effect. Plausibility checks can often be employed to discover such errors in a dataset. The effects from random errors can be mitigated by increasing the amount of data collected. The second types of errors are the systematic errors which refer to biases across the population and mainly depend on the patient. These include memory problems, lack of correct estimation in the frequency of foods consumed and portion size (under or over-reporting of intake) and social desirability bias. The latter describes a situation in which the interviewees report the more desirable or socially accepted response (Hebert et al. [4]). In Section 4.1, we will describe how ML can be used to improve the accuracy of dietary records.

2.2. Nonlinearities

The simplest models assume a linear association between exposure and outcome. However, most of the times a nonlinear relationship exists. A variety of approaches have been developed for dealing with nonlinearities, but they are usually ad hoc solutions based on experience, needing domain expertise and careful assessment and interpretation of the results. Some factors are specifically crucial in jeopardising the detection of nonlinear effects, which in turn cause misinterpretation of the results (May and Bigelow [5]). Some of these factors are: (i) sample size: If the data are not large enough, detecting the true relationship between exposure and outcome is troublesome. Additionally, if the range of exposure is small, any inference of nonlinearity can pass unnoticed, together with possible threshold effects; (ii): Individual susceptibility is an important factor that can contribute to a disease. In this case, adding data from genetic studies could be helpful although laborious and challenging to realise; (iii): wrong models: Sometimes, authors just do not use the right model to estimate the true function between exposure and outcome. A quick fix for such a problems is to approach the modelling process by using multiple approaches in an exploratory phase until the right model is found. For this reason, results from a specific study should be judged in the context of their specific application. In Section 3 and Section 4.2.1 we will describe how complex ML models have been shown to deal particularly well with nonlinear data.

2.3. Confounding

Confounding is one of the main sources of systematic errors that can appear whenever assessing causality in epidemiologic studies (Greenland and Morgenstern [6]). Confounding represents an alternative explanation for the association between the exposure and the outcome, and it is introduced by a third variable called a confounder. Confounding is a critical error in nutritional epidemiology. For it to occur, the confounder must be associated, but not measured, with both the exposure and the outcome. In this case, statistical adjustment in analyses can be performed to control for some of the effects of confounding. One example is entering confounders as covariates in analytic models (Zeraatkar et al. [7]). However, complete removal of the influence of confounding cannot be assured. ML approaches to mitigate confounding are described in Section 4.3.

2.4. Missing Data

After an epidemiologic study, data that were collected are then recorded and inserted into a database. Here, it is crucial to consider which information was not provided by the interviewees, that is, missing data. There are different techniques that deal with missing data (Sangra and Codina [8]), the most common being: (i) listwise method, which completely deletes the interviewee presenting missing data, however it affects the power of the tests; (ii): pairwise deletion, which keeps the subject whenever such missing values do not affect the analysis, otherwise entirely deletes a subject; (iii) imputation, which replaces missing values with the mean of other values, or though regression, etc. This approach can however introduce random bias in the later modelling stage, which is not taken into account by current methods. Similar to the errors related to measurement methods, missing data can be imputed with ML models or can be dealt with during training of a ML model, as explained in Section 5.
An exciting area of future research regards ML, which has the promise of mitigating or solving some of these challenges. ML is presented in the Section 3.

3. Machine Learning

ML supports domain experts by automatically learning from data, thus removing the need for manual analytical model building. Such techniques are more flexible than the classical statistical model approaches (Ciavatta et al. [9]) because they can take advantage of data-rich applications. ML models are then integrated into different downstream tasks and service applications to provide data insights and support decision making. Clearly, data are an essential component of ML. The first 2 decades of the 21st century have witnessed a sharp increase in the volume of digitally available data, which in turn has allowed the development and further deployment of ML in a wide range of applications in healthcare, finTech, cybersecurity, robotics, predictive maintenance to improve processes, customer experiences, and decision outcomes. The topic of digital data correlates with big data, a terminology that is usually found in the AI field. Such a term refers to data coming in large amounts. An enabling technology for big data has been the ever increasing amount of low-cost smaller electronics, computing devices, sensors, and the Internet of Things (IoT). IoT can be considered as a system of such physical devices connected through an area network and communicating with each other. An IoT device can also be a smartphone. This interaction can generate an enormous quantity of data, which in turn can be used to create new services and applications through ML (Shanthamallu et al. [10], Mahdavinejad et al. [11]). In healthcare applications, IoT and big data analytics through ML is now allowing applications which facilitate the collection and smart usage of patients’ data, improving treatments and services. This can also be used for population healthcare projects and clinical research.

3.1. Training and Evaluation of Machine Learning Models

The main goal of using ML is to produce a model that enables the prediction of a value y ^ for any observation represented as a d-dimensional input vector x which has the values for d variables (features). Model training in ML is referred to as the process by which a mathematical model ϕ is identified by means of data selected for this task. During model training, the model’s parameters are identified based on the training dataset (Hastie et al. [12]). For ML models to understand how to provide predictions, the training datasets are fed into the ML algorithm; this is then followed by validation datasets which ensure that the model is interpreting this data accurately and it is not overfitting. Overfitting happens whenever a ML model learns how to exactly reproduce the pattern of the training data. When this happens, the algorithm unfortunately cannot generalise well on new, unseen data. Broadly speaking, the more data are provided to the ML system, the better that model can learn to generalise and improve its performance. Model performance is generally defined as the model’s generalisation capacity, i.e., the ability of a training model to correctly predict new values for previously unseen data points. The data can be represented in different ways: (i) We can use a collection of observations, each describing a state y. For example, we can have different biomarker measurements for the same patient which might express that observation as belonging to the class y denoting cancer. (ii) Time-series data representing the temporal changes of the variables under study might be used. These data are usually used for regression problems, where the goal is to predict what the next values will be, given the previous ones. (iii) Image data, mainly used recently in computer vision applications, might also be employed.

3.2. Machine-Learning Techniques

Based on the availability of target labels, ML models can be broadly divided in supervised and unsupervised models.
Supervised models are trained using the training data with associated target labels y i (Caruana and Niculescu-Mizil [13]). The model ϕ is trained with the given dataset D = { ( x i , y i ) } i = 1 N to predict the associated label. These predictions are compared with the target labels, and the parameters of the model are learned by means of the loss function and the optimisation algorithm. The loss function evaluates how well the model is fitting on the given data. The optimisation algorithm is used to find the values of the model’s parameters that minimise the loss function. Most typically, the labels are provided by a human expert. After training, the models can be used to predict the target label for each new unseen data point. In unsupervised techniques, the model scores the data solely based on the patterns in the training dataset without any target label (Hastie et al. [14]). In this case, the training dataset D = { x i } i = 1 N consists of N data points where ( i = 1 , , N ) . ML techniques can also be divided in classification applications, where y i can only acquire discrete label values associated to a class. A class denotes a set of data having common characteristics. Regression problems involve learning the underlying function f ( x ) of input–output, which means predicting a continuous value y i .

3.3. Neural Networks and Deep Learning

Between the most employed ML models, special mention goes to artificial neural networks (ANNs) due to their vast usage (Hassoun et al. [15]). ANNs compute the predicted output value by means of a network of simple yet nonlinear unit operations (neurons). Each neuron has its own set of parameters. While ANNs are essentially nonlinear regression models, they are extremely flexible due to the ability to string an almost arbitrary number of neurons together, specifically by using large numbers of layers between data samples and predicted labels (hidden layers), each including many neurons. ANNs are used in both supervised regression and classification tasks. A deep neural network (DNN) (LeCun et al. [16]) falls under the deep learning umbrella and can be considered as multiple ANNs composed of several layers that have the ability of learning very complex functions. In fact, in the mathematical theory of ANNs, the universal approximation theorem (Winkler and Le [17]) states that ANNs and specifically DNNS, have the capabilities of universal approximators, i.e., no matter what underlying function represents the data, there is always a network that can approximately approach the result. Of course, this is an ideal situation that depends on the architectural choice and on the quality (and quantity) of input data. Clearly, this can be proven to be very useful in modelling complex diet–disease processes. There are different types of DNNs, and their main difference lies in the types of neurons (nodes) used (LeCun et al. [16]). For example, convolutional neural networks (CNNs) are mostly used for image recognition and for their ability to learn hierarchical feature representations. Recurrent neural networks are also another type of DNNs, where the connections between nodes form a graph along a temporal sequence and are mainly using for time-series data and speech recognition.
To conclude this section, we summarise in Table 1 the different learning types, tasks and ML models that are most commonly found in the literature. Note that some ML models can be used for more than one technique (e.g., ANNs can be applied in both classification and regression tasks). For a deeper dive into these, we refer the reader to Hastie et al. [14], LeCun et al. [16], Murphy [18].

4. Applications and Common Pitfalls

In this section, we describe different applications of ML in the field of nutritional epidemiology, highlighting their strengths and weaknesses.

4.1. Health and Dietary Input Data

4.1.1. Increasing the Amount of Data

Nutritional digital data can be generated through multiple means, thanks to the ubiquity of Internet-connected computers and smartphones. For example, ML models can now successfully leverage the entire contents of the electronic health records (Morgenstern et al. [19]). In wearable technology (Phillips et al. [20]), dietary assessment can be conducted by using wearable devices containing gyroscopes and/or accelerometers. These can track wrist movements and indirectly record lifting of hands and cutlery as an approximation to account for calorie intake and eating patterns (Vu et al. [21]). Of course, such systems can also be used to monitor daily physical activity. Wearable sensors can also be used to continuously monitor glucose (Cappon et al. [22]) for diabetes treatment and self-management in downstream ML-based decision-support systems (Contreras et al. [23], Kavakiotis et al. [24]).
Additionally, mobile calorie counting apps (Limketkai et al. [25]) can record large amounts of data from the users, self-reporting daily portions and calorie intakes. Nowadays, it is estimated that there exist about 165,000 publicly available mobile health apps related to health and wellness (Kao and Liebovitz [26]). Such systems can represent a large opportunity for nutritional epidemiologic studies since the data does not need to be collected inside the clinics. Rather, it can be collected remotely at home and on the go, without suffering from recollection bias. Furthermore, compared to traditional data collection methods (e.g., phone interviews etc.), mobile self-reporting is also less time demanding. A simple and inexpensive yet powerful way to track dietary habits is through grocery purchases in smartphone-tracking applications. These records have been used in several public health nutrition research studies to evaluate interventions (Bandy et al. [27]). Another way to decrease the burdensome task of collecting dietary data is to use ML-based natural language processing techniques, which convert speech into text or input variables, e.g., by using a smartwatch combined with an audio-based detector of chews and swallows to identify eating behaviours (Kalantarian and Sarrafzadeh [28]). Dietary data can also be automatically “mined” through ML approaches from social media for ecological studies to study community-level health outcomes (Shah et al. [29]). In another study (Gerina et al. [30]), a ML model was used to detect cooking activities based on air quality sensor data. The model’s output was integrated with a social robot that, by interacting with the participant, supported filling in a food diary.
Such approaches can facilitate more longitudinal, repeated dietary measurements, allowing for expanded sample sizes thereby increasing statistical power. However, as sample size grows, it decreases the quality of data aggregated through wearable devices, apps or social media due to the lack of supervision by a clinical researcher during collection.

4.1.2. Improving Data Quality

By its very nature, a ML model is sensitive to the quality of the data used during its training and validation. The rule of “garbage in, garbage out” (Grimes [31]) applies here, where even small errors or bias in the training data can lead to unexpected consequences in the model’s prediction.
To increase and improve the accuracy of self-reported dietary records during food logging, an active area of research has recently seen ML and deep learning models used to classify foods and calories from pictures of users (Lo et al. [32], Tay et al. [33], Sahoo et al. [34], Lo et al. [35]). Accuracy was reported to be as high as 90%. Another approach is to evaluate food sizes (Ege et al. [36]) from food images and use the resulting value to evaluate caloric intake. Such approaches can be applied to address systematic and random errors. However, image-based dietary assessment relies heavily on computing algorithms and storage facilities, which can greatly increase the processing time. For example in (Puri et al. [37]), the authors report an image-based assessment time of 33 s. Furthermore, such analysis can be troublesome if the dataset to train (and test) the model is limited, such as in (Zhu et al. [38], Woo et al. [39]), which could make it necessary to semi-automate the process of food recognition, as presented in (Jia et al. [40]). To mitigate this issue, large datasets of food images are becoming increasingly available (Min et al. [41], Aguilar et al. [42]). The idea here is that deep learning models for image recognition such as CNNs can be pretrained on such large datasets and then, thanks to a technique called transfer learning, can be used for online applications in a continuously learning manner (He and Zhu [43]), further decreasing prediction times.
Low quality input data can also lead to a biased ML model. These raise ethical questions whether we can safely trust decisions taken based on ML models’ predictions. In fact, it is critical to ensure that the data is representative of the population under study and not skewed or imbalanced. Large amounts of data would usually suffice to prevent biases of the ML model, but even large datasets need to be of high quality. For example, patients with a low socioeconomic status usually have limited access to healthcare facilities (Arpey et al. [44]). They therefore have insufficient information in their electronic health records, leading to a biased ML model presenting disparities (Gianfrancesco et al. [45]). A possible solution to avoid such bias is to make sure that features that are biased, such as ethnicity and social determinants of health, are also included in algorithms (Gianfrancesco et al. [45]). Related to model bias is model transparency (i.e., why a certain prediction was given) which is necessary for real-world implementations since special care needs to be take whenever the ML model will be used for downstream tasks such as patient care. We will discuss the aspects of transparency (interpretability and explainability) in Section 5.4.

4.2. Modelling of Dietary Variables

4.2.1. Non-Linearities

The diet–disease relationship in epidemiologic research is mostly conducted in its simplest form by considering linear models of association between exposure and outcome (Boeing [46]). However, this assumed relationship is generally not correct as nonlinear relations exist (Ioannidis [47]). For instance, several studies related to cardiovascular diseases have shown that salt, carbohydrates and fats possibly present a U- or J-shaped relation with such disease (Kong et al. [48], Investigators et al. [49]), largely due to confounding factors. As a consequence, if incorrect linear assumptions are taken during the modelling phase, this can cause spurious associations and biased effect estimates (Bodnar et al. [50]).
Complex ML and deep learning algorithms such as ANNs and DNNs can overcome these limitations due to their ability to model complex relationships between input variables. For example, in (de Cos Juez et al. [51]), an ANN was implemented to relate 38 diet and lifestyle variables to bone mineral density in post-menopausal women. The sample size, however, was relatively small, with only 200 patients. No information was given about using a validation dataset to prevent overfitting of the model. In a cross-sectional survey (Zeng et al. [52]), the risk of hyperuricemia based on dietary information was estimated using an ANN. In (Chew [53]) DNNs were used to classify images of patients with age-related macular degeneration supporting the authors in evaluating the importance of nutritional supplements. Other examples of DNNs for modelling nonlinear diet–disease relationships are still scarce in the nutritional epidemiology literature. We suspect this is due to the generally low availability of large datasets. In fact, in other broader disciplines such us clinical and epidemiologic research, where data availability is improved, we found that several studies employ DNNs, e.g., a deep learning model for the detection of breast cancer (Puvanesarajah et al. [54]); a CNN model for non-imaging diagnostic to predict for skin cancer (Vivot et al. [55]); or for diabetic retinopathy screening (Wong and Bressler [56]). In medical and epidemiologic research, it is also reported that DNNs perform better than ML approaches (Byeon [57], Xiong et al. [58], VoPham et al. [59]) due to their exceptional capabilities to model nonlinear relationships and because they need virtually no feature engineering of the input data. Because of this, we believe that a large amount of structured data is still needed for nutritional epidemiology to unlock the true power of DNN models.

4.2.2. Dimensionality Reduction

Curse of dimensionality refers to a phenomenon that arises when data in high-dimensional spaces (i.e., with a high number of variables/features) are used for analysis or modelling purposes. Given the large amount of possible explanatory variables, modelling studies in nutritional epidemiology have become difficult to conduct, and the task of identifying the most predictive ones is extremely challenging (Bodnar et al. [50]). In addition, common statistical models cannot easily deal with a large number of variables. In this case, ML can be used as a dimensionality reduction technique. Such techniques can be split into (1) feature extraction and (2) feature selection methods.
Feature extraction works by finding a combination of new features from the original ones. Such algorithms simplify modelling, thereby partially overcoming the curse of dimensionality. Depending on the technique that is used, these methods find the best linear or nonlinear transformation that reduces the number of dimensions with a minimum loss of information. Examples of linear methods are principal component analysis (PCA) and linear discriminant analysis. In (Hoffmann et al. [60]), three linear methods (PCA, reduced rank regression and partial least squares) were employed for feature extraction to derive dietary patterns from 49 food groups related to type 2 diabetes. The authors in (Zhang et al. [61]), used PCA to derive dietary patterns predictive of cardiovascular disease risk. In (Santos et al. [62]), PCA was applied on 34 variables expressing the mean food intake of 1102 individuals from a population-based study. Nonlinear dimensionality reduction techniques are deep autoencoders and manifold learning (Morgenstern et al. [19]). Deep autoencoders are especially interesting since they are DNNs and combine the advantages of being both unsupervised and nonlinear approaches, allowing for the embedding of data into a low-dimensional representation while conserving its properties (Falissard et al. [63], Wang et al. [64]). Clustering techniques (e.g., Gaussian mixture models or k-means clustering) can also be included into feature extraction techniques as they can eliminate noisy variables. For example, (Kwon et al. [65]) employed k-means clustering to pinpoint risk factors for low muscle mass based on nutritional factors. The algorithm generated clusters of patients based on their dietary and health-related data, where patients within the same cluster had similar attributes. ML logistic regression was then applied to find risk factors in each cluster.
Feature selection methods aim at reducing the dimensionality in a large dataset. Rather than transforming or grouping the variables into a new representation, feature selection methods can restrict the number of variables to a smaller subset by applying a selective filter. The selected variables are chosen based on how informative they are for the prediction of the outcome (Walter and Tiemeier [66]). For example, techniques such as permutation feature importance (Altmann et al. [67]) can be used to train the ML model recursively with different variables, thereby finding the ones providing best model performance. In (Zeevi et al. [68]), this technique was used to find the most important variable related to glycemic responses in a regression algorithm, which was found to be related to microbiota. In (Dipnall et al. [69]), a methodology for feature selection using ML on a large epidemiological dataset was successfully implemented for detecting 3 out of 67 biomarkers (red cell distribution width, serum glucose and total bilirubin) associated with depression.
The main downside of such dimensionality reduction approaches is that they are usually difficult to automate and apply to a different dataset because of their complexity and application-specific fine-tuning. Additionally, they require extensive domain expertise, experience, and process understanding (Russo et al. [70]).

4.3. ML Approaches to Confounding

In nutritional epidemiologic studies, confounding and multicollinearity are systematic errors that can cause misinterpretation of the results. That is because they generate a false relationship between the dietary variable and the disease risk. This aspect is particular challenging in nutritional data since nutrients from different foods are not only correlated with each other but also with different outcomes (diseases) (Trepanowski and Ioannidis [71]). Confounding is usually controlled by a statistical adjustment during data analyses. This approach, however, requires domain confidence and expertise and making assumptions about the measured variables.
A way to deal with confounding through ML is by creating “high-capacity models”. In ML, capacity is a term representing model complexity, i.e., a model with higher capacity is expected to be able to model more complex relationships between the variables. Usually, high-capacity models are ANNs or deep-learning models presenting an extreme number of parameters (in terms of millions). Training high-capacity deep-learning models with a high number of observations and variables has been shown to better deal with confounding in clinical research (Brisk et al. [72], Badgeley et al. [73]). That is because such models have the capability of processing data through their several layers and create entirely new types of variables. This approach not only adjusts for confounding, but also for any existing nonlinearity (Morgenstern et al. [19]) by taking advantage of the data richness in number of observations and variables. High-capacity models of this type however are prone to overfitting and errors in the testing phase. In the next section we will discuss the issue of overfitting together with different methods used to adjust for it.

5. Practical Recommendations

5.1. Data Preparation

The most important step after data—especially big data—collection, is to preprocess (i.e., clean and/or put in the right format), store, and make it available for research (García et al. [74]). Additionally, metadata should also be provided. Finally, each of the above data preparation steps should also be well documented, ideally providing the end user with all the versions of the data, from raw to clean. Addressing the points above can be challenging and time demanding, but it is fundamental. In fact, the success of ML techniques is dependent on the quality of the data that they operate on (Kotsiantis et al. [75]).
At this stage, missing data should be handled correctly. Imputation of missing data through ML techniques has been extensively studied (Lakshminarayan et al. [76], Richman et al. [77], Batista and Monard [78]) and has shown to outperform common imputation statistical methods (Jerez et al. [79]). Different ML algorithms can be applied, such as k-nearest neighbours and self-organisation maps. Recently, nonlinear techniques, such as ANNs, are also employed (Al-Milli and Almobaideen [80]).
Dimensionality reduction techniques and feature engineering are also part of this stage. Feature engineering is the process of adding or creating new features with the goal of supporting and increasing the information provided to the ML model (Heaton [81]). Such a step is usually cumbersome and requires domain expertise. For example, in a study exploring the predictive power of nutrients for cardiovascular disease using ML (Morgenstern et al. [82]), the authors combined features concerning cultural and racial origin and household income, and derived variables based on smoking as well as on the participants’ immigrant status, age, and years since immigration. Compared to ML algorithms, an advantage of deep learning models is that they do not require feature engineering due to their abilities to extract meaningful information from data automatically (LeCun et al. [16]).
During data preparation, labelling is also conducted. This is the process of assigning a target to each observation. Data labelling is necessary for a supervised ML model to learn from the training dataset to predict the associated label (target). Labels are also needed in both the supervised and unsupervised case for performance evaluation (Russo et al. [83]). While labelling an entire dataset is tedious work that requires a dedicated team of experts, correct data labelling is crucial for any ML model. For these reasons, repeated-labelling strategies can been employed (Sheng et al. [84]). For example, a minimum of two people is required to label the data, and target labels are then computed by merging the provided results to avoid bias due to decision fatigue. Additionally, a user interface designer is needed to provide a simple, intuitive system to label data.

5.2. Data Quality and Quantity

“How much data is enough for a ML model”? There is no correct answer to this question, as it usually depends on the type of model chosen; the number of variables (the higher the dimension of the dataset, the more observations are needed for the model to create an input–output function); and the data quality (Gudivada et al. [85]). Data quality refers to the presence of noise and missing data, but it also deals with how representative of the population under study it is. Unfortunately, the presence of low-quality observations in the data will negatively impact any ML method. A way to address such problems is through specialised loss functions (Wang et al. [86]) which reduce the weights of low-quality (or imputed) observations during training so that the model focuses more on the most informative samples. For example, in (Tran et al. [87]) a focal loss is employed for improving early detection and classification of pulmonary nodules.
On the other hand, if the only limitation is the amount of data (i.e., data shows a signal-to-noise ratio good enough to learn meaningful input–output relationships), techniques such as regularisation and k-fold cross-validation can be used to support correct training of ML models. We will discuss these techniques below.

5.3. Avoiding Overfitting

Overfitting is most probably one of the main errors that happen when training a ML model. Common causes of overfitting are related to the quantity of data (the fewer the data, the less the model generalises well) and using a model that is too complex (we mentioned before the term high capacity). As a general rule, we are in the presence of overfitting whenever the performance of a trained ML model on an unseen test set is considerably lower than the training set. In general, we should beware of results that are too good to be true.
A method used to monitor the correct training of ML models is k-fold cross-validation (Rodriguez et al. [88]). This is a technique that uses k numbers of held-out validation sets to constantly evaluate whether the model is overfitting on the training data during the training process. Such a technique is especially convenient for small datasets (which is usually the case for nutritional epidemiology data). In this process, the model is tested during training on an unseen data set. If the performance of the model on the held-out set is much worse than the ones on the training data, it means the model is overfitting and: (i) training should be stopped earlier; or (ii) regularisation techniques should be used.
Regularisation is one of the most important concepts of ML and is a technique that prevents the model from overfitting by discouraging learning a more complex or flexible model (Hastie et al. [12]). Regularisation comes into play during model training, i.e., by shrinking the coefficient estimates towards zero or by preventing them from rising too high.

5.4. Dealing with Biased Data

ML models are used to generate predictions which are then implemented in downstream tasks such as diagnosis, causality assessment, as well as decision making in self-driving cars or the automation industry. There are some instances, however, where an ML-based prediction could be incorrect and cannot be relied upon because the predictions are biased towards a particular class.
In fact, while it is nearly impossible to point out why a particular decision was made because ML models are considered “black boxes”, a large field of research focuses on interpretable and explainable AI which aims at providing model transparency (Holzinger et al. [89], Gunning et al. [90]). Specifically, interpretability is about understanding the cause and effect within the ML system, that is, how much are we able to predict what is going to happen, given an input. Explainability, on the other hand, is the extent to which the internal architecture and mechanics of a ML model can be explained in human terms. Another approach also focuses on what has been called Bayesian deep learning (Kendall and Gal [91]), which aims at obtaining a realistic, well-calibrated expression of when and how much a model is uncertain about its own prediction and therefore how trustworthy such a prediction is. A Bayesian neural network employs a prior distribution for each weight of a neural network, then posterior inference is applied. Interpretability and explainability in ML are still an open research question, and special attention should be paid to how much a ML model’s predictions are trusted to make final decisions.

5.5. Performance Metrics

Whenever a ML model is used during training or testing, it is critical to evaluate its predictive performance. For regression tasks, these are usually measured in terms of R 2 or explained variance, while classification tasks generally require an accuracy metric, that is:
Accuracy = t p + t n t p + t n + f p + f n ,
where true positives (tp) are the number of correctly identified members of the disease class, true negatives (tn) the number of correctly classified members of the no-disease class, false negatives (fn) and false positives (fp) are, respectively, the number of incorrectly classified no-disease and disease (this equation refers to a case with two classes only). Clearly, an accuracy close to one represents a ML model with powerful predictive capabilities. However, the accuracy metric works best only when the two classes (no-disease and disease) are balanced, which is usually not the case. Generally in fact, the data can be skewed towards the no-disease class which will result in a prediction bias towards the majority class. In this case, the accuracy metric will still return a value close to one, but the model is actually unable to correctly predict the patients with disease. This will cause misleading results and incorrect conclusions. For example, in (Batterham et al. [92]) the authors employed and compared different ML models to classify participants achieving at least 10,000 steps per day in a nutrition-related intervention study. A metric called area under the curve AUC was used for performance assessment. However, the number of participants reaching the target steps was 79%, resulting in an imbalanced dataset, where such a performance metric is not ideal (for example, precision–recall curves or a F1 score would have been more feasible).
Therefore, whenever the data present imbalanced classes, it is recommended to compute the F1-score as:
F 1 = 2 p r e c i s i o n · r e c a l l p r e c i s i o n + r e c a l l ,
where
p r e c i s i o n = t p t p + f p r e c a l l = t p t p + f n
Generally, there is an inverse relationship between precision and recall (recall is also sometimes called sensitivity): as one increases, the other decreases. This is called the precision–recall trade-off. While recall expresses the ability to find all relevant cases in the dataset (what proportion of actual diseases were identified correctly), precision shows the proportion of the data points the model identifies as diseases were real (e.g., a model that produces no false positives has a precision  = 1 ) (Davis and Goadrich [93]). In healthcare applications, recall is usually preferred over precision as it is more important to identify the disease class at the cost of having a higher number of false positives.

5.6. Skilled Personnel

Finally, applying ML in nutritional epidemiology needs teams of experts from a large variety of disciplines. That is because best practices and approaches need to be followed to build a productive ML team (Schelter et al. [94]). Necessary figures include: (1) a data engineer responsible for building data pipelines architectures and infrastructure; (2) a data scientist to identify cases that can be solved with ML and develop custom ML models; (3) a ML scientist/researcher to optimise and deploy models to production and conduct research for novel ML use cases and applications. Building such a team in small enterprises or for short-term research projects, however, is a challenge and can lead to lack of adequate personnel and scientific credibility of conducted research. Usage of consultancy firms can be helpful, although expensive, in this case.

6. Conclusions

6.1. Critical Points for the Application of ML

This study tried to reveal current and possible future applications of machine learning (ML) to the field of nutritional epidemiology. The current literature is, however, still almost devoid of practical and successful examples of ML in this sector. We have identified some critical points that need to be addressed for future integration and development of these fields (Diebolt et al. [95]).
  • Most of the studies in the literature are limited to few models and small datasets, therefore not showing the real advantages of one method compared to another. Systematic comparisons and benchmark datasets are therefore needed.
  • It is important to take advantage of the datasets already collected in different studies. That means an organised system of aggregation of the data is essential, together with a regulatory framework for ensuring data privacy and trustworthiness. Extensive work is needed to ensure that research projects collect and publish datasets in a well-organised manner and with robust security.
  • In addition, availability of technical skills in the use of ML, as well as access to high-performance computing, is needed to produce clear, quantifiable demonstrations of the benefits of ML in nutritional epidemiology research. This can be reached thanks to collaborations and large investments in the training of personnel and infrastructures.
  • Whenever dealing with data with a low signal to noise ratio, such as survival rate or readmission rate in hospitals, several epidemiologic studies have shown that ML algorithms provide improved performance compared with traditional statistical models (Feng et al. [96], Mortazavi et al. [97]). On the other hand, the situation is overturned for data with higher signal to noise ratio, such as risk prediction of major chronic diseases or depression. In this case, ML models have been surpassed by conventional statistical models (Nusinovici et al. [98], Gravesteijn et al. [99]). We expect a similar situation to occur in nutritional epidemiology, although the high correlation between nutritional variables could also play a big role in favour of ML models.

6.2. Limitations of Current Work

The main challenges found while reviewing the applications of ML in nutritional epidemiology are not only the limited literature examples but also their broad range of applications. For example, studies employing deep-learning models range from modelling input–output relationships to computer vision-based food detection. To compensate for this, research papers employing ML in clinical research were also included in the present review with the goal of showing the prospective applications of ML. If that was the case, it was explicitly specified in the text.
This review focused on current applications and pitfalls of ML in nutritional epidemiology. An aspect of nutritional epidemiology which was not discussed relates to the existing biorepositories and national surveys in public health nutrition such as (Rosso et al. [100], Riboli et al. [101]). Similarly, we did not delve into the aspect of specialised medicine and nutrition as application fields for ML (Zeevi et al. [68]). The reason for these choices lies in the objective of this study, which is to provides guidelines to avoid common pitfalls for practitioners applying data-driven models to nutritional data, leaving the above topics outside the scope of our work.
Similar to our work is (Sak and Suchodolska [102]), where the authors explore the applications of AI in nutrients science research. Different than this review, however, the authors focused on a broad range of applications areas: biomedical nutrients, clinical nutrients and nutritional epidemiology. As for the latter, the study mainly focuses on using ML for data integration such as dietary assessment and IoT systems, without mentioning modelling capabilities and providing practical recommendations. To the best of our knowledge, our work is the first review that provides a state-of-the-art snapshot of the current literature, critically evaluating diverse results while revealing possible inconsistencies in the body of research. Additionally, this work provides for the first time practical recommendations and possible future directions in the field of ML and nutritional epidemiology.

6.3. Future Perspectives

We believe that in the future a complete integration of ML into the field of nutritional epidemiology might be accomplished by data-driven approaches, the likes of which are seen today in artificial neural networks and deep learning, despite the challenges discussed in this work. Such techniques not only have powerful modelling capacities, requiring minimal data preprocessing, but also have promising applications within the area of data augmentation. This is mainly due to their capability in the computer vision field, of increasingly extracting abstract features from images (Goodfellow et al. [103]), which has been already applied for estimating dietary intake from food pictures. Although few initial works on deep learning in nutritional epidemiology can be found in the body of research, to really reap the power of such techniques, larger datasets and efforts are still needed.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

This work is an extension of the M. Sc. project entitled “Artificial Intelligence in Nutritional Epidemiology” completed in fulfillment of the degree in Nutritional Science at San Raffaele University Roma.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Satija, A.; Yu, E.; Willett, W.C.; Hu, F.B. Understanding nutritional epidemiology and its role in policy. Adv. Nutr. 2015, 6, 5–18. [Google Scholar] [CrossRef] [Green Version]
  2. Illner, A.; Freisling, H.; Boeing, H.; Huybrechts, I.; Crispim, S.; Slimani, N. Review and evaluation of innovative technologies for measuring diet in nutritional epidemiology. Int. J. Epidemiol. 2012, 41, 1187–1203. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Thornton, K.; Villamor, E. Nutritional Epidemiology. In Encyclopedia of Food and Health; Caballero, B., Finglas, P.M., Toldrá, F., Eds.; Academic Press: Oxford, UK, 2016; pp. 104–107. [Google Scholar] [CrossRef]
  4. Hebert, J.R.; Clemow, L.; Pbert, L.; Ockene, I.S.; Ockene, J.K. Social desirability bias in dietary self-report may compromise the validity of dietary intake measures. Int. J. Epidemiol. 1995, 24, 389–398. [Google Scholar] [CrossRef] [PubMed]
  5. May, S.; Bigelow, C. Modeling nonlinear dose-response relationships in epidemiologic studies: Statistical approaches and practical challenges. Dose-Response 2005, 3. [Google Scholar] [CrossRef] [PubMed]
  6. Greenland, S.; Morgenstern, H. Confounding in health research. Annu. Rev. Public Health 2001, 22, 189–212. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Zeraatkar, D.; Cheung, K.; Milio, K.; Zworth, M.; Gupta, A.; Bhasin, A.; Bartoszko, J.J.; Kiflen, M.; Morassut, R.E.; Noor, S.T.; et al. Methods for the selection of covariates in nutritional epidemiology studies: A meta-epidemiological review. Curr. Dev. Nutr. 2019, 3, nzz104. [Google Scholar] [CrossRef]
  8. Sangra, R.A.; Codina, A.F. The identification, impact and management of missing values and outlier data in nutritional epidemiology. Nutr. Hosp. 2015, 31, 189–195. [Google Scholar]
  9. Ciavatta, S.; Pastres, R.; Lin, Z.; Beck, M.; Badetti, C.; Ferrari, G. Fault detection in a real-time monitoring network for water quality in the lagoon of Venice (Italy). Water Sci. Technol. 2004, 50, 51–58. [Google Scholar] [CrossRef] [PubMed]
  10. Shanthamallu, U.S.; Spanias, A.; Tepedelenlioglu, C.; Stanley, M. A brief survey of machine learning methods and their sensor and IoT applications. In Proceedings of the 2017 8th International Conference on Information, Intelligence, Systems & Applications (IISA), Larnaca, Cyprus, 27–30 August 2017; pp. 1–8. [Google Scholar]
  11. Mahdavinejad, M.S.; Rezvan, M.; Barekatain, M.; Adibi, P.; Barnaghi, P.; Sheth, A.P. Machine learning for Internet of Things data analysis: A survey. Digit. Commun. Netw. 2018, 4, 161–175. [Google Scholar] [CrossRef]
  12. Hastie, T.; Tibshirani, R.; Friedman, J.H.; Friedman, J.H. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer: Berlin/Heidelberg, Germany, 2009; Volume 2. [Google Scholar]
  13. Caruana, R.; Niculescu-Mizil, A. An empirical comparison of supervised learning algorithms. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 161–168. [Google Scholar]
  14. Hastie, T.; Tibshirani, R.; Friedman, J. Unsupervised learning. In The Elements of Statistical Learning; Springer: Berlin/Heidelberg, Germany, 2009; pp. 485–585. [Google Scholar]
  15. Hassoun, M.H. Fundamentals of Artificial Neural Networks; MIT Press: Cambridge, MA, USA, 1995. [Google Scholar]
  16. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  17. Winkler, D.A.; Le, T.C. Performance of deep and shallow neural networks, the universal approximation theorem, activity cliffs, and QSAR. Mol. Inform. 2017, 36, 1600118. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Murphy, K.P. Machine Learning: A Probabilistic Perspective; MIT Press: Cambridge, MA, USA, 2012. [Google Scholar]
  19. Morgenstern, J.D.; Rosella, L.C.; Costa, A.P.; de Souza, R.J.; Anderson, L.N. Perspective: Big data and machine learning could help advance nutritional epidemiology. Adv. Nutr. 2021, 12, 621–631. [Google Scholar] [CrossRef] [PubMed]
  20. Phillips, S.M.; Cadmus-Bertram, L.; Rosenberg, D.; Buman, M.P.; Lynch, B.M. Wearable technology and physical activity in chronic disease: Opportunities and challenges. Am. J. Prev. Med. 2018, 54, 144. [Google Scholar] [CrossRef] [PubMed]
  21. Vu, T.; Lin, F.; Alshurafa, N.; Xu, W. Wearable food intake monitoring technologies: A comprehensive review. Computers 2017, 6, 4. [Google Scholar] [CrossRef]
  22. Cappon, G.; Acciaroli, G.; Vettoretti, M.; Facchinetti, A.; Sparacino, G. Wearable continuous glucose monitoring sensors: A revolution in diabetes treatment. Electronics 2017, 6, 65. [Google Scholar] [CrossRef] [Green Version]
  23. Contreras, I.; Vehi, J. Artificial intelligence for diabetes management and decision support: Literature review. J. Med. Internet Res. 2018, 20, e10775. [Google Scholar] [CrossRef]
  24. Kavakiotis, I.; Tsave, O.; Salifoglou, A.; Maglaveras, N.; Vlahavas, I.; Chouvarda, I. Machine learning and data mining methods in diabetes research. Comput. Struct. Biotechnol. J. 2017, 15, 104–116. [Google Scholar] [CrossRef]
  25. Limketkai, B.N.; Mauldin, K.; Manitius, N.; Jalilian, L.; Salonen, B.R. The Age of Artificial Intelligence: Use of Digital Technology in Clinical Nutrition. Curr. Surg. Rep. 2021, 9, 20. [Google Scholar] [CrossRef]
  26. Kao, C.K.; Liebovitz, D.M. Consumer mobile health apps: Current state, barriers, and future directions. PM&R 2017, 9, S106–S115. [Google Scholar]
  27. Bandy, L.; Adhikari, V.; Jebb, S.; Rayner, M. The use of commercial food purchase data for public health nutrition research: A systematic review. PLoS ONE 2019, 14, e0210192. [Google Scholar] [CrossRef]
  28. Kalantarian, H.; Sarrafzadeh, M. Audio-based detection and evaluation of eating behavior using the smartwatch platform. Comput. Biol. Med. 2015, 65, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Shah, N.; Srivastava, G.; Savage, D.W.; Mago, V. Assessing Canadians health activity and nutritional habits through social media. Front. Public Health 2020, 7, 400. [Google Scholar] [CrossRef] [PubMed]
  30. Gerina, F.; Pes, B.; Reforgiato Recupero, D.; Riboni, D. Toward supporting food journaling using air quality data mining and a social robot. In Proceedings of the European Conference on Ambient Intelligence, Rome, Italy, 13–15 November 2019; pp. 318–323. [Google Scholar]
  31. Grimes, D.A. Epidemiologic research using administrative databases: Garbage in, garbage out. Obstet. Gynecol. 2010, 116, 1018–1019. [Google Scholar] [CrossRef] [PubMed]
  32. Lo, F.P.W.; Sun, Y.; Qiu, J.; Lo, B. Image-based food classification and volume estimation for dietary assessment: A review. IEEE J. Biomed. Health Inform. 2020, 24, 1926–1939. [Google Scholar] [CrossRef] [PubMed]
  33. Tay, W.; Kaur, B.; Quek, R.; Lim, J.; Henry, C.J. Current developments in digital quantitative volume estimation for the optimisation of dietary assessment. Nutrients 2020, 12, 1167. [Google Scholar] [CrossRef]
  34. Sahoo, D.; Hao, W.; Ke, S.; Xiongwei, W.; Le, H.; Achananuparp, P.; Lim, E.P.; Hoi, S.C. FoodAI: Food image recognition via deep learning for smart food logging. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 2260–2268. [Google Scholar]
  35. Lo, F.P.W.; Sun, Y.; Qiu, J.; Lo, B. Food volume estimation based on deep learning view synthesis from a single depth map. Nutrients 2018, 10, 2005. [Google Scholar] [CrossRef] [Green Version]
  36. Ege, T.; Ando, Y.; Tanno, R.; Shimoda, W.; Yanai, K. Image-based estimation of real food size for accurate food calorie estimation. In Proceedings of the 2019 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), San Jose, CA, USA, 28–30 March 2019; pp. 274–279. [Google Scholar]
  37. Puri, M.; Zhu, Z.; Yu, Q.; Divakaran, A.; Sawhney, H. Recognition and volume estimation of food intake using a mobile device. In Proceedings of the 2009 Workshop on Applications of Computer Vision (WACV), Snowbird, UT, USA, 7–8 December 2009; pp. 1–8. [Google Scholar]
  38. Zhu, F.; Bosch, M.; Woo, I.; Kim, S.; Boushey, C.J.; Ebert, D.S.; Delp, E.J. The use of mobile devices in aiding dietary assessment and evaluation. IEEE J. Sel. Top. Signal Process. 2010, 4, 756–766. [Google Scholar]
  39. Woo, I.; Otsmo, K.; Kim, S.; Ebert, D.S.; Delp, E.J.; Boushey, C.J. Automatic portion estimation and visual refinement in mobile dietary assessment. In Computational Imaging VIII; International Society for Optics and Photonics: Bellingham, WA, USA, 2010; Volume 7533, p. 75330O. [Google Scholar]
  40. Jia, W.; Yue, Y.; Fernstrom, J.D.; Yao, N.; Sclabassi, R.J.; Fernstrom, M.H.; Sun, M. Imaged based estimation of food volume using circular referents in dietary assessment. J. Food Eng. 2012, 109, 76–86. [Google Scholar] [CrossRef] [Green Version]
  41. Min, W.; Wang, Z.; Liu, Y.; Luo, M.; Kang, L.; Wei, X.; Wei, X.; Jiang, S. Large scale visual food recognition. arXiv 2021, arXiv:2103.16107. [Google Scholar]
  42. Aguilar, E.; Bolaños, M.; Radeva, P. Regularized uncertainty-based multi-task learning model for food analysis. J. Vis. Commun. Image Represent. 2019, 60, 360–370. [Google Scholar] [CrossRef]
  43. He, J.; Zhu, F. Online continual learning for visual food classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 2337–2346. [Google Scholar]
  44. Arpey, N.C.; Gaglioti, A.H.; Rosenbaum, M.E. How socioeconomic status affects patient perceptions of health care: A qualitative study. J. Prim. Care Community Health 2017, 8, 169–175. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Gianfrancesco, M.A.; Tamang, S.; Yazdany, J.; Schmajuk, G. Potential biases in machine learning algorithms using electronic health record data. JAMA Intern. Med. 2018, 178, 1544–1547. [Google Scholar] [CrossRef] [PubMed]
  46. Boeing, H. Nutritional epidemiology: New perspectives for understanding the diet-disease relationship? Eur. J. Clin. Nutr. 2013, 67, 424–429. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Ioannidis, J.P. The challenge of reforming nutritional epidemiologic research. JAMA 2018, 320, 969–970. [Google Scholar] [CrossRef] [PubMed]
  48. Kong, Y.W.; Baqar, S.; Jerums, G.; Ekinci, E.I. Sodium and its role in cardiovascular disease—The debate continues. Front. Endocrinol. 2016, 7, 164. [Google Scholar] [CrossRef] [Green Version]
  49. Investigators, S.; Dehghan, M.; Mente, A.; Zhang, X.; Swaminathan, S.; Li, W.; Mohan, V.; Iqbal, R.; Kumar, R.; Wentzel-Viljoen, E.; et al. Associations of fats and carbohydrate intake with cardiovascular disease and mortality in 18 countries from five continents (PURE): A prospective cohort study. Lancet 2017, 390, 2050–2062. [Google Scholar]
  50. Bodnar, L.M.; Cartus, A.R.; Kirkpatrick, S.I.; Himes, K.P.; Kennedy, E.H.; Simhan, H.N.; Grobman, W.A.; Duffy, J.Y.; Silver, R.M.; Parry, S.; et al. Machine learning as a strategy to account for dietary synergy: An illustration based on dietary intake and adverse pregnancy outcomes. Am. J. Clin. Nutr. 2020, 111, 1235–1243. [Google Scholar] [CrossRef]
  51. de Cos Juez, F.J.; Suárez-Suárez, M.; Lasheras, F.S.; Murcia-Mazón, A. Application of neural networks to the study of the influence of diet and lifestyle on the value of bone mineral density in post-menopausal women. Math. Comput. Model. 2011, 54, 1665–1670. [Google Scholar] [CrossRef]
  52. Zeng, J.; Zhang, J.; Li, Z.; Li, T.; Li, G. Prediction model of artificial neural network for the risk of hyperuricemia incorporating dietary risk factors in a Chinese adult study. Food Nutr. Res. 2020, 64, 3712. [Google Scholar] [CrossRef] [Green Version]
  53. Chew, E.Y. Age-related Macular Degeneration: Nutrition, Genes and Deep Learning—The LXXVI Edward Jackson Memorial Lecture. Am. J. Ophthalmol. 2020, 217, 335–347. [Google Scholar] [CrossRef]
  54. Puvanesarajah, S.; Hodge, J.M.; Evans, J.L.; Seo, W.; Yi, M.; Fritz, M.M.; Macheski-Preston, M.; Gansler, T.; Gapstur, S.M.; Gaudet, M.M. Unsupervised deep-learning to identify histopathological features among breast cancers in the Cancer Prevention Study-II Nutrition Cohort. Cancer Res. 2019, 79, 2417. [Google Scholar]
  55. Vivot, A.; Grégory, J.; Porcher, R. Application of Basic Epidemiologic Principles and Electronic Health Records in a Deep Learning Prediction Model. JAMA Dermatol. 2020, 156, 472–473. [Google Scholar] [CrossRef] [PubMed]
  56. Wong, T.Y.; Bressler, N.M. Artificial intelligence with deep learning technology looks into diabetic retinopathy screening. JAMA 2016, 316, 2366–2367. [Google Scholar] [CrossRef] [PubMed]
  57. Byeon, H. Is Deep Learning Better than Machine Learning to Predict Benign Laryngeal Disorders? Int. J. Adv. Comput. Sci. Appl. 2021, 12, 112–117. [Google Scholar] [CrossRef]
  58. Xiong, H.; Lin, P.; Yu, J.G.; Ye, J.; Xiao, L.; Tao, Y.; Jiang, Z.; Lin, W.; Liu, M.; Xu, J.; et al. Computer-aided diagnosis of laryngeal cancer via deep learning based on laryngoscopic images. EBioMedicine 2019, 48, 92–99. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  59. VoPham, T.; Hart, J.E.; Laden, F.; Chiang, Y.Y. Emerging trends in geospatial artificial intelligence (geoAI): Potential applications for environmental epidemiology. Environ. Health 2018, 17, 40. [Google Scholar] [CrossRef] [PubMed]
  60. Hoffmann, K.; Schulze, M.B.; Schienkiewitz, A.; Nöthlings, U.; Boeing, H. Application of a new statistical method to derive dietary patterns in nutritional epidemiology. Am. J. Epidemiol. 2004, 159, 935–944. [Google Scholar] [CrossRef] [PubMed]
  61. Zhang, F.; Tapera, T.M.; Gou, J. Application of a new dietary pattern analysis method in nutritional epidemiology. BMC Med. Res. Methodol. 2018, 18, 119. [Google Scholar] [CrossRef] [PubMed]
  62. Santos, R.d.O.; Gorgulho, B.M.; Castro, M.A.d.; Fisberg, R.M.; Marchioni, D.M.; Baltar, V.T. Principal component analysis and factor analysis: Differences and similarities in nutritional epidemiology application. Rev. Bras. Epidemiol. 2019, 22, e190041. [Google Scholar] [CrossRef] [Green Version]
  63. Falissard, L.; Fagherazzi, G.; Howard, N.; Falissard, B. Deep clustering of longitudinal data. arXiv 2018, arXiv:1802.03212. [Google Scholar]
  64. Wang, Y.; Yao, H.; Zhao, S. Auto-encoder based dimensionality reduction. Neurocomputing 2016, 184, 232–242. [Google Scholar] [CrossRef]
  65. Kwon, Y.J.; Kim, H.S.; Jung, D.H.; Kim, J.K. Cluster analysis of nutritional factors associated with low muscle mass index in middle-aged and older adults. Clin. Nutr. 2020, 39, 3369–3376. [Google Scholar] [CrossRef] [PubMed]
  66. Walter, S.; Tiemeier, H. Variable selection: Current practice in epidemiological studies. Eur. J. Epidemiol. 2009, 24, 733–736. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  67. Altmann, A.; Toloşi, L.; Sander, O.; Lengauer, T. Permutation importance: A corrected feature importance measure. Bioinformatics 2010, 26, 1340–1347. [Google Scholar] [CrossRef] [PubMed]
  68. Zeevi, D.; Korem, T.; Zmora, N.; Israeli, D.; Rothschild, D.; Weinberger, A.; Ben-Yacov, O.; Lador, D.; Avnit-Sagi, T.; Lotan-Pompan, M.; et al. Personalized nutrition by prediction of glycemic responses. Cell 2015, 163, 1079–1094. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  69. Dipnall, J.F.; Pasco, J.A.; Berk, M.; Williams, L.J.; Dodd, S.; Jacka, F.N.; Meyer, D. Fusing data mining, machine learning and traditional statistics to detect biomarkers associated with depression. PLoS ONE 2016, 11, e0148195. [Google Scholar] [CrossRef] [Green Version]
  70. Russo, S.; Li, G.; Villez, K. Automated model selection in principal component analysis: A new approach based on the cross-validated ignorance score. Ind. Eng. Chem. Res. 2019, 58, 13448–13468. [Google Scholar] [CrossRef]
  71. Trepanowski, J.F.; Ioannidis, J.P. Perspective: Limiting dependence on nonrandomized studies and improving randomized trials in human nutrition research: Why and how. Adv. Nutr. 2018, 9, 367–377. [Google Scholar] [CrossRef]
  72. Brisk, R.; Bond, R.; Finlay, D.; McLaughlin, J.; Piadlo, A.; Leslie, S.J.; Gossman, D.E.; Menown, I.B.; McEneaney, D.J.; Warren, S. The effect of confounding data features on a deep learning algorithm to predict complete coronary occlusion in a retrospective observational setting. Eur. Heart J.-Digit. Health 2021, 2, 127–134. [Google Scholar] [CrossRef]
  73. Badgeley, M.A.; Zech, J.R.; Oakden-Rayner, L.; Glicksberg, B.S.; Liu, M.; Gale, W.; McConnell, M.V.; Percha, B.; Snyder, T.M.; Dudley, J.T. Deep learning predicts hip fracture using confounding patient and healthcare variables. NPJ Digit. Med. 2019, 2, 31. [Google Scholar] [CrossRef] [Green Version]
  74. García, S.; Ramírez-Gallego, S.; Luengo, J.; Benítez, J.M.; Herrera, F. Big data preprocessing: Methods and prospects. Big Data Anal. 2016, 1, 9. [Google Scholar] [CrossRef] [Green Version]
  75. Kotsiantis, S.B.; Kanellopoulos, D.; Pintelas, P.E. Data preprocessing for supervised leaning. Int. J. Comput. Sci. 2006, 1, 111–117. [Google Scholar]
  76. Lakshminarayan, K.; Harp, S.A.; Goldman, R.P.; Samad, T. Imputation of Missing Data Using Machine Learning Techniques. In Proceedings of the KDD, Portland, OR, USA, 2–4 August 1996; Volume 96. [Google Scholar]
  77. Richman, M.B.; Trafalis, T.B.; Adrianto, I. Missing data imputation through machine learning algorithms. In Artificial Intelligence Methods in the Environmental Sciences; Springer: Berlin/Heidelberg, Germany, 2009; pp. 153–169. [Google Scholar]
  78. Batista, G.E.; Monard, M.C. An analysis of four missing data treatment methods for supervised learning. Appl. Artif. Intell. 2003, 17, 519–533. [Google Scholar] [CrossRef]
  79. Jerez, J.M.; Molina, I.; García-Laencina, P.J.; Alba, E.; Ribelles, N.; Martín, M.; Franco, L. Missing data imputation using statistical and machine learning methods in a real breast cancer problem. Artif. Intell. Med. 2010, 50, 105–115. [Google Scholar] [CrossRef]
  80. Al-Milli, N.; Almobaideen, W. Hybrid neural network to impute missing data for IoT applications. In Proceedings of the 2019 IEEE Jordan International Joint Conference on Electrical Engineering and Information Technology (JEEIT), Amman, Jordan, 9–11 April 2019; pp. 121–125. [Google Scholar]
  81. Heaton, J. An empirical analysis of feature engineering for predictive modeling. In Proceedings of the SoutheastCon 2016, Amman, Jordan, 9–11 April 2016; pp. 1–6. [Google Scholar]
  82. Morgenstern, J.D.; Rosella, L.C.; Costa, A.P.; Anderson, L.N. Development of Machine Learning Prediction Models to Explore Nutrients Predictive of Cardiovascular Disease Using Canadian Linked Population-Based Data. Appl. Physiol. Nutr. Metab. 2022. [Google Scholar] [CrossRef]
  83. Russo, S.; Besmer, M.D.; Blumensaat, F.; Bouffard, D.; Disch, A.; Hammes, F.; Hess, A.; Lürig, M.; Matthews, B.; Minaudo, C.; et al. The value of human data annotation for machine learning based anomaly detection in environmental systems. Water Res. 2021, 206, 117695. [Google Scholar] [CrossRef]
  84. Sheng, V.S.; Provost, F.; Ipeirotis, P.G. Get another label? Improving data quality and data mining using multiple, noisy labelers. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 614–622. [Google Scholar]
  85. Gudivada, V.; Apon, A.; Ding, J. Data quality considerations for big data and machine learning: Going beyond data cleaning and transformations. Int. J. Adv. Softw. 2017, 10, 1–20. [Google Scholar]
  86. Wang, Q.; Ma, Y.; Zhao, K.; Tian, Y. A comprehensive survey of loss functions in machine learning. Ann. Data Sci. 2020, 9, 187–212. [Google Scholar] [CrossRef]
  87. Tran, G.S.; Nghiem, T.P.; Nguyen, V.T.; Luong, C.M.; Burie, J.C. Improving accuracy of lung nodule classification using deep learning with focal loss. J. Healthc. Eng. 2019, 2019, 5156416. [Google Scholar] [CrossRef]
  88. Rodriguez, J.D.; Perez, A.; Lozano, J.A. Sensitivity analysis of k-fold cross validation in prediction error estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 569–575. [Google Scholar] [CrossRef]
  89. Holzinger, A.; Biemann, C.; Pattichis, C.S.; Kell, D.B. What do we need to build explainable AI systems for the medical domain? arXiv 2017, arXiv:1712.09923. [Google Scholar]
  90. Gunning, D.; Stefik, M.; Choi, J.; Miller, T.; Stumpf, S.; Yang, G.Z. XAI—Explainable artificial intelligence. Sci. Robot. 2019, 4, eaay7120. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  91. Kendall, A.; Gal, Y. What uncertainties do we need in bayesian deep learning for computer vision? In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  92. Batterham, M.; Neale, E.; Martin, A.; Tapsell, L. Data mining: Potential applications in research on nutrition and health. Nutr. Diet. 2017, 74, 3–10. [Google Scholar] [CrossRef] [PubMed]
  93. Davis, J.; Goadrich, M. The relationship between Precision-Recall and ROC curves. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 233–240. [Google Scholar]
  94. Schelter, S.; Biessmann, F.; Januschowski, T.; Salinas, D.; Seufert, S.; Szarvas, G. On Challenges in Machine Learning Model Management. 2018. Available online: http://sites.computer.org/debull/A18dec/p5.pdf (accessed on 17 March 2022).
  95. Diebolt, V.; Azancot, I.; Boissel, F.H.; Adenot, I.; Balague, C.; Barthelemy, P.; Boubenna, N.; Coulonjou, H.; Fernandez, X.; Habran, E.; et al. “Artificial intelligence”: Which services, which applications, which results and which development today in clinical research? Which impact on the quality of care? Which recommendations? Therapies 2019, 74, 155–164. [Google Scholar] [CrossRef] [PubMed]
  96. Feng, J.Z.; Wang, Y.; Peng, J.; Sun, M.W.; Zeng, J.; Jiang, H. Comparison between logistic regression and machine learning algorithms on survival prediction of traumatic brain injuries. J. Crit. Care 2019, 54, 110–116. [Google Scholar] [CrossRef] [PubMed]
  97. Mortazavi, B.J.; Downing, N.S.; Bucholz, E.M.; Dharmarajan, K.; Manhapra, A.; Li, S.X.; Negahban, S.N.; Krumholz, H.M. Analysis of machine learning techniques for heart failure readmissions. Circ. Cardiovasc. Qual. Outcomes 2016, 9, 629–640. [Google Scholar] [CrossRef] [Green Version]
  98. Nusinovici, S.; Tham, Y.C.; Yan, M.Y.C.; Ting, D.S.W.; Li, J.; Sabanayagam, C.; Wong, T.Y.; Cheng, C.Y. Logistic regression was as good as machine learning for predicting major chronic diseases. J. Clin. Epidemiol. 2020, 122, 56–69. [Google Scholar] [CrossRef]
  99. Gravesteijn, B.Y.; Nieboer, D.; Ercole, A.; Lingsma, H.F.; Nelson, D.; Van Calster, B.; Steyerberg, E.W.; Åkerlund, C.; Amrein, K.; Andelic, N.; et al. Machine learning algorithms performed no better than regression models for prognostication in traumatic brain injury. J. Clin. Epidemiol. 2020, 122, 95–107. [Google Scholar] [CrossRef]
  100. Rosso, N.; Giabbanelli, P. Accurately inferring compliance to five major food guidelines through simplified surveys: Applying data mining to the UK National Diet and Nutrition Survey. JMIR Public Health Surveill. 2018, 4, e9536. [Google Scholar] [CrossRef]
  101. Riboli, E.; Hunt, K.; Slimani, N.; Ferrari, P.; Norat, T.; Fahey, M.; Charrondiere, U.; Hemon, B.; Casagrande, C.; Vignat, J.; et al. European Prospective Investigation into Cancer and Nutrition (EPIC): Study populations and data collection. Public Health Nutr. 2002, 5, 1113–1124. [Google Scholar] [CrossRef]
  102. Sak, J.; Suchodolska, M. Artificial Intelligence in Nutrients Science Research: A Review. Nutrients 2021, 13, 322. [Google Scholar] [CrossRef] [PubMed]
  103. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
Table 1. Summary of the main machine-learning categories.
Table 1. Summary of the main machine-learning categories.
Learning TypeTechniqueModels
SupervisedClassificationRandom Forest, Naïve Bayes,
Support Vector Machine,
k-Nearest Neighbour, ANN
RegressionLinear Regression
Logistic Regression
Random Forest, ANN
UnsupervisedFeature extractionPCA
Deep Autoencoders
Manifold Learning
ClusteringGaussian Mixture Models
k-Means
Deep Neural Networks
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Russo, S.; Bonassi, S. Prospects and Pitfalls of Machine Learning in Nutritional Epidemiology. Nutrients 2022, 14, 1705. https://doi.org/10.3390/nu14091705

AMA Style

Russo S, Bonassi S. Prospects and Pitfalls of Machine Learning in Nutritional Epidemiology. Nutrients. 2022; 14(9):1705. https://doi.org/10.3390/nu14091705

Chicago/Turabian Style

Russo, Stefania, and Stefano Bonassi. 2022. "Prospects and Pitfalls of Machine Learning in Nutritional Epidemiology" Nutrients 14, no. 9: 1705. https://doi.org/10.3390/nu14091705

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop