Next Article in Journal
Formation and Expansion of Memory B Cells against Coronavirus in Acutely Infected COVID-19 Individuals
Next Article in Special Issue
Engineered Cell Line Imaging Assay Differentiates Pathogenic from Non-Pathogenic Bacteria
Previous Article in Journal
Role of Cellular Metabolism during Candida-Host Interactions
Previous Article in Special Issue
Beyond the List: Bioagent-Agnostic Signatures Could Enable a More Flexible and Resilient Biodefense Posture Than an Approach Based on Priority Agent Lists Alone
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparison of Infectious Disease Forecasting Methods across Locations, Diseases, and Time

by
Samuel Dixon
1,
Ravikiran Keshavamurthy
1,2,
Daniel H. Farber
1,
Andrew Stevens
1,
Karl T. Pazdernik
1,3 and
Lauren E. Charles
1,2,*
1
Pacific Northwest National Laboratory, Richland, WA 99354, USA
2
Paul G. Allen School for Global Health, Washington State University, Pullman, WA 99164, USA
3
Department of Statistics, North Carolina State University, Raleigh, NC 27695, USA
*
Author to whom correspondence should be addressed.
Pathogens 2022, 11(2), 185; https://doi.org/10.3390/pathogens11020185
Submission received: 20 December 2021 / Revised: 23 January 2022 / Accepted: 27 January 2022 / Published: 29 January 2022
(This article belongs to the Special Issue Advances in Biosurveillance for Human, Animal, and Plant Health)

Abstract

:
Accurate infectious disease forecasting can inform efforts to prevent outbreaks and mitigate adverse impacts. This study compares the performance of statistical, machine learning (ML), and deep learning (DL) approaches in forecasting infectious disease incidences across different countries and time intervals. We forecasted three diverse diseases: campylobacteriosis, typhoid, and Q-fever, using a wide variety of features (n = 46) from public datasets, e.g., landscape, climate, and socioeconomic factors. We compared autoregressive statistical models to two tree-based ML models (extreme gradient boosted trees [XGB] and random forest [RF]) and two DL models (multi-layer perceptron and encoder–decoder model). The disease models were trained on data from seven different countries at the region-level between 2009–2017. Forecasting performance of all models was assessed using mean absolute error, root mean square error, and Poisson deviance across Australia, Israel, and the United States for the months of January through August of 2018. The overall model results were compared across diseases as well as various data splits, including country, regions with highest and lowest cases, and the forecasted months out (i.e., nowcasting, short-term, and long-term forecasting). Overall, the XGB models performed the best for all diseases and, in general, tree-based ML models performed the best when looking at data splits. There were a few instances where the statistical or DL models had minutely smaller error metrics for specific subsets of typhoid, which is a disease with very low case counts. Feature importance per disease was measured by using four tree-based ML models (i.e., XGB and RF with and without region name as a feature). The most important feature groups included previous case counts, region name, population counts and density, mortality causes of neonatal to under 5 years of age, sanitation factors, and elevation. This study demonstrates the power of ML approaches to incorporate a wide range of factors to forecast various diseases, regardless of location, more accurately than traditional statistical approaches.

1. Introduction

Infectious diseases have a significant negative impact on veterinary and public health, both at the regional and global levels. The emergence of novel infectious agents, the reemergence of infectious agents previously under control, the constant burden of the endemic pathogens, and the development of antimicrobial resistance have complicated control and prevention initiatives. The threat posed by these diseases varies widely in terms of mortalities, morbidity, social, and economic disruptions. These threats are further magnified by anthropogenic and ecological factors, such as rapidly increasing population, globalization, urbanization, climate change, administrative conflicts, weak health systems, and the changing nature of pathogen transmission between human and animal populations [1]. Additionally, the wide variety of potential causes for such disease occurrences has made preparedness and timely response a challenge. Due to these reasons and despite a significant improvement in control and prevention efforts over the past few decades, infectious diseases continue to pose a major challenge, causing millions of deaths each year worldwide, especially in low-income countries [2].
Advanced knowledge of disease outbreaks can help public health decision-makers prevent, mitigate, and reduce infections and epidemic severity. Infectious disease forecasting aims at synthesizing epidemiological information related to the disease occurrence in advance for better allocating public health resources at appropriate places and times. Recent disease forecasting methods have taken advantage of the increasing abundance of publicly accessible data and advanced algorithms for better prediction of infectious disease outbreaks. These open access data streams are routinely updated, cover a wide variety of topics, and are increasingly detailed and complete, which could be used to increase the accuracy of the forecasting models [3,4].
Time series models have a long history of being used for disease forecasting. Traditionally, disease forecasting has relied on statistical techniques, such as autoregressive integrated moving average with exogenous variables (ARIMAX) and generalized linear autoregressive moving average models (GLARMA). These models are widely popular because they can be easily implemented and interpreted in retrospective studies. However, they can only process a limited amount of feature data and complexity to generate disease outcomes. Machine learning (ML) models, which have now been applied to a wide variety of time series regression tasks [5,6,7], are being used in disease forecasting to detect cryptic patterns arising from interactions between multiple features, which are difficult, and impossible at times, to uncover with conventional statistical methods. ML-based ensemble models, such as random forest (RF) and extreme gradient boost (XGB), are particularly popular in healthcare research because of their ease of implementation and interpretation [8,9]. These models are based on a decision tree approach that repeatedly splits the regression space non-linearly and can be visualized in the form of a tree. The RF uses an ensemble of hundreds of trees that are generated by bagging (random selection of a subset of data) and random resampling of a small number of features, while XGB makes use of a gradient boosting framework where trees are added sequentially to make predictions. These models have been successfully used in forecasting infectious diseases, such as infectious diarrhea [10], influenza [11,12], brucellosis [13,14], and dengue [15].
Deep learning (DL) models are ML models consisting of a minimum of one hidden layer of nodes between the input and output layer. A fully connected multi-layer perceptron (MLP) is the simplest architecture DL model, comprised of two or more layers of nodes, all of which are unidirectionally connected. A sequence-to-sequence encoder–decoder model (Enc–Dec) is a more complex architecture DL model, which consists of two subsections: the first subsection, called the encoder, translates a variable-length source input into a numerical embedding, and the second subsection, called the decoder, converts this numerical representation back to an output of the original form [16]. By creating this numerical vector representation as an intermediate, the Enc–Dec has been shown to better incorporate both short-term and long-term temporal features into forecasts [17]. DL models have outperformed ML models in several tasks, including time-series predictions [18]. However, DL models’ increase in predictive power relative to simpler ML models comes at a serious cost. DL algorithms are not readily interpretable, although there are recent advances in developing such techniques [19]. These models generally require more computational resources and experience to train [20,21] as well as more training data to generalize well [22].
In this study, we compare the performance of ML, DL, and statistical models to forecast three human infectious diseases: typhoid fever, Q-fever, and campylobacteriosis, across multiple geographic regions within Australia, Israel, and the United States. These diseases were selected because they are internationally notifiable diseases with disease cycles and prevalence rates that vary across regions and time, providing a broad range of scenarios to test the performance of the forecasting models. We demonstrate the utility of a traditional statistical time series model (GLARMA), ML regression trees (RF and XGB), and DL models (MLP and Enc–Dec) in forecasting diseases, thereby providing insights into how the models perform given different disease lifecycles and incidence across different geographic landscapes and climates as well as region-specific population demography and socioeconomic factors. Finally, we compare the performance of these disease models to accurately predict case counts at different times, from one month to eight months into the future.

2. Results

2.1. Datasets

The mean infections of campylobacteriosis in 2009–2018 by month and region was greatest in Australia (i.e., 190 cases), nearly five times the next highest mean infection counts in the United States and Israel (Table 1). In addition, there were many more campylobacteriosis cases on average per month than either Q-fever or typhoid (range 0–4 mean cases/month) with Australia having the largest mean number of case counts for all of the explored diseases. There were more cases of Q-fever than typhoid in Australia and Israel, which was reversed in the US. None of the target countries showed a consistent seasonal disease pattern over the full time period, which increases the difficulty of forecasting. However, the differences observed in case counts and disease life cycles, as well as country features and incidence, enabled this study to investigate the performance advantages and disadvantages of the given forecasting methods under a variety of conditions.
All ML models were trained using the 2009–2017 data for all seven countries where available (Figure 1, Figure 2 and Figure 3). However, there was missing or erroneous data in 2018 for at least one disease in Finland, Japan, Norway, and Sweden. Therefore, models were evaluated for prediction performance on 2018 regional data only in Australia, Israel, and the United States. Australia is the only country where the full dataset of cases for all three diseases is present for the full time period. Israel and the United States are missing early years for campylobacteriosis, and Israel is also missing some for Q-fever. For all regions except Australia, there is a visual periodicity in average regional case counts of campylobacteriosis between 2009–2018 with a spike towards the end of each year (Figure 1). For Q-fever and typhoid, the annual pattern is less clear in part because of the low number of case counts (Figure 2 and Figure 3).

2.2. Overall Model Performance

The tree-based ML models (e.g., RF and XGB) were most accurate in forecasting campylobacteriosis, typhoid, and Q-fever than the DL or statistical models (Figure 4). The Alt-XGB (i.e., the XGB model without including country:region as a feature) outperformed all models for all diseases in MAE, RMSE, and Poisson deviance. On the other end of the spectrum, the MLP models performed the worst for all diseases and performance metrics. The statistical GLARMA model and DL Enc–Dec produced similar results across diseases and metrics, except MAE in typhoid where GLARMA performed the worst. Compared to the tree-based models, these models were in line for typhoid and minutely worse for the other diseases. The random forest models (RF and Alt-RF) had nearly identical performance across diseases and were only a few tenths away from the XGB models. The relative scale of the performance metrics is different between diseases; since the case counts are much higher all around for campylobacteriosis, the values ranged from 20.3–78.3, whereas the values for Q-fever (range 0.3–1.8) and typhoid (range 0.2–0.8) are much lower. Note that for the statistical time series models, the GLARMA model was fit with the negative binomial distribution and parameter estimation converged in all regions. For the tree-based ML models, the standard average pooling provided the best results out of all pooling methods attempted and, therefore, results are only shown using that method.

2.3. Model Performance by Country

The performance of each model by country and disease had similar overall performance trends regardless of metrics (Figure 5). For Australia, the MLP model was by far the worst performer. For campylobacteriosis and Q-fever, the tree-based models performed the best, notably the XGB (Alt and not) were the top performers. For typhoid, however, while XGB was best when assessed by MAE, the GLARMA model outperformed all other methods when assessed by RMSE and deviance. For the United States, the model performance results were all very close regardless of disease. Model performances were essentially the same for Q-fever; the tree-based models, namely RF, were slightly better for typhoid; and both Alt-XGB and GLARMA were slightly better for campylobacteriosis. For Israel, all models performed the same for typhoid. XGB slightly outperformed Alt-XGB as the top model for campylobacteriosis in Israel. Similarly, Q-fever performance in Israel depends on the metric used with Enc–Dec slightly better than Alt-XGB by MAE, GLARMA slightly better than Alt-XGB by RMSE, and Alt-XGB slightly better than Enc–Dec by deviance.

2.4. Model Performance by Time Interval

The performance of the models by time interval (i.e., nowcast [1 month], short-term forecast [2–5 months], and long-term forecast [6–8 months]) varied by disease and location (Table 2; Figure 6). For nowcasting overall, XGB-based models performed the best for campylobacteriosis, RF-based for Q-fever, and Enc–Dec for typhoid. In general, for short-term forecast best models, XGB-based models were the best for campylobacteriosis, and all the tree-based models were comparable for Q-fever, whereas XGB (Alt and not) and/or GLARMA were best for typhoid based on metric. Namely, for typhoid, XGB and GLARMA performed similarly for deviance, Alt-XGB was best for MAE, and GLARMA for RMSE. For overall long-term forecast performance, there were more discrepancies in performance metrics. For campylobacteriosis, MAE and RMSE show the best models were RF (Alt and not) and Alt-XGB, whereas deviance favors nearly all models except MLP with Alt-XGB being minutely better. For Q-fever, Alt-XGB and XGB are comparable by MAE and RMSE with Alt-XGB performing the best by deviance. MLP, which performed the worst in any other circumstance, performed the best for long-term typhoid forecasting using all three metrics. When comparing relative performance across the time scales for the best performing model by disease, there are not any consistent patterns (Figure 6). For campylobacteriosis, short-term forecasting is most accurate followed by nowcasting, and then long-term forecasts. For Q-fever, the order of best to least performance is by time, i.e., nowcast, short-term forecasts, and then long-term forecasts. Typhoid, again different, has the best performance with nowcasting, then long-term forecasting, and lastly, short-term forecasting. In general, error bars are larger for long-term vs. short-term forecasting.

2.5. Model Performance by Disease Incidence

In 2018, across all regions in Australia, Israel, and the United States, there were a wide range of case counts from 0 to nearly 8000 per month. More specifically, the monthly case count ranges were 0 (25 regions) to 7742 (Queensland, Australia) for campylobacteriosis; 0 (10 regions) to 19 (Victoria, Australia) for typhoid; and 0 (62 regions) to 78 (Queensland, Australia) for Q-fever. The disease forecasting model performance increased with a decreasing number of cases, i.e., all models for the 20 regions with the lowest case counts outperformed those 20 regions with the highest case counts (Figure 7). For the regions with the lowest case counts, all models correctly predicted 0 case counts for typhoid and Q-fever, except GLARMA. However, for the lowest campylobacteriosis case counts, all models forecasted positive case counts in at least 15 of the 20 regions without a case count, with Alt-XGB and GLARMA performing the best and MLP performing the worst. For the highest case count regions, tree-based models (especially Alt-XGB) are optimal for campylobacteriosis and Q-fever, but GLARMA performs the best for typhoid. Overall, the DL models were the worst performers with regions that had high case counts. It is interesting to note that for counties that never had the disease, all models correctly predicted zero.

2.6. Summary of Top Model Performance by Disease

Overall, the top-performing models for forecasting campylobacteriosis, Q-fever, and typhoid vary by the way the resulting metrics are split (Table 3). For example, overall, the top-performing model for any disease, location, or forecasted month is Alt-XGB. However, when looking specifically at different data splits, other models outperformed or equally performed as well as Alt-XGB.

2.7. Feature Importance

The relative importance of the 46 explanatory factors varied only slightly in each tree-based ML disease model (Figure 8). Greater than 90% of all model performance based on these features can be contributed (in descending order) to previous case counts, country-region for non-Alt models, population counts, population density, mortality of neonatal to under 5 years of age, and sanitation with elevation also included for Alt models. In general, RF (Alt and not) models tend to weigh the previous case counts most importantly for all diseases, whereas XGB equally uses location when predicting campylobacteriosis and almost five times as often as case counts for typhoid predictions. Whether or not the location is present, the top 20 features remain the same for all models with slight order variation depending on the exact model and disease.

3. Discussion

Infectious diseases have the potential to cause significant veterinary, public health, and socioeconomic losses throughout the world. Hence, prior intelligence regarding their occurrences is crucial for their timely control and prevention. To the best of our knowledge, this is the first study comparing and forecasting the performance of ML, DL, and statistical models to forecast three notifiable human infectious diseases; typhoid, Q-fever, and campylobacteriosis for region-level predictions for three countries over three continents.
Datasets. Since ML and DL models perform better with more training data, we used disease case count and feature data from seven countries to train our ML and DL models, while performing prediction on only the three countries with complete datasets. While all available data for the four other countries was used in training, none of these countries had consistent case count reporting and, in particular, had gaps in the forecasted year, 2018. By augmenting with empirical data, we boosted our training dataset numbers with observations of the diseases under different feature conditions without having to use simulation or imputation techniques that may or may not represent reality. The result was reasonably large training sets for each disease, with the largest set for campylobacteriosis, then typhoid, and lastly Q-fever.
Model performance. Averaged across all countries, months, and regions over eight months, the best ML forecasts outperformed conventional statistical approaches (i.e., GLARMA) by MAE, RMSE, and Poisson deviance for every disease. Comparing between the ML models, the tree-based models (i.e., XGB and RF) outperformed the DL models (i.e., Dec–Enc and MLP) by all metrics as well, with the XGB-based models routinely providing the best performance. Averaging performance within different subsets of the data (by country, time interval, or disease incidence) sometimes provided conflicting results on which methodology performed best. However, a general pattern was observed, with the number of conflicting metric results increasing as disease incidence decreased. In only one instance each, regardless of the metric, the GLARMA model, for regions with the highest incidence of typhoid, and the Enc–Dec, for nowcasting of typhoid, performed the best. But even in these scenarios, the difference in error metrics between these models and the best tree-based ML model was extremely small. In addition, the corresponding error metrics were so low in magnitude that the practical implications of using one model over the other are minimal. In summary, if a single methodology is preferred for simplicity, our results suggest that XGB-based models are the best option across the study diseases and locations.
Model comparison. Tree-based ML models and statistical approaches are interpretable in terms of the relative importance each explanatory variable plays in prediction. However, the ML models each demonstrate significant advantages over the statistical approach. These models can incorporate more exogenous variables and more complex interactions between variables than the univariate GLARMA, which resulted in improved predictions for the ML models. While the GLARMA model can incorporate some explanatory variables and their interactions, they require manual construction and pruning, which can become unmanageable as the number of features grows large. While DL models can also accommodate many explanatory variables, they generally performed the worst and the importance of the variables are not interpretable. The DL approaches here could have suffered from overfitting due to a variety of potential causes, most common being learning the noise of the training data [23]. DL models generally perform better with more data and, therefore, the Enc–Dec model may be able to outperform the ML approaches with more training data and/or greater temporal and/or spatial extents of the data. However, simple interpretation of the model via variable importance still remains a benefit of the ML tree-based models over the DL models. Given the evidence of DL models outperforming ML in other applications [24] and recent developments in hierarchical DL models, in which basal features are learned and then inform higher-level classification [25], there remains the potential for hierarchical DL to improve performance in disease forecasting while maintaining interpretability.
Feature importance. Across all diseases and models, the most significant feature contributing to tree-based ML model forecasting is previous case counts, which emphasizes the criticality of timely, reliable, and accurate disease surveillance data. The date, number of months forecasting out, and country-region location are all very accessible information that also play key roles in prediction. While the country-region location was critical to the non-Alt models, the fact that the Alt models did, on average, perform better suggests that the other features that were included were sufficient to predict future case counts. Notably, when the country-region location had a high degree of importance, it often was accompanied by a reduction in relative importance of previous case counts. These features are often correlated, as certain regions historically have a higher or lower incidence of different diseases. Comparing the RF and XGB models, country/region was more important to the XGB models which could be due to its greedy iterative approach of tree construction. At a high-level, the next most important features for all diseases included population factors, neonatal mortality, sanitation metrics, and water resources. The other major contributors identified are climate, seasonality, livestock consumption, GDP, and waste. However, it is important to note that including correlated features splits the importance between these features potentially decreasing their true importance and making the final interpretation of results challenging. By grouping the individual features together at a high-level, we hope to overcome this issue to some extent.
Campylobacteriosis, being one of the most common infectious disease of humans globally, is mainly transmitted through contaminated food and water [26]. Similarly, typhoid is an important public health concern transmitted largely via sewage contamination of food or water sources. Given the preponderance of food and water borne transmission of these diseases, it is not surprising that the most important features in forecasting, aside from case counts, were largely related to population factors and sanitation metrics. Furthermore, the typhoid tree-based models in comparison with Q-fever or campylobacteriosis tree-based models relied less on the previous case counts relative to other factors, suggesting that the typhoid model learned better from these additional features. Q-fever is an important zoonotic disease, which has the potential to cause large outbreaks in humans and livestock [27]. Q-fever has a complex disease cycle compared to other two diseases in our study and a range of risk factors associated with it. Additionally, the disease has been often neglected by health personnel, resulting in a poor understanding of its epidemiology [28,29]. Due to these reasons, Q-fever forecasting is a challenge, even in countries with good disease surveillance infrastructure. In our study, the previous case counts dominated the most important features in forecasting Q-fever, contrasted with the relatively wide range of contributing explanatory variables in forecasting typhoid and campylobacteriosis. Nevertheless, population, seasonality, health, and sanitation metrics were still among the most important features. Additionally, the majority of the disease forecasting models were more accurate when country/region was not incorporated (Alt model), suggesting the models were learning features of the regions rather than the regions themselves.
This study has several strengths as it represents a significant advancement in disease forecasting in a few key respects. Firstly, this study engineered a single data set containing 46 explanatory variables spanning seven different countries both at the regional spatial scale and monthly temporal scale which could be used to create and evaluate forecasts for almost any disease and model type. Secondly, we applied several different ML models (RF, XGB, Enc–Dec, and MLP) and a statistical model (GLARMA) using the same feature inputs for three internationally notifiable diseases (campylobacteriosis, Q-fever, and typhoid), which have highly divergent epidemic processes and incidences. These models were compared to each other using three metrics, MAE, RMSE, and Poisson deviance, which provided slightly differing analyses of each model’s performance. Thirdly, we used an interpretable ML approach to determine the relative importance of each explanatory variable in predicting future incidences of each of the three diseases. The most important features were retrieved from the tree-based ML models with and without one-hot encoded country/region vectors (RF and XGB versus Alt-RF and Alt-XGB) to provide a perspective of which data are most predictive with and without knowledge of specific country/regions. The study also had many limitations. Many data sets were only available for a subset of the locations studied and were, therefore, not incorporated into the model. As more data becomes available, we expect the models we examined here could incorporate them readily and improve the performance. Additionally, these models could be applied to additional locations and at finer spatiotemporal resolutions. While this data set is novel in the breadth of the explanatory variables considered, these disease forecasting models were limited by the available biosurveillance data and would benefit greatly from improved diagnostics and reporting of each disease as there is relatively limited labeled data to build a training dataset. There is very likely a bias undercounting disease incidence in rural and impoverished regions with less access to the most accurate diagnostic tools [30,31]. This bias and underreporting underscore the countries included in this study, which are higher in many socioeconomic metrics as compared with the median country and compared with the countries with the highest incidences of these diseases. Additionally, a future direction would be to consider a spatial component, as nearby regions appear to potentially be linked.
In conclusion, we have demonstrated the efficacy of forecasting campylobacteriosis, Q-fever, and typhoid, three internationally notifiable diseases of significant public health impact, using ML, DL, and statistical models. Overall, we found the tree-based ML models to be more accurate in forecasting compared to DL and statistical autoregressive models. Previous case counts were among the most important features in forecasting all diseases, with population density and factors relating to clean water and sanitation also very important across diseases. This study demonstrates the application of ML approaches in forecasting disease, and the wide range of data that can be incorporated into such models. This work has the potential to greatly improve the lives of people living in these areas and the safety of travelers. These ML models, particularly the XGB model, can offer a significant improvement in predicting emerging outbreaks over traditional surveillance methods, as its predictions can be updated rapidly by retraining the model in a matter of hours as digital data streams are updated [32]. Further use of ML models may improve our ability to mitigate future epidemics and may aid our understanding of which combinations of factors are likely to lead to outbreaks.

4. Methods

4.1. Data Collection

This study focused on three human infectious diseases: campylobacteriosis, Q-fever, and typhoid. Case count data of each disease from 2009–2018 were collected from EpiArchive [33]. The regional data, across seven countries from 2009–2017, were used to build the models and the final model performances were evaluated on available 2018 data, which was consistent only from Australia, Israel, and United States through August (Table 4).
Identification of potential explanatory variables was informed by a literature review of the diseases. Only publicly accessible data available for all countries of interest were considered and collected (Table 5). However, some variables contained missing data for some time periods within a region of interest. These data were imputed, as described in Section 4.3.

4.2. Data Engineering

Relevant features were extracted from each data source for the countries of interest. This raw data contained different file types, organization, spatial extents, spatial resolutions, temporal extents, and temporal resolutions. Data were resampled to a uniform monthly temporal and regional-level spatial resolution using GDAL [34]. The ML algorithms required tabular data without missing values. In instances where feature data were not continuously available every month through 2017, missing values were linearly extrapolated from the last date available [35].
Separate training data matrices were engineered for each disease. To build these matrices, all feature data types were combined into a single data set, in which each row denoted a single disease/location/region/time bin with corresponding explanatory variables. To enable the use of region-country as a feature in ML algorithms, each region-country was one-hot encoded, i.e., the categorical variable is represented by several columns equal to the number of entries; the rows corresponding to a particular entry contain a “1” in that entry’s column and a “0” in all other entry columns [36].

4.3. Data Analysis

A separate model was created in python for each disease for each algorithm type: moving average models, tree-based ML models, and DL models, described below. The response variable of each model was disease case counts for the countries/regions and dates available (Table 2). The same explanatory variables were included as features for all diseases (Table 3).

4.3.1. Moving Average Models

Moving average models served as a baseline statistical time-series analysis. ARIMA and ARIMAX were considered but were not viable solutions for this dataset because the data did not meet the normality requirement and struggled to converge with many explanatory variables. However, GLARMA with a negative binomial distribution was robust and useful for comparison in this study.

4.3.2. Tree-Based ML Models

RF [37] and XGB [38] are the ensemble, tree-based ML regression methods used in the analyses. Our XGB models were built using the XGBoost python package [39], while our tree-based models, evaluation and tuning methods were created using the scikit-learn framework [40]. Two variations of each of the disease’s model were included, one containing all explanatory features collected (Table 2), and a second model, “Alt,” which lacked the one-hot encoded country/region data. The Alt-XGB and Alt-RF models enabled the comparison of model performance in situations when the country/region data are unknown. The data were split into test and training subsets using nested K-fold time-series-split cross-validation with five folds. The models were retrained 10 times across each fold, varying the hyper-parameters. The final model was trained on all of the data through 2017 using the best hyper-parameters.
Each tree-based ML model output generated 12 predictions for each region-month interval by generating predictions from data one to 12 months before the prediction interval. For example, 12 predicted case counts in Wisconsin (USA) for January 2018 were generated, each one based on all explanatory variables in Wisconsin from a single month from January 2017 to December 2017. To extract a single prediction for each region and month, these 12 predictions were pooled via four approaches: maximum prediction, minimum prediction, standard average and average weighted by the inverse of the number month forward being predicted to reflect a decrease in confidence in longer-term predictions.

4.3.3. DL Models

The MLP and Enc–Dec models were created using pytorch [41] and tuned using available data from years 2009–2016 and validated on the year 2017. Since the computational time required for an exhaustive search of all hyperparameters was unreasonable for the DL models, we choose to tune on learning rate and optimizer (namely, Adam and stochastic gradient descent). Different methods of regularization (i.e., dropout and batch normalization) and model architecture (i.e., depth and breadth) were also explored in the early stages of model development. The final models were trained on all of the available data from 2009–2017. The encoder portion of the Enc–Dec model uses a gated recurrent unit (GRU) model to create a vector representation of the previous year of data by processing it sequentially. The decoder then takes that representation as input to a separate GRU and sequentially decodes it into predictions for the next year.

4.4. Metrics

Model performance was assessed by mean absolute error (MAE), root mean square error (RMSE), and by the Poisson deviance (deviance). Note that MAE and deviance are preferred metrics for skewed and highly discrete data, and that MAE, in particular, offers a more natural measure of average error which (unlike RMSE) is unambiguous [42]. However, we included all metrics (i.e., MAE, RMSE, and deviance) for completeness.
The deviance statistic for the Poisson distribution from each disease/location/time case count prediction was computed and averaged over each disease. The aggregate score was calculated by first averaging over time, then region, and finally country. Each model’s performance was assessed in generating disease incidence predictions over three time intervals: nowcast, i.e., the proximal month after the training data (January 2018); short-term forecast, i.e., 2–5 months after the training data (February 2018–May 2018); and long-term forecast, i.e., 6–8 months after the training data (June 2018–August 2018). Additionally, each model’s performance was assessed within individual countries and within individual regions.

Author Contributions

Conceptualization and methodology, L.E.C., K.T.P. and S.D.; data analysis, S.D., D.H.F. and A.S.; data collection, D.H.F., A.S.; validation, L.E.C., K.T.P. and R.K.; writing–original draft preparation, L.E.C., S.D., R.K. and D.H.F.; writing–review and editing, L.E.C., S.D., R.K. and K.T.P.; project management, L.E.C.; funding acquisition, L.E.C., K.T.P. and S.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Defense Threat Reduction Agency (project number CB11029).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data used in the study are publicly available with website links in the Methods section, including Table 5.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

In-Text AbbreviationDescription
-Alt a model trained without country-region as a feature
ARIMA auto-regressive integrated moving average
ARIMAX auto-regressive integrated moving average with exogenous
variables
DE encoder-decoder model
DeviancePoisson deviance
DL deep learning
DLR dynamic linear regression
GLARMA generalized linear autoregressive moving averages
GRU gated recurrent unit
MAE mean absolute error
ML machine learning
MLP multi-layer perceptron
RF random forest
RMSEroot mean squared error
RNN recurrent neural networks
SARIMA seasonal auto-regressive integrated moving average
SIR susceptible, infectious, and removed
SVM support vector machines
XGB extreme gradient boosted trees

References

  1. Bloom, D.E.; Cadarette, D. Infectious Disease Threats in the Twenty-First Century: Strengthening the Global Response. Front. Immunol. 2019, 10, 549. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Neglected Tropical Diseases. Available online: https://www.who.int/news-room/q-a-detail/neglected-tropical-diseases (accessed on 29 September 2021).
  3. Hartley, D.M.; Nelson, N.P.; Arthur, R.R.; Barboza, P.; Collier, N.; Lightfoot, N.; Linge, J.P.; van der Goot, E.; Mawudeku, A.; Madoff, L.C.; et al. An overview of internet biosurveillance. Clin. Microbiol. Infect. 2013, 19, 1006–1013. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. O’Shea, J. Digital Disease Detection: A Systematic Review of Event-Based Internet Biosurveillance Systems. Int. J. Med. Inform. 2017, 101, 15–22. [Google Scholar] [CrossRef] [PubMed]
  5. Bontempi, G.; ben Taieb, S.; le Borgne, Y.A. Machine Learning Strategies for Time Series Forecasting. In European Business Intelligence Summer School; Lecture Notes in Business Information Processing; Springer: Berlin/Heidelberg, Germany, 2012; pp. 62–77. [Google Scholar] [CrossRef]
  6. Voyant, C.; Notton, G.; Kalogirou, S.; Nivet, M.L.; Paoli, C.; Motte, F.; Fouilloy, A. Machine learning methods for solar radiation forecasting: A review. Renew. Energy 2017, 105, 569–582. [Google Scholar] [CrossRef]
  7. Makridakis, S.; Spiliotis, E.; Assimakopoulos, V. Statistical and Machine Learning Forecasting Methods: Concerns and Ways Forward. PLoS ONE 2018, 13, e0194889. [Google Scholar] [CrossRef] [Green Version]
  8. Rashidi, H.H.; Tran, N.K.; Betts, E.V.; Howell, L.P.; Green, R. Artificial Intelligence and Machine Learning in Pathology: The Present Landscape of Supervised Methods. Acad. Pathol. 2019, 6, 2374289519873088. [Google Scholar] [CrossRef]
  9. Stiglic, G.; Kocbek, P.; Fijacko, N.; Zitnik, M.; Verbert, K.; Cilar, L. Interpretability of Machine Learning-Based Prediction Models in Healthcare. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2020, 10, e1379. [Google Scholar] [CrossRef]
  10. Fang, X.; Liu, W.; Ai, J.; He, M.; Wu, Y.; Shi, Y.; Shen, W.; Bao, C. Forecasting Incidence of Infectious Diarrhea Using Random Forest in Jiangsu Province, China. BMC Infect. Dis. 2020, 20, 222. [Google Scholar] [CrossRef]
  11. Kane, M.J.; Price, N.; Scotch, M.; Rabinowitz, P. Comparison of ARIMA and Random Forest Time Series Models for Prediction of Avian Influenza H5N1 Outbreaks. BMC Bioinform. 2014, 15, 276. [Google Scholar] [CrossRef]
  12. Cheng, H.Y.; Wu, Y.C.; Lin, M.H.; Liu, Y.L.; Tsai, Y.Y.; Wu, J.H.; Pan, K.H.; Ke, C.J.; Chen, C.M.; Liu, D.P.; et al. Applying machine learning models with an ensemble approach for accurate real-time influenza forecasting in Taiwan: Development and validation study. J. Med. Internet Res. 2020, 22, e15394. [Google Scholar] [CrossRef]
  13. Bagheri, H.; Tapak, L.; Karami, M.; Amiri, B.; Cheraghi, Z. Epidemiological Features of Human Brucellosis in Iran (2011-2018) and Prediction of Brucellosis with Data-Mining Models. J. Res. Health Sci. 2019, 19, e00462. [Google Scholar] [PubMed]
  14. Alim, M.; Ye, G.H.; Guan, P.; Huang, D.S.; Zhou, B.S.; Wu, W. Comparison of ARIMA Model and XGBoost Model for Prediction of Human Brucellosis in Mainland China: A Time-Series Study. BMJ Open 2020, 10, 39676. [Google Scholar] [CrossRef] [PubMed]
  15. Ong, J.; Liu, X.; Rajarethinam, J.; Kok, S.Y.; Liang, S.; Tang, C.S.; Cook, A.R.; Ng, L.C.; Yap, G. Mapping Dengue Risk in Singapore Using Random Forest. PLoS Negl. Trop. Dis. 2018, 12, e0006587. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  17. Du, S.; Li, T.; Horng, S.J. Time Series Forecasting Using Sequence-to-Sequence Deep Learning Framework. In Proceedings of the 2018 9th International Symposium on Parallel Architectures, Algorithms and Programming (PAAP), Taipei, Taiwan, 26–28 December 2018. [Google Scholar]
  18. Bouktif, S.; Fiaz, A.; Ouni, A.; Serhani, M.A. Optimal Deep Learning Lstm Model for Electric Load Forecasting Using Feature Selection and Genetic Algorithm: Comparison with Machine Learning Approaches. Energies 2018, 11, 1636. [Google Scholar] [CrossRef] [Green Version]
  19. Gagne, D.J., II; Haupt, S.E.; Nychka, D.W.; Thompson, G. Interpretable deep learning for spatial analysis of severe hailstorms. Mon. Weather. Rev. 2019, 147, 2827–2845. [Google Scholar] [CrossRef]
  20. García-Martín, E.; Rodrigues, C.F.; Riley, G.; Grahn, H. Estimation of Energy Consumption in Machine Learning. J. Parallel Distrib. Comput. 2019, 134, 75–88. [Google Scholar] [CrossRef]
  21. DeFries, R.S.; Chan, J.C.W. Multiple Criteria for Evaluating Machine Learning Algorithms for Land Cover Classification from Satellite Data. Remote Sens. Environ. 2000, 74, 503–515. [Google Scholar] [CrossRef]
  22. Wanas, N.; Auda, G.; Kamel, M.S.; Karray, F. On the Optimal Number of Hidden Nodes in a Neural Network. Can. Conf. Electr. Comput. Eng. 1998, 2, 918–921. [Google Scholar] [CrossRef]
  23. Davis, R.; Holan, S.; Lund, R.; Ravishanker, N. Handbook of Discrete-Valued Time Series; CRC Press: Boca Raton, FL, USA, 2016. [Google Scholar]
  24. Gasparrini, A.; Scheipl, F.; Armstrong, B.; Kenward, M.G. A Penalized Framework for Distributed Lag Non-Linear Models. Biometrics 2017, 73, 938–948. [Google Scholar] [CrossRef]
  25. Hase, P.; Chen, C.; Li, O.; Rudin, C. Interpretable Image Recognition with Hierarchical Prototypes. Proc. AAAI Conf. Hum. Comput. Crowdsourcing 2019, 7, 32–40. [Google Scholar]
  26. Kaakoush, N.O.; Castaño-Rodríguez, N.; Mitchell, H.M.; Man, S.M. Global Epidemiology of Campylobacter Infection. Clin. Microbiol. Rev. 2015, 28, 687–720. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Guatteo, R.; Seegers, H.; Taurel, A.F.; Joly, A.; Beaudeau, F. Prevalence of Coxiella Burnetii Infection in Domestic Ruminants: A Critical Review. Vet. Microbiol. 2011, 149, 1–16. [Google Scholar] [CrossRef] [PubMed]
  28. Keshavamurthy, R.; Singh, B.B.; Kalambhe, D.G.; Aulakh, R.S.; Dhand, N.K. Identification of Risk Factors Associated with Coxiella Burnetii Infection in Cattle and Buffaloes in India. Prev. Vet. Med. 2020, 181, 105081. [Google Scholar] [CrossRef] [PubMed]
  29. Njeru, J.; Henning, K.; Pletz, M.W.; Heller, R.; Neubauer, H. Q Fever Is an Old and Neglected Zoonotic Disease in Kenya: A Systematic Review. BMC Public Health 2016, 16, 297. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Jha, A.K.; DesRoches, C.M.; Kralovec, P.D.; Joshi, M.S. A Progress Report on Electronic Health Records In U.S. Hospitals. Health Aff. 2017, 29, 1951–1957. [Google Scholar] [CrossRef] [Green Version]
  31. DesRoches, C.M.; Worzala, C.; Joshi, M.S.; Kralovec, P.D.; Jha, A.K. Small, Nonteaching, And Rural Hospitals Continue To Be Slow In Adopting Electronic Health Record Systems. Health Aff. 2017, 31, 1092–1099. [Google Scholar] [CrossRef] [Green Version]
  32. Desai, A.N.; Kraemer, M.U.; Bhatia, S.; Cori, A.; Nouvellet, P.; Herringer, M.; Cohn, E.L.; Carrion, M.; Brownstein, J.S.; Madoff, L.C.; et al. Real-time epidemic forecasting: Challenges and opportunities. Health Secur. 2019, 17, 268–275. [Google Scholar] [CrossRef]
  33. Khalsa, H.S.; Cordova, S.R.; Generous, N.; Khalsa, P.S. Epi Archive: Automated Synthesis of Global Notifiable Disease Data. Online J. Public Health Inform. 2019, 11, e280. [Google Scholar] [CrossRef]
  34. GDAL/OGR Geospatial Data Abstraction Software Library-Google Scholar. Available online: https://scholar.google.com/scholar?hl=en&as_sdt=0%2C48&q=GDAL%2FOGR+geospatial+data+abstraction+software+library&btnG= (accessed on 29 September 2021).
  35. Little, R.; Rubin, D. Statistical Analysis with Missing Data; John Wiley & Sons: Hoboken, NJ, USA, 2019. [Google Scholar]
  36. Hancock, J.T.; Khoshgoftaar, T.M. Survey on Categorical Data for Neural Networks. J. Big Data 2020, 7, 28. [Google Scholar] [CrossRef] [Green Version]
  37. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  38. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar] [CrossRef]
  39. XGBoost Documentation—Xgboost 1.5.2 Documentation. Available online: https://xgboost.readthedocs.io/en/stable/index.html (accessed on 19 December 2021).
  40. Pedregosa, F.; Michel, V.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Vanderplas, J.; Cournapeau, D.; Pedregosa, F.; Varoquaux, G.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  41. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury Google, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. Adv. Neural Inf. Processing Syst. 2019, 32, 8026–8037. [Google Scholar]
  42. Willmott, C.; Matsuura, K. Advantages of the Mean Absolute Error (MAE) over the Root Mean Square Error (RMSE) in Assessing Average Model Performance. Clim. Res. 2005, 30, 79–82. [Google Scholar] [CrossRef]
Figure 1. Mean campylobacteriosis case counts per region for available data between January 2009 and August 2018 for all seven countries with shaded areas representing two standard deviations from the mean. Note: only data from 2018 in Australia, Israel, and the United States was held out as the test set.
Figure 1. Mean campylobacteriosis case counts per region for available data between January 2009 and August 2018 for all seven countries with shaded areas representing two standard deviations from the mean. Note: only data from 2018 in Australia, Israel, and the United States was held out as the test set.
Pathogens 11 00185 g001
Figure 2. Mean Q-fever case counts per region for available data between January 2009 and August 2018 with shaded areas representing two standard deviations from the mean.
Figure 2. Mean Q-fever case counts per region for available data between January 2009 and August 2018 with shaded areas representing two standard deviations from the mean.
Pathogens 11 00185 g002
Figure 3. Time series of mean typhoid case counts per region for all available data between January 2009 and August 2018 with shaded areas representing two standard deviations from the mean.
Figure 3. Time series of mean typhoid case counts per region for all available data between January 2009 and August 2018 with shaded areas representing two standard deviations from the mean.
Pathogens 11 00185 g003
Figure 4. Average forecasting model performance by region for all predicted months during 2018 for campylobacteriosis, Q-fever, and typhoid as assessed by mean absolute error (MAE), root mean squared error (RMSE), and Poisson deviance (deviance) in Australia, Israel, and the United States.
Figure 4. Average forecasting model performance by region for all predicted months during 2018 for campylobacteriosis, Q-fever, and typhoid as assessed by mean absolute error (MAE), root mean squared error (RMSE), and Poisson deviance (deviance) in Australia, Israel, and the United States.
Pathogens 11 00185 g004
Figure 5. Mean performance of ML and statistical models by mean absolute error (MAE), root mean squared error (RMSE), and Poisson deviance (deviance) for Australia, United States, and Israel in forecasting monthly regional disease case counts in 2018.
Figure 5. Mean performance of ML and statistical models by mean absolute error (MAE), root mean squared error (RMSE), and Poisson deviance (deviance) for Australia, United States, and Israel in forecasting monthly regional disease case counts in 2018.
Pathogens 11 00185 g005
Figure 6. Comparison of model performance by metric in forecasting regional monthly case counts for January to August 2018 in Australia, Israel, and the United States. Error bar shows highest and lowest error by month.
Figure 6. Comparison of model performance by metric in forecasting regional monthly case counts for January to August 2018 in Australia, Israel, and the United States. Error bar shows highest and lowest error by month.
Pathogens 11 00185 g006
Figure 7. Comparison of model performance in forecasting monthly case counts from 2018 by metric for the 20 highest case count regions (highest) and lowest case count regions (lowest) across Australia, Israel, and the United States.
Figure 7. Comparison of model performance in forecasting monthly case counts from 2018 by metric for the 20 highest case count regions (highest) and lowest case count regions (lowest) across Australia, Israel, and the United States.
Pathogens 11 00185 g007
Figure 8. Top 20 features grouped in categories with their relative importance by disease and tree-based ML model-type shaded by contribution. Features are sorted by decreasing average relative importance of all Alt-models.
Figure 8. Top 20 features grouped in categories with their relative importance by disease and tree-based ML model-type shaded by contribution. Features are sorted by decreasing average relative importance of all Alt-models.
Pathogens 11 00185 g008
Table 1. The total case counts for each disease by country for 2009–2018, including mean, minimum, and maximum of the monthly regional cases. All available data through 2017 for the seven countries were included in the ML training sets but only highlighted countries were used in performance evaluations. NA means not available.
Table 1. The total case counts for each disease by country for 2009–2018, including mean, minimum, and maximum of the monthly regional cases. All available data through 2017 for the seven countries were included in the ML training sets but only highlighted countries were used in performance evaluations. NA means not available.
Campylobacteriosis
Case Counts
Q-Fever
Case Counts
Typhoid
Case Counts
Mean Min–MaxTotalMeanMin–MaxTotalMeanMin–MaxTotal
Australia190.100–880182,4983.920–3237681.170–141122
Israel40.640–36129,4291.120–218090.060–763
Japan NA NA NA NA 0–2 10 0.07 0–7 264
Norway 10.36 0–18422,839 0.2010–114 0.08 0-8 169
Sweden 31.55 0–42716,380 NA NA NA ~0.000–1 5
United States42.040–79498,7200.060–34230.190–291251
Table 2. The top-performing disease model by location and forecast length. For the model, if all metrics match, one model is listed; if the best model by metric did not match, the cell contains ‘MAE; RMSE; deviance’ and if more than one model produced the same value for the metric, they are listed together and separated by commas.
Table 2. The top-performing disease model by location and forecast length. For the model, if all metrics match, one model is listed; if the best model by metric did not match, the cell contains ‘MAE; RMSE; deviance’ and if more than one model produced the same value for the metric, they are listed together and separated by commas.
DiseaseLocationNowcastShort-TermLong-Term
Campylo-bacteriosisAll countriesAlt-XGBAlt-XGBAlt-XGB; RF (Both); Alt-XGB
AustraliaXGB (Both)Alt-XGBAlt-XGB
IsraelMLP; MLP; XGBXGBXGB; GLARMA; XGB
USAlt-XGB; GLARMA; Alt-XGBGLARMA; GLARMA; Alt-XGBAlt-XGB; Alt-RF; Alt-XGB
Q-FeverAll countriesRFAlt-RF; Alt-XGB; Alt-XGBAlt-XGB
AustraliaGLARMAAlt-XGBXGB
IsraelMLPEnc–Dec; GLARMA; Enc–DecAlt-XGB
USAll Models All Tree-based ML (Alt-XGB *)All Tree-based ML(Alt-XGB *)
TyphoidAll countriesEnc–DecAlt-XGB; GLARMA; XGB (Both)MLP
AustraliaGLARMA; XGB(Both), GLARMA; Enc–DecGLARMAMLP; Alt-RF; MLP
IsraelAll ModelsAll ModelsAll Models
USMLPAll Tree-based MLRF; MLP; RF
* Smallest error range. Note: “Both” refers to the Alt and Not versions of the same model.
Table 3. The top-performing disease models by data split. For the model, if MAE, RMSE, and deviance matched, there is one model listed; if the best model by metric did not match, the cell contains ‘MAE; RMSE; deviance’ and if greater than one model produced the same value for all the metrics, they are listed together and separated by commas.
Table 3. The top-performing disease models by data split. For the model, if MAE, RMSE, and deviance matched, there is one model listed; if the best model by metric did not match, the cell contains ‘MAE; RMSE; deviance’ and if greater than one model produced the same value for all the metrics, they are listed together and separated by commas.
Number of CasesCountry Over All MonthsForecast Time Over All Locations
DiseaseOverallHigh CasesLow Cases (Zero)AustraliaIsraelUSNowcastingShort TermLong Term
Campylo-
bacteriosis
Alt-XGBAlt-XGBGLARMAAlt-XGBXGB (Both)GLARMA, Alt-XGBXGBXGBRF(Both); RF(Both); Alt-XGB
Q-feverAlt-XGBAlt-XGBTree-based, DLAlt-XGB GLARMA; Enc–Dec; Alt-XGBAllRFTree-basedAlt-XGB
TyphoidAlt-XGBGLARMAAllAlt-XGB; GLARMA; GLARMAAllRFEnc–DecAlt-XGB; GLARMA; XGB (Both)MLP
Note in this table that Both refers to the Alt and Not versions of the same model.
Table 4. Summary of case count data used in the analysis by country and disease name for the study period of January 2009 through August 2018.
Table 4. Summary of case count data used in the analysis by country and disease name for the study period of January 2009 through August 2018.
CountryCampylobacteriosisQ-FeverTyphoid
Date Range# RegionsDate Range# RegionsDate Range# Regions
Australia2009–201882009–201882009-20188
Finland2009–201718NA0NA0
Israel2012–201862012–201862009–20186
JapanNA02012–2017472012–201747
Norway2009–2017182009–2017182009–201718
Sweden2009–201721NA02009–201721
United States2015–2018512009–2018512009–201851
“NA” means no data available in EpiArchive for the specified diseases and countries. “#” means number of.
Table 5. Explanatory variable data types, website for public access, individual features by name, geographic location, geographic resolution, time period, and periodicity.
Table 5. Explanatory variable data types, website for public access, individual features by name, geographic location, geographic resolution, time period, and periodicity.
Data TypeWebsiteIndividual
Features
Geographic
Location
Geographic ResolutionTime PeriodPeriodicity
Case Countsepiarchive.bsvgateway.org accessed on 28 May 2019Incidences of select human diseases.Countries of interestRegion-level2009–2018Daily
Political Bordersgadm.org accessed on 28 May 2019Geopolitical borders (country and within country)Countries of interestRegion-level2018Single instance
Climatedisc.gsfc.nasa.gov; earthdata.nasa.gov accessed on 28 May 2019air temperature, humidity, precipitation, soil moisture, and wind speedGlobalGridded 0.25° × 0.25°,
1° × 1°
2012–2018Monthly
Gross Domestic Productwww.ers.usda.gov accessed on 28 May 2019Gross Domestic ProductGlobalCountry-levelVariesYearly
Elevationwww.diva-gis.org accessed on 28 May 2019Digital Elevation MapGlobal43,200 × 17,200 (30 arc seconds)NANA
Mortalitywww.who.int accessed on 28 May 2019Deaths by country, year, sex, age group, and cause of death.GlobalCountry-level2009–2018Yearly
Municipal wastestats.oecd.org accessed on 28 May 2019Municipal waste generation and treatmentCountries of interestCountry-level2009–2017Yearly
Socio-political
and Physical data
www.naturalearthdata.com accessed on 28 May 2019Country and internal administrative borders; socioeconomic and political attributesGlobalVaries by country; 1: 10 m–110 m2019Single instance
Populationpopulation.un.org accessed on 28 May 2019Population by age intervals by locationGlobalCountry-level2009–2015Every 5 years
Population Densitysedac.ciesin.columbia.edu accessed on 28 May 2019Population densityGlobal30 arc-seconds2009–2015Every 5 years
Water Potability and Treatmentstats.oecd.org accessed on 28 May 2019Freshwater resources, available water, wastewater treatment plant capacity, surface waterCountries of interestCountry-level2009–2017Yearly
“NA” means no data available.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dixon, S.; Keshavamurthy, R.; Farber, D.H.; Stevens, A.; Pazdernik, K.T.; Charles, L.E. A Comparison of Infectious Disease Forecasting Methods across Locations, Diseases, and Time. Pathogens 2022, 11, 185. https://doi.org/10.3390/pathogens11020185

AMA Style

Dixon S, Keshavamurthy R, Farber DH, Stevens A, Pazdernik KT, Charles LE. A Comparison of Infectious Disease Forecasting Methods across Locations, Diseases, and Time. Pathogens. 2022; 11(2):185. https://doi.org/10.3390/pathogens11020185

Chicago/Turabian Style

Dixon, Samuel, Ravikiran Keshavamurthy, Daniel H. Farber, Andrew Stevens, Karl T. Pazdernik, and Lauren E. Charles. 2022. "A Comparison of Infectious Disease Forecasting Methods across Locations, Diseases, and Time" Pathogens 11, no. 2: 185. https://doi.org/10.3390/pathogens11020185

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop