Next Article in Journal
YOLOv3-Based Matching Approach for Roof Region Detection from Drone Images
Next Article in Special Issue
Meteorological Drivers of Permian Basin Methane Anomalies Derived from TROPOMI
Previous Article in Journal
An Overview of Requirements, Procedures and Current Advances in the Calibration/Validation of Radar Altimeters
Previous Article in Special Issue
Spatial Assessment of Health Economic Losses from Exposure to Ambient Pollutants in China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Imputing Satellite-Derived Aerosol Optical Depth Using a Multi-Resolution Spatial Model and Random Forest for PM2.5 Prediction

1
Department of Biostatistics and Bioinformatics, Rollins School of Public Health, Emory University, Atlanta, GA 30322, USA
2
Gangarosa Department of Environmental Health, Rollins School of Public Health, Emory University, Atlanta, GA 30322, USA
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(1), 126; https://doi.org/10.3390/rs13010126
Submission received: 9 December 2020 / Revised: 22 December 2020 / Accepted: 27 December 2020 / Published: 1 January 2021
(This article belongs to the Special Issue Satellite Remote Sensing for Air Quality and Health)

Abstract

:
A task for environmental health research is to produce complete pollution exposure maps despite limited monitoring data. Satellite-derived aerosol optical depth (AOD) is frequently used as a predictor in various models to improve PM2.5 estimation, despite significant gaps in coverage. We analyze PM2.5 and AOD from July 2011 in the contiguous United States. We examine two methods to aid in gap-filling AOD: (1) lattice kriging, a spatial statistical method adapted to handle large amounts data, and (2) random forest, a tree-based machine learning method. First, we evaluate each model’s performance in the spatial prediction of AOD, and we additionally consider ensemble methods for combining the predictors. In order to accurately assess the predictive performance of these methods, we construct spatially clustered holdouts to mimic the observed patterns of missing data. Finally, we assess whether gap-filling AOD through one of the proposed ensemble methods can improve prediction of PM2.5 in a random forest model. Our results suggest that ensemble methods of combining lattice kriging and random forest can improve AOD gap-filling. Based on summary metrics of performance, PM2.5 predictions based on random forest models were largely similar regardless of the inclusion of gap-filled AOD, but there was some variability in daily model predictions.

1. Introduction

Ambient outdoor air pollution, particularly particulate matter less than 2.5 micrometers in aerodynamic diameter (PM2.5), poses a substantial risk to human health [1,2,3]. Air pollution monitors that can directly measure pollution concentrations are placed at a limited set of locations, resulting in large areas without direct measurements of ground-level pollution exposure. Aerosol optical depth (AOD) measures the amount of aerosol in the atmosphere and can be remotely sensed by satellite instruments at various spatial resolutions [4]. A growing literature has developed for using satellite-derived AOD as a proxy and predictor for PM2.5 concentrations, often in conjunction with land-use and meteorological variables, using a range of model types such as geographically weighted regression, linear mixed effect models, and machine learning methods [5,6,7].
However, AOD itself has substantial missingness, complicating the process of predicting PM2.5 concentrations. Gaps in AOD coverage are a result of cloud cover, snow cover, and surface brightness; for the moderate resolution imaging spectroradiometer (MODIS) 10 km product, on average each grid cell has no AOD available on approximately 70% of days, with substantial variation across regions [8,9]. AOD’s patterns of missingness are also not random for the purpose of PM2.5 prediction; cloud and snow cover may plausibly be related to PM2.5 concentrations [9,10]. At the scale of the continental United States, research suggests that missingness as a result of cloud cover is not likely to greatly bias monthly and yearly PM2.5, although there is regional and seasonal variation [11]. However, Liang et al. [12] show that long-term PM2.5 estimates in China are substantially biased as result of missing AOD observations. Furthermore, for health effects research, the relevant geographic scale is small and more impacted by missingness. When using PM2.5 estimates based on satellite-derived AOD with substantial missingness, time series studies will miss many days and lose statistical power, and cohort studies will use potentially biased exposure estimates, resulting in a loss of statistical power. A number of approaches have been proposed for handling missing AOD observations when estimating PM2.5 [7]. One approach has been to combine different AOD retrievals, although this will still result in incomplete coverage; e.g., Geng et al. [13] combine AOD measurements from Terra and Aqua satellites using linear regression. Other approaches have used AOD where available, but otherwise bypassed the need for gap-filling AOD [14,15,16].
Many recent studies use multi-stage approaches, where AOD is gap-filled, and then a model relating PM2.5 to the gap-filled AOD and other land-use and meteorological variables is fit. These gap-filling models may use land-use and meteorological terms, as well as chemical transport model (CTM) estimates. Hu et al. [17] forego a statistical modeling procedure for gap-filling AOD, saving computational time, and they replace missing AOD values with CTM (GEOS-Chem) estimates of AOD. Xiao et al. [18] and Huang et al. [19] use linear models that include cloud fraction estimates, meteorological and land-use data together with smoothing splines, to account for spatial correlation for imputing AOD. Lv et al. [20] gap-fill AOD using a model that relates AOD to the ratio of daily and seasonal averages of PM2.5 multiplied by the average seasonal AOD values for the grid cell for each city under study; a second stage then uses ordinary Kriging to fill in the remaining gaps. Because of the computational costs of ordinary Kriging, this method will not scale well to large datasets, but previous studies suggest that smoothing splines may not perform as well as Kriging in some settings [21]. Chen et al. [22] use a mixed effect model to first combine Terra and Aqua AOD measurements, and interpolate missing AOD values using inverse-distance weighting (IDW). IDW with a maximum distance will not be able to provide full coverage for AOD, however, as there are large missing areas with no observed data. Random forest (RF) is a popular machine learning method used for gap-filling, due to the fast implementations available and its ability to account for complex non-linear interactions of features [7,23]. Bi et al. [10] uses a two-stage model with RF being used to impute AOD using a number of relevant variables, including MODIS cloud and snow fractions. Several other recent papers impute AOD as part of a multi-stage process using RF (e.g., see Stafoggia et al. [24], Huang et al. [25], and Zhang et al. [26]). However, judging performance based on “out-of-bag” measures or random holdouts of observed data may be misleading in spatial prediction problems with large contiguous areas of missing data. Furthermore, when a strong spatial pattern is present as in AOD, it is unclear how RF performs compared to spatial statistical models. Jiang et al. [27] use RF to gap-fill AOD in China and evaluate performance using aerosol robotic network (AERONET) measurements [28,29]; however, there are relatively few locations with AERONET measurements on which to validate performance on any given day.
Importantly, models for gap-filling AOD are generally more costly to fit than models for estimating PM2.5, due to the much larger number of daily observations. For example, in our case study, using a modeling grid of 12 km spatial resolution over the contiguous United States entails over 50,000 daily cells. While several studies have used machine learning methods for AOD gap-filling to overcome the computational costs, traditional spatial statistical methods like Kriging are not well-suited to handle large datasets due to the need to invert the spatial covariance matrix. Over the course of the last decade or so, several spatial statistical methods have been developed to handle big data [30,31]. Although considerable attention has been given to using ensemble and hybrid approaches for estimating PM2.5 (e.g., [32,33,34,35]), AOD gap-filling for large areas has received less focus, possibly due to the greater computational cost.
To our knowledge, studies have not thus far considered ensemble methods for combining large-scale spatial statistical methods with machine learning methods for gap-filling AOD. In this study, we focus on a particular spatial statistical method, lattice kriging (LK) [36], together with RF for AOD gap-filling. Our study considers both RF and LK models for gap-filling MODIS AOD, as well as ensemble methods for combining these predictions following the super learner methodology [37,38]. Our case study focuses on the contiguous United States using daily data for the month of July 2011. We focus on a single month for computational reasons as each AOD model is fit daily to tens of thousands of observations, and we perform 10-fold cross-validation on each day as part of the ensemble method construction, resulting in an additional 10 models fit per day for LK and RF. We assess performance using spatially clustered holdouts for AOD gap-filling models that may more accurately measure performance than more commonly used approaches. Finally, we assess whether the imputed AOD product using ensemble methods improves PM2.5 estimation in a random forest model. Broadly, we find that ensemble methods can be effective for AOD gap-filling, but there is less evidence to suggest an ultimate benefit for PM2.5 estimation.

2. Materials and Methods

2.1. Study Area

The study area of interest is the contiguous United States, consisting of 48 states, and Washington DC, using daily data for the month of July 2011. We focused our analysis on July, as summer months generally have less AOD missingness than other months on average, while retaining substantial day-to-day variability in missingness. Descriptions of the data sources follow the work of Hu et al. [17].

2.2. PM2.5 Measurements

We obtain measurements of PM2.5 from the U.S. Environmental Protection Agency (EPA) Air Quality System (AQS) (https://www.epa.gov/outdoor-air-quality-data). We used 24h averaged concentrations collected from 1248 federal reference method samplers.

2.3. MODIS AOD

For the purpose of this study, we utilize Collection 6 level 2 Aqua MODIS retrievals at 550 nm wavelength using the MYD04_L2 product [4,39]. High-confidence AOD retrievals from the combined deep-blue/dark target parameter were used [8]. AOD at 550 nm is the main product provided by the various mainstream aerosol instruments, and 550 nm is the most common wavelength used in research applications for aerosols (https://atmosphere-imager.gsfc.nasa.gov/faqs/aerosol). Following previous work [17], these retrievals at a resolution of 10 km are regridded to 12 km × 12 km community multi-scare air quality (CMAQ) grids. We consider daily MODIS AOD data from July 2011, for a total of 53,807 daily cells in the contiguous United States. The proportion of cells in which daily AOD is observed ranges from a minimum of 26.33% to 54.63%, with an average of 41.08%. The top row of Figure 1 demonstrates two days with the least and most missing observed AOD points. We used MODIS AOD rather than a finer scaled product (e.g., 1 km2 products), in part because our goal was to explore large-scale variation in the national map for AOD, and in part due to computational costs.

2.4. GEOS-Chem AOD

GEOS-Chem is a “global 3-D model of atmospheric chemistry driven by assimilated meteorological observations from the Goddard Earth Observing System (GEOS) of the NASA Global Modeling Assimilation Office (GMAO)” (http://acmg.seas.harvard.edu/geos/) [40]. We utilize version 10.1 of the model, using GEOS-5 meteorological data for 2011, with total column AOD calculated as the sum of 6 AOD parameters (sulfate-nitrate-ammonium, black carbon, organic carbon, accumulation-mode sea-salt, coarse-mode sea-salt, and total dust) over 37 vertical layers (from the surface up to ≈20 km) [17,41].

2.5. Meteorological Variables

We obtained meteorological data from the North American Land Data Assimilation System phase 2 (NLDAS-2) (https://ldas.gsfc.nasa.gov/nldas/) [42,43]. These data have a spatial resolution of approximately 13 km and are available hourly. For this analysis, we use pressure at surface (pa), u- and v-direction wind speed (m/s), temperature (K), relative humidity (%), precipitation (kg/m2), fraction of total precipitation that is convective (no units), convective available potential energy (J/kg), surface DW shortwave radiation flux (W/m2), surface DW longwave radiation flux (W/m2), and potential evaporation (kg/m2). Measurements are averaged from 10 a.m. to 4 p.m. local time to construct daily daytime observations, roughly coinciding with the Aqua overpass time (about 1:30 p.m.).

2.6. Land Use

We include elevation obtained from the National Elevation Dataset at 30 m spatial resolution (https://viewer.nationalmap.gov/basic/). We obtained total length of highways (m), total length of limited-access road (m), and total length of local road (m) from ESRI StreetMap USA (Environmental Systems Research Institute, Redlands, CA, USA). Forest cover (unitless) and impervious surface (%) are derived from the National Land Cover Database (https://www.mrlc.gov/). In addition, we include point emissions data for PM2.5 and PM10 combined (in tons) from the EPA 2011 National Emissions Inventory report (https://www.epa.gov/air-emissions-inventories). Population density is obtained from the 2010 Census at the tract level (population/km2).

2.7. Data Integration

Data were projected into a common coordinate system using the U.S. Lambert conformal conic projection. For each 12 km × 12 km grid cell, forest cover, impervious surface, and elevation were averaged, while road length and point emission values were summed. Meteorological variables and population density were assigned based on nearest distance. Grid cells containing multiple PM2.5 monitors for a day were averaged.

2.8. Statistical Methods

We conduct AOD gap-filling using two distinct methods from the spatial statistical and machine learning fields, respectively, as well as methods for combining them.

2.8.1. Lattice Kriging

Lattice kriging (LatticeKrig or LK) is a multi-resolution Gaussian process model [36]. At a high-level, LK models the spatial process using several levels of two-dimensional basis functions, which are laid out on a grid and approximately double with each successive layer. These basis functions are compact, which means that for a particular point, only a small number of basis functions are used to make the prediction. The coefficients associated with the basis functions are assumed to be correlated, and this structure can flexibly model observed spatial covariance structures. Estimation proceeds through a likelihood-based approach after specifying various tuning parameters.
Following Nychka et al. [36], we observe { y i } at locations { x i } for i = 1 , . . , n . We assume that { y i } follow an additive model consisting of a mean function based on covariates, a spatial process, and a measurement error term:
y i = Z i T d + g ( x i ) + ϵ i ,
where d is a p × 1 vector of fixed coefficients associated with the covariates Z i , and g ( x i ) denotes the spatial process. The mean-zero error terms ϵ i are presumed to be independent and identically distributed, i.e., ϵ N ( 0 , σ 2 I ) , where ϵ = ( ϵ 1 , . . . , ϵ n ) T .
The overall spatial process g ( x i ) can be written as a sum of L independent spatial processes g l ( x i ) :
g ( x i ) = l = 1 L g l ( x i ) = l = 1 L j = 1 m ( l ) c j l ϕ j , l ( x i ) ,
where ϕ j , l denotes the the lth level of resolution’s jth basis function, and c j l denotes the coefficient associated with this basis function. Although the basis functions and number of levels are fixed (i.e., chosen), the coefficients for each level l, c l = ( c 1 l , . . . , c m ( l ) l ) T are assumed to follow a multivariate normal with mean zero and covariance ρ Q l 1 :
c l N ( 0 , ρ Q l 1 ) .
Each level’s spatial process is independent with marginal variance ρ α l subject to the constraint l = 1 L α l = 1 , so that the marginal variance of the overall spatial process g ( x i ) is ρ . The main parameters that need to be selected are the number of levels of resolution (L), the number of grid points along the largest dimension at the coarsest level of resolution (which in turn determines the total number of basis functions), the relative weight of each spatial level’s process (parameterized by ν ), and the scale/range parameter (a). We provide a more thorough description of the model in the Supplemental Materials, and we also refer readers to the originating paper [36], the comparison paper by Heaton et al. [30], and the documentation for the R implementation (version 8.4) [44].

2.8.2. Random Forest

Random forest (RF) consists of constructing a large number of regression trees [23]. At a high level, regression trees search for the best (as determined by mean-squared error) binary split among the covariates (i.e., features), and then split the data accordingly. This process continues until some condition is met (e.g., there is only 1 observation left, so no further split can be made). Two key components of RF are: (1) bagging, or bootstrap aggregation, wherein each tree is fit to a random sample (with replacement) of the original sample; and (2) at each node, the algorithm considers only a random subset m of the original p predictors for deciding on the best split. A single decision tree would likely overfit to the data. Averaging many trees that implement these two components results in reducing variance, while maintaining a low bias from the procedure.
The key parameters are the number of trees (B), the number of predictors to randomly select for each split (m, or mtry) from the original p, and, to a lesser extent, the node size ( n s i z e ). In general, we choose the number of trees B to be large, and we use validation data to help select m and n s i z e . For this study, we use the ranger package (version 0.12.1) in R [45].

2.8.3. Super Learner Methods

Super learners (SL), related to stacked generalization and stacked regression methods [46], use a potentially large and diverse set of algorithms by weighting their predictions optimally according to some risk measure such as squared error loss. Although a large number of algorithms are recommended in practice, we use just RF and LK as our algorithms in order to demonstrate the use of SL and to maintain focus on the cross-validation approach. The process for super learners is as follows [37,38,47,48]:
  • Divide observed data into k folds.
  • For each fold k, let the kth fold be the validation data, and the remainder be the training data. Fit each algorithm or model m to the training data and make predictions on the kth fold.
  • Stack all predictions y ^ m for each algorithm.
  • Estimate the weights α m for algorithm m = 1 , , M using the model formulation
    y i = m = 1 M α m y ^ i , m + ϵ i ,
    where α m 0 and m = 1 M α m = 1 . α m can be estimated by non-negative least-squares methods and then normalizing the weights to sum to 1.
After these α m model weights are estimated through the cross-validation process, each algorithm is fit to the full observed data, and test data predictions are made by using these weights for combining predictions. Davies and van der Laan [48] provide a discussion of extending SL theory to the case of spatial data. Murray et al. [35] use a similar stacked regression approach for determining weights in combining separate models for PM2.5 prediction.

2.9. AOD Gap-Filling Analysis

From the observed AOD data, we consider a spatially clustered approach for creating a testing dataset on which to evaluate the results. In the proposed method, ten random AOD observations are selected from the observed AOD values for each day in July 2011. These observations and any other observations within a 250 km radius are then held out as the test dataset. We selected ten random points, in particular, to ensure that different areas of the observed data had a sufficient chance of being held out for the testing data, so that the performance of the different methods for a particular day was not likely to be judged solely on predictions for one area of the observed map. We also observed that this procedure resulted in close to a 70/30 split between training and testing data, on average. Figure 1 demonstrates the observed and training data on two particular days. This approach to creating testing data is an attempt to mimic the actual observed pattern of AOD data, where large contiguous areas are missing and require imputation. In particular, for many missing observations, there are likely no points nearby to aid in prediction. In our analysis, we consider each day separately for model fitting and prediction.
For RF and LK, we used validation data taken from the training data following a similar approach. For RF, we use 2000 trees with a node size of 5 and m = 7 . We include 22 variables: the projected centroid coordinates, elevation, the 2011 emission inventory, forest cover, impervious surface, total lengths of highway, limited-access road, and local road, population density, potential evaporation, surface DW longwave radiation flux, surface DW shortwave radiation flux, convective available potential energy, fraction of total precipitation that is convective, precipitation, relative humidity, temperature, u- and v-direction wind speed, pressure at surface, and GEOS-Chem AOD. For LK, we set the number of levels of resolution to 5, a = 12 , ν = 0.1 , and the number of grid points along the largest dimension at the coarsest level of resolution to 15. By default, the LatticeKrig package includes the spatial coordinates as fixed predictors in Z . In addition, we include the interaction between the coordinates, GEOS-Chem, the interaction between each coordinate and GEOS-Chem, and elevation as the fixed predictors in the model. No variable selection was performed—we instead focused on tuning the spatial aspect of the model.
In addition to comparing RF and LK, we aim to consider SL approaches for combining these distinct methods. For each day, we construct 10 cross-validation folds using the blockCV R package (version 2.1.1) [49]. This constructs spatial blocks for the validation dataset, so that performance more accurately mimics the task of gap-filling AOD. In the Supplementary Materials (Figure S6), we provide a full set of the maps showing these spatial block cross-validation folds. Sarafian et al. [50], Murray et al. [35] (for PM2.5 prediction) and Young et al. [51] (for NO2) also consider spatially clustered cross-validation approaches for assessing model performance. Based on the cross-validation folds, we stack LK and RF validation predictions as discussed in the previous section. We assess 4 different methods for combining LK and RF:
  • Average. We construct a simple average of RF and LK predictions. Cross-validation data are not used in this approach.
  • SL: overall. After stacking all of the cross-validation predictions for all days together, we produce a single set of optimal weights with (4) for making predictions.
  • SL: daily. We stack cross-validation predictions for each day separately, and we obtaining a daily set of optimal weights with (4).
  • SL: distance-based. For each cross-validation fold on each day, we determine the closest distance between each point in the cross-validation fold and the training data. We then stack all of the cross-validation predictions across days together with these nearest-neighbor distances. We bin these stacked predictions according to nearest-neighbor distances with bin widths of 25 km from 0 to 300 km and higher. Using (4), we estimate the optimal weights for LK and RF for each binned distance grouping. We then fit a simple loess model relating interval mid-point distance and optimal weight, and we use these fitted optimal weights for combining LK and RF for predictions. The motivation for this last technique is that the further the unobserved point is from the observed data, the more different algorithms may be in predictions. If there is strong spatial correlation, then LK may perform better; in contrast, if there is limited range in the spatial correlation and the covariates are more important, then RF may produce better fits based on the relationship between the covariates and response.
In all cases, we restrict weights to be between 0.1 and 0.9. We primarily assess model performance on the basis of root mean square error (RMSE) and the coefficient of determination (R2), as well as the intercept and slope from a linear model relating the true AOD observations to the predicted values.

2.10. PM2.5 Analysis

We compare several random forest models for PM2.5 concentration estimation to determine whether including gap-filled AOD as a predictor can improve prediction performance. Our gap-filled AOD is based on the super learner distance-based method. There are five variations on the random forest models’ features:
  • M1: Includes neither AOD nor GEOS-Chem.
  • M2: Includes GEOS-Chem.
  • M3: Includes gap-filled AOD. This variable is defined to be observed AOD where available, and otherwise the predicted AOD value based on the super-learner distance-based method. GEOS-Chem is also included as a separate feature.
  • M4: Includes AOD by replacing missing values of AOD with GEOS-Chem (as in Hu et al. [17]).
  • M5: Includes observed AOD, and training solely on observations where AOD is observed. For predictions, missing values of AOD are replaced with the gap-filled AOD. GEOS-Chem is also included as a separate feature.
The other features included in the random forest models are the same as those in the AOD analysis. All models except M5 additionally include an indicator variable for whether AOD was observed at the location, and all models include a so-called convolution layer of PM2.5. Several analyses [17,34] have demonstrated that a weighted-average of nearby PM2.5 observations can aid in model prediction for PM2.5. Briefly, for each location, the convolution layer of PM2.5 is a weighted average of all other training PM2.5 observations, not including the location itself. The weights are inversely proportional to the squared distance between locations (less distant observations in the training data are weighted more). The procedure for creating the convolution PM2.5 layer must be repeated for each training/validation split for each day.
We consider three distinct 10-fold cross-validation approaches for assessing performance:
  • Random: Cross-validation folds are constructed by selecting observed PM2.5 monitors at random on a daily basis.
  • Constant spatial clusters: Cross-validation folds are constructed by creating spatially clustered areas that are constant across all days. A particular area of the map will be assigned to the same cross-validation fold for each day.
  • Varying spatial clusters: Cross-validation folds are constructed by creating spatial clusters at random for each day.
Spatial clusters are constructed using the blockCV package (version 2.1.1) in R with block widths of 150 km. Figure S8 in the Supplemental Materials displays the constant spatial construction by color-coding monitor locations.
Models were fit both for each day separately, as well as for all of the days in July 2011 together. In the latter spatio-temporal random forest model, day of the year and day of the week are additionally included as integer predictor variables. Our primary metrics of interest are RMSE, R2, and the full prediction maps, but we also present the intercept/slope estimates from fitting a linear regression model with the true PM2.5 values as the dependent variable and the random forest prediction as the independent variable. For all models, we set the number of trees to 2000. We varied m (mtry) between values of 4, 8, 12, and 16 and presented the best results for each model and cross-validation fold type on the basis of RMSE. The full maps and feature importance results are based on m = 4 . The Supplemental Materials include additional figures and tables for m = 8 .

3. Results

3.1. AOD Gap-Filling

We highlight several results from our analysis. First, daily results show that any of the average/super learner approaches match or exceed performance from either LK or RF alone on a majority of days on the basis of RMSE (Table 1, Figure S2). While there are some days where either LK or RF does particularly well, there are also days where these two methods perform worse than any of the other methods. The ensemble methods are the best or close to the best in terms of RMSE and R2 on the majority of days. The distance-based SL method performs best on more days (10 out of 31 days) than any of the other approaches considered. Based on our described approach, the distance-based SL prediction weights LK greater for testing points that are close to training data, while for points further away from the training data, it weights RF more, relying on a combination of location, land use, and meteorological features (see Figure S7 in the Supplemental Materials).
Evaluating test predictions across all 31 days together, LK and RF have R2 values of 0.644 and 0.619, respectively. The average and SL methods improve to R2 values in the range of 0.657 to 0.659. Compared to LK alone, the RMSE is reduced 2.34% and 2.30% in the distance-based and overall SL models, respectively. A simple average of the LK and RF predictions also provides most of these gains. The super learner method based on the daily construction of weights performs well but is marginally worse than the other ensemble methods. In explaining this small difference in performance, we posit that the daily weight estimates may be slightly over-fitted to the daily training data relative to the other ensemble methods, which benefit from pooling cross-validation predictions across days together for the estimation of the super learner weights.
Both LK and RF methods diverge substantially from using just the observed AOD data in the July 2011 averages (Figure 2). LK and RF predictions are notably different from each other in the Appalachian Mountains and in parts of the Southwest, where LK predicts higher values relative to RF (Figure 3). There are also apparent edge effects in some areas such as south Florida, where LK predictions will tend to diverge more substantially from those of RF. These may be partly due to issues with LK’s coefficient estimation in areas where there are few data for a particular day (see Figures S3 and S4 in the Supplemental Materials for plots of daily AOD predictions and daily differences between LK and RF).

3.2. PM2.5 Prediction

The gap-filled AOD used in the PM2.5 analyses is based on the SL distance-based method for combining RF and LK predictions on the basis of the results in the AOD analysis. We first highlight a few notable results from the daily random forest models. First, the daily random forest models suggest that RMSE is improved consistently but only marginally by including the imputed AOD predictor vs. the four alternatives (Table 2, M3a). The outlier model is M5, which trains solely on observations where AOD is observed and predicts using the imputed AOD where AOD is missing. The results from model M5 are substantially worse than the other models, with a relatively biased prediction map (Figure 4d). For the remainder of the results, we omit discussion of this model. Generally, the gain in RMSE for the model using gap-filled AOD (M3) is small and ranges from 0.01 to 0.03 μg m−3 against the other models. The results for R2 are similar, with small gains of approximately 0.002 to 0.006. Second, the daily random forest models tended to have better PM2.5 prediction in locations where AOD was not observed as compared to areas where AOD was observed, regardless of the features included. Third, cross-validated RMSE is substantially larger in spatial cross-validation settings than in the case with folds consisting of randomly selected locations.
The spatio-temporal random forest results in the second set of columns in Table 2 show somewhat different patterns. RMSE and R2 are generally improved over the daily models for the random and varying spatially clustered cross-validation analyses, but there is no longer a benefit to including imputed AOD. On the contrary, the model predictions tend to do better when neither AOD nor GEOS-Chem are included on the basis of R2 and RMSE. The exception to these results are in the constant spatially clustered cross-validation setting—here, there is some very marginal improvement from including imputed AOD over the other models. We posit that in spatio-temporal models, multiple days’ observations in the same area as where we intend to make a prediction on a different day can largely diminish the predictive power of AOD. However, when the same spatial area is consistently missing, the model can no longer rely on other days’ observations for the same area to improve prediction accuracy. Notably, this setting mirrors qualities of producing full maps of PM2.5 observations. Given that there is a fixed network of monitors (not all of which operate on every day), PM2.5 prediction is primarily focused on areas where there is never a monitor present. We emphasize that the percentage decrease in RMSE from including the imputed AOD in this constant spatially clustered cross-validation setting for the spatio-temporal random forest models is small at 1.25%, 0.67%, and 0.77% compared to the models with no AOD or GEOS-Chem, just GEOS-Chem, or the combined AOD/GEOS-Chem variable, respectively. Notably in this case, daily models slightly outperform the performance of the spatio-temporal models. In the other cross-validation settings, RMSE and R2 both improve substantially from fitting a full spatio-temporal model over a series of daily models.
Results for RMSE and R2 by region (as defined by NOAA) and cross-validation setting are also provided in Table 3 and Table 4. RMSE results tend to be worst in the West, Southwest, and Central regions across cross-validation settings. Notably, although RMSE is quite low for the Northwest region, the R2 for this area is comparatively low, suggesting low pollution levels but also poor prediction performance from the models. The spatio-temporal models improve the regional RMSE and R2 as compared to the daily models, except for the constant spatially clustered cross-validation setting, where there is variability in changes in regional performance. Variable importance metrics for the m = 4 setting based on the spatio-temporal models are presented in Table S1 using the permutation-based method [23]. Briefly, this importance metric denotes the increase in mean-squared error on the OOB sample for each tree after permuting the values of the feature. On this basis, the convolution layer of PM2.5 is the most important predictor for these models. When imputed AOD is included, the relative importance of several other variables is slightly diminished. While imputed AOD is not the most important feature, it appears substantively important on the basis of a mean decrease in accuracy. Additional feature importance tables are provided in the Supplemental Materials for m = 8 (Table S2). In general, for larger m, the convolution layer of PM2.5 becomes more important—it is more likely to be selected as the optimal feature for splitting a node as m increases, and it is a particularly strong predictor.
When comparing model M3 to models M1, M2, and M4, the average of monthly mean differences are close to 0 μg m−3, but the monthly mean differences are apparently spatially correlated (Figure 4). The model trained only on points where AOD is observed (M5) leads to over-estimated average monthly values of PM2.5 relative to the model using gap-filled AOD, with an average difference of 0.25 μg m−3. The standard deviation of daily differences for all grid cells for July 2011 is 0.44 μg m−3 for M3 and M1, 0.32 μg m−3 for M3 and M2, and 0.38 μg m−3 for M3 and M4, suggesting small but meaningful variability in the daily model predictions. Some of the largest daily differences between models tend to be in areas with sparse air pollution monitors (Figure S11). Figure 5 shows the July 2011 averaged values from M3 with the gap-filled AOD as a feature in the spatio-temporal random forest model.
2

4. Discussion

We highlight the main findings of this study in three points. First, we emphasize the importance of constructing testing and cross-validation data that mimic the missing data patterns for both AOD and PM2.5 prediction. Previously reported metrics for AOD gap-filling using RF may be over-stated if using out-of-bag (OOB) metrics [10,24], as using large contiguous areas for testing suggests substantially lower R2. Different cross-validation settings for PM2.5 model evaluation also suggest that performance varies considerably based on the manner of holdout, echoing the findings of previous studies that “spatial” cross-validation performance metrics are typically worse than random cross-validation metrics. In our study, our spatial cross-validation procedures leave out spatially clustered sets of monitors as in several recent studies [35,50,51]. Our results show roughly similar performance metrics for PM2.5 estimation compared to previous RF results when using the random cross-validation setting, with ≈0.80 (≈2.99) vs. 0.81 (2.78) R2 (RMSE) for summer 2011 in Hu et al. [17]. The small difference in performance in our model as compared to that of Hu et al. [17] is likely because we fit the model solely to data in July 2011 rather than the full year, and because we did not include additional variables such as convolutional layers for land use terms, which have been shown to improve overall performance.
Second, we demonstrate how super learner approaches combining a large-scale spatial statistical method and machine learning predictions can improve upon the performance of each constituent predictor, and how the super learner method can be further modified for the particular task of AOD gap-filling. Future work should examine extensions to more machine learning and spatial statistical methods. For example, several recent studies have highlighted a number of spatial statistical methods with promising predictive performance and low computational costs [30,31,52,53,54,55,56,57,58], and using these in an ensemble approach may provide further improvements. Spatial data present additional theoretical challenges for super learner methods, given that the training data and testing data will generally not be independent of each other [48]. A limitation of the current study is the limited time frame and the use of daily rather than spatio-temporal AOD gap-filling models. We focused on July 2011 as there was, on average, less missingness in AOD in the summer than in other months, and we limited our analysis to a single month due to the high computational cost of fitting daily models with 10-fold cross-validation for the super learner. Future studies may consider expanding the time frame of spatial prediction beyond a month and including spatio-temporal statistical models that may better utilize the available observed data.
Finally, we demonstrate that imputed AOD using our proposed ensemble method can have a very small impact on particular RF models for estimating PM2.5 concentrations, depending on the cross-validation setting. With a convolution layer of PM2.5 and a rich set of other features, we generally find that AOD (imputed or not) is not strictly needed for good prediction of PM2.5 in RF models as judged by R2 and RMSE. However, population-level metrics like R2 and RMSE may be misleading in masking improved small-scale predictions, and we find subtle differences in the predicted values between models, with and without gap-filled AOD as a predictor. Similarly, Huang et al. [25] find meaningful differences in daily PM2.5 predictions in models with and without AOD as a predictor, particularly in areas of the map with sparse PM2.5 monitors and on high-pollution days. Thus, although gap-filling AOD in the proposed ensemble approach adds to the computational costs of making PM2.5 predictions, as compared to fitting a model without AOD as a predictor, there may be some marginal benefit to the quality of PM2.5 predictions. A limitation of the current study is the lack of certain variables for AOD gap-filling and PM2.5 estimation; previous work has found that the inclusion of cloud and snow fractions may improve AOD gap-filling and produce meaningful visual improvements in PM2.5 estimation [10]. Moreover, finer resolution AOD products such as multi-angle implementation of atmospheric correction (MAIAC)-derived AOD may provide greater predictive power for PM2.5 [59] at the expense of increasing the computational costs of gap-filling. Future studies should examine the computational costs and benefits of using the proposed method to gap-fill AOD for predicting PM2.5 at the 1 km resolution.

5. Conclusions

This article considered a recently developed large-scale spatial statistical method (LK), a popular machine learning method (RF), and various combinations of these methods for gap-filling daily MODIS AOD on a 12 km × 12 km grid in the contiguous United States in July 2011. Using large contiguous areas as holdouts for our performance comparison, we found that ensemble approaches can improve daily AOD gap-filling as compared to the individual methods on their own, on the basis of RMSE and R2. The ultimate goal of making improvements in gap-filling AOD is to improve the performance of PM2.5 prediction models and to provide complete spatial coverage. For this task, we compared several daily and spatio-temporal random forest models, with and without the inclusion of gap-filled AOD as a predictor. These results were mixed, showing very small but consistent improvements to RMSE and R2 by including the gap-filled AOD in the daily models, but varying results in the spatio-temporal models. However, in the more realistic spatially clustered cross-validation setting, where spatial clusters of the observed locations are assigned to the same cross-validation fold for all days, we find small improvements for PM2.5 from including the gap-filled AOD predictor. Furthermore, although the differences in RMSE and R2 are small, daily prediction maps suggest some meaningful differences between the considered models, particularly in areas with sparse monitoring locations. Future research should extend this work by considering computationally efficient spatio-temporal statistical approaches for gap-filling AOD (as compared to daily gap-filling models), increasing the time frame of study, and considering gap-filling at finer resolutions (e.g., MAIAC AOD at 1 km resolution).

Supplementary Materials

All codes for replicating the analyses in this paper are provided at https://github.com/bzki/aodpm25_paper. Additional figures, tables, and details are available online at https://www.mdpi.com/2072-4292/13/1/126/s1. Figure S1: Daily split between training and testing data for AOD experiments, Figure S2: RMSE and R2 across days for various methods in AOD experiments, Figure S3: Daily observed and predicted AOD values, Figure S4: Daily differences between LK and RF AOD predictions, Figure S5: Differences in average predictions for various methods, Figure S6: Spatial cross-validation folds for PM2.5, Figure S7: Comparison of LK and RF at difference distances between test and training data, Figure S8: Constant spatial clustering cross-validation map for PM2.5, Figure S9: Average PM2.5 predicted map, Figure S10: Average differences from gap-filled AOD RF model, Figure S11: Daily differences from gap-filled AOD RF model, Figures S12–S14: Scatter plots of predicted and observed PM2.5, Tables S1 and S2: Feature importance tables from RF models for PM2.5, Tables S3–S5: Intercept and slope estimates from PM2.5 cross-validation, Section S3: Additional LatticeKrig modeling details.

Author Contributions

Conceptualization, B.K. and H.H.C.; methodology, B.K. and H.H.C.; software, B.K.; validation, B.K.; formal analysis, B.K.; investigation, B.K.; resources, H.H.C.; data curation, B.K. and H.H.C.; writing—original draft preparation, B.K.; writing—review and editing, B.K., H.H.C. and Y.L.; visualization, B.K., H.H.C. and Y.L.; supervision, H.H.C. and Y.L.; project administration, H.H.C. and Y.L.; funding acquisition, H.H.C. and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

The work of Y. Liu was supported by the NASA Applied Sciences Program (Grant # 80NSSC19K0191) and the MAIA science team at the JPL, California Institute of Technology (Subcontract #1588347). This research was also supported by the National Institute of Environmental Health Sciences of the National Institutes of Health under award number R01-ES027892. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. See the Materials and Methods section for details.

Acknowledgments

We thank Xuefei Hu for sharing the dataset and Douglas Nychka for generously responding to questions about the LatticeKrig package.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AODAerosol optical depth
PM2.5Particulate matter less than 2.5 micrometers in aerodynamic diameter
RFRandom forest
LKLattice kriging
SLSuper learner
RMSERoot mean square error
IDWInverse distance weighting
CTMChemical transport model
MODISModerate Resolution Imaging Spectroradiometer
MAIACMulti-angle implementation of atmospheric correction
OOBOut-of-bag

References

  1. WHO. Ambient (Outdoor) Air Pollution. 2018. Available online: https://web.archive.org/web/20200824220508/https%3A%2F%2Fwww.who.int%2Fnews-room%2Ffact-sheets%2Fdetail%2Fambient-%2528outdoor%2529-air-quality-and-health (accessed on 24 August 2020).
  2. Lim, S.S.; Vos, T.; Flaxman, A.D.; Danaei, G.; Shibuya, K.; Adair-Rohani, H.; AlMazroa, M.A.; Amann, M.; Anderson, H.R.; Andrews, K.G.; et al. A comparative risk assessment of burden of disease and injury attributable to 67 risk factors and risk factor clusters in 21 regions, 1990–2010: A systematic analysis for the Global Burden of Disease Study 2010. Lancet 2012, 380, 2224–2260. [Google Scholar] [CrossRef] [Green Version]
  3. Lelieveld, J.; Evans, J.S.; Fnais, M.; Giannadaki, D.; Pozzer, A. The contribution of outdoor air pollution sources to premature mortality on a global scale. Nature 2015, 525, 367–371. [Google Scholar] [CrossRef] [PubMed]
  4. Levy, R.; Mattoo, S.; Munchak, L.; Remer, L.; Sayer, A.; Patadia, F.; Hsu, N. The Collection 6 MODIS aerosol products over land and ocean. Atmos. Meas. Tech. 2013, 6, 2989. [Google Scholar] [CrossRef] [Green Version]
  5. Sorek-Hamer, M.; Just, A.; Kloog, I. Satellite remote sensing in epidemiological studies. Curr. Opin. Pediatr. 2016, 28, 228. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Chu, Y.; Liu, Y.; Li, X.; Liu, Z.; Lu, H.; Lu, Y.; Mao, Z.; Chen, X.; Li, N.; Ren, M.; et al. A review on predicting ground PM2.5 concentration using satellite aerosol optical depth. Atmosphere 2016, 7, 129. [Google Scholar] [CrossRef] [Green Version]
  7. Shin, M.; Kang, Y.; Park, S.; Im, J.; Yoo, C.; Quackenbush, L.J. Estimating ground-level particulate matter concentrations using satellite-based data: A review. GIScience Remote Sens. 2020, 57, 174–189. [Google Scholar] [CrossRef]
  8. Belle, J.H.; Liu, Y. Evaluation of Aqua MODIS Collection 6 AOD Parameters for Air Quality Research over the Continental United States. Remote Sens. 2016, 8, 815. [Google Scholar] [CrossRef] [Green Version]
  9. Belle, J.H.; Chang, H.H.; Wang, Y.; Hu, X.; Lyapustin, A.; Liu, Y. The potential impact of satellite-retrieved cloud parameters on ground-level PM2.5 mass and composition. Int. J. Environ. Res. Public Health 2017, 14, 1244. [Google Scholar] [CrossRef] [Green Version]
  10. Bi, J.; Belle, J.H.; Wang, Y.; Lyapustin, A.I.; Wildani, A.; Liu, Y. Impacts of snow and cloud covers on satellite-derived PM2.5 levels. Remote Sens. Environ. 2019, 221, 665–674. [Google Scholar] [CrossRef]
  11. Christopher, S.A.; Gupta, P. Satellite remote sensing of particulate matter air quality: The cloud-cover problem. J. Air Waste Manag. Assoc. 2010, 60, 596–602. [Google Scholar] [CrossRef]
  12. Liang, F.; Xiao, Q.; Huang, K.; Yang, X.; Liu, F.; Li, J.; Lu, X.; Liu, Y.; Gu, D. The 17-y spatiotemporal trend of PM2.5 and its mortality burden in China. Proc. Natl. Acad. Sci. USA 2020, 117, 25601–25608. [Google Scholar] [CrossRef] [PubMed]
  13. Geng, G.; Murray, N.L.; Tong, D.; Fu, J.S.; Hu, X.; Lee, P.; Meng, X.; Chang, H.H.; Liu, Y. Satellite-Based Daily PM2.5 Estimates During Fire Seasons in Colorado. J. Geophys. Res. Atmos. 2018, 123, 8159–8171. [Google Scholar] [CrossRef] [PubMed]
  14. Kloog, I.; Koutrakis, P.; Coull, B.A.; Lee, H.J.; Schwartz, J. Assessing temporally and spatially resolved PM2.5 exposures for epidemiological studies using satellite aerosol optical depth measurements. Atmos. Environ. 2011, 45, 6267–6275. [Google Scholar] [CrossRef]
  15. Kloog, I.; Nordio, F.; Coull, B.A.; Schwartz, J. Incorporating local land use regression and satellite aerosol optical depth in a hybrid model of spatiotemporal PM2.5 exposures in the Mid-Atlantic states. Environ. Sci. Technol. 2012, 46, 11913–11921. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Lee, M.; Kloog, I.; Chudnovsky, A.; Lyapustin, A.; Wang, Y.; Melly, S.; Coull, B.; Koutrakis, P.; Schwartz, J. Spatiotemporal prediction of fine particulate matter using high-resolution satellite images in the Southeastern US 2003–2011. J. Expo. Sci. Environ. Epidemiol. 2016, 26, 377–384. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Hu, X.; Belle, J.H.; Meng, X.; Wildani, A.; Waller, L.A.; Strickland, M.J.; Liu, Y. Estimating PM2.5 concentrations in the conterminous United States using the random forest approach. Environ. Sci. Technol. 2017, 51, 6936–6944. [Google Scholar] [CrossRef]
  18. Xiao, Q.; Wang, Y.; Chang, H.H.; Meng, X.; Geng, G.; Lyapustin, A.; Liu, Y. Full-coverage high-resolution daily PM2.5 estimation using MAIAC AOD in the Yangtze River Delta of China. Remote Sens. Environ. 2017, 199, 437–446. [Google Scholar] [CrossRef]
  19. Huang, K.; Xiao, Q.; Meng, X.; Geng, G.; Wang, Y.; Lyapustin, A.; Gu, D.; Liu, Y. Predicting monthly high-resolution PM2.5 concentrations with random forest model in the North China Plain. Environ. Pollut. 2018, 242, 675–683. [Google Scholar] [CrossRef]
  20. Lv, B.; Hu, Y.; Chang, H.H.; Russell, A.G.; Bai, Y. Improving the accuracy of daily PM2.5 distributions derived from the fusion of ground-level measurements with aerosol optical depth observations, a case study in North China. Environ. Sci. Technol. 2016, 50, 4752–4759. [Google Scholar] [CrossRef]
  21. Laslett, G.M. Kriging and splines: An empirical comparison of their predictive performance in some applications. J. Am. Stat. Assoc. 1994, 89, 391–400. [Google Scholar] [CrossRef]
  22. Chen, Z.Y.; Zhang, T.H.; Zhang, R.; Zhu, Z.M.; Yang, J.; Chen, P.Y.; Ou, C.Q.; Guo, Y. Extreme gradient boosting model to estimate PM2.5 concentrations with missing-filled satellite data in China. Atmos. Environ. 2019, 202, 180–189. [Google Scholar] [CrossRef]
  23. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  24. Stafoggia, M.; Bellander, T.; Bucci, S.; Davoli, M.; De Hoogh, K.; De’Donato, F.; Gariazzo, C.; Lyapustin, A.; Michelozzi, P.; Renzi, M.; et al. Estimation of daily PM10 and PM2.5 concentrations in Italy, 2013–2015, using a spatiotemporal land-use random-forest model. Environ. Int. 2019, 124, 170–179. [Google Scholar] [CrossRef]
  25. Huang, K.; Bi, J.; Meng, X.; Geng, G.; Lyapustin, A.; Lane, K.J.; Gu, D.; Kinney, P.L.; Liu, Y. Estimating daily PM2.5 concentrations in New York City at the neighborhood-scale: Implications for integrating non-regulatory measurements. Sci. Total Environ. 2019, 697, 134094. [Google Scholar] [CrossRef] [PubMed]
  26. Zhang, R.; Di, B.; Luo, Y.; Deng, X.; Grieneisen, M.L.; Wang, Z.; Yao, G.; Zhan, Y. A nonparametric approach to filling gaps in satellite-retrieved aerosol optical depth for estimating ambient PM2.5 levels. Environ. Pollut. 2018, 243, 998–1007. [Google Scholar] [CrossRef]
  27. Jiang, T.; Chen, B.; Nie, Z.; Ren, Z.; Xu, B.; Tang, S. Estimation of hourly full-coverage PM2.5 concentrations at 1-km resolution in China using a two-stage random forest model. Atmos. Res. 2021, 248, 105146. [Google Scholar] [CrossRef]
  28. Holben, B.N.; Eck, T.F.; Slutsker, I.A.; Tanre, D.; Buis, J.; Setzer, A.; Vermote, E.; Reagan, J.A.; Kaufman, Y.; Nakajima, T.; et al. AERONET—A federated instrument network and data archive for aerosol characterization. Remote Sens. Environ. 1998, 66, 1–16. [Google Scholar] [CrossRef]
  29. Sayer, A.; Munchak, L.; Hsu, N.; Levy, R.; Bettenhausen, C.; Jeong, M.J. MODIS Collection 6 aerosol products: Comparison between Aqua’s e-Deep Blue, Dark Target, and “merged” data sets, and usage recommendations. J. Geophys. Res. Atmos. 2014, 119, 13–965. [Google Scholar] [CrossRef]
  30. Heaton, M.J.; Datta, A.; Finley, A.O.; Furrer, R.; Guinness, J.; Guhaniyogi, R.; Gerber, F.; Gramacy, R.B.; Hammerling, D.; Katzfuss, M.; et al. A case study competition among methods for analyzing large spatial data. J. Agric. Biol. Environ. Stat. 2019, 24, 398–425. [Google Scholar] [CrossRef] [Green Version]
  31. Bradley, J.R.; Cressie, N.; Shi, T. A comparison of spatial predictors when datasets could be very large. Stat. Surv. 2016, 10, 100–131. [Google Scholar] [CrossRef]
  32. Shao, Y.; Ma, Z.; Wang, J.; Bi, J. Estimating daily ground-level PM2.5 in China with random-forest-based spatiotemporal kriging. Sci. Total Environ. 2020, 740, 139761. [Google Scholar] [CrossRef] [PubMed]
  33. Xiao, Q.; Chang, H.H.; Geng, G.; Liu, Y. An ensemble machine-learning model to predict historical PM2.5 concentrations in China from satellite data. Environ. Sci. Technol. 2018, 52, 13260–13269. [Google Scholar] [CrossRef] [PubMed]
  34. Di, Q.; Amini, H.; Shi, L.; Kloog, I.; Silvern, R.; Kelly, J.; Sabath, M.B.; Choirat, C.; Koutrakis, P.; Lyapustin, A.; et al. An ensemble-based model of PM2.5 concentration across the contiguous United States with high spatiotemporal resolution. Environ. Int. 2019, 130, 104909. [Google Scholar] [CrossRef] [PubMed]
  35. Murray, N.L.; Holmes, H.A.; Liu, Y.; Chang, H.H. A Bayesian ensemble approach to combine PM2.5 estimates from statistical models using satellite imagery and numerical model simulation. Environ. Res. 2019, 178, 108601. [Google Scholar] [CrossRef] [PubMed]
  36. Nychka, D.; Bandyopadhyay, S.; Hammerling, D.; Lindgren, F.; Sain, S. A multiresolution Gaussian process model for the analysis of large spatial datasets. J. Comput. Graph. Stat. 2015, 24, 579–599. [Google Scholar] [CrossRef] [Green Version]
  37. van der Laan, M.J.; Polley, E.C.; Hubbard, A.E. Super learner. Stat. Appl. Genet. Mol. Biol. 2007, 6. [Google Scholar] [CrossRef]
  38. Naimi, A.I.; Balzer, L.B. Stacked generalization: An introduction to super learning. Eur. J. Epidemiol. 2018, 33, 459–464. [Google Scholar] [CrossRef]
  39. Levy, R.; Hsu, C. MODIS Atmosphere L2 Aerosol Product. NASA MODIS Adaptive Processing System; Goddard Space Flight Center: Greenbelt, MD, USA, 2015; Vulume 10. [CrossRef]
  40. Bey, I.; Jacob, D.J.; Yantosca, R.M.; Logan, J.A.; Field, B.D.; Fiore, A.M.; Li, Q.; Liu, H.Y.; Mickley, L.J.; Schultz, M.G. Global modeling of tropospheric chemistry with assimilated meteorology: Model description and evaluation. J. Geophys. Res. Atmos. 2001, 106, 23073–23095. [Google Scholar] [CrossRef]
  41. Li, S.; Garay, M.J.; Chen, L.; Rees, E.; Liu, Y. Comparison of GEOS-Chem aerosol optical depth with AERONET and MISR data over the contiguous United States. J. Geophys. Res. Atmos. 2013, 118, 11–228. [Google Scholar] [CrossRef]
  42. Cosgrove, B.A.; Lohmann, D.; Mitchell, K.E.; Houser, P.R.; Wood, E.F.; Schaake, J.C.; Robock, A.; Marshall, C.; Sheffield, J.; Duan, Q.; et al. Real-time and retrospective forcing in the North American Land Data Assimilation System (NLDAS) project. J. Geophys. Res. Atmos. 2003, 108. [Google Scholar] [CrossRef] [Green Version]
  43. Mitchell, K.E.; Lohmann, D.; Houser, P.R.; Wood, E.F.; Schaake, J.C.; Robock, A.; Cosgrove, B.A.; Sheffield, J.; Duan, Q.; Luo, L.; et al. The multi-institution North American Land Data Assimilation System (NLDAS): Utilizing multiple GCIP products and partners in a continental distributed hydrological modeling system. J. Geophys. Res. Atmos. 2004, 109. [Google Scholar] [CrossRef] [Green Version]
  44. Nychka, D.; Hammerling, D.; Sain, S.; Lenssen, N. LatticeKrig: Multiresolution Kriging Based on Markov Random Fields. R Package Version 8.4, 2016. Available online: https://cran.r-project.org/web/packages/LatticeKrig/index.html (accessed on 30 December 2020). [CrossRef]
  45. Wright, M.N.; Ziegler, A. ranger: A Fast Implementation of Random Forests for High Dimensional Data in C++ and R. J. Stat. Softw. 2017, 77. [Google Scholar] [CrossRef] [Green Version]
  46. Breiman, L. Stacked regressions. Mach. Learn. 1996, 24, 49–64. [Google Scholar] [CrossRef] [Green Version]
  47. Polley, E.C.; van der Laan, M.J. Super learner in prediction. In U.C. Berkeley Division of Biostatistics Working Paper Series. Working Paper 266; U.C. Berkeley: Berkeley, CA, USA, 2010. [Google Scholar]
  48. Davies, M.M.; van der Laan, M.J. Optimal spatial prediction using ensemble machine learning. Int. J. Biostat. 2016, 12, 179–201. [Google Scholar] [CrossRef] [Green Version]
  49. Valavi, R.; Elith, J.; Lahoz-Monfort, J.J.; Guillera-Arroita, G. blockCV: An R package for generating spatially or environmentally separated folds for k-fold cross-validation of species distribution models. Methods Ecol. Evol. 2019, 10, 225–232. [Google Scholar] [CrossRef] [Green Version]
  50. Sarafian, R.; Kloog, I.; Just, A.C.; Rosenblatt, J.D. Gaussian Markov Random Fields versus Linear Mixed Models for satellite-based PM2.5 assessment: Evidence from the Northeastern USA. Atmos. Environ. 2019, 205, 30–35. [Google Scholar] [CrossRef] [Green Version]
  51. Young, M.T.; Bechle, M.J.; Sampson, P.D.; Szpiro, A.A.; Marshall, J.D.; Sheppard, L.; Kaufman, J.D. Satellite-based NO2 and model validation in a national prediction model based on universal kriging and land-use regression. Environ. Sci. Technol. 2016, 50, 3686–3694. [Google Scholar] [CrossRef] [Green Version]
  52. Cressie, N.; Shi, T.; Kang, E.L. Fixed rank filtering for spatio-temporal data. J. Comput. Graph. Stat. 2010, 19, 724–745. [Google Scholar] [CrossRef] [Green Version]
  53. Lindgren, F.; Rue, H.; Lindström, J. An explicit link between Gaussian fields and Gaussian Markov random fields: The stochastic partial differential equation approach. J. R. Stat. Soc. Ser. B Stat. Methodol. 2011, 73, 423–498. [Google Scholar] [CrossRef] [Green Version]
  54. Datta, A.; Banerjee, S.; Finley, A.O.; Gelfand, A.E. Hierarchical nearest-neighbor Gaussian process models for large geostatistical datasets. J. Am. Stat. Assoc. 2016, 111, 800–812. [Google Scholar] [CrossRef] [Green Version]
  55. Datta, A.; Banerjee, S.; Finley, A.O.; Hamm, N.A.S.; Schaap, M. Nonseparable dynamic nearest neighbor Gaussian process models for large spatio-temporal data with an application to particulate matter analysis. Ann. Appl. Stat. 2016, 10, 1286–1316. [Google Scholar] [CrossRef] [PubMed]
  56. Bradley, J.R. What is the best predictor that you can compute in five minutes using a given Bayesian hierarchical model? arXiv 2019, arXiv:1912.04542. [Google Scholar]
  57. Katzfuss, M. A multi-resolution approximation for massive spatial datasets. J. Am. Stat. Assoc. 2017, 112, 201–214. [Google Scholar] [CrossRef] [Green Version]
  58. Appel, M.; Pebesma, E. Spatiotemporal multi-resolution approximations for analyzing global environmental data. Spat. Stat. 2020, 38, 100465. [Google Scholar] [CrossRef]
  59. Goldberg, D.L.; Gupta, P.; Wang, K.; Jena, C.; Zhang, Y.; Lu, Z.; Streets, D.G. Using gap-filled MAIAC AOD and WRF-Chem to estimate daily PM2.5 concentrations at 1 km resolution in the Eastern United States. Atmos. Environ. 2019, 199, 443–452. [Google Scholar] [CrossRef]
Figure 1. Moderate Resolution Imaging Spectroradiometer (MODIS) aerosol optical depth (AOD) on 12 km grid; full observed data for (a) 1 July and (b) 12 July 2011; training data for (c) July 1 and (d) 12 July 2011. Notably, 1 July has the least missingness, and 12 July has the most missingness in July 2011. Grid cells with observed AOD values greater than 1 are excluded from display.
Figure 1. Moderate Resolution Imaging Spectroradiometer (MODIS) aerosol optical depth (AOD) on 12 km grid; full observed data for (a) 1 July and (b) 12 July 2011; training data for (c) July 1 and (d) 12 July 2011. Notably, 1 July has the least missingness, and 12 July has the most missingness in July 2011. Grid cells with observed AOD values greater than 1 are excluded from display.
Remotesensing 13 00126 g001
Figure 2. July 2011 average of observed and predicted daily AOD: (a) observed AOD; (b) lattice kriging (LK); (c) random forest (RF); (d) average of LK and RF; (e) Super-learner (SL): overall; (f) SL: daily; (g) SL: distance-based. Grid cells with observed AOD values greater than 1 are excluded from display.
Figure 2. July 2011 average of observed and predicted daily AOD: (a) observed AOD; (b) lattice kriging (LK); (c) random forest (RF); (d) average of LK and RF; (e) Super-learner (SL): overall; (f) SL: daily; (g) SL: distance-based. Grid cells with observed AOD values greater than 1 are excluded from display.
Remotesensing 13 00126 g002
Figure 3. Difference between lattice kriging and random forest averaged daily AOD predictions for July 2011.
Figure 3. Difference between lattice kriging and random forest averaged daily AOD predictions for July 2011.
Remotesensing 13 00126 g003
Figure 4. Differences between the random forest (RF) model that includes gap-filled AOD as a predictor (M3) and alternative RF models for average July 2011 PM2.5 predictions (μg m−3): (a) M1: no AOD or GEOS-Chem included as predictors; (b) M2: GEOS-Chem included as a predictor; (c) M4: Replacing missing values of AOD with GEOS-Chem; (d) M5: Training on observed AOD, and predicting by replacing missing AOD values with the gap-filled values.
Figure 4. Differences between the random forest (RF) model that includes gap-filled AOD as a predictor (M3) and alternative RF models for average July 2011 PM2.5 predictions (μg m−3): (a) M1: no AOD or GEOS-Chem included as predictors; (b) M2: GEOS-Chem included as a predictor; (c) M4: Replacing missing values of AOD with GEOS-Chem; (d) M5: Training on observed AOD, and predicting by replacing missing AOD values with the gap-filled values.
Remotesensing 13 00126 g004
Figure 5. Average July 2011 PM2.5 predicted map (μg m−3) using the spatio-temporal random forest model that includes gap-filled AOD as a predictor (model M3).
Figure 5. Average July 2011 PM2.5 predicted map (μg m−3) using the spatio-temporal random forest model that includes gap-filled AOD as a predictor (model M3).
Remotesensing 13 00126 g005
Table 1. Summary statistics on the combined test AOD predictions across all days of July 2011. The two right columns denote the number of days in which the method performed best or worst on the basis of root-mean-square error (RMSE).
Table 1. Summary statistics on the combined test AOD predictions across all days of July 2011. The two right columns denote the number of days in which the method performed best or worst on the basis of root-mean-square error (RMSE).
Method # of Days Ranked
R2RMSE (×100)InterceptSlopeBestWorst
LatticeKrig0.6446.66−0.010.94712
Random Forest0.6196.90−0.010.92319
LK-RF Average0.6586.52−0.010.9740
SL: Overall0.6596.51−0.010.9740
SL: Daily0.6576.52−0.010.9630
SL: Distance-based0.6596.50−0.010.96100
Table 2. R2 and root mean square error (RMSE) result from the daily and spatio-temporal random forest models for three different 10-fold cross-validation settings. Summary metrics are evaluated on all observations as well as separately on the basis of AOD’s missingness status (observed or missing). The M3a and M3b random forest models include gap-filled AOD as a predictor.
Table 2. R2 and root mean square error (RMSE) result from the daily and spatio-temporal random forest models for three different 10-fold cross-validation settings. Summary metrics are evaluated on all observations as well as separately on the basis of AOD’s missingness status (observed or missing). The M3a and M3b random forest models include gap-filled AOD as a predictor.
DailySpatio-Temporal
SettingAOD StatusM1aM2aM3aM4aM5aM1bM2bM3bM4bM5b
RMSE (μg m−3)
RandomAll3.303.293.283.303.682.962.982.992.983.20
RandomMissing3.303.283.273.293.842.993.003.023.013.29
RandomObserved3.313.303.293.313.412.922.942.952.943.04
Constant clusterAll3.663.643.623.643.993.683.663.633.663.75
Constant clusterMissing3.613.583.563.594.093.633.603.573.613.74
Constant clusterObserved3.743.733.723.733.823.763.753.723.733.76
Varying clusterAll3.663.653.633.654.003.333.343.353.353.56
Varying clusterMissing3.613.583.573.604.113.343.333.343.343.61
Varying clusterObserved3.753.743.723.743.833.333.353.353.353.47
R2 (×100)
RandomAll75.375.575.775.469.780.480.279.880.077.1
RandomMissing75.976.276.376.168.180.580.479.980.176.5
RandomObserved73.873.974.173.872.379.979.779.379.578.0
Constant clusterAll69.970.170.470.164.169.669.970.369.968.6
Constant clusterMissing71.571.872.271.763.671.271.672.171.569.8
Constant clusterObserved66.866.967.266.965.266.466.767.166.866.6
Varying clusterAll69.770.070.370.063.975.675.675.475.572.1
Varying clusterMissing71.471.872.171.763.276.376.476.276.372.2
Varying clusterObserved66.566.667.166.765.174.173.973.973.771.8
Table 3. Regional root-mean-square error (RMSE) results (μg m−3) for daily and spatio-temporal random forest model for different 10-fold cross-validation settings. The M3a and M3b random forest models include gap-filled AOD as a predictor.
Table 3. Regional root-mean-square error (RMSE) results (μg m−3) for daily and spatio-temporal random forest model for different 10-fold cross-validation settings. The M3a and M3b random forest models include gap-filled AOD as a predictor.
DailySpatio-Temporal
SettingAOD StatusM1aM2aM3aM4aM5aM1bM2bM3bM4bM5b
RandomCentral3.703.673.653.684.153.563.563.583.563.78
East North Central3.173.173.163.183.523.043.033.013.023.15
Northeast3.313.273.263.293.742.972.963.002.973.23
Northwest1.481.491.491.491.821.231.241.241.241.35
South2.352.352.332.352.772.112.132.152.152.39
Southeast3.553.543.533.564.223.393.413.403.433.72
Southwest4.134.114.094.104.433.583.613.623.543.94
West4.414.424.414.404.473.583.643.703.673.76
West North Central2.912.922.922.923.082.342.372.352.352.50
Constant clusterCentral4.384.344.324.344.674.544.494.464.504.60
East North Central3.363.363.343.363.673.463.413.383.433.48
Northeast3.553.513.493.533.903.543.493.463.503.60
Northwest1.481.481.501.491.841.461.461.461.461.53
South2.472.472.462.472.902.492.472.472.482.64
Southeast3.913.883.833.904.543.943.923.853.934.08
Southwest4.294.274.264.274.604.344.344.354.324.47
West5.205.215.215.195.285.105.115.085.085.08
West North Central3.043.043.043.043.183.093.083.083.083.09
Varying clusterCentral4.394.344.324.344.704.294.264.264.274.50
East North Central3.343.353.343.343.743.303.293.283.293.37
Northeast3.533.493.493.513.943.263.233.273.253.49
Northwest1.511.511.521.531.851.291.301.311.311.39
South2.482.492.482.492.932.312.332.342.332.59
Southeast3.923.893.843.924.543.763.743.703.754.03
Southwest4.454.444.424.424.763.974.024.074.024.30
West5.165.175.145.165.184.184.234.264.254.42
West North Central2.972.972.972.973.112.402.432.452.452.57
Table 4. Regional R2 (×100) results for daily and spatio-temporal random forest model for different 10-fold cross-validation settings. The M3a and M3b random forest models include gap-filled AOD as a predictor.
Table 4. Regional R2 (×100) results for daily and spatio-temporal random forest model for different 10-fold cross-validation settings. The M3a and M3b random forest models include gap-filled AOD as a predictor.
DailySpatio-Temporal
SettingAOD StatusM1aM2aM3aM4aM5aM1bM2bM3bM4bM5b
RandomCentral61.161.762.161.551.164.764.763.864.159.3
East North Central69.469.669.769.363.372.873.272.672.570.7
Northeast77.077.677.877.470.881.982.081.381.678.4
Northwest44.644.244.044.528.261.260.960.060.153.6
South68.068.168.468.057.974.574.073.273.167.5
Southeast70.470.670.870.259.573.273.172.872.368.3
Southwest46.647.347.747.442.461.160.659.361.051.9
West50.750.550.850.949.268.767.765.766.464.8
West North Central39.239.138.938.833.361.560.460.660.755.3
Constant clusterCentral48.049.049.449.240.145.046.046.745.943.5
East North Central66.066.166.666.260.864.465.666.365.165.1
Northeast73.874.474.774.168.174.074.775.174.573.3
Northwest45.445.044.345.029.145.645.645.345.543.7
South64.564.664.964.553.564.364.965.064.661.5
Southeast64.464.965.964.553.564.164.465.664.262.5
Southwest41.942.542.842.635.140.540.440.340.937.5
West34.634.634.535.132.036.636.637.337.237.5
West North Central34.034.134.134.028.832.132.432.632.431.6
Varying clusterCentral47.348.548.948.639.050.851.451.551.545.4
East North Central66.566.566.766.559.168.569.169.369.067.4
Northeast74.374.874.974.667.679.079.378.779.175.3
Northwest42.542.442.041.926.758.157.757.056.852.1
South64.364.264.464.152.770.269.969.669.863.1
Southeast64.364.865.864.253.268.068.369.068.163.5
Southwest37.537.838.538.530.653.052.050.952.843.7
West33.833.634.434.132.858.857.757.357.552.4
West North Central36.636.736.836.631.459.858.858.158.453.4
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kianian, B.; Liu, Y.; Chang, H.H. Imputing Satellite-Derived Aerosol Optical Depth Using a Multi-Resolution Spatial Model and Random Forest for PM2.5 Prediction. Remote Sens. 2021, 13, 126. https://doi.org/10.3390/rs13010126

AMA Style

Kianian B, Liu Y, Chang HH. Imputing Satellite-Derived Aerosol Optical Depth Using a Multi-Resolution Spatial Model and Random Forest for PM2.5 Prediction. Remote Sensing. 2021; 13(1):126. https://doi.org/10.3390/rs13010126

Chicago/Turabian Style

Kianian, Behzad, Yang Liu, and Howard H. Chang. 2021. "Imputing Satellite-Derived Aerosol Optical Depth Using a Multi-Resolution Spatial Model and Random Forest for PM2.5 Prediction" Remote Sensing 13, no. 1: 126. https://doi.org/10.3390/rs13010126

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop