Next Article in Journal
Empirical and Comparative Validation for a Building Energy Model Calibration Methodology
Next Article in Special Issue
Comparisons of Laboratory and On-Road Type-Approval Cycles with Idling Emissions. Implications for Periodical Technical Inspection (PTI) Sensors
Previous Article in Journal
Robust Wavelength Selection Using Filter-Wrapper Method and Input Scaling on Near Infrared Spectral Data
Previous Article in Special Issue
Wildfire Smoke Adjustment Factors for Low-Cost and Professional PM2.5 Monitors with Optical Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Machine Learning for the in-Field Correction of a PM2.5 Low-Cost Sensor Network

by
Wen-Cheng Vincent Wang
1,
Shih-Chun Candice Lung
1,2,3,* and
Chun-Hu Liu
1
1
Research Center for Environmental Changes, Academia Sinica, Nangang, Taipei 115, Taiwan
2
Department of Atmospheric Sciences, National Taiwan University, Taipei 106, Taiwan
3
Institute of Environmental Health, National Taiwan University, Taipei 106, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(17), 5002; https://doi.org/10.3390/s20175002
Submission received: 12 July 2020 / Revised: 27 August 2020 / Accepted: 31 August 2020 / Published: 3 September 2020
(This article belongs to the Collection Sensors for Air Quality Monitoring)

Abstract

:
Many low-cost sensors (LCSs) are distributed for air monitoring without any rigorous calibrations. This work applies machine learning with PM2.5 from Taiwan monitoring stations to conduct in-field corrections on a network of 39 PM2.5 LCSs from July 2017 to December 2018. Three candidate models were evaluated: Multiple linear regression (MLR), support vector regression (SVR), and random forest regression (RFR). The model-corrected PM2.5 levels were compared with those of GRIMM-calibrated PM2.5. RFR was superior to MLR and SVR in its correction accuracy and computing efficiency. Compared to SVR, the root mean square errors (RMSEs) of RFR were 35% and 85% lower for the training and validation sets, respectively, and the computational speed was 35 times faster. An RFR with 300 decision trees was chosen as the optimal setting considering both the correction performance and the modeling time. An RFR with a nighttime pattern was established as the optimal correction model, and the RMSEs were 5.9 ± 2.0 μg/m3, reduced from 18.4 ± 6.5 μg/m3 before correction. This is the first work to correct LCSs at locations without monitoring stations, validated using laboratory-calibrated data. Similar models could be established in other countries to greatly enhance the usefulness of their PM2.5 sensor networks.

1. Introduction

Millions of premature deaths worldwide can be attributed to particulate matter with an aerodynamic diameter less than or equal to 2.5 μm (PM2.5) [1,2], which is one of the human carcinogens classified by the International Agency for Research on Cancer [3]. Rising PM2.5 levels in the ambient air and their associated health impacts are important environmental health issues that concern the general public, especially in developing countries [4,5]. In eastern Asia during 1998–2000, 51% of the population lived in areas with annual mean PM2.5 levels above the recommended guideline of the World Health Organization (35 μg/m3). This percentage increased to 70% during 2010–2012 [6], showing the deterioration of the air quality in this region.
In resource-limited Asian countries, there are insufficient numbers of regulatory monitoring stations in urban areas with high population densities. The purpose of the PM2.5 monitoring stations of Environmental Protection Administrations (EPAs) worldwide is to assess the well-mixed ambient pollutant levels. Therefore, such monitors are situated at a height of 10–15 m above the ground. However, the intensive emissions of community pollution sources in Asia, such as restaurants and temples, result in high PM2.5 levels in the immediate living environments of citizens at street level [7,8,9,10]. Even living in the same airshed, residents from different communities with different emission sources are exposed to different PM2.5 levels, demonstrating the need for community air monitoring. In Taiwan island, there are 57 regular air monitoring stations, covering an area of 35,887 km2 [11]. Excluding the two-thirds of mountainous areas with little population, on average, every 300 km2 of land area has only one air quality station. Even in the capital area (Taipei metropolitan), every 141 km2 has only one air quality station on average. Thus, there are still large monitoring gaps between these stations. These gaps can be filled by the observations provided by low-cost sensors (LCSs) for ambient-level and street-level air quality monitoring.
The United Stations Environmental Protection Agency (USEPA) noted that LCSs can be used to fill the observational gaps of environmental monitoring after appropriate quality assurance and quality control procedures [12]. Currently, there are more than 4000 PM2.5 LCSs in the citizen air quality network (CAQN [13]) in Taiwan as a result of the collaboration of citizens, private enterprises, and non-governmental organizations (NGOs) with partial government funding. The PM2.5 sensor network on the Island of Taiwan is possibly one of the most densely distributed sensor networks in the world [14,15], with (on average) three LCSs per 1 km2. CAQN has increased the environmental awareness of Taiwanese citizens dramatically. However, these sensors installed by citizens or NGOs are not calibrated, and data accuracy is not assured. Thus, their readings may be 2–3 times the actual concentrations [10]. Such limitations in data quality have restrained the application of these sensors in research and policy development.
LCSs need to be calibrated, since their readings usually deviate from those of research-grade instruments [10]. The actual cut-off points of particle sizes for LCSs are not consistent with the statements declared by the manufactures [16]; individual correction equations are, thus, needed for each sensor, even those from the same batch [10]. Most environmental scientists evaluate the performance of LCSs with side-by-side comparisons against research-grade instruments in the laboratory or in the field. Research-grade instruments, which are considered the golden standard for such comparisons, include instruments with measuring principles based on light scattering, oscillating tapered elements, β-attenuation, and filter-weighing, such as GRIMM instruments [10,17,18,19,20], TSI SidePake AM510 [21,22], tapered element oscillating microbalance [18,23], β-attenuation monitor [17,18,19,24,25], and ultrasonic personal aerosol sampler [26]. Correction equations are established to adjust the LCS readings to those of research-grade measurements. Mathematical models, such as linear regression [17,18,19,20,27] or multiple linear regression (MLR) models [23,25,28], have been used to establish these correction equations. The latter models usually consider ambient environmental factors.
To obtain accurate PM2.5 levels for LCSs, data correction is a necessary procedure. For thousands of PM2.5 sensors already installed by citizens in the field, an innovative method for correction is needed, since traditional laboratory or field evaluations with side-by-side comparisons require a great deal of human resources and time. Furthermore, it is impractical to correct thousands of sensors. Innovative data science techniques, such as machine learning, could also be used for correcting LCS readings [29]. Machine learning techniques have been used for air pollution issues for different purposes. Cheng et al. applied transfer learning to correct the PM2.5 values of the citizen network in Beijing, China using the PM2.5 levels from public environmental monitoring stations [30]. Pandey et al. predicted the levels of ultrafine particles in China through a machine learning approach [31]. Hsieh et al. used an artificial neural network and deep learning to analyze the proper position to install air pollution stations in Beijing [32]. Zheng et al. analyzed and predicted the air quality in different areas through data mining and machine learning using data from 43 cities in China [33]. Paas et al. used an artificial neural network to acquire the mass concentrations and numbers of different sizes of particles in Germany [34]. Peng et al. forecasted the levels of ozone, PM2.5, and NO2 in Canada using multiple linear regression and multi-layer perceptron neural networks [35]. Moreover, machine learning methods may provide a significant accuracy improvement for gaseous and particle sensors [23,36].
Therefore, we proposed a hybrid method combining laboratory evaluations and data science to ensure that the LCS networks provide accurate PM data [20]. First, the LCS data are corrected through side-by-side laboratory comparisons for “seed” LCS devices, which can be installed strategically in areas without EPA stations; secondly, statistical or machine learning methods are applied to adjust nearby uncalibrated LCS devices using data from the EPA stations or the seed LCS devices wherever available. In this way, readings from uncalibrated LCS devices in the CAQN can be corrected to nearly research-grade observations. The first part of obtaining reliable and robust correction equations is to convert the readings of LCS devices to research-grade (or FEM-comparable) measurements via side-by-side comparisons with research-grade instruments in the laboratory, as presented in [20].
The current work focuses on the second part of this process: Applying machine learning to correct data from the LCS network with the PM2.5 values from Taiwan EPA stations. The objectives of this work are: (1) To establish data correction models based on machine learning techniques with the PM2.5 data from the Taiwan EPA to correct the readings of the sensor network; (2) to evaluate the model performance with data from the same sensor network calibrated with laboratory evaluations; and (3) to explore the best data correction models using choices of computing efficiency and day/night periods. As can be seen in the later section, using in-field corrections with machine learning techniques, the PM2.5 data quality of LCSs can be greatly improved. Furthermore, this is the first work to introduce the use of a nighttime dataset instead of a whole-day dataset for the establishment of a data correction model with machine learning in order to prevent interference from local emissions during the daytime. This method can be further used to conduct in-field corrections for CAQN in Taiwan, as well as for other sensor networks in other countries.

2. Materials and Methods

2.1. Sensor Network Introduction

The LCS network corrected by Taiwan EPA data in this work consists of LCS devices designed for research purposes, namely, AS-LUNG-O. LCS devices integrate LCS, power, and data transmission components. AS-LUNG-O is an LCS device integrated by our team and designed for long-term outdoor monitoring for scientific research [10]. AS stands for Academia Sinica (the research institute that supports its development), while LUNG indicates the human organ most commonly affected by air pollutants, and O indicates the “outdoor” version. AS-LUNG-O (~650 USD basic manufacturing costs) incorporates sensors for PM2.5 (PMS3003, Plantower, Beijing, China), CO2 (S8, Senseair AB, Sweden), temperature/humidity (SHT31, SENSIRION, Staefa ZH, Switzerland), and Global Positioning System (GPS, u-blox, Switzerland). The PM2.5 sensor, PMS3003, has been evaluated by several research teams in laboratory environments. For example, Kelly et al. [18] obtained an R2 of 0.73–0.97 in wind tunnels, and Sayahi et al. [37] obtained an R2 > 0.978 for 242 sets of PMS3003 in a controlled chamber. These results indicated the good performance of PMS3003 compared to research-grade instruments.
The sensors for AS-LUNG-O are placed in a waterproof shelter connected to a solar panel with backup batteries for the power supply, with the option to use household electricity, where easily accessible. The size of the whole set is roughly 60 cm (W) × 50 cm (D) × 50 cm (H), with a weight of approximately 4.8 kg. Data can be transmitted wirelessly using the built-in 4G modules to a cloud database with one-min intervals. An SD card was added as a complement to avoid data loss during wireless transmission. Currently, most of the LCSs used in the CAQN (http://www.aqmd.gov/aq-spec/product/edimax) in Taiwan are PMS5003 (Plantower, Beijing, China), which is also a Plantower LCS.
AS-LUNG-O is a versatile LCS device capable of operating under the harsh weather conditions in subtropical Taiwan, which experiences high humidity (e.g., The mean relative humidity (RH) was 74% in the year of 2016 [10]) and frequent Typhoons [38]. AS-LUNG-O can be installed on the light poles in the streets and was used in a small town in a mountainous area to fill the data gaps of PM2.5 monitoring. The incremental PM2.5 concentration increases due to different community sources were, thus, quantified using AS-LUNG-O [10]. Therefore, other communities without EPA monitoring stations can also use AS-LUNG-O to acquire the PM2.5 levels.
This work uses data from the AS-LUNG-O network, including 39 AS-LUNG-O sets installed in different communities since July 2017 around Taiwan (Figure 1). Twenty-eight sets were installed in urban communities in Taipei’s metropolitan areas with high population densities in northern Taiwan, while two sets were placed in the suburban communities of Taipei. In addition, 6, 2, and 1 set(s) were installed in central, southern, and eastern Taiwan, respectively. Out of the 28 sets in the urban communities in Taipei, 27 were installed near certain community sources such as traffic, restaurants, temples, night markets, etc., as described in [10]. These 27 sets were set-up at street-level on light poles around 2–2.5 m above the ground. The other 12 sets were set-up at a high-level (around 10–15 m above the ground) on the rooftops of elementary schools or government buildings to assess the PM2.5 in ambient air.
The AS-LUNG-O network is considered a research-grade sensor network, since the data of AS-LUNG-O were corrected using correction equations based on laboratory evaluations with side-by-side comparisons against a research-grade instrument for every AS-LUNG-O reading [10]. The research-grade instrument used in these laboratory evaluations was GRIMM 1.109 (GRIMM Aerosol Technik GmbH and Co. KG, Ainring, Germany). The data from GRIMM 1.109 were in excellent agreement (R2 = 0.999, with a bias of roughly ±11%) with the data from an EDM-180 (GRIMM Aerosol Technik Ainring GmbH and Co, Ainring, German) [20], an FEM instrument designated by the USEPA for PM2.5. The mean values of R2 for the correction equations were 0.97, with ranges from 0.82 to 0.99 for these 39 sets. Without data correction, AS-LUNG-O overestimates PM2.5 by about 1.5–2.9 times [10,20]. PM2.5 observations with 1 min resolutions from AS-LUNG-O were converted to GRIMM comparable measurements according to the correction equations and saved in the cloud database for this AS-LUNG-O network.
This research-grade AS-LUNG-O network provides a great opportunity to evaluate the feasibility and performance of the correction models based on machine learning techniques. The raw PM2.5 readings of AS-LUNG-O corrected by the laboratory correction equations are “GRIMM-calibrated PM2.5”, while those corrected by the machine learning techniques are “model-corrected PM2.5”. The performance of the machine learning correction models can be evaluated by comparing the GRIMM-calibrated PM2.5 with model-corrected PM2.5.

2.2. The Data Correction Models

This work applied machine learning techniques to correct the raw readings of AS-LUNG-O sets with those of Taiwan EPA monitoring stations. Only EPA stations within 3 km of the AS-LUNG-O sets were selected in our work. Out of the 57 regular Taiwan EPA stations, 11 stations were selected, including 6 stations in the north, 2 in central Taiwan, 2 in the south, and 1 in the east. Hourly measurements of 24 h during July 2017 and December 2018 from these stations were used to establish the data correction models. The instruments used in Taiwan EPA [39] are Met One BAM-1020 (Met One, Inc., Grants Pass, OR, USA) and VEREWA-F701 (VEREWA, Ltd., Germany).
To establish in-field data correction models, two machine learning techniques (introduced below) were used in this work and compared with a correction model established using traditional MLR. These three models used PM2.5 data from EPA stations as their simulation targets to adjust the raw readings of AS-LUNG-O sets within a 3 km radius. These models were constructed using a personal computer environment with an Intel® CoreTM i7-8700 and 32 GB RAM.
The inputs for these models were: (a) The raw PM2.5 readings, (b) the temperature, RH, latitude, and longitude of the AS-LUNG-O sets, (c) the PM2.5 levels of the nearest EPA station, and (d) the distance between AS-LUNG-O and the EPA station. These inputs (from July 2017 to December 2018) were used for 10-fold cross-validation (90% of data randomly selected for the training set; the others for the validation set) and holdout validation (50% of data randomly selected for the training set; the others for the validation set) tests to evaluate the robustness of these three models. Since there were only 1.5 years of data, we used the holdout method to generate the correction models; therefore, most of the data could be kept to validate the models, which can avoid to overvaluing the performance of the models under the situation of only using less data (10% of data) for the evaluation. Leaving more data in the validation set (50% of data in the holdout validation in this study) could increase the power of the model estimation. [40,41]. Data correction models were constructed using the training dataset, and the validation dataset was used to evaluate the correction accuracy of the models built. The MLR and machine learning models are introduced below.
The MLR model is established as follows:
P M 2.5 t a r g e t = β 0 + β 1 × L C S P M 2.5 r a w + β 2 × T + β 3 × R H + β 4 × M o n t h + β 5 × D a y     + β 6 × H r + β 7 × l a t + β 8 × l o n + β 9 × D
where β 0 is the intercept; β 1 β 9 are the regression coefficients; P M 2.5 t a r g e t is the simulation target of the correction model, EPA PM2.5; L C S P M 2.5 r a w is the raw readings from AS-LUNG-O (μg/m3); T is the temperature (°C); RH is relative humidity (%); Month, Day, and Hr (hour) are the time values of the observations; lat and lon are the latitude and longitude of the AS-LUNG-O sets; and D is the distance of the AS-LUNG-O and the nearest EPA station (km).
Two machine learning techniques used in this work were support vector machine (SVM) and random forest (RF). SVM is based on the generalized portrait algorithm developed in the 1960s by Russian mathematicians and is a supervised learning algorithm used for classification [42,43]. The SVM algorithm is a popular machine learning tool that offers solutions for both classification and regression problems. The objective of SVM is to build an optimal hyperplane as a classifier in high dimensional space, and the data points closest to the hyperplane are called support vectors. New data are then divided by that classifier and predicted to belong to a category based on the hyperplane [44]. Our present work applies SVM to construct the support vector regression (SVR) model.
The random forest model is an ensemble learning method for classification and regression that builds a multitude of decision trees during the training process and constructs the modes of the classes or the mean predictions for classification and regression [45,46,47]. Using the random subspace method to build decision trees was first proposed by Ho et al. [45]. Breiman [46] further proposed to use the bagging algorithm to generate random forest to avoid over-fitting in the decision trees. The learning targets are numerical variables rather than class labels [46,47]. Our present work applies a random forest to construct a random forest regression (RFR) model.

2.3. Evaluation of the Correction Models

Figure 2 shows a flow chart of the data correction process. Raw PM2.5 readings with a 1 min resolution were averaged to their hourly means to match the hourly observations from the nearest EPA stations within a 3 km radius. If the numbers for the raw PM2.5 in one hour were less than 45, this hourly mean was discarded. After collecting all aforementioned input data, data correction models with three different methods can be established. A model-corrected PM2.5 based on the optimal correction model can be obtained and then compared with the GRIMM-calibrated PM2.5 corrected based on traditional laboratory evaluations. In this way, the performance of the PM2.5 correction model can be evaluated accordingly.
Since the differences between the AS-LUNG-O readings and EPA observations may be affected by the community sources (of which emission activities change over time), the correction model with the best performance will be constructed based on whole-day (24 h) or nighttime (00:00–06:00) periods. The latter period was chosen because most of the community PM2.5 sources associated with human activities were minimal during this period. The optimal correction model built with data from nighttime patterns can be used to obtain the systematic relationships of data from AS-LUNG-O and EPA instruments without interference from nearby sources around the locations of the AS-LUNG-O sets.
The indicators used for evaluating model performance are root mean square error (RMSE), Pearson correlation coefficient (r), and coefficient of determination (R2). R2 is used to assess the predictive or explanatory ability of the model and should be close to 1, while r shows the correlations between two variables. The equation of RMSE is as follows:
RMSE = i = 1 n ( Y i M i ) 2 n
The values of RMSE represent the difference between the model-corrected PM2.5 (Mi) and referenced PM2.5 levels (Yi) (EPA PM2.5 used in the selection evaluation for the machine learning methods; GRIMM-corrected PM2.5 used in the performance evaluation of the selected correction model). Thus, the closer these values are to zero, the better the model performs. Additionally, for the final model, mean absolute errors (MAEs) were also calculated for comparison with those from literature.

3. Results

3.1. Measurements of AS-LUNG-O Sets and EPA Stations

Table 1 shows a summary of the raw PM2.5 of the AS-LUNG-O sets and the PM2.5 observations of the nearest EPA stations during July 2017 and December 2018, as classified by different seasons. The range of PM2.5 for the EPA stations during this period is 2.0–135.0 µg/m3. It can be seen that the highest PM2.5 means and the maximum PM2.5 occurred during winter for both the AS-LUNG-O and EPA PM2.5 levels. The raw PM2.5 values of the AS-LUNG-O sets were, on average, higher than those from the EPA by about 1.9–2.2 fold. These data were used to establish and evaluate the three data correction models.

3.2. Performance Evaluation of the Correction Models

We conducted 10-fold cross-validation and holdout validation tests to evaluate the robustness of the models. The results of the 10-fold cross-validation test for MLR, SVR, and RFR were based on the same training and validation datasets. The average values of RMSE and R2 for the results of the 10-fold cross-validation test were 6.88 ± 0.10 μg/m3 and 0.76, 5.23 ± 0.08 μg/m3 and 0.86, and 4.36 ± 0.06 μg/m3 and 0.91, for MLR, SVR, and RFR, respectively. The results of the holdout evaluation test were presented in Figure 3. The differences between the RMSEs of the 10-fold cross-validation and the holdout validation tests were about averagely 0.03, 0.30, and 0.34 μg/m3 for MLR, SVR, and RFR, respectively. Figure 3a–f show the distribution of model-corrected PM2.5 from the AS-LUNG-O sets and EPA PM2.5 data in the training and validation sets with three different data correction models. Based on the same training set of 63,190 data points, the computation time is 0.01, 8.16, and 0.23 minutes for building models MLR, SVR, and RFR, respectively. The R2 of 0.76–0.99 for these three models shows these models have good explanatory abilities. In terms of RMSE, RFR is the best model (1.73) (Figure 3a,c,e). To further evaluate whether these models perform well for new datasets (Figure 3b,d,f), 63,191 data points from the validation sets were used to input these models. The RMSEs for RFR were 35% and 85% lower than those for SVR in evaluations of the training set and validation set, respectively. The results show that the R2 values of these three models are 0.76–0.89, with the lowest RMSE value (4.7) in RFR for the validation sets. Based on the above evaluation, RFR is chosen as the best data correction model to be used for further applications.

3.3. Sensitivity Analysis of RFR

To optimize the computing efficiency of the RFR model, a sensitivity analysis of RFR was conducted with 50 to 1000 decision trees (with 50-tree increases in each simulation) to assess the changes in modeling efficiency. Figure 4 shows that RFR offers good model performance (RMSE = 1.81 and R2 = 0.9843) when there are only 50 trees. As the decision trees increase, the modeling efficiency is enhanced most significantly before the number of decision trees reaches 300. When the decision trees number is 300, the RMSE is 1.73, and the R2 is 0.9858. Afterward, the efficiency enhancement is not significantly altered by adding more decision trees, which takes more computing time. In the overall evaluations, RFR with 300 decision trees was chosen as the model with the best efficiency.

3.4. Comparison of the Model-Corrected PM2.5 and GRIMM-Calibrated PM2.5

3.4.1. RFR with Whole-Day and Nighttime Patterns

A performance evaluation was further conducted for the RFR with a whole-day pattern using a whole-day dataset based on four seasons. Table 2 shows that the RMSE is the lowest in the summer model (mean: 5.4 μg/m3, ranging from 3.1 to 11.2 μg/m3), followed by the fall model (mean: 6.1 μg/m3, ranging from 3.7 to 10.0 μg/m3), while the RMSE values are slightly higher in the winter and spring models (mean: 6.8 and 7.3 μg/m3, respectively). For certain AS-LUNG-O sets at community locations, the r is as low as 0.33 between the model-corrected PM2.5 and GRIMM-calibrated PM2.5. This discrepancy is possibly caused by some nearby community sources that could not be detected by EPA monitoring stations. These community sources, such as traffic or restaurants, likely generate PM2.5 in the daytime. Therefore, AS-LUNG-O may have different PM2.5 trends from the nearby EPA station, leading to low correlations between these observations.
To focus on the systematic difference of the LCS and EPA observations, a data correction model was established for the nighttime dataset only. Since emissions from community sources resulting from human activity usually reached the lowest levels between 00:00 and 06:00, the data from this period were used to establish the model. Afterward, the established RFR with a nighttime pattern was used to correct the raw PM2.5 for all datasets (including both daytime and nighttime). This way, the readings of AS-LUNG-O were adjusted according to the systematic differences between the AS-LUNG-O and EPA instruments, while the extra PM2.5 increases due to community sources in the daytime could also be retained. Table 2 shows that the r-values of the model-corrected PM2.5 and GRIMM-calibrated PM2.5 were enhanced for the overall datasets, including both street-level and high-level AS-LUNG-O sets (including both daytime and nighttime). In the seasonal models, the r-values for RFR with a whole-day pattern were 0.83, 0.82, 0.85, and 0.90, which were improved to 0.92, 0.88, 0.88, and 0.94 for the RFR with a nighttime pattern for the spring, summer, fall, and winter models, respectively.
The model improvement is most obvious in spring for certain street-level AS-LUNG-O sets. Compared with the whole-day model with an r-value of 0.34, the r-value of the nighttime model is enhanced to 0.81 (Table 3). Nevertheless, the r-values of certain AS-LUNG-O sets were not improved with the nighttime models. It is possible that the trends of PM2.5 concentrations at these AS-LUNG-O locations were different from those in the EPA stations, regardless of whether it was during the day or night.
Furthermore, the results of the model evaluation were categorized as street-level and high-level (Table 3 and Table 4, respectively). For data correction at a high-level, the r-values between the model-corrected PM2.5 and GRIMM-calibrated PM2.5 are all above 0.68 in the whole-day models, while the nighttime models between them are all above 0.75. These results indicate that the PM2.5 levels at high-level AS-LUNG-O locations are moderately correlated with certain deviations from those of the EPA stations in a 3 km radius. This correlation was enhanced in the nighttime model. The minimum r-values of high-level AS-LUNG-O sets were improved from 0.68–0.83 to 0.75–0.86 (Table 4). This phenomenon was also observed in the PM2.5 correction of certain street-level AS-LUNG-O sets; the minimum r-values were improved from 0.33–0.79 to 0.57–0.89 (Table 3). However, the improvement of street-level correlations was not as good as those of high-level AS-LUNG-O sets. The PM2.5 levels sensed by street-level AS-LUNG-O locations are affected by local community sources, resulting in different PM2.5 patterns from those of the EPA stations. It is important to keep these local features while correcting LCS data with the systematic differences between LCS and research-grade instruments in the correction procedures. Thus, based on the evaluations of RFR with whole-day patterns and nighttime patterns, the latter were selected to correct the raw PM2.5 of AS-LUNG-Os.

3.4.2. PM2.5 Corrections by RFR

Figure 5a shows the RMSE values between the GRIMM-calibrated PM2.5 and model-corrected PM2.5 using RFR with a nighttime pattern. The RMSEs were 2.6–10.9 (mean 6.7), 2.8–11.3 (mean 5.7), 3.2–9.9 (mean 5.7), and 2.4–14.4 (mean 6.1) μg/m3 for spring, summer, fall, and winter, respectively (Table 2). Before the model correction, the RMSE values were 15.4–32.7 (mean 23.3), 6.1–19.4 (mean 13.4), 5.3–31.2 (mean 17.4), and 5.2–40.6 (mean 21.6) μg/m3 for spring, summer, fall, and winter, respectively. Clearly, RFR can greatly reduce the RMSE values. The significant differences in RMSEs among different seasons almost disappeared after correction with RFR (the nighttime pattern).
In addition, the MAE is used to evaluate the performance of RFR compared to the models in the literature. Figure 5b shows MAE values between the model-corrected PM2.5 and GRIMM-calibrated PM2.5 with RFR (nighttime pattern). The MAEs were 1.9–9.6 (mean 5.7), 2.4–9.0 (mean 4.9), 2.4–8.3 (mean 4.8), and 1.9–10.3 (mean 4.9) μg/m3 for spring, summer, fall, and winter, respectively. Before the model correction, the MAE values were 14.6–30.1 (mean 21.5), 5.3–16.9 (mean 11.6), 4.4–27.6 (mean 15.2), and 4.1–36.6 (mean 18.8) μg/m3 for spring, summer, fall, and winter, respectively. In summary, after correction, the RMSE was improved from 18.4 ± 6.5 to 5.9 ± 2.0 μg/m3, and the MAE was improved from 16.2 ± 6.0 to 5.0 ± 1.8 μg/m3. This demonstrates that the RFR model greatly reduces MAEs, lowers the overestimation of AS-LUNG-O raw data, and improves the agreements between AS-LUNG-O and EPA PM2.5 levels.
Figure 6 shows a time series plot of the raw PM2.5 for AS-LUNG-O, the model-corrected PM2.5 with RFR (nighttime pattern), and the GRIMM-calibrated PM2.5 for the whole year of 2018. After learning from the PM2.5 observations of nearby EPA stations, the model-corrected PM2.5 levels were close to the GRIMM-calibrated PM2.5. In addition, after being corrected by machine learning techniques, the overestimation of the raw PM2.5 was greatly reduced. However, with nighttime models, the peak values of model-corrected PM2.5 were retained (as shown in the graph) to preserve the contributions of local community sources. After all, the purpose of community air quality monitoring is to assess the contributions of local sources; these important local features need to be preserved during the data correction processes.

4. Discussion

4.1. Comparison of in-Field PM2.5 Correction Models

Typically, environmental research groups apply monitoring instruments calibrated by the manufacturers. For LCS sensors, Rai et al. [48] proposed a two-stage calibration process with laboratory calibration performed by the manufacturers and calibration checks performed by the end-users. This process would be ideal if the manufacturers followed the suggestions. However, demanding manufacturers to calibrate LCSs may be unrealistic, since LCSs are made in larger quantities with much lower costs than more expensive instruments. Therefore, calibration needed to be carried out by the end-users as described in the introduction; most studies use laboratory evaluations before installing LCSs to establish correction equations. Moreover, only a few studies developed in-field data correction models for PM2.5 accuracy correction. Previously, we proposed a hybrid method combining laboratory evaluations and data science to ensure that LCS networks provide accurate PM data [20]. Statistical or machine learning methods were applied to adjust uncalibrated LCS devices with research-grade data within 3 km of the radius from either nearby EPA stations or calibrated seed LCS devices.
The focus of this work was to establish and evaluate three in-field data correction models for a PM2.5 LCS network—namely, the MLR, SVR, and RFR models. The AS-LUNG-O network, with individual correction equations for each LCS in the laboratory, offers a great opportunity to assess the performance of the data correction models established by machine learning techniques. Based on the results of the 10-fold cross-validation and holdout validation tests, there was only a little difference between the RMSEs of the 10-fold cross-validation and holdout validation tests for MLR, SVR, and RFR. Thus, to avoid overvaluing the performance of the models, we used the holdout validation to establish our models. Among them, RFR is the best model, with an RMSE of 1.73 and an R2 of 0.99 based on 63,190 raw hourly data of 39 AS-LUNG-O sets corrected with the PM2.5 levels of EPA stations located within a 3 km radius. In the validation set, RFR was not overfitted, and the data corrected with the RFR model agreed well with the EPA observations. Thus, RFR was chosen as the data correction model. This work demonstrated the applicability of RFR in correcting LCS networks. Previous studies used statistical and machine learning models with data from regulatory stations to correct side-by-side LCS sets on the same locations [15,30]. This work is the first to use data from monitoring stations to correct the data of the LCS network in other locations without monitoring stations. This work is unique in providing LCS data (GRIMM-calibrated PM2.5) based on laboratory evaluations for comparison in other locations.
Among the few in-field correction studies, one study applied the generalized additive model (GAM) to correct LCSs installed at three Taiwan EPA stations during December 2017 [15]. The RMSE values were reduced from 15.55–31.34 μg/m3 to 4.88–9.66 μg/m3 after correction. These results are similar to our data correction model results, showing RMSE values of 5.2–40.6 μg/m3 that were reduced to 4.3–9.6 μg/m3 (high-level PM2.5 corrected by RFR with nigh-time patterns in winter, see Table 4). Our correction model used similar input data. However, while an individual GAM model needed to be established for each LCS at least once every day [15], our model, with cumulative data of more than one year, was constructed as a one-time effort, which saved much computing time. Another study applied transfer learning to correct 10 months of PM2.5 data for LCSs located at seven public environmental monitoring stations in Beijing, China, and obtained MAEs of 7–12 μg/m3 [30]. The RFR model in our work obtained MAEs of 1.9–10.3 μg/m3 and performed at least comparable to or even better than those previous in-field data correction models for PM2.5. Furthermore, our methodology is also suitable to be applied to PM2.5 sensor networks in other countries.
One may concern that the upwind or downwind locations may also affect the relationship of AS-LUNG-O readings and EPA measurements. Since the prevailing wind in Taiwan changes with the season, the seasonal air flow variations were considered in the correction model by the variable of “season”. This variable “season” also considered all other seasonal factors.
In addition, in real-world applications, the computing efficiency of a model is a key issue to determine feasibility. The impressive computer efficiency of RFR is another advantage of the model, as it can handle huge datasets for the in-field correction of sensor networks. Based on the same training dataset, RFR needed only 0.23 minutes for training a model, while SVR needed approximately 8.16 minutes. In comparison, RFR was 35 times faster. With such fast computations, this method has great potential to be expanded to CAQN in Taiwan, which includes more than 4000 uncalibrated LCSs. Decreasing run time in the real-time corrections of CAQN can be carried out by increasing the computational capacity with a larger CPU and RAM. Moreover, parallel processing can be applied in model coding to speed up the correction task. Due to its fast computing efficiency, RFR is an excellent model for big data analytics for any data applications in sensor networks.
It should be noted that the RFR can only be applied for sensor networks with PM2.5 LCSs, which have good R2 with research-grade instruments. If the precision of PM2.5 LCS is not good, the RFR cannot adjust this inherent disadvantage. One previous study only enhanced the r-value between LCSs and research-grade instruments from 0.53 to 0.63 with RFR, since they used LCSs with R2 of 0.07 to 0.27 (compared to research-grade instruments) [49,50]. On the contrary, our present work used LCS devices with good R2 (>0.95) [10,20], so that our RFR model showed good performance in data correction for sensor networks. Thus, LCS with good precision is a prerequisite for a good data correction by RFR.

4.2. Limitations of This Work

There are some limitations to this work. RFR could greatly improve data accuracy, as shown by the reduction of RMSEs and MAEs. However, there is still an average bias of 4.8–5.7 μg/m3 for four seasons after correction. These deviations may come from the inherent differences in ambient (high-level) PM2.5 levels and street PM2.5 levels. The latter is affected by various community sources, while the former is measured by EPA monitoring stations purposely assessing well-mixed ambient PM2.5 levels free from any local emission interference. The AS-LUNG-O sets were located within 3 km radius of EPA stations. Theoretically, the PM2.5 levels were uniformly distributed within these distances. Nevertheless, there were usually multiple localized sources (home factories, restaurants, traffic, etc.) within 100–500 m in Taiwanese communities resulting in significant spatial variations. Therefore, the PM2.5 levels at AS-LUNG-O locations were different from those in EPA stations. This further demonstrates the necessity of establishing LCS networks to assess local ambient air.
On the other hand, the heights of AS-LUNG-O sets and EPA stations might be another variable for the correction model. Since there were only two types of height (ground level and high level) of AS-LUNG-O sets, we did not consider the altitude in the current correction model. This variable could be considered in the future.
Moreover, the performance of RFR may be improved by multiple year inputs, which may cover more environmental conditions for building the decision trees. This could be evaluated further after the accumulation of observations from the AS-LUNG-O network. Additionally, the co-localization of some AS-LUNG-O sets with EPA stations could be conducted in the future for comparison under the same environmental conditions.

5. Conclusions

Current data correction methods for PM2.5 sensor networks are mostly established in the laboratory and in the field before LCS installation. For citizen PM2.5 sensor networks without calibration, this work has developed an in-field data correction model with machine learning to adjust the accuracy deviations of the LCS network to enhance the data applicability of these networks. With the RFR model, the RMSEs and MAEs of the model-corrected PM2.5 and GRIMM-calibrated PM2.5 are greatly reduced. The contributions of local community sources to street-level PM2.5 concentrations are also preserved by RFR with a nighttime pattern (00:00 to 06:00). This work provides a feasible method for the in-field data correction of uncalibrated PM2.5 sensor networks with machine learning techniques. In addition, this work demonstrates the great potential of machine learning to enhance the agreement of LCSs and research-grade instruments, and thus, expands the applications of machine learning in the field of environmental monitoring.
Previously, we proposed a hybrid method combining laboratory evaluations and data science to ensure that LCS networks provide accurate PM data. Statistical or machine learning methods were applied to adjust uncalibrated LCS devices with data from nearby EPA stations or seed LCS devices that have been corrected by laboratory side-by-side comparisons and installed strategically in areas without EPA stations. The current work focuses on applying machine learning to correct the LCS network with PM2.5 from the Taiwan EPA. Under the trend of the smart city movement, increasingly more sensors will be installed in our living surroundings for air quality monitoring. Thus, accurate data are essential to avoid false impressions of better or worse air quality leading to ineffective air pollution control strategies. With the established in-field data correction models presented in this work and the calibrated seed LCS devices, accurate PM2.5 data from the sensor networks can be further applied to citizen science, public education, environmental research, and policymaking. Similar correction models can be established in other countries based on this example to greatly enhance the applicability and usefulness of PM2.5 sensor networks worldwide.

Author Contributions

Conceptualization, W.-C.V.W., S.-C.C.L., and C.-H.L.; methodology, W.-C.V.W., S.-C.C.L., and C.-H.L.; validation, W.-C.V.W. and C.-H.L.; formal analysis, W.-C.V.W. and C.-H.L.; investigation, W.-C.V.W., and C.-H.L.; resources, S.-C.C.L.; data curation, W.-C.V.W. and C.-H.L.; writing—original draft preparation, W.-C.V.W. and S.-C.C.L.; writing—review and editing, W.-C.V.W., S.-C.C.L., and C.-H.L.; visualization, W.-C.V.W.; supervision, S.-C.C.L.; project administration, W.-C.V.W. and C.-H.L.; funding acquisition, S.-C.C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by ACADEMIA SINICA, Taipei, Taiwan under “Integrated Multi-source and High-resolution Heat Wave Vulnerability Assessment of Taiwan (AS-104-SS-A02)” and “Trans-disciplinary PM2.5 Exposure Research in Urban Areas for Health-oriented Preventive Strategies (AS-SS-107-03)”.

Acknowledgments

We would like to acknowledge the team in the Environmental Health Assessment Laboratory at the Research Center for Environmental Changes for their efforts to set up and maintain the AS-LUNG-O network.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Forouzanfar, M.H.; Afshin, A.; Alexander, L.T.; Anderson, H.R.; Bhutta, Z.A.; Biryukov, S.; Brauer, M.; Burnett, R.; Cercy, K.; Charlson, F.J.; et al. Global, regional, and national comparative risk assessment of 79 behavioural, environmental and occupational, and metabolic risks or clusters of risks, 1990–2015: A systematic analysis for the Global Burden of Disease Study 2015. Lancet 2016, 388, 1659–1724. [Google Scholar] [CrossRef] [Green Version]
  2. Lelieveld, J.; Evans, J.S.; Fnais, M.; Giannadaki, D.; Pozzer, A. The contribution of outdoor air pollution sources to premature mortality on a global scale. Nature 2015, 525, 367–371. [Google Scholar] [CrossRef] [PubMed]
  3. Loomis, D.; Grosse, Y.; Lauby-Secretan, B.; El Ghissassi, F.; Bouvard, V.; Benbrahim-Tallaa, L.; Guha, N.; Baan, R.; Mattock, H.; Straif, K.; et al. The carcinogenicity of outdoor air pollution. Lancet Oncol. 2013, 14, 1262–1263. [Google Scholar] [CrossRef]
  4. Betha, R.; Behera, S.N.; Balasubramanian, R. 2013 Southeast Asian smoke haze: Fractionation of particulate-bound elements and associated health risk. Environ. Sci. Technol. 2014, 48, 4327–4335. [Google Scholar] [CrossRef]
  5. Ji, X.; Yao, Y.X.; Long, X.L. What causes PM2.5 pollution? Cross-economy empirical analysis from socioeconomic perspective. Energ. Policy 2018, 119, 458–472. [Google Scholar] [CrossRef]
  6. van Donkelaar, A.; Martin, R.V.; Brauer, M.; Boys, B.L. Use of satellite observations for long-term exposure assessment of global concentrations of fine particulate matter. Environ. Health Persp. 2015, 123, 135–143. [Google Scholar] [CrossRef] [Green Version]
  7. Chen, C.; Zeger, S.; Breysse, P.; Katz, J.; Checkley, W.; Curriero, F.C.; Tielsch, J.M. Estimating indoor PM2.5 and CO concentrations in households in southern Nepal: The Nepal cookstove intervention trials. PLoS ONE 2016, 11. [Google Scholar] [CrossRef]
  8. Lung, S.C.C.; Hsiao, P.K.; Wen, T.Y.; Liu, C.H.; Fu, C.B.; Cheng, Y.T. Variability of intra-urban exposure to particulate matter and CO from Asian-type community pollution sources. Atmos. Environ. 2014, 83, 6–132014. [Google Scholar] [CrossRef] [Green Version]
  9. Lung, S.C.C.; Mao, I.F.; Liu, L.J.S. Residents’ particle exposures in six different communities in Taiwan. Sci. Total Environ. 2007, 377, 81–92. [Google Scholar] [CrossRef]
  10. Lung, S.C.C.; Wang, W.C.V.; Wen, T.Y.J.; Liu, C.H.; Hu, S.C. A versatile low-cost sensing device for assessing PM2.5 spatiotemporal variation and quantifying source contribution. Sci. Total Environ. 2020, 716, 137145. [Google Scholar] [CrossRef]
  11. Taiwan Environmental Protection Agency (Taiwan EPA), Taiwan Air Quality Monitoring Network. Available online: https://airtw.epa.gov.tw/ENG/default.aspx (accessed on 30 June 2020).
  12. USEPA, Air Sensor Guidebook. United States Environmental Protection Agency (USEPA). 2014. Available online: https://www.epa.gov/air-sensor-toolbox/how-use-air-sensors-air-sensor-guidebook (accessed on 20 June 2020).
  13. The Citizen Air Quality Network (CAQN). Available online: https://airbox.edimaxcloud.com (accessed on 20 June 2020).
  14. Chen, L.J.; Ho, Y.H.; Lee, H.C.; Wu, H.C.; Liu, H.M.; Hsieh, H.H.; Huang, Y.T.; Lung, S.C.C. An Open Framework for Participatory PM2.5 Monitoring in Smart Cities. IEEE Access 2017, 5, 14441–14454. [Google Scholar] [CrossRef]
  15. Lee, C.H.; Wang, Y.B.; Yu, H.L. An efficient spatiotemporal data calibration approach for the low-cost PM2.5 sensing network: A case study in Taiwan. Environ. Int. 2019, 130, 104838. [Google Scholar] [CrossRef] [PubMed]
  16. Kuula, J.; Makela, T.; Aurela, M.; Teinila, K.; Varjonen, S.; Gonzalez, O.; Timonen, H. Laboratory evaluation of particle-size selectivity of optical low-cost particulate matter sensors. Atmos. Meas. Tech. 2020, 13, 2413–2423. [Google Scholar] [CrossRef]
  17. Holstius, D.M.; Pillarisetti, A.; Smith, K.R.; Seto, E. Field calibrations of a low-cost aerosol sensor at a regulatory monitoring site in California. Atmos. Meas. Tech. 2014, 7, 1121–1131. [Google Scholar] [CrossRef] [Green Version]
  18. Kelly, K.E.; Whitaker, J.; Petty, A.; Widmer, C.; Dybwad, A.; Sleeth, D.; Martin, R.; Butterfield, A. Ambient and laboratory evaluation of a low-cost particulate matter sensor. Environ. Pollut. 2017, 221, 491–500. [Google Scholar] [CrossRef]
  19. Mukherjee, A.; Stanton, L.G.; Graham, A.R.; Roberts, P.T. Assessing the utility of low-cost particulate matter sensors over a 12-week period in the Cuyama Valley of California. Sensors 2017, 17, 1805. [Google Scholar] [CrossRef] [Green Version]
  20. Wang, W.C.V.; Lung, S.C.C.; Liu, C.H.; Shui, C.K. Laboratory evaluation of correction equations with multiple choices for seed low-cost particle sensing devices in sensor networks. Sensors 2020, 20, 3661. [Google Scholar] [CrossRef]
  21. Dacunto, P.J.; Klepeis, N.E.; Cheng, K.C.; Acevedo-Bolton, V.; Jiang, R.T.; Repace, J.L.; Ott, W.R.; Hildemann, L.M. Determining PM2.5 calibration curves for a low-cost particle monitor: Common indoor residential aerosols. Environ. Sci. Process Impacts 2015, 17, 1959–1966. [Google Scholar] [CrossRef]
  22. Wang, Y.; Li, J.Y.; Jing, H.; Zhang, Q.; Jiang, J.K.; Biswas, P. Laboratory evaluation and calibration of three low-cost particle sensors for particulate matter measurement. Aerosol Sci. Technol. 2015, 49, 1063–1077. [Google Scholar] [CrossRef]
  23. Liu, H.Y.; Schneider, P.; Haugen, R.; Vogt, M. Performance Assessment of a low-cost PM2.5 sensor for a near four-month period in Oslo, Norway. Atmosphere 2019, 10, 41. [Google Scholar] [CrossRef] [Green Version]
  24. Zamora, M.L.; Xiong, F.L.Z.; Gentner, D.; Kerkez, B.; Kohrman-Glaser, J.; Koehler, K. Field and laboratory evaluations of the low-cost Plantower particulate matter sensor. Environ. Sci. Technol. 2019, 53, 838–849. [Google Scholar] [CrossRef] [PubMed]
  25. Zheng, T.S.; Bergin, M.H.; Johnson, K.K.; Tripathi, S.N.; Shirodkar, S.; Landis, M.S.; Sutaria, R.; Carlson, D.E. Field evaluation of low-cost particulate matter sensors in high-and low-concentration environments. Atmos. Meas. Tech. 2018, 11, 4823–4846. [Google Scholar] [CrossRef] [Green Version]
  26. Tryner, J.; L’Orange, C.; Mehaffy, J.; Miller-Lionberg, D.; Hofstetter, J.C.; Wilson, A.; Volckens, J. Laboratory evaluation of low-cost PurpleAir PM monitors and in-field correction using co-located portable filter samplers. Atmos. Environ. 2020, 220, 117067. [Google Scholar] [CrossRef]
  27. Cavaliere, A.; Carotenuto, F.; Di Gennaro, F.; Gioli, B.; Gualtieri, G.; Martelli, F.; Matese, A.; Toscano, P.; Vagnoli, C.; Zaldei, A. Development of low-cost air quality stations for next generation monitoring networks: Calibration and validation of PM2.5 and PM10 sensors. Sensors 2018, 18, 2843. [Google Scholar] [CrossRef] [Green Version]
  28. Jiao, W.; Hagler, G.; Williams, R.; Sharpe, R.; Brown, R.; Garver, D.; Judge, R.; Caudill, M.; Rickard, J.; Davis, M.; et al. Community Air Sensor Network (CAIRSENSE) project: Evaluation of low-cost sensor performance in a suburban environment in the southeastern United States. Atmos. Meas. Tech. 2016, 9, 5281–5292. [Google Scholar] [CrossRef] [Green Version]
  29. Morawska, L.; Thai, P.K.; Liu, X.T.; Asumadu-Sakyi, A.; Ayoko, G.; Bartonova, A.; Bedini, A.; Chai, F.H.; Christensen, B.; Dunbabin, M.; et al. Applications of low-cost sensing technologies for air quality monitoring and exposure assessment: How far have they gone? Environ. Int. 2018, 116, 286–299. [Google Scholar] [CrossRef]
  30. Cheng, Y.; He, X.; Zhou, Z.; Thiele, L. ICT: In-field Calibration Transfer for Air Quality Sensor Deployments. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019, 3, 6. [Google Scholar] [CrossRef]
  31. Pandey, G.; Zhang, B.; Jian, L. Predicting submicron air pollution indicators: A machine learning approach. Environ. Sci. Process. Impacts 2013, 15, 996–1005. [Google Scholar] [CrossRef]
  32. Hsieh, H.P.; Lin, S.D.; Zheng, Y. Inferring air quality for station location recommendation based on urban big data. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 10–13 August 2015; ACM: Sydney, Australia, 2015; pp. 437–446. [Google Scholar]
  33. Zheng, Y.; Yi, X.; Li, M.; Li, R.; Shan, Z.; Chang, E.; Li, T. Forecasting fine-grained air quality based on big data. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 10–13 August 2015; ACM: Sydney, Australia, 2015; pp. 2267–2276. [Google Scholar]
  34. Paas, B.; Stienen, J.; Vorlander, M.; Schneider, C. Modelling of urban near-road atmospheric pm concentrations using an artificial neural network approach with acoustic data input. Environments 2017, 4, 26. [Google Scholar] [CrossRef]
  35. Peng, H.P.; Lima, A.R.; Teakles, A.; Jin, J.; Cannon, A.J.; Hsieh, W.W. Evaluating hourly air quality forecasting in Canada with nonlinear updatable machine learning methods. Air Qual. Atmos. Health 2017, 10, 195–211. [Google Scholar] [CrossRef]
  36. Zimmerman, N.; Presto, A.A.; Kumar, S.P.N.; Gu, J.; Hauryliuk, A.; Robinson, E.S.; Robinson, A.L.; Subramanian, R. A machine learning calibration model using random forests to improve sensor performance for lower-cost air quality monitoring. Atmos. Meas. Tech. 2018, 11, 291–313. [Google Scholar] [CrossRef] [Green Version]
  37. Sayahi, T.; Kaufman, D.; Becnel, T.; Kaur, K.; Butterfield, A.E.; Collingwood, S.; Zhang, Y.; Gaillardon, P.E.; Kelly, K.E. Development of a calibration chamber to evaluate the performance of low-cost particulate matter sensors. Environ. Pollut. 2019, 255, 113131. [Google Scholar] [CrossRef] [PubMed]
  38. Taiwan Central Weather Bureau (Taiwan CWB). Available online: https://www.cwb.gov.tw/eng (accessed on 28 June 2020).
  39. Introduction to Air Quality Monitoring Stations of Taiwan EPA. Available online: https://airtw.epa.gov.tw/ENG/EnvMonitoring/Central/article_station.aspx (accessed on 30 June 2020).
  40. Carrico, C.; Gennings, C.; Wheeler, D.C.; Factor-Litvak, P. Characterization of weighted quantile sum regression for highly correlated data in a risk analysis setting. JABES 2015, 20, 100–120. [Google Scholar] [CrossRef]
  41. Tanner, E.M.; Bornehag, C.G.; Gennings, C. Repeated holdout validation for weighted quantile sum regression. MethodsX 2019, 6, 2855–2860. [Google Scholar] [CrossRef] [PubMed]
  42. Vapnik, V.N.; Lerner, A. Pattern recognition using generalized portrait method. Autom. Remote Control 1963, 24, 774–780. [Google Scholar]
  43. Vapnik, V.N.; Chervonenkis, A. A note on one class of perceptrons. Autom. Remote Control 1964, 25, 821–837. [Google Scholar]
  44. Smola, A.J.; Scholkopf, B. A tutorial on support vector regression. Stat. Comput. 2004, 14, 199–222. [Google Scholar] [CrossRef] [Green Version]
  45. Ho, T.K. The random subspace method for constructing decision forests. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 832–844. [Google Scholar]
  46. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  47. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef] [Green Version]
  48. Rai, A.C.; Kumar, P.; Pilla, F.; Skouloudis, A.N.; Di Sabatino, S.; Ratti, C.; Yasar, A.; Rickerby, D. End-user perspective of low-cost sensors for outdoor air pollution monitoring. Sci. Total Environ. 2017, 607, 691–705. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Borrego, C.; Costa, A.M.; Ginja, J.; Amorim, M.; Coutinho, M.; Karatzas, K.; Sioumis, T.; Katsifarakis, N.; Konstantinidis, K.; De Vito, S.; et al. Assessment of air quality microsensors versus reference methods: The EuNetAir joint exercise. Atmos. Environ. 2016, 147, 246–263. [Google Scholar] [CrossRef] [Green Version]
  50. Borrego, C.; Ginja, J.; Coutinho, M.; Ribeiro, C.; Karatzas, K.; Sioumis, T.; Katsifarakis, N.; Konstantinidis, K.; De Vito, S.; Esposito, E.; et al. Assessment of air quality microsensors versus reference methods: The EuNetAir Joint Exercise—Part II. Atmos. Environ. 2018, 193, 127–142. [Google Scholar] [CrossRef]
Figure 1. Distribution of AS-LUNG-O sets, Environmental Protection Administration (EPA) stations, and sensors of the citizen air quality network (CAQN) in (a) the whole of Taiwan island, (b) the Taipei metropolitan area from July 2017 to December 2018, and (c) an AS-LUNG-O set at street level.
Figure 1. Distribution of AS-LUNG-O sets, Environmental Protection Administration (EPA) stations, and sensors of the citizen air quality network (CAQN) in (a) the whole of Taiwan island, (b) the Taipei metropolitan area from July 2017 to December 2018, and (c) an AS-LUNG-O set at street level.
Sensors 20 05002 g001aSensors 20 05002 g001b
Figure 2. Flow chart of the data correction process.
Figure 2. Flow chart of the data correction process.
Sensors 20 05002 g002
Figure 3. Model performance of (a) multiple linear regression (MLR) with a training set, (b) MLR with a validation set, (c) support vector regression (SVR) with a training set, (d) SVR with a validation set, (e) random forest regression (RFR) with a training set, and (f) RFR with a validation set. The RMSE, R2, and n are listed in the graphs.
Figure 3. Model performance of (a) multiple linear regression (MLR) with a training set, (b) MLR with a validation set, (c) support vector regression (SVR) with a training set, (d) SVR with a validation set, (e) random forest regression (RFR) with a training set, and (f) RFR with a validation set. The RMSE, R2, and n are listed in the graphs.
Sensors 20 05002 g003
Figure 4. Model efficiency of the random forest regression model.
Figure 4. Model efficiency of the random forest regression model.
Sensors 20 05002 g004
Figure 5. Correction results of RFR with a nighttime pattern for different seasons with (a) RMSE and (b) MAE as performance indicators.
Figure 5. Correction results of RFR with a nighttime pattern for different seasons with (a) RMSE and (b) MAE as performance indicators.
Sensors 20 05002 g005
Figure 6. Time-series of raw PM2.5, model-corrected PM2.5, and GRIMM-calibrated PM2.5.
Figure 6. Time-series of raw PM2.5, model-corrected PM2.5, and GRIMM-calibrated PM2.5.
Sensors 20 05002 g006
Table 1. Seasonal mean with the standard deviation (SD), data range, and sample size (n) of the hourly raw PM2.5 of AS-LUNG-O sets and EPA PM2.5 (µg/m3).
Table 1. Seasonal mean with the standard deviation (SD), data range, and sample size (n) of the hourly raw PM2.5 of AS-LUNG-O sets and EPA PM2.5 (µg/m3).
Raw PM2.5 of AS-LUNG EPA PM2.5n
Mean (SD)Range (Min, Max)Mean (SD)Range (Min, Max)
Spring48.0 (20.6)(3.1, 295.9)24.3 (13.6)(2.0, 100.0)19,924
Summer28.4 (21.3)(1.0, 249.8)12.9 (13.0)(2.0, 75.0)37,638
Fall36.8 (15.3)(1.0, 223.6)19.3 (8.1)(2.0, 127.0)43,624
Winter51.5 (29.2)(1.0, 309.8)27.0 (18.2)(2.0, 135.0)25,195
Table 2. Performance evaluation of the random forest regression model (RFR) with whole-day and nighttime patterns in overall AS-LUNG-O sets (street-level and high-level).
Table 2. Performance evaluation of the random forest regression model (RFR) with whole-day and nighttime patterns in overall AS-LUNG-O sets (street-level and high-level).
Overall SeasonRMSE 1Pearson Correlationn
Mean (SD 2)Range (Min, Max)Mean (SD)Range (Min, Max)
RFR with whole-day patternsSpring7.3 (2.6)(4.1, 14.1)0.83 (0.15)(0.34, 0.96)19,924
Summer5.4 (1.7)(3.1, 11.2)0.82 (0.11)(0.33, 0.93)37,638
Fall6.1 (1.6)(3.7, 10.0)0.85 (0.08)(0.53, 0.94)43,624
Winter6.8 (2.3)(3.5, 12.9)0.90 (0.04)(0.79, 0.97)25,195
RFR with nighttime patternsSpring6.7 (2.4)(2.6, 10.9]0.92 (0.05)(0.80, 0.98)19,924
Summer5.7 (1.7)(2.8, 11.3)0.88 (0.07)(0.57, 0.95)37,638
Fall5.7 (1.6)(3.2, 9.9)0.88 (0.08)(0.68, 0.96)43,624
Winter6.1 (2.3)(2.4, 14.4)0.94 (0.03)(0.86, 0.98)25,195
1 RMSE, root mean square error; 2 SD, standard deviation.
Table 3. Performance evaluation of the random forest regression model (RFR) with whole-day and nighttime patterns in street-level AS-LUNG-O sets.
Table 3. Performance evaluation of the random forest regression model (RFR) with whole-day and nighttime patterns in street-level AS-LUNG-O sets.
Street-LevelSeasonRMSE 1Pearson Correlationn
Mean (SD 2)Range (Min, Max)Mean (SD)Range (Min, Max)
RFR with whole-day patternsSpring7.1 (2.7)(4.1, 14.1)0.84 (0.15)(0.34, 0.96)17,255
Summer5.4 (1.8)(3.1, 11.2)0.83 (0.11)(0.33, 0.93)30,710
Fall5.8 (1.5)(3.7, 10.0)0.85 (0.09)(0.53, 0.94)32,606
Winter6.5 (2.2)(3.5, 12.9)0.91 (0.04)(0.79, 0.97)19,448
RFR with nighttime patternsSpring6.5 (2.5)(2.6, 10.9)0.93 (0.04)(0.81, 0.98)17,255
Summer5.6 (1.8)(2.8, 11.3)0.89 (0.07)(0.57, 0.95)30,710
Fall5.7 (1.7)(3.2, 9.9)0.89 (0.08)(0.68, 0.96)32,606
Winter5.9 (2.4)(2.4, 14.4)0.94 (0.02)(0.89, 0.98)19,448
1 RMSE, root mean square error; 2 SD, standard deviation.
Table 4. Performance evaluation of the random forest regression model (RFR) with whole-day and nighttime patterns in high-level AS-LUNG-O sets.
Table 4. Performance evaluation of the random forest regression model (RFR) with whole-day and nighttime patterns in high-level AS-LUNG-O sets.
High-LevelSeasonRMSE 1Pearson Correlationn
Mean (SD 2)Range (Min, Max)Mean (SD)Range (Min, Max)
RFR with whole-day patternsSpring8.8 (2.3)(7.2, 10.4)0.78 (0.15)(0.68, 0.89)2669
Summer5.7 (1.3)(3.6, 7.6)0.78 (0.07)(0.70, 0.88)6928
Fall7.3 (2.0)(4.6, 9.9)0.84 (0.07)(0.75, 0.91)11,018
Winter8.0 (2.9)(4.6, 12.9)0.88 (0.04)(0.83, 0.94)5747
RFR with nighttime patternsSpring8.2 (1.9)(6.8, 9.5)0.87 (0.09)(0.80, 0.94)2669
Summer5.8 (1.3)(4.1, 7.9)0.85 (0.06)(0.76, 0.94)6928
Fall6.2 (1.7)(4.9, 9.3)0.88 (0.09)(0.75, 0.95)11,018
Winter6.7 (2.2)(4.3, 9.6)0.92 (0.03)(0.86, 0.94)5747
1 RMSE, root mean square error; 2 SD, standard deviation.

Share and Cite

MDPI and ACS Style

Wang, W.-C.V.; Lung, S.-C.C.; Liu, C.-H. Application of Machine Learning for the in-Field Correction of a PM2.5 Low-Cost Sensor Network. Sensors 2020, 20, 5002. https://doi.org/10.3390/s20175002

AMA Style

Wang W-CV, Lung S-CC, Liu C-H. Application of Machine Learning for the in-Field Correction of a PM2.5 Low-Cost Sensor Network. Sensors. 2020; 20(17):5002. https://doi.org/10.3390/s20175002

Chicago/Turabian Style

Wang, Wen-Cheng Vincent, Shih-Chun Candice Lung, and Chun-Hu Liu. 2020. "Application of Machine Learning for the in-Field Correction of a PM2.5 Low-Cost Sensor Network" Sensors 20, no. 17: 5002. https://doi.org/10.3390/s20175002

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop