Next Article in Journal
A High-Gain Circularly Polarized Magnetoelectric Dipole Antenna Array for Millimeter-Wave Applications
Previous Article in Journal
Fusion Ranging Method of Monocular Camera and Millimeter-Wave Radar Based on Improved Extended Kalman Filtering
Previous Article in Special Issue
Algorithms and Resources for the Monitoring of Very-Low-Frequency Signal Deviations Due to Solar Activity Using a Web-Based Software-Defined Radio-Distributed Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluating PurpleAir Sensors: Do They Accurately Reflect Ambient Air Temperature?

Department of Landscape Architecture and Environmental Planning, University of California, Berkeley, CA 94720, USA
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 3044; https://doi.org/10.3390/s25103044
Submission received: 22 March 2025 / Revised: 22 April 2025 / Accepted: 3 May 2025 / Published: 12 May 2025
(This article belongs to the Special Issue Sensor Network Applications for Environmental Monitoring)

Abstract

:

Highlights

What are the main findings?
  • PurpleAir sensors exhibit strong temperature overestimations with an MAE of 4.71 °C and RMSE of 6.30 °C.
  • Sensor performance demonstrates nonlinear behavior with significant seasonal and diurnal variations.
What is the implication of the main finding?
  • Calibrated PurpleAir sensors have the potential to advance hyperlocal heat mapping and multi-hazard vulnerability assessments.

Abstract

Low-cost sensors (LCSs) emerge as a popular tool for urban micro-climate studies by offering dense observational coverage. This study evaluates the performance of PurpleAir (PA) sensors for ambient temperature monitoring—a key but underexplored aspect of their use. While widely used for particulate matter, PA sensors’ temperature data remain underutilized and lack thorough validation. For the first time, this research evaluates their accuracy by comparing PA temperature measurements with collocated high-precision temperature data loggers across a dense urban network in a humid subtropical U.S. county. Results show a moderate correlation with reference data (r = 0.86) but an average overestimation of 3.77 °C, indicating PA sensors are better suited for identifying temperature trends but not for precise applications like extreme heat events. We also developed and compared eight calibration methods to create a replicable model using readily available crowdsourced data. The best-performing model reduced RMSE and MAE by 51% and 47%, respectively, and achieved an R2 of 0.89 compared to the uncalibrated scenario. Finally, the practical application of PA temperature data for identifying heat wave events was investigated, including an assessment of associated uncertainties. In sum, this work provides a crucial evaluation of PA’s temperature monitoring capabilities, offering a pathway for improved heat mapping, multi-hazard vulnerability assessments, and public health interventions in the development of climate-resilient cities.

1. Introduction

The growing demand for high spatial-temporal urban environmental data, driven by climate-resilient city initiatives, has rapidly advanced the development of low-cost sensors (LCSs). Traditionally, localized environmental data were acquired from regulatory stations with spatial-temporal limitations in capturing complex urban phenomena [1]. LCS offers a promising alternative by filling gaps in regulatory monitoring networks, bringing transformational benefits [2,3], such as motivating scientific communities to better understand hyperlocal problems [4,5] and empowering public environmental awareness [6].
Among the various applications of LCS, air quality monitoring constitutes a significant portion of large-scale projects such as AirCasting, Citi-Sense, AirVisual, and City Health Outlook [7,8]. PurpleAir (PA), a leading air quality LCS vendor, stands out with over 30,000 citizen scientists contributing to its real-time map (as of July 2024). With over 18 times more devices than government-operated weather stations in the U.S. [9], PA enables unprecedented block-level environment characterization [10,11,12]. Unlike remote sensing technologies that provide proxy data [13,14], PA provides high temporal resolution ground-level in situ measurements reflecting direct human environmental exposure.
While most studies focus on PA sensors for particulate matter measurements, these devices simultaneously collect valuable meteorological data—including temperature, relative humidity (RH), and pressure—which have primarily been used for detecting sensor malfunctions and supporting field calibration [15]. As demand grows for integrated heat health information systems, large-scale sensor networks for measuring ambient air temperature have become critical for improving urban heat island (UHI) monitoring and modeling to support the development of climate-resilient communities [16]. Ongoing efforts to enhance UHI mapping and develop a comprehensive national heat–health information system position PA sensors as a potentially pivotal platform for raising public awareness and strategizing interventions. Furthermore, the concurrent collection of air quality and temperature data enables researchers to explore the multi-disaster vulnerability in urban systems, such as the intricate relationship between UHI and urban pollution islands and their mechanisms [17,18]. As the majority of these studies relied on remote sensing products or in situ observations from sparely distributed stations [19], PA sensors present the opportunity to overcome this limitation and reveal hyperlocal patterns for these urban phenomena.
Despite their significant implications in urban climatology and public health, data quality remains an indispensable concern in LCS applications [20,21,22]. Many studies use field calibration techniques by collocating LCS with regulatory instruments to adjust biases and improve accuracy [23,24,25]. While there is a growing interest in PA’s air quality monitoring performance, thorough assessments of PA’s temperature data remain limited, with only a few intercomparison studies [26], which differ from field evaluations. Couzo et al. evaluated PA temperature measurements using a collocated research-grade instrument in Asheville, North Carolina, and developed a simple linear regression correction model [27]. However, their study did not evaluate the spatial and temporal variability in sensor performance. This information is critical for holistically understanding influential factors of measurement biases to improve field calibration methods. Furthermore, PA recommends a constant correction of −4.4 °C, hereinafter PA-suggested calibration, to align with ambient conditions [28]. To the best of our knowledge, this black box factory adjustment has not been empirically tested in the ambient environment. The performance of electrical devices can be sensitive to various mechanical and ambient factors [29], necessitating a comprehensive understanding of PA’s performances and the underlying influencing factors. As emerging UHI studies increasingly rely on PA temperature data to supplement satellite or regulatory station observations [30], addressing these uncertainties has become critical.
This study offers a first-hand evaluation of PA’s temperature monitoring capabilities and develops a calibration approach through three key objectives: (1) to assess PA’s performance in estimating ambient temperature using collocated high-accuracy temperature data loggers across a dense sensor network; (2) to evaluate PA temperature data uncertainties through the application of heat wave detection; and (3) to develop and compare calibration methods to create a replicable model based on widespread crowdsourced data.

2. Materials and Methods

2.1. Collocated Temperature Sensor Network and Data Preprocessing

The PA sensors and reference data were collected from an LCS network of 51 sites in Denton County, Texas, where each site was equipped with collocated PA-II-SD sensors and Onset HOBO MX2301A temperature/relative humidity (RH) outdoor data loggers. PA-II-SD sensors use Plantower PMS5003 laser-based particulate sensors to measure PM1, PM2.5, and PM10 concentrations, along with a BOSCH BME280 sensor for pressure, temperature, and humidity. Data are transmitted via a 2.4 GHz 802.11 b/g/n wireless network, with an SD card available for local storage. The Onset HOBO MX2301A is a rugged, battery-powered temperature and RH data logger designed for outdoor environmental monitoring. The temperature sensor has a reported accuracy of ±0.2 °C and ±2% and a drift of less than 0.01 °C and 1% per year [31], making it a reliable reference for this study. MX2301A uses Bluetooth low energy for data retrieval via mobile devices. We configured the HOBO data loggers to record temperature at 30 min intervals, balancing data resolution needs and battery life, from March 2022 to August 2023. This period includes the record-hot summer of 2023, offering key insights into PA sensor performance during prolonged extreme heat. All collocated sensors were installed at 1.5 to 2 m above ground level to ensure comparable measurables that are representative of the local scale [32]. PA sensor mounting locations were carefully selected to prevent overheating of optical components, ensure adequate airflow, and maintain stable access to power and WiFi. In residential areas, sensors were typically mounted on structures like yard posts that are fully shaded yet positioned slightly away from walls to minimize interference from wall material reflectance. HOBO data loggers, housed in solar radiation shields, are more flexible in placement as they can be exposed to direct or reflect sunlight. They were collocated as closely as possible to the PA sensors, with the only constraint being a minimum distance of 2 m from nearby tree canopies to reduce the influence of evapotranspiration (Figure S1).
Denton County’s landscape comprises predominantly grass (50.6%), tree canopy (20.4%), and water bodies (8.2%), with an elevation ranging from approximately 130 to 300 m. The region’s humid subtropical climate is characterized by humid, hot summers and cool winters, with year-round precipitation. To ensure optimal spatial representation, we employed a stratified sampling approach and used 300 × 300 m grid cells as the basic spatial unit, dividing the area into six urban strata based on proximity to major traffic and the percentage of impervious surfaces (Table S1) [5]. The 51 sites were distributed across 4 Rural High, 12 Rural Low, 1 Suburban High, 16 Suburban Low, 6 Urban High, and 12 Urban Low sites, ensuring comprehensive spatial coverage (Figure 1).

2.2. Sensor Data Preprocessing

The raw data from PA and HOBO sensors were first averaged to hourly measurements, represented by TPA and THOBO, to capture the diurnal temperature cycles. Datasets were then matched by sensor locations and timestamps for pairwise comparisons, yielding 282,355 pairwise measurements. After that, z-scores were calculated for every PA and HOBO temperature observation in the joint dataset, and a z-score threshold of two was applied to all measurements across all time to exclude outliers in HOBO and PA measurements. This ensured valid measurements from both instrument types during the study period to avoid biased comparison due to sensor malfunctions or measurement errors from either collocated sensor. Finally, two sites with limited sample sizes were removed to ensure statistical significance, yielding a total number of pairwise observations of 261,310 measured from 47 sites.

2.3. Spatial-Temporal Variations in Sensor Performance

We performed spatial-temporal analysis on sensor performance to better understand its behavior across diverse urban environments and seasonal conditions. Spatially, measurements were grouped based on the six strata to identify patterns of systematic differences, accounting for varying impacts from anthropogenic heat output, urban heat intensity, and modification of airflow. We calculated the inter- and intra-group differences using temperature anomalies (dT), which are expressed as:
d T i = T P A , i T H O B O , i
where T P A , i and T H O B O , i are pairwise measurements from PA and HOBO sensors for observation i, with i ranging from 1 to n = 261,310.
We tested the significance of inter- and intra-strata differences using the Kruskal–Wallis H-test, chosen for its suitability with non-normal distributions and variance dissimilarities identified via the Kolmogorov–Smirnov test and Levene’s test [33]. The null hypothesis is that there is no in-class variability across the strata due to similar thermal climates. The Suburban High group was omitted from the intra-group test due to insufficient sensor samples. The inter-group evaluation was followed up by a nonparametric post hoc pairwise test (Dunn’s test) to identify specific disparities.
Temporally, all sensor data were grouped by month of the year and hour of the day to examine seasonal and diurnal variations, accounting for potential nonlinear behavior at extreme temperatures caused by ambient heat and sensor self-heating during operation [34].

2.4. Performance Metrics

We selected the Pearson correlation coefficient (r), coefficient of determination (R2), mean bias error (MBE), mean absolute error (MAE), and root mean square error (RMSE) as performance metrics. MBE measures the average differences between predicted and observed values, suggesting overestimation or underestimation. MAE indicates the total error values, whereas RMSE is more sensitive to larger deviations.
M B E = 1 n i = 1 n d T i
M A E = 1 n i = 1 n d T i
R M S E = 1 n i = 1 n ( d T i ) 2

2.5. Independent Variables

We developed field calibration models for PA temperature sensors based on the protocol used for PM sensors [26,35]. Field calibration involves comparing collocated data collected from ambient environment and using data-driven approaches, such as statistical or machine learning models, to minimize discrepancies in sensor readings. While some studies employ complex machine learning models [36,37], we chose linear regression and multiple linear regression (MLR) for their model transparency, interpretability, and replicability, and they have proven reliable in studies requiring minimal parameter tuning and model selection [29].
We selected a set of meteorological factors as independent variables to account for the influence of ambient environmental conditions (Table 1). PA relative humidity (RHPA) has been proven as a useful factor in particulate matter measurements calibration to adjust for the hygroscopicity of particles [38,39]. Wind speed (WNDS) can induce biases in PA’s particle readings either high or low by a factor of 1.6 [40], but its influence on heat exchange between the environment and LCS has not been thoroughly explored yet. Hourly wind speed observations were acquired at the Denton Airport South ambient air monitoring station administered by the Texas Commission on Environmental Quality. Furthermore, radiative fluxes estimate incoming and outgoing solar radiation, serving as proxies for the time of day and sky conditions. Hence, both downwelling longwave (LW) and shortwave (SW) irradiance were included to account for the effects of cloudy days on temperature variability [41]. Hourly LW and SW were acquired from the GOES-EAST Surface Solar Irradiance product provided by the EUMETSAT Ocean and Sea Ice Satellite Application Facility [42]. We extracted LW and SW values at each site based on its coordinates.

2.6. Developing Calibration Models

We first conducted overall and monthly bivariate analyses between hourly dT and each meteorological factor in Table 1 for variable selection. The correlation coefficient (r) was used to assess the directions and strength of the relationship and the p value determined their significance. Only variables with a significance level of p < 0.001 were included as inputs for the calibration model development.
Eight models were developed using different combinations of additive terms and interactive terms to account for interdependence between variables [35]. These models were evaluated using the performance metrics described in Section 2.3 and the “leave one out” approach by splitting 20% of the data as a test set. The best calibration model was determined by considering the R2, RMSE, MAE, and Akaike information criterion (AIC) score to balance the model performance and complexity. An AIC score measures the goodness of fit while penalizing model complexity to reduce the risk of overfitting and underfitting, making it an ideal metric for model selection.
  • Model 1: Simple linear regression
    T H O B O = s 1 T P A + ε
  • Model 2: MLR with an additive RHPA term
    T H O B O = s 1 T P A + s 2 R H P A + ε
  • Model 3: MLR with an additive SW term
    T H O B O = s 1 T P A + s 2 S W + ε
  • Model 4: MLR with an additive LW term
    T H O B O = s 1 T P A + s 2 L W + ε
  • Model 5: MLR with additive SW and LW terms
    T H O B O = s 1 T P A + s 2 L W + s 3 S W + ε
  • Model 6: MLR with additive RHPA, LW, and SW terms
    T H O B O = s 1 T P A + s 2 R H + s 3 L W + s 4 S W + ε
  • Model 7: MLR with additive RHPA, LW, SW, and WNDS terms
    T H O B O = s 1 T P A + s 2 R H + s 3 L W + s 4 S W + s 5 W N D S + ε
  • Model 8: MLR with additive and multiplicative TPA and RHPA terms
    T H O B O = s 1 T P A + s 2 R H P A + s 3 T P A R H P A + ε
Since dT is likely to exhibit varying temporal patterns due to the oscillating influence of meteorological factors, we assess how much of its variance can be explained by the meteorological factors, after accounting for their deviations from observed monthly and diurnal patterns. We define the monthly diurnal deviations of dT as the anomaly dT (dT′), calculated as:
d T m h ¯ = 1 N m h k = 1 N m h d T m n ( K )
d T = d T d T m h ¯
where Nmh is the number of observations in month m at hour h and d T m n ( K ) is the value of dT at the kth observations in month m at hour h.
Similarly, we define THOBO, RHHOBO, SW′, LW′, and WNDS′ as the anomalies of their respective variables, which are used as the independent variables in the MLR to model dT′. This anomaly model helps us determine the effectiveness of anomaly variables in calibrating TPA. We then created Model 9 by incorporating the anomaly corresponding to the variables in the best-performing model from Model 1 to Model 8. This adjustment accounts for unexplained errors when relying solely on the hourly observed values of the meteorological variables in Table 1. Finally, we compared the performance of Model 9 and the best model from Model 1 to Model 8 to determine the final calibration model. A variance inflation factor (VIF) was calculated for each independent variable in the final model to detect the severity of multicollinearity.

2.7. Performance in Apparent Temperature Calculation

Other than air temperature, we evaluated PA’s performance in measuring apparent temperature, commonly represented by the Heat Index (HI). HI is a better indicator of human heat exposure, as it reflects the temperature perceived by humans by combining air temperature, R H , and sometimes wind speed [43]. Given the significant implications of HI in public health and urban heat research [44,45], we anticipate that HI will be a key metric in studies involving PA sensors in public health or related fields. Here, we present the first assessment of PA sensors’ performance in estimating HI. We used an extended Heat Index that extends the calculation of the Heat Index to a wider range of temperature and R H [46], accommodating the high heat and humidity over Texas in the summer of 2023. HOBO- and PA-derived HIs were calculated using their corresponding meteorological data at the hourly interval. Subsequently, we compared their proportions in each HI class as defined by the National Weather Service to exemplify potential flaws in a potential application of PA temperature measurements.

3. Results and Discussion

3.1. Evaluation of Uncalibrated PA Temperature Measurements

The overall MBE is 3.77 °C during our study period, meaning that PA measurements on average overestimate ambient temperature by 3.77 °C (Figure 2). The comparison of hourly mean time series shows that TPA closely follows THOBO but remains consistently higher, indicating a consistent overestimation. A LOWESS (locally weighted smoothing) curve has been added as a reference using air temperature measurements reported from Denton Airport South monitoring station, which has an average air temperature of 21.08 °C. Both TPA and THOBO demonstrate similar seasonal variations with the smoothed reference temperature trend.
Figure 3 plots dTi as a function of the percentile of TPA,i for all n observations. Over 79% of hourly dT values fall between −2 °C and 10 °C. The anomalies are particularly pronounced at the extremes, as expected, ranging from −20 °C at the lower percentiles to over +30 °C at the higher percentiles. While positive dT values are present in every percentile, there are no negative dT values at the highest percentiles. This suggests that underestimation is very rare when PA sensors report high-temperature observations (>40 °C), but underestimation and overestimation may exist when temperature is not extremely high.
TPA exhibits a moderately strong correlation with THOBO (r = 0.86), but with an MAE and RMSE of 4.71 °C and 6.30 °C, respectively. The distribution of dT in most months is right-skewed and unimodal, except for July, which exhibits a bimodal pattern (Figure S2). January, February, August, September, and October have peaks near 0 °C, while the remaining months peak between 2 and 6 °C. This result contrasts with a previous study in Asheville, North Carolina, which reported a very strong agreement of PA (r = 0.99) with the reference instruments and a low overall RMSE of 2.8 °C [27].
The performance of PA in measuring air temperature varies significantly by month, with r values ranging from 0.28 to 0.80 (Figure 4 left). Late summer months (July to September) exhibit a greater magnitude of overestimations when compared with winter months (December to February). May shows the highest correlation (r = 0.80), followed by June (0.79) and April (0.77). Conversely, September has the lowest correlation (r = 0.28), followed by July (r = 0.34). The decline in sensor performance coincides with the seasonal temperature cycle (Figure 4 right), peaking from July to September, suggesting that overheating is a primary contributor to bias in sensor readings. October has the greatest RMSE (8.07 °C) and MAE (5.69 °C), while December has the lowest RMSE (5.27 °C) and MAE (3.98 °C). The highest MBE (4.26 °C) is observed in June and September, while the lowest (2.90 °C) is observed in December. The MBE in each month stays positive, suggesting that PA tends to overestimate temperature across all seasons. The significant differences between MAE and MBE in late winter and late summer indicate the strong presence of underestimation, leading to a smaller MBE. Both RMSE and MAE demonstrate a significant increase starting from April, followed by a sharp decline between October and December, while MBE begins to increase in January and starts to decline after September. This pattern aligns with the seasonal temperature cycle, especially between spring and summer when temperature begins to increase.
Conversely, the diurnal performance metrics exhibit a strong antiphase relationship with diurnal temperature variations, with peak MAE (9.80 °C) and RMSE (11.39 °C) occurring at 6:00 AM local time, and the lowest MAE (2.47 °C) and RMSE (3.17 °C) at 2:00 PM local time (Figure 5 left). The measurement agreement does not vary as much as the monthly agreement, with r values ranging from 0.76 to 0.96 only. The MBEs are positive, except from 3:00 to 5:00 PM local time, with negative values ranging from −0.10 °C to −0.30 °C. A possible explanation for the antiphase relationship is internal heating from the WiFi module. During nighttime and early morning, when ambient temperatures are low, internal heating amplifies baseline overestimation, leading to larger temperature differences. In the afternoon, when ambient temperatures are high and PA sensors tend to underestimate the temperature, internal heating offsets this bias, resulting in higher accuracy. Regardless, PA still underestimates temperatures during peak temperature hours, leading to negative MBEs. Our findings highlight the nonlinear behavior of sensor performance and its diurnal and seasonal patterns, underscoring the limitation of applying a constant correction factor.
Spatially, no systematic patterns of dT are identified across the six strata groups (Figure 6). All groups show similar intra-group variability except for Urban High, which has a smaller interquartile range and the highest median dT (3.40 °C). On the other hand, Rural Low has the lowest median dT (2.69 °C), followed by Urban Low (2.98 °C). The results of the Kruskal–Wallis H-test for inter- and intra-strata comparisons are statistically significant (p < 0.001), indicating significant differences between and within each strata group. The post hoc test shows significant differences for all strata pairs except for Rural Low and Suburban High groups. These findings suggest that sensor performance is more likely to be influenced by site-specific factors.

3.2. Factors Influencing Termpretuare Anomaly

All five meteorological variables are significantly related (p < 0.001) to temperature anomaly but do not exhibit strong overall correlations (|r| < 0.35) (Figure 7). Among them, RHHOBO has the strongest overall correlation (r = 0.33), while LW has the weakest (r = −0.08). At the monthly scale, all variables maintain a consistent relationship with dT across months. Except for WNDS, all meteorological variables display a downward-sloping best-fit line, indicating that PA sensors tend to overestimate temperature at lower values, with this effect diminishes as the values increase. The degree of overestimation varies by month for each variable. THOBO has the strongest relationship in September (r = −0.55), while the flattest slope is observed in December (r = −0.25) (Figure S3). Other summer and early fall months (July, August, and October) show a similar trend, with r values ranging from −0.49 to −0.51.
In contrast, RHHOBO is positively correlated with dT across all months and follows a similar seasonal pattern (Figure S4). It has a relatively strong correlation from July to October, with r values ranging from 0.45 to 0.57, and the weakest correlation is found in April and June (r = 0.22). The similarity of the seasonal response in temperature and RH corresponds to the humid subtropical climate in Denton. In terms of solar radiation, PA sensors exhibit a high level of discrepancies when values of SW and LW are low (Figures S5 and S6). Similarly to the impact of THOBO, they also manifest seasonal variation, with summer months showing the strongest negative correlation. From July to October, LW has r values ranging from −0.31 to −0.40, while SW has r values ranging from −0.21 to −0.33. These results are counterintuitive since we expect PA sensors to be more susceptible to radiative heating, which would lead to positive relationships with temperature anomalies. Lastly, the impact of WNDS also varies significantly in different months but does not exhibit a profound seasonal pattern, with r values ranging from −0.05 to −0.30 (Figure S7). The evident improvement in sensor performance under high wind speed conditions implies its mitigating effect on internal heating.
For the anomaly model, although the model is statistically significant (p < 0.001), it explains only 6.4% of the variance in dT′ with an R2 of 0.064. All five independent variables are significantly associated with dT′. WNDS′, LW′, and THOBO show negative correlations, while RHHOBO and SW′ exhibit positive correlations. Their low predictive power suggests the external factors contribute to the variability of dT′, beyond the unexplained patterns of meteorological factors.

3.3. Comparison of Calibration Models for PA Temperature Sensors

The performance of the eight models generally improves incrementally as the model complexity increases (Table 2). All models have a very small MBE, meaning that the models do not systematically overestimate or underestimate the true values. Model 1 uses only TPA and has the worst model performance. Among the models with two additive terms (Models 2–4), Model 4 achieves the best performance with the highest R2 value (0.82) and lowest error (RMSE = 3.93 °C and MAE = 3.16 °C), outperforming models that use RHPA or SW. Model 7, with the most additive terms, achieves the best overall performance, followed closely by Models 5 and 6. The minimal performance improvement between Models 6 and 7, despite their added complexity with RHPA and WNDS terms, suggests that these predictors have limited predictive power when LW and SW additive terms are present. Model 8, which incorporates a multiplicative term to account for the collinearity between TPA and RHPA, only offers a slight improvement over Model 2, with an RMSE of 4.56 °C and MAE of 3.45 °C. Since Model 5 has very similar performance metric values as Model 6 and Model 7 but with fewer additive terms, we compared this model with Model 9, which adds its corresponding anomaly terms (LW′ and SW′) to maintain the model simplicity. Model 9 shows a higher agreement (R2 = 0.89) and lower error (RMSE = 3.10 °C and MAE = 2.46 °C) compared to Model 5. All variables in Model 9 have a VIF score between 1 to 5, meaning no significant multicollinearity is found. Therefore, Model 9 is selected as the final calibration model, which generates the following equation:
T H O B O = 0.2839 T P A + 0.117 L W + 0.0048 S W 0.0403 L W + 0.0094 S W 29.41
By comparing the calibrated TPA with the uncalibrated scenario, our model shows a reduction in RMSE and MAE by at least 51% and 47%, respectively (Figure 8). In contrast, the PA-suggested calibration only reduces RMSE and MAE by 19% and 18%. The drastic differences in their model performance highlight the strength of our MLR model in adjusting raw TPA measurements to more accurately represent true temperature conditions.
The percentage change in sensor performance following the MLR-based calibration exhibits spatial variability, with a few notable outliers (Figure 9). Overall, all sites show improvement across all performance metrics, with MAE demonstrating the most substantial enhancement, highlighting the effectiveness of the MLR-based calibration at the site level. The percentage change in MBE ranges from −36.1% to −134.5%, while RMSE and MAE show similar but much narrower ranges. Two locations in the southern part of the study area exhibit relatively small reductions in MBE (less than 60%) compared to other sites. Additionally, one urban site within the City of Denton shows only modest improvements, with a 14.5% reduction in RMSE and a 17.5% reduction in MAE. No obvious spatial patterns are observed in the distribution of the performance improvement. These outliers suggest the influence of site-specific factors contributing to sensor biases that are not captured by the current calibration model.

3.4. Evalutation in the Context of the Heat Index

The hourly performance metrics of TPA-derived HI has a peak MAE and RMSE value recorded at 6:00 AM local time, and the lowest at 1:00 PM local time (Figure S8), demonstrating similar diurnal patterns with the performance of TPA. However, there are strong differences between their monthly performances, particularly in sensor agreement in August, in which TPA-derived HI (r = 0.29) demonstrates a much weaker agreement than TPA (r = 0.42) (Figure 10). Similar to the monthly performance pattern of TPA, there is a steady increase in the magnitude of overestimation from May to September, with the highest RMSE (8.52 °C), MAE (6.03 °C), and MBE (4.32 °C) in September. However, the sharp improvement in sensor performance begins in September instead of October, with the lowest RMSE (4.94 °C) and MAE (3.71 °C), and MBE (3.71 °C) observed in December. HI shows a greater magnitude of overestimation with the largest RMSE (0.89 °C) and MAE difference (0.46 °C) with TPA performance metrics in September. The results exemplify the possibility of having a compounding effect where the mixture of RHPA and TPA measurements amplifies the magnitude of overestimation.
Table 3 summarizes the proportion of hourly temperature measurements in each heat class after converting hourly THOBO and TPA to hourly HI. Similar to ambient temperature, PA tends to overestimate apparent temperature, and it is most significant in the Extreme Caution and Danger class, in which they are 6.3% and 5.7% higher than THOBO-derived HI respectively. However, the discrepancy is relatively small in the Caution and Extreme Danger class, illustrating its vulnerability in certain heat classes. The differences highlight the shortcomings of a potential application of PA meteorological data as it may raise a false alarm for heat events due to inherent temperature overestimation.

4. Limitations

Generality is a key criteria for a calibration model, as it ensures the model can operate across different geographical locations and seasons [47]. To maintain generality and simplicity, we considered only a linear relationship. While most regression model results in our study area show a high degree of linearity, this may not apply to other regions. For example, some studies discovered a nonlinear effect of RHPA [48] on PA particulate matter measurements and applied a nonlinear empirical correction equation [49]. Therefore, it is necessary to test various regression functions to account for nonlinearity. Additionally, this study used temperature data from only one county. As a result, the model’s transferability may not be applicable to other climate zones with significantly different local climates.
While the sensor distribution covered the study area well and a stratified sampling approach was used to reduce bias, some strata were oversampled and others were under-sampled. For instance, Suburban High only had one site (3.6% of the dataset), but Suburban Low had 16 sites (36.9% of the dataset). This imbalance was also constrained by the recruitment timeline. Furthermore, this stratified sampling approach only considered two urban factors to characterize the built environment. However, the urban thermal environments are heterogeneous in both horizontal and vertical dimensions, influenced by complex anthropogenic and natural factors [50]. Consequently, other classification systems, such as the Local Climate Zone [51], may be more effective in representing different local climates, capturing systematic patterns, and optimizing model development. Moreover, the spatial analysis of PA temperature performance was limited to the strata grouping approach, as site-specific analysis was not feasible due to data incompleteness and significant imbalances in sample sizes across sites. A more robust and complete dataset is needed to enable reliable spatial assessments at the local level.
In terms of independent variable selection, this study focused on PA sensitivity to meteorological factors and did not consider other mechanical factors, such as sensor age. Since the life expectancy of a PA sensor is about two years, aging sensors can degrade and cause significant inaccuracies. To address this, some studies have incorporated the sensor’s total operating time and the consecutive operation time in their calibration models to account for sensor aging and operational stability [37]. Including these factors could improve the calibration model’s performance, but since not all sensors have this information available, we opted not to include them to reduce complexity.

5. Conclusions

This study provides a firsthand and comprehensive assessment of PA temperature measurements and proposes a linear calibration model. Uncalibrated PA temperature measurements had a moderate agreement with reference data (r = 0.86) but overestimated temperature by 3.77 °C on average. The results reveal seasonal and diurnal variations in PA sensor performance, influenced by local conditions. The selected calibration model outperformed the PA-suggested calibration method, achieving MAE = 2.45 °C and RMSE = 3.08 °C, providing a simple yet effective adjustment for PA data. While all eight models performed well, the improvements in accuracy were relatively small compared to the additional effort required to collect more variables—particularly radiation variables. For general use, it is important to balance the benefits of slightly improved accuracy with the practicality of data collection. The Heat Index application highlights PA’s potential in heat monitoring, though data accuracy is crucial to avoid overestimating heat events. As rising temperatures threaten our public health and urban livability, reliable in situ temperature data from LCSs, like PA, become valuable. This work underscores the importance of leveraging PA’s network for heat monitoring and offers guidance for integrating such data into research and urban decision-making processes.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s25103044/s1, Table S1: Classification rules of each sampling strata; Figure S1: Collocation of a PurpleAir sensor and a HOBO data logger. Figure S2: Monthly histograms of temperature anomaly; Figure S3: Monthly scatterplots of hourly temperature anomaly and hourly HOBO temperature; Figure S4: Monthly scatterplots of hourly temperature anomaly and hourly HOBO relative humidity measurements; Figure S5: Monthly scatterplots of hourly temperature anomaly and hourly downwelling longwave irradiance observations; Figure S6: Hourly temperature anomaly and hourly downwelling shortwave irradiance observations; Figure S7: Monthly scatterplots of hourly temperature anomaly and hourly wind speed measurements; Figure S8: Monthly TPA-derived Heat Index performance and THOBO-derived Heat Index temperature cycle.

Author Contributions

Conceptualization, L.L.; methodology, J.T. and L.L.; validation, J.T.; formal analysis, J.T.; investigation, J.T. and L.L.; resources, L.L.; data curation, J.T.; writing—original draft preparation, J.T.; writing—review and editing, L.L.; visualization, J.T.; supervision, L.L.; project administration, L.L.; funding acquisition, L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by the National Science Foundation (BCS-2117505) and a grant from The Katherine and James Lau Foundation for Climate Equity and Environmental Justice in the Built Environment.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The field calibration code and data for PurpleAir temperature measurements are available at https://github.com/lu-liang-geo/PurpleAir-Temp-Calibration/tree/main (accessed on 5 May 2025).

Acknowledgments

The authors would like to thank student volunteers at the University of North Texas for setting up the calibration system at the Texas site, deploying sensors, and collecting data. We thank David M. Romps Department of Earth and Planetary Science, at the University of California, Berkeley for providing surface irradiance data and his feedback on manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LCSLow-cost sensors
PAPurpleAir
MAEMean absolute error
RMSERoot mean square error
AICAkaike information criterion
RHRelative humidity
WNDSWind speed
LWLongwave downwelling surface irradiance
SWShortwave downwelling surface irradiance
UHIUrban heat island
HIHeat Index
MLRMultiple linear regression
VIFVariance inflation factor

References

  1. Muller, C.L.; Chapman, L.; Johnston, S.; Kidd, C.; Illingworth, S.; Foody, G.; Overeem, A.; Leigh, R.R. Crowdsourcing for Climate and Atmospheric Sciences: Current Status and Future Potential. Int. J. Climatol. 2015, 35, 3185–3203. [Google Scholar] [CrossRef]
  2. Snyder, E.G.; Watkins, T.H.; Solomon, P.A.; Thoma, E.D.; Williams, R.W.; Hagler, G.S.W.; Shelow, D.; Hindin, D.A.; Kilaru, V.J.; Preuss, P.W. The Changing Paradigm of Air Pollution Monitoring. Environ. Sci. Technol. 2013, 47, 11369–11377. [Google Scholar] [CrossRef] [PubMed]
  3. Chapman, L.; Bell, S.; Randall, S. Can Crowdsourcing Increase the Durability of an Urban Meteorological Network? Urban Clim. 2023, 49, 101542. [Google Scholar] [CrossRef]
  4. Chapman, L.; Bell, C.; Bell, S. Can the Crowdsourcing Data Paradigm Take Atmospheric Science to a New Level? A Case Study of the Urban Heat Island of London Quantified Using Netatmo Weather Stations. Int. J. Climatol. 2017, 37, 3597–3605. [Google Scholar] [CrossRef]
  5. Liang, L.; Daniels, J.; Bailey, C.; Hu, L.; Phillips, R.; South, J. Integrating Low-Cost Sensor Monitoring, Satellite Mapping, and Geospatial Artificial Intelligence for Intra-Urban Air Pollution Predictions. Environ. Pollut. 2023, 331, 121832. [Google Scholar] [CrossRef]
  6. Kumar, P.; Morawska, L.; Martani, C.; Biskos, G.; Neophytou, M.; Di Sabatino, S.; Bell, M.; Norford, L.; Britter, R. The Rise of Low-Cost Sensing for Managing Air Pollution in Cities. Environ. Int. 2015, 75, 199–205. [Google Scholar] [CrossRef]
  7. Morawska, L.; Thai, P.K.; Liu, X.; Asumadu-Sakyi, A.; Ayoko, G.; Bartonova, A.; Bedini, A.; Chai, F.; Christensen, B.; Dunbabin, M.; et al. Applications of Low-Cost Sensing Technologies for Air Quality Monitoring and Exposure Assessment: How Far Have They Gone? Environ. Int. 2018, 116, 286–299. [Google Scholar] [CrossRef]
  8. Liang, L.; Gong, P.; Cong, N.; Li, Z.; Zhao, Y.; Chen, Y. Assessment of Personal Exposure to Particulate Air Pollution: The First Result of City Health Outlook (CHO) Project. BMC Public Health 2019, 19, 711. [Google Scholar] [CrossRef]
  9. Dodds, P.S.; Mitchell, L.; Reagan, A.J.; Danforth, C.M. Tracking Climate Change through the Spatiotemporal Dynamics of the Teletherms, the Statistically Hottest and Coldest Days of the Year. PLoS ONE 2016, 11, e0154184. [Google Scholar] [CrossRef]
  10. Lu, Y.; Giuliano, G.; Habre, R. Estimating Hourly PM2.5 Concentrations at the Neighborhood Scale Using a Low-Cost Air Sensor Network: A Los Angeles Case Study. Environ. Res. 2021, 195, 110653. [Google Scholar] [CrossRef]
  11. Yadav, N.; Sharma, C. Spatial Variations of Intra-City Urban Heat Island in Megacity Delhi. Sustain. Cities Soc. 2018, 37, 298–306. [Google Scholar] [CrossRef]
  12. Amegah, A.K.; Dakuu, G.; Mudu, P.; Jaakkola, J.J.K. Particulate Matter Pollution at Traffic Hotspots of Accra, Ghana: Levels, Exposure Experiences of Street Traders, and Associated Respiratory and Cardiovascular Symptoms. J. Expo. Sci. Environ. Epidemiol. 2022, 32, 333–342. [Google Scholar] [CrossRef] [PubMed]
  13. Zhou, D.; Xiao, J.; Bonafoni, S.; Berger, C.; Deilami, K.; Zhou, Y.; Frolking, S.; Yao, R.; Qiao, Z.; Sobrino, J.A. Satellite Remote Sensing of Surface Urban Heat Islands: Progress, Challenges, and Perspectives. Remote Sens. 2018, 11, 48. [Google Scholar] [CrossRef]
  14. Paciorek, C.J.; Liu, Y. Limitations of Remotely Sensed Aerosol as a Spatial Proxy for Fine Particulate Matter. Environ. Health Perspect. 2009, 117, 904–909. [Google Scholar] [CrossRef]
  15. Liang, L. Calibrating Low-Cost Sensors for Ambient Air Monitoring: Techniques, Trends, and Challenges. Environ. Res. 2021, 197, 111163. [Google Scholar] [CrossRef] [PubMed]
  16. Zabow, M.; Jones, H.; Trtanj, J.; McMahon, K.; Schramm, P. The National Integrated Heat Health Information System: A Federal Approach to Addressing Extreme Heat and Health. In Proceedings of the 103rd AMS Annual Meeting, Denver, CO, USA, 8–12 January 2023. [Google Scholar]
  17. Li, H.; Sodoudi, S.; Liu, J.; Tao, W. Temporal Variation of Urban Aerosol Pollution Island and Its Relationship with Urban Heat Island. Atmos. Res. 2020, 241, 104957. [Google Scholar] [CrossRef]
  18. Li, H.; Meier, F.; Lee, X.; Chakraborty, T.; Liu, J.; Schaap, M.; Sodoudi, S. Interaction between Urban Heat Island and Urban Pollution Island during Summer in Berlin. Sci. Total Environ. 2018, 636, 818–828. [Google Scholar] [CrossRef]
  19. Ulpiani, G. On the Linkage between Urban Heat Island and Urban Pollution Island: Three-Decade Literature Review towards a Conceptual Framework. Sci. Total Environ. 2021, 751, 141727. [Google Scholar] [CrossRef]
  20. Johnson, K.; Holder, A.; Frederick, S.; Hagler, G.; Clements, A. PurpleAir PM2.5 Performance across the U.S.#2, 2020. In Proceedings of the Meeting between ORD, OAR/AirNow, and USFS, Research Triangle Park, NC, USA, 3 February 2020. [Google Scholar]
  21. Feenstra, B.; Papapostolou, V.; Hasheminassab, S.; Zhang, H.; Boghossian, B.D.; Cocker, D.; Polidori, A. Performance Evaluation of Twelve Low-Cost PM2.5 Sensors at an Ambient Air Monitoring Site. Atmos. Environ. 2019, 216, 116946. [Google Scholar] [CrossRef]
  22. Castell, N.; Dauge, F.R.; Schneider, P.; Vogt, M.; Lerner, U.; Fishbain, B.; Broday, D.; Bartonova, A. Can Commercial Low-Cost Sensor Platforms Contribute to Air Quality Monitoring and Exposure Estimates? Environ. Int. 2017, 99, 293–302. [Google Scholar] [CrossRef]
  23. Ardon-Dryer, K.; Dryer, Y.; Williams, J.N.; Moghimi, N. Measurements of PM2.5 with PurpleAir under Atmospheric Conditions. Atmos. Meas. Tech. 2020, 13, 5441–5458. [Google Scholar] [CrossRef]
  24. Campmier, M.J.; Gingrich, J.; Singh, S.; Baig, N.; Gani, S.; Upadhya, A.; Agrawal, P.; Kushwaha, M.; Mishra, H.R.; Pillarisetti, A.; et al. Seasonally Optimized Calibrations Improve Low-Cost Sensor Performance: Long-Term Field Evaluation of PurpleAir Sensors in Urban and Rural India. Atmos. Meas. Tech. 2023, 16, 4357–4374. [Google Scholar] [CrossRef]
  25. Barkjohn, K.K.; Holder, A.L.; Frederick, S.G.; Clements, A.L. Correction and Accuracy of PurpleAir PM2.5 Measurements for Extreme Wildfire Smoke. Sensors 2022, 22, 9669. [Google Scholar] [CrossRef]
  26. Stavroulas, I.; Grivas, G.; Michalopoulos, P.; Liakakou, E.; Bougiatioti, A.; Kalkavouras, P.; Fameli, K.; Hatzianastassiou, N.; Mihalopoulos, N.; Gerasopoulos, E. Field Evaluation of Low-Cost PM Sensors (Purple Air PA-II) Under Variable Urban Air Quality Conditions, in Greece. Atmosphere 2020, 11, 926. [Google Scholar] [CrossRef]
  27. Couzo, E.; Valencia, A.; Gittis, P. Evaluation and Correction of PurpleAir Temperature and Relative Humidity Measurements. Atmosphere 2024, 15, 415. [Google Scholar] [CrossRef]
  28. PurpleAir. PurpleAir Sensors Functional Overview. Available online: https://community.purpleair.com/t/purpleair-sensors-functional-overview/150 (accessed on 15 December 2024).
  29. Liang, L.; Daniels, J. What Influences Low-Cost Sensor Data Calibration?—A Systematic Assessment of Algorithms, Duration, and Predictor Selection. Aerosol Air Qual. Res. 2022, 22, 220076. [Google Scholar] [CrossRef]
  30. McBroom, B.D.; Rahn, D.A.; Brunsell, N.A. Urban Fraction Influence on Local Nocturnal Cooling Rates from Low-Cost Sensors in Dallas-Fort Worth. Urban Clim. 2024, 53, 101823. [Google Scholar] [CrossRef]
  31. Onset Computer Corporation. HOBO MX1101 Data Logger. Available online: https://www.onsetcomp.com/products/data-loggers/mx1101 (accessed on 15 December 2024).
  32. Oke, T. Initial Guidance to Obtain Representative Meteorological Observations at Urban Sites; World Meteorological Organization: Geneva, Switzerland, 2006. [Google Scholar]
  33. Nayak, B.; Hazra, A. How to Choose the Right Statistical Test? Indian J. Ophthalmol. 2011, 59, 85. [Google Scholar] [CrossRef] [PubMed]
  34. Giordano, M.R.; Malings, C.; Pandis, S.N.; Presto, A.A.; McNeill, V.F.; Westervelt, D.M.; Beekmann, M.; Subramanian, R. From Low-Cost Sensors to High-Quality Data: A Summary of Challenges and Best Practices for Effectively Calibrating Low-Cost Particulate Matter Mass Sensors. J. Aerosol Sci. 2021, 158, 105833. [Google Scholar] [CrossRef]
  35. Barkjohn, K.K.; Gantt, B.; Clements, A.L. Development and Application of a United States-Wide Correction for PM2.5 Data Collected with the PurpleAir Sensor. Atmos. Meas. Tech. 2021, 14, 4617–4637. [Google Scholar] [CrossRef]
  36. Raheja, G.; Nimo, J.; Appoh, E.K.-E.; Essien, B.; Sunu, M.; Nyante, J.; Amegah, M.; Quansah, R.; Arku, R.E.; Penn, S.L.; et al. Low-Cost Sensor Performance Intercomparison, Correction Factor Development, and 2+ Years of Ambient PM2.5 Monitoring in Accra, Ghana. Environ. Sci. Technol. 2023, 57, 10708–10720. [Google Scholar] [CrossRef] [PubMed]
  37. Bi, J.; Wildani, A.; Chang, H.H.; Liu, Y. Incorporating Low-Cost Sensor Measurements into High-Resolution PM2.5 Modeling at a Large Spatial Scale. Environ. Sci. Technol. 2020, 54, 2152–2162. [Google Scholar] [CrossRef]
  38. Malings, C.; Tanzer, R.; Hauryliuk, A.; Saha, P.K.; Robinson, A.L.; Presto, A.A.; Subramanian, R. Fine Particle Mass Monitoring with Low-Cost Sensors: Corrections and Long-Term Performance Evaluation. Aerosol. Sci. Technol. 2020, 54, 160–174. [Google Scholar] [CrossRef]
  39. Patel, M.Y.; Vannucci, P.F.; Kim, J.; Berelson, W.M.; Cohen, R.C. Towards a Hygroscopic Growth Calibration for Low-Cost PM2.5 Sensors. Atmos. Meas. Tech. 2024, 17, 1051–1060. [Google Scholar] [CrossRef]
  40. Malm, W.C.; Schichtel, B.A.; Prenni, A.J.; Day, D.; Perkins, R.J.; Sullivan, A.; Tigges, M. The Dependence of PurpleAir Sensors on Particle Size Distribution and Meteorology and Their Relationship to Integrating Nephelometers. Atmos. Environ. 2024, 337, 120757. [Google Scholar] [CrossRef]
  41. Wild, M. The Global Energy Balance as Represented in CMIP6 Climate Models. Clim. Dyn. 2020, 55, 553–577. [Google Scholar] [CrossRef]
  42. GOES-E SSI. OSI SAF GOES-East Surface Solar Irradiance 2021-Onwards, OSI-306-b. EUMETSAT Ocean and Sea Ice Satellite Application Facility. Available online: https://osi-saf.eumetsat.int/products/osi-306-b (accessed on 5 May 2025).
  43. Ho, H.C.; Knudby, A.; Xu, Y.; Hodul, M.; Aminipouri, M. A Comparison of Urban Heat Islands Mapped Using Skin Temperature, Air Temperature, and Apparent Temperature (Humidex), for the Greater Vancouver Area. Sci. Total Environ. 2016, 544, 929–938. [Google Scholar] [CrossRef]
  44. Romps, D.M. Heat Index Extremes Increasing Several Times Faster than the Air Temperature. Environ. Res. Lett. 2024, 19, 041002. [Google Scholar] [CrossRef]
  45. Méndez-Lázaro, P.; Muller-Karger, F.E.; Otis, D.; McCarthy, M.J.; Rodríguez, E. A Heat Vulnerability Index to Improve Urban Public Health Management in San Juan, Puerto Rico. Int. J. Biometeorol. 2018, 62, 709–722. [Google Scholar] [CrossRef]
  46. Lu, Y.-C.; Romps, D.M. Extending the Heat Index. J. Appl. Meteorol. Climatol. 2022, 61, 1367–1383. [Google Scholar] [CrossRef]
  47. Concas, F.; Mineraud, J.; Lagerspetz, E.; Varjonen, S.; Liu, X.; Puolamäki, K.; Nurmi, P.; Tarkoma, S. Low-Cost Outdoor Air Quality Monitoring and Sensor Calibration: A Survey and Critical Analysis. ACM Trans. Sens. Netw. 2021, 17, 1–44. [Google Scholar] [CrossRef]
  48. Wallace, L.; Bi, J.; Ott, W.R.; Sarnat, J.; Liu, Y. Calibration of Low-Cost PurpleAir Outdoor Monitors Using an Improved Method of Calculating PM. Atmos. Environ. 2021, 256, 118432. [Google Scholar] [CrossRef]
  49. Zheng, T.; Bergin, M.H.; Johnson, K.K.; Tripathi, S.N.; Shirodkar, S.; Landis, M.S.; Sutaria, R.; Carlson, D.E. Field Evaluation of Low-Cost Particulate Matter Sensors in High- and Low-Concentration Environments. Atmos. Meas. Tech. 2018, 11, 4823–4846. [Google Scholar] [CrossRef]
  50. Oke, T.R. The Energetic Basis of the Urban Heat Island. Q. J. R. Meteorol. Soc. 1982, 108, 1–24. [Google Scholar] [CrossRef]
  51. Stewart, I.D.; Oke, T.R. Local Climate Zones for Urban Temperature Studies. Bull. Am. Meteorol. Soc. 2012, 93, 1879–1900. [Google Scholar] [CrossRef]
Figure 1. The distribution of sensors. Each site has a pair of a collocated PurpleAir sensor and HOBO data logger.
Figure 1. The distribution of sensors. Each site has a pair of a collocated PurpleAir sensor and HOBO data logger.
Sensors 25 03044 g001
Figure 2. Time series of hourly mean TPA and THOBO averaged over all stations are shown as solid curves. Dashed lines display means of the TPA and THOBO time series. DTO represents a LOWESS curve of hourly air temperature measurements from the Denton Airport South ambient monitoring station.
Figure 2. Time series of hourly mean TPA and THOBO averaged over all stations are shown as solid curves. Dashed lines display means of the TPA and THOBO time series. DTO represents a LOWESS curve of hourly air temperature measurements from the Denton Airport South ambient monitoring station.
Sensors 25 03044 g002
Figure 3. dT as a function of PA temperature percentile. The black dashed line represents the zero-error scenario.
Figure 3. dT as a function of PA temperature percentile. The black dashed line represents the zero-error scenario.
Sensors 25 03044 g003
Figure 4. Performance metrics of hourly PA measurements for each month (left). Monthly mean (dot) and interquartile range (vertical line) of HOBO temperature for each month (right).
Figure 4. Performance metrics of hourly PA measurements for each month (left). Monthly mean (dot) and interquartile range (vertical line) of HOBO temperature for each month (right).
Sensors 25 03044 g004
Figure 5. Changes in performance metrics of hourly PA measurements across different hours of the day (left). The black dashed line is added as a reference for overestimation and underestimation. Hourly mean (dot) and interquartile range (vertical line) of HOBO temperature for each hour (right).
Figure 5. Changes in performance metrics of hourly PA measurements across different hours of the day (left). The black dashed line is added as a reference for overestimation and underestimation. Hourly mean (dot) and interquartile range (vertical line) of HOBO temperature for each hour (right).
Sensors 25 03044 g005
Figure 6. Boxplots of hourly temperature anomaly in each strata group. The median value of each group is annotated at the top.
Figure 6. Boxplots of hourly temperature anomaly in each strata group. The median value of each group is annotated at the top.
Sensors 25 03044 g006
Figure 7. Bivariate correlation between paired hourly observations of temperature anomaly dT′ and (a) HOBO temperature, (b) HOBO relative humidity, (c) downwelling surface longwave radiation, (d) downwelling surface shortwave radiation, and (e) wind speed. The black dashed line represents the zero-error scenario, and the orange solid line is the best-fit line. *** indicates a p-value of <0.001.
Figure 7. Bivariate correlation between paired hourly observations of temperature anomaly dT′ and (a) HOBO temperature, (b) HOBO relative humidity, (c) downwelling surface longwave radiation, (d) downwelling surface shortwave radiation, and (e) wind speed. The black dashed line represents the zero-error scenario, and the orange solid line is the best-fit line. *** indicates a p-value of <0.001.
Sensors 25 03044 g007
Figure 8. Comparison of agreement in temperature measurements with no calibration (a), PA-suggested calibration (b), and our MLR-based calibration (c). The black dashed line represents the perfect fit line, and the orange solid line is the best-fit line.
Figure 8. Comparison of agreement in temperature measurements with no calibration (a), PA-suggested calibration (b), and our MLR-based calibration (c). The black dashed line represents the perfect fit line, and the orange solid line is the best-fit line.
Sensors 25 03044 g008
Figure 9. Comparison of the spatial variations of percentage change in RMSE (a), MAE (b), and MBE (c) after applying the MLR-based calibration. The grey lines are the census unit.
Figure 9. Comparison of the spatial variations of percentage change in RMSE (a), MAE (b), and MBE (c) after applying the MLR-based calibration. The grey lines are the census unit.
Sensors 25 03044 g009
Figure 10. Monthly TPA-derived HI performance and THOBO-derived HI cycle. Performance metrics of hourly TPA-derived HI measurements for each month (left). Hourly mean (dot) and interquartile range (vertical line) of THOBO-derived HI for each month (right).
Figure 10. Monthly TPA-derived HI performance and THOBO-derived HI cycle. Performance metrics of hourly TPA-derived HI measurements for each month (left). Hourly mean (dot) and interquartile range (vertical line) of THOBO-derived HI for each month (right).
Sensors 25 03044 g010
Table 1. Descriptions of meteorological variables used in calibration model development.
Table 1. Descriptions of meteorological variables used in calibration model development.
Variable (Acronym)SourceUnitSpatial
Resolution
Temporal
Resolution
Wind speed (WNDS)Texas Commission on
Environmental Quality
m/s/Hourly
Longwave downwelling
surface irradiance (LW)
GOES-East Surface Solar IrradianceW/m20.05°
Shortwave downwelling
surface irradiance (SW)
GOES-East Surface Solar IrradianceW/m20.05°
Relative Humidity (RHPA)PurpleAir Sensor%/
Air Temperature (TPA)PurpleAir Sensor°C/
Table 2. Comparison of calibration model performance.
Table 2. Comparison of calibration model performance.
ModelNumber of Training SampleR2RMSE (°C)MAE (°C)MBE (°C)
1209,0480.734.733.540.01
2209,0480.754.593.490.01
3209,0480.774.413.380.01
4208,9880.823.933.16−0.02
5208,9880.863.382.730
6208,9880.863.382.730
7207,9990.873.372.720.01
8209,0480.754.563.450.01
9208,9880.893.102.46−0.01
Table 3. Comparison of Heat Index classification derived from PA and HOBO data.
Table 3. Comparison of Heat Index classification derived from PA and HOBO data.
ClassRange of HI (°F)HOBO HI (% of Time)PA HI (% of Time)
Caution80–9017.618.8
Extreme Caution90–10314.020.3
Danger103–1243.99.6
Extreme Danger 12500.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tse, J.; Liang, L. Evaluating PurpleAir Sensors: Do They Accurately Reflect Ambient Air Temperature? Sensors 2025, 25, 3044. https://doi.org/10.3390/s25103044

AMA Style

Tse J, Liang L. Evaluating PurpleAir Sensors: Do They Accurately Reflect Ambient Air Temperature? Sensors. 2025; 25(10):3044. https://doi.org/10.3390/s25103044

Chicago/Turabian Style

Tse, Justin, and Lu Liang. 2025. "Evaluating PurpleAir Sensors: Do They Accurately Reflect Ambient Air Temperature?" Sensors 25, no. 10: 3044. https://doi.org/10.3390/s25103044

APA Style

Tse, J., & Liang, L. (2025). Evaluating PurpleAir Sensors: Do They Accurately Reflect Ambient Air Temperature? Sensors, 25(10), 3044. https://doi.org/10.3390/s25103044

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop