Next Article in Journal
A Comparative Study of X Data About the NHS Using Sentiment Analysis
Previous Article in Journal
A Critical Analysis of Government Communication via X (Twitter)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interpretable Emotion Estimation in Indoor Remote Work Environments via Environmental Sensor Data

1
Graduate School of Science and Engineering, Chiba University, Chiba 263-8522, Japan
2
Digital Transformation Enhancement Council, Chiba University, Chiba 263-8522, Japan
*
Author to whom correspondence should be addressed.
Big Data Cogn. Comput. 2025, 9(10), 243; https://doi.org/10.3390/bdcc9100243
Submission received: 21 August 2025 / Revised: 13 September 2025 / Accepted: 22 September 2025 / Published: 23 September 2025

Abstract

Indoor environmental factors such as CO2 concentration, temperature, and humidity can significantly influence individuals’ emotional states and productivity. This study continuously collected environmental data using wireless sensors and emotional data from wearable devices in an office-like remote-work setting. Machine learning models, including Random Forest and Gradient Boosting Decision Tree, were developed and interpreted using SHAP (Shapley Additive Explanations). The proposed models achieved estimation accuracies above 85%. SHAP analysis revealed that CO2 concentration, temperature, and humidity were influential factors in predicting pleasant or unpleasant states. These findings demonstrate the feasibility of real-time, data-driven emotion estimation and provide insights into the design of indoor environments that foster comfort and mental well-being.

1. Introduction

The diversification of modern work styles, particularly the rise of remote work, has highlighted the critical role of indoor environments in affecting a worker’s mental state. Unfavorable environmental conditions can negatively impact productivity and lead to an increase in human error. Numerous studies have shown that indoor factors such as temperature, humidity, lighting, ventilation, and acoustics significantly influence mental health and productivity [1]. For instance, noise and lighting have been reported to affect comfort and work performance among office workers [2], with a series of studies confirming the indoor environment’s impact on well-being [3,4,5,6,7,8,9,10,11]. However, most of this research has relied on questionnaires or controlled laboratory tests, which struggle to capture the complex, dynamic relationship between environmental factors and an individual’s mental state in real-world settings over time.
In the fields of psychology and cognitive science, methods for understanding mental states can be broadly categorized into two groups: those based on subjective self-reports (e.g., questionnaires) and those based on objective physiological data (e.g., heart rate, brain waves, or perspiration). While questionnaires can effectively capture subjective psychological aspects, they often lack objectivity and quantitativeness. In contrast, methods using physiological data provide an objective and quantitative basis for analyzing the relationships between physiological indicators and mental states [12,13,14,15,16].
Previous research has explored the impact of indoor factors, such as elevated CO2 levels, on human performance [17,18,19] and physiological functions [20] through controlled experiments. However, such studies often rely on indirect methods like self-reported symptoms of sick building syndrome [21] or standard sleepiness scales [22], which may not accurately reflect unconscious mental states. Furthermore, studies conducted in highly controlled climate chambers [23,24,25] have provided foundational insights but faced challenges in capturing the complex interplay of multiple environmental factors that occur in realistic, dynamic settings.
To address these limitations, our study introduces a comprehensive system that extends our previous work [26,27,28]. The system is designed for multimodal, real-time monitoring of both environmental factors and an individual’s emotional state in a realistic indoor environment. We utilize wearable devices to acquire biometric data for emotion estimation and a wireless sensor network to collect environmental data. This approach allows us to overcome the limitations of traditional methods by providing an objective, continuous, and integrated dataset.
A key contribution of this work is the application of interpretable machine learning—specifically, SHAP (SHapley Additive exPlanations)—to analyze the developed machine learning models. Unlike previous black-box models that could only infer mental states from environmental data with high accuracy [26,27,28], SHAP enables us to quantify and interpret the specific influence of each environmental factor on an individual’s emotional state. This allows us to move beyond simple estimation and provide direct, actionable insights into which factors affect pleasantness or unpleasantness.
In summary, this study addresses the gap in existing research by offering a system for the continuous collection of multimodal data in a real-world setting. The primary contributions of this work include the following:
  • A novel system for continuous, multimodal data collection on both indoor environmental factors and human emotional states;
  • The application of an interpretable machine learning approach (SHAP) to identify and quantify the specific environmental factors that influence a person’s sense of pleasantness or unpleasantness;
  • A comprehensive analysis of the dynamic relationship between indoor factors, such as CO2 concentration, temperature, and humidity, and emotional well-being.

2. Materials and Methods

An overview of this study is illustrated in Figure 1, which is structured into three distinct parts:
  • Data Acquisition: Various indoor environmental data related to human cognitive functions and mental states (pleasantness/unpleasantness) are acquired by edge computersvia wireless sensor networks. The data are then segmented into training and test datasets for the machine learning model.
  • Model Construction and Evaluation: Utilizing the training data, a machine learning model is developed, where the environmental data serve as the input and the mental state (pleasantness/unpleasantness) serves as the output. Post construction, the accuracy of the model is assessed using the test data.
  • SHAP Analysis: SHAP is employed to construct a regression model from the training model. This model aims to determine which indoor environmental values impact emotions (pleasantness/unpleasantness) in the original machine learning model, as represented by SHAP values, thereby adding interpretability to the model.
The novelty of this study lies in its unified approach, which continuously collects environmental and emotional data from multiple sensors. By applying SHAP to analyze the machine learning model, this research objectively and interpretably reveals the relationships between dynamic environmental factors and individual emotional states in an indoor work environment. This approach contributes to the quantitative clarification of key influencing factors, a feat that goes beyond the limitations of conventional subjective questionnaires or one-off experiments, which often fail to capture the continuous, nuanced, and multifaceted nature of these relationships.

2.1. Environmental Sensor

The multiple types of environmental data (units) relevant to human cognitive functions used in constructing the estimation model are listed in Table 1, along with the number of sensors deployed.
The following measures were implemented to ensure reliable sensor data acquisition. A wireless sensor network was configured in a star topology using Xbee, enabling real-time data collection at 10 s intervals. In parallel, the NEC Emotion Analysis Solutions system [29] was used to ensure consistency in measurements of mental states.

2.2. Emotional Recognition

To measure participants’ emotions, the “Emotion Analysis Solutions” systemof NEC [29] was utilized. This system interprets pulse and barometric data from wearable devices to visualize mental states [30,31], allowing for the determination of subjects’ arousal levels and valence values from heart-rate variation data. The study integrates “Silmee TMW22”, a wristband-type wearable device, with a cloud-based analysis platform featuring emotion recognition technology, enabling the real-time collection and storage of subjects’ barometric data. In this research, “Silmee TMW22” is employed for barometric data acquisition, and the valence values (pleasantness/unpleasantness) derived from the “Emotion Snalysis Solutions” system of NEC are utilized as the emotional data.

2.3. Model Building

In this study, machine learning techniques are employed to develop an emotion estimation model. Specifically, random forest (RF) and gradient boosting decision tree (GBDT) are utilized as machine learning models, with all parameters, except those stated, maintained at their default settings in the respective libraries. RF and GBDT were selected due to their superior estimation performance. Furthermore, these tree-based methods are particularly compatible with SHAP analysis, enabling an intuitive interpretation of feature contributions.

2.3.1. RF

RF operates by constructing multiple decision trees based on training data and making predictions of new data (test data) by averaging the outcomes of each decision tree. For this experiment, we used sklearn, an open-source Python 3.13 library. The RF parameters determined through the grid search are listed in Table 2. In particular, max_depth represents the depth of the decision trees, whereas n_estimator indicates their quantity.

2.3.2. GBDT

GBDT is an algorithm that generates multiple decision trees from training data, sequentially corrects errors in these trees, and iteratively builds a model. This experiment employed LightGBM, another open-source Python library. The GBDT parameters obtained from grid search are outlined in Table 3.

2.4. SHAP

The proposed system analyzes the developed machine learning model using SHAP (SHapley Additive exPlanations), as introduced by Lundberg et al. at NIPS 2017 [32]. SHAP is a game theory-based method that provides interpretable insights into machine learning model outputs through additive feature attribution. In this method, the output of the model is expressed as the sum of the contributions of each feature, thereby allowing for an understanding of how each feature influences the prediction.
Given an input vector ( x = ( x 1 , x 2 , , x a ) ) and a model ( f ( x ) ), the SHAP framework approximates an output with an interpretable model ( g ( x ) ) using a simplified input ( x ), expressed as follows:
f ( x ) = g ( x ) = ϕ 0 + i = 1 M ϕ i x i
where ϕ i represents the contribution of each feature to the prediction. These SHAP values are uniquely determined based on the desirable properties of local accuracy, missingness, and consistency.
In this study, we apply TreeSHAP, a variant of SHAP optimized for tree-based models such as Random Forest (RF) and Gradient Boosting Decision Trees (GBDT), which are used in our system. This enables us to estimate the influence of indoor environmental features on individuals’ emotional states (pleasantness or unpleasantness).
By examining the SHAP-value distributions corresponding to environmental features, we identify which features have the greatest impact on mental state and approximate the threshold values where their effect becomes significant.
For detailed mathematical formulation of SHAP, including its derivation from cooperative game theory and full expression of the Shapley value calculation, please refer to Appendix A.

2.5. Experiment

2.5.1. Participants and Experiment Design

This study was conducted in a residential house within the Chiba University experimental facility. A total of ten healthy, non-smoking participants (five couples), aged 20 and over, were recruited to reside in the facility for 3 nights and 4 days. Participants were selected based on a strict set of inclusion and exclusion criteria to ensure a representative and controlled sample for the study.
  • Inclusion Criteria:
    • A healthy couple or partnershipaged 20 or older.
    • Available for a 3-night, 4-day stay at the experimental house at Chiba University.
    • Capable of working indoors (remote working) during the experiment (participants or their partners could also be registered as participants).
    • Low sensitivity to chemical substances, as determined by the Quick Environmental Exposure and Sensitivity Inventory (QEESI) questionnaire.
    • No prior history of symptoms related to sick building syndrome, as determined by the questionnaire.
    • No known allergies.
  • Exclusion Criteria:
    • A history of smoking (if either partner had a history of smoking, the couple was excluded).
    • A diagnosis of chemical sensitivity (if either partner had a diagnosis, the couple was excluded).
Prior to the experiment, all participants provided written informed consent. All procedures were approved by the Institutional Review Board of Chiba University (Receipt No: M10032).

2.5.2. Experimental Environment and Data Collection

This study utilized a hybrid system combining wearable devices and wireless sensor networks to continuously collect both personal vital data and indoor environmental data. This setup was designed to capture the dynamic interplay between a participant’s emotional state and the surrounding environment during extended indoor work periods. Figure 2 displays the layout of this facility, with red circles marking the locations of the wireless environmental sensors in each room. These sensors collected 12 types of indoor environmental data, as detailed in Table 1 and Table 4.
For personal emotional data, we used the Silmee TMW22 wearable device, which estimates a person’s emotional state based on fluctuations in their pulse and skin temperature. A central server was used to collect, manage, and synchronize all sensor and wearable data. To facilitate data matching, participants were required to fill out a behavior chart, which was used to correlate their location and activities within the facility with the time-stamped sensor data. Data collection was paused during periods of sleep, outings, and bathing to focus on emotional states during active, indoor periods.

3. Results

3.1. Model Development and Validation Method

For both GBDT and RF, models were constructed using training data labeled with the subjects’ emotions and the 12 types of indoor environmental data detailed in Table 1. Emotional labeling was primarily conducted using the Emotions Analysis Solutions method [29], which estimates four types of emotion (HAPPY, STRESSED, RELAXED, and SAD) based on fluctuations in pulse and skin temperature.
The dataset, as a whole, may not be strictly time-series data; however, because it includes time-series components, the potential temporal dependencies must be considered in the validation method. In this study, the dataset is divided into segments of fixed length (e.g., groups of five consecutive samples), and these segments serve as the unit of partitioning. Specifically, 80% of the segments are allocated to the training set, while the remaining 20% are assigned to the test set. Crucially, the chronological order of the data within and across segments is maintained to ensure that temporal neighbors of the test data are excluded from the training set.

3.2. Emotion Estimation

Table 5 and Table 6 present the accuracy of the emotion estimation models for each participant ID, including precision, recall, F1 score, accuracy, and the number of samples. While the primary evaluation focuses on test data performance, results for both training and testing are included to assess potential overfitting.
The average accuracy of the training data was 95.2% for RF and 95.3% for GBDT, whereas the average accuracy of the test data was 87.8% for RF and 86.5% for GBDT. Despite potential overfitting, both models demonstrated high accuracy as personal emotion estimation models. Furthermore, the precision, recall, and F1 score of the test data were approximately 0.90, signifying the requirement of high reliability.

3.3. Identification of Critical Environmental Factors Using SHAP

Figure 3 and Figure 4 illustrate the distribution of SHAP values for each environmental factor across the two models. Each subgraph (a–f) corresponds to a participant (ID1–ID6). The horizontal axis represents SHAP values: larger positive values indicate a stronger contribution toward unpleasant states, while negative values indicate contributions toward pleasant states. The color coding (red for high, blue for low) represents the magnitude of each environmental variable. The order of features, sorted by average absolute SHAP values, reflects their overall importance.

3.3.1. Key Environmental Factors

Table 7 summarizes the frequency with which each factor ranked among the top four in terms of SHAP importance. CO2 concentration and temperature consistently ranked highest across participants and models, indicating their dominant influence. Humidity and barometric pressure were also found to be significant but exhibited stronger individual variation. Factors such as light intensity, odor, and sound showed weaker or more inconsistent associations.

3.3.2. CO2 Concentration

Figure 5 and Figure 6 show the relationship between SHAP values and CO2 levels. Across most participants, SHAP values were minimal, ranging between 400 and 600 ppm, suggesting a stable emotional state. Above 600 ppm, SHAP values increased, indicating a tendency toward negative emotions. This implies that lower CO2 levels support psychological comfort, whereas elevated levels may impair well-being.

3.3.3. Temperature

Figure 7 and Figure 8 demonstrate the association between temperature and SHAP values. Stable, near-zero SHAP values were observed within the 23–26 °C range, indicating optimal comfort. Beyond this range, particularly above 26 °C, SHAP values increased, suggesting potential negative effects on emotional states. This threshold is consistent with prior findings on thermal comfort.

3.3.4. Humidity

Figure 9 and Figure 10 present the humidity–SHAP relationship. Unlike CO2 and temperature, no consistent trend was observed, and individual differences were pronounced. Some participants exhibited high SHAP values at low humidity (dry conditions), whereas others showed sensitivity to higher humidity. This variation suggests that while humidity is a relevant factor, its psychological impact is highly dependent on the individual.

4. Discussion

4.1. Key Findings and Comparison with Prior Work

Our findings confirm that indoor environmental factors, particularly CO2 concentration, temperature, and humidity, significantly influence a person’s mental state. While previous studies [25] have explored these relationships under highly controlled conditions, our work extends this research by developing a system for continuous, real-time monitoring in a more realistic remote work environment. This approach allowed us to capture the dynamic interplay of multiple factors that are challenging to measure with traditional methods like questionnaires and one-off experiments.
The models developed in this study achieved high estimation accuracy (>85% for both Random Forest and GBDT), validating our approach. More importantly, the SHAP analysis provided transparent insights into the decision-making process of these black-box models. This is a significant contribution, as it moves beyond simple prediction to provide a causality-like understanding of how specific environmental changes influence emotional well-being.
These findings underscore the strength of SHAP-based interpretation for clarifying the contribution of environmental parameters to emotional states. Conventional machine learning models typically offer only relative feature importance, making it difficult to quantify contributions or interpret correlations. In contrast, SHAP enables a more interpretable, participant-specific understanding of how indoor environmental conditions—particularly CO2, temperature, and humidity—influence mental states.
Our results are consistent with prior studies on environmental factors and emotions [20,33,34,35,36,37,38,39,40]. For example, Bombail demonstrated that odors induce stress responses [35], while Shin et al. showed that lighting conditions influence emotions and brain activity [40]. Choi et al. [41] investigated how changes in indoor environmental quality affect task performance, whereas our study directly links environmental parameters to emotional states. By enabling multimodal, real-time monitoring, our approach provides a practical framework for designing adaptive work environments that support psychological well-being.

4.2. Interpretation of Key Findings

Our analysis, supported by SHAP values, showed that both CO2 concentration and temperature have a consistent and predictable impact on mental state across most participants. Specifically, higher CO2 concentrations and non-optimal temperatures were strongly correlated with predictions of “Unpleasantness.” This is consistent with physiological findings that high CO2 levels can reduce cognitive performance [19] and evoke physical symptoms, which, in turn, affect mood.
In contrast, the influence of humidity was found to be highly inconsistent, with no clear trend across all participants. As demonstrated in Figure 9 and Figure 10, some individuals showed a negative correlation with high humidity, while others were more sensitive to low humidity. This suggests that humidity’s impact is highly subjective and may depend on individual physiological differences or personal preferences. This finding highlights a critical limitation of one-size-fits-all environmental guidelines and underscores the necessity of personalized environmental control systems.

4.3. Limitations and Future Work

This study is not without limitations. The sample size of six participants, while sufficient for a detailed case study, limits the generalizability of our findings. Future research should aim to validate these results with a larger and more diverse cohort. Furthermore, while we simplified emotional states to two categories (Pleasantness and Unpleasantness), incorporating a wider range of emotional states, such as stress or relaxation, could provide a more nuanced understanding.
In the future, we plan to deploy our system in a wider range of remote work and office environments. This will allow us to gather a more comprehensive dataset and develop models that can adapt to various building designs and occupant behaviors. We also aim to develop a real-time feedback system that can recommend personalized environmental adjustments to improve comfort and mental well-being.

5. Conclusions

This study introduced a system designed to analyze the impact of environmental data on individual pleasantness or unpleasantness. Through experimentation, we developed a model to estimate an individual’s mental state from indoor environmental data. By applying SHAP to approximate this model, we identified and analyzed the environmental factors that influence an individual’s sense of pleasantness or unpleasantness.
The experimental outcomes affirmed that each environmental factor plays a pivotal role in understanding the correlation and contribution to an individual’s mental state. These insights are of substantial importance for enhancing mental health and fostering comfortable working environments.
The study, while insightful, is limited by a small sample size and potential overfitting in machine learning models, indicating the need for broader participant samples in future research. Further exploration of individual environmental sensitivities and application across various settings would enhance the model’s universality.
Despite these limitations, the research significantly contributes to understanding of the impact of indoor environments on mental health, offering practical implications for the design of healthier, more productive spaces and underscoring its relevance in environmental psychology and cognitive science.
Our study advances the field by demonstrating the feasibility of continuously estimating mental states through an environmental monitoring system that captures multimodal environmental data. Unlike conventional approaches, which lack interpretability in understanding the relationship between environmental factors and human mental states and fail to identify the specific parameters involved, our system integrates machine learning and SHAP-based analysis to deliver real-time, interpretable insights into occupant well-being.
Based on our findings, it is recommended that indoor environments incorporate adaptive ventilation and temperature regulation systems to enhance occupants’ mental states.
Furthermore, given the limitations of our study, such as limited sample diversity, future work should involve research conducted in more diverse settings and over extended monitoring periods.

Author Contributions

Conceptualization, N.K.; methodology, Y.T., T.I. and N.K.; formal analysis, Y.T., T.I. and N.K.; investigation, Y.T., T.I. and N.K.; data curation, N.K.; writing—original draft preparation, Y.T. and N.K.; writing—review and editing, T.I. and N.K.; project administration, N.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Ethics Committee of the Graduate School of Medicine, Chiba University (M10032), on 8 September 2021.

Informed Consent Statement

Informed consent for participation was obtained from all subjects involved in the study.

Data Availability Statement

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Conflicts of Interest

All authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Abbreviations

The following abbreviations are used in this manuscript:
NECNippon Electric Company, Limited
RFRandom Forest
GBDTGradient Boosting Decision Tree

Appendix A. Detailed Formulation of SHAP

This appendix provides the theoretical background and detailed mathematical formulation of SHAP (SHapley Additive exPlanations).
Consider an input variable ( x = ( x 1 , x 2 , , x a ) ) and a model ( f ( x ) ). SHAP approximates f ( x ) using an interpretable model ( g ( x ) ) defined over simplified inputs ( x ):
f ( x ) = g ( x ) = ϕ 0 + i = 1 M ϕ i x i
x = h x ( x )
where M is the number of input features, h x is a mapping function that relates x and x , and ϕ i represents the contribution of the i-th feature. SHAP values ( ϕ i ) satisfy the following three properties:
  • Local accuracy: Ensures f ( x ) = g ( x ) ;
  • Missingness: If a feature is missing, i.e., x i = 0 , then ϕ i = 0 ;
  • Consistency: If the contribution of a feature increases in a new model, its SHAP value will not decrease.
The unique solution satisfying these properties is derived as the Shapley value from cooperative game theory:
ϕ i ( f , x ) = z x | z | ! ( M | z | 1 ) ! M ! f x ( z ) f x ( z i )
where | z | is the number of non-zero elements in z and z i denotes the vector with the i-th element removed.
SHAP also incorporates conditional expectations:
f x ( z ) f ( h x ( z ) ) E [ f ( z ) | z S ]
where S represents the index set of features present in z and E [ f ( z ) | z S ] is the expected value of f conditioned on those features.
This approach allows for meaningful model interpretation by attributing prediction outcomes to individual features, and it supports model-agnostic and model-specific variants such as TreeSHAP, as used in our study.

References

  1. Zhang, X.; Du, J.; Chow, D. Association between perceived indoor environmental characteristics and occupants’ mental well-being, cognitive performance, productivity, satisfaction in workplaces: A systematic review. Build. Environ. 2023, 246, 360–1323. [Google Scholar] [CrossRef]
  2. Lamb, S.; Kowk, K.S.C. A longitudinal investigation of work environment stressors on the performance and wellbeing of office workers. Appl. Ergon. 2016, 52, 104–111. [Google Scholar] [CrossRef] [PubMed]
  3. Standard 55.66; Thermal Comfort Conditions. ASHRAE: New York, NY, USA, 1966.
  4. Fields, J.M.; De Jong, R.; Gjestland, T.; Flindell, I.H.; Job, R.F.S.; Kurra, S.; Lercher, P.; Vallet, M.; Yano, T.; Guski, R.; et al. Standardized general-purpose noise reaction questions for community noise surveys: Research and a recommendation. J. Sound Vib. 2001, 242, 641–679. [Google Scholar] [CrossRef]
  5. Pransky, G.; Finkelstein, S.; Berndt, E.; Kyle, M.; Mackell, J.; Tortorice, D. Objective and self-report work performance measures: A comparative analysis. Int. J. Prod. Perform. Manag. 2006, 55, 390–399. [Google Scholar] [CrossRef]
  6. Stroop, R. Studies of interference in serial verbal reactions. J. Exp. Psychol. 1935, 18, 643–662. [Google Scholar] [CrossRef]
  7. Thach, T.Q.; Mahirah, D.; Sauter, C.; Roberts, A.C.; Dunleavy, G.; Nazeha, N.; Rykov, Y.; Zhang, Y.; Christopoulos, G.I.; Soh, C.K.; et al. Associations of perceived indoor environmental quality with stress in the workplace. Indoor Air 2020, 30, 1166–1177. [Google Scholar] [CrossRef]
  8. Thatcher, A.; Milner, K. Changes in productivity, psychological wellbeing and physical wellbeing from working in a ’green’ building. Work 2014, 49, 381–393. [Google Scholar] [CrossRef]
  9. Cheung, S.S.; Lee, J.K.; Oksa, J. Thermal stress, human performance, and physical employment standards. Appl. Physiol. Nutr. Metab. 2016, 41, 148–164. [Google Scholar] [CrossRef]
  10. Alkharabsheh, O.H.M.; Jaaffar, A.H.; Chou, Y.C.; Rawati, E.; Fong, P.W. The Mediating Effect of Fatigue on the Nature Element, Organisational Culture and Task Performance in Central Taiwan. Int. J. Environ. Res. Public Health 2022, 19, 8759. [Google Scholar] [CrossRef]
  11. Sakellaris, I.A.; Saraga, D.E.; Mandin, C.; Roda, C.; Fossati, S.; de Kluizenaar, Y.; Carrer, P.; Dimitroulopoulou, S.; Mihucz, V.G.; Szigeti, T.; et al. Perceived Indoor Environment and Occupants’ Comfort in European “Modern” Office Buildings: The OFFICAIR Study. Int. J. Environ. Res. Public Health 2016, 13, 444. [Google Scholar] [CrossRef] [PubMed]
  12. Schultz, W. Behavioral theories and the neurophysiology of reward. Annu. Rev. Psychol. 2006, 57, 87–155. [Google Scholar] [CrossRef] [PubMed]
  13. Healey, J.; Picard, R. Smart Car: Detecting driver stress. In Proceedings of the 15th International Conference on Pattern Recognition, Barcelona, Spain, 3–7 September 2000; Volume 4, pp. 218–221. [Google Scholar]
  14. Singh, R.R.; Conjeti, S.; Banerjee, R. A comparative evaluation of neural network classifiers for stress level analysis of automotive drivers using physiological signals. Biomed. Signal Process. Control 2013, 8, 740–754. [Google Scholar] [CrossRef]
  15. Abbasi, S.F.; Jamil, H.; Chen, W. EEG-based neonatal sleep stage classification using ensemble learning. Comput. Mater. Contin 2022, 70, 4619–4633. [Google Scholar]
  16. Tomoya, K.; Takao, T.; Shinya, K.; Ken, N.; Koji, N.; Toshio, O. Physical stimuli and emotional stress-induced sweat secretions in the human palm and forehead. Anal. Chim. Acta 1998, 365, 319–326. [Google Scholar] [CrossRef]
  17. Snow, S.; Amy, S.B.; Karlien, H.W.P.; Hannah, G.; Marco-Felipe, K.; Janet, B.; Catherine, J.N.; Schraefel, M.C. Exploring the physiological, neurophysiological and cognitive performance effects of elevated carbon dioxide concentrations indoors. Build. Environ. 2022, 156, 243–252. [Google Scholar] [CrossRef]
  18. Zhang, X.; Wargocki, P.; Lian, Z.; Thyregod, C. Effects of exposure to carbon dioxide and bioeffluents on perceived air quality, self-assessed acute health symptoms, and cognitive performance. Indoor Air 2017, 27, 45–64. [Google Scholar] [CrossRef]
  19. Maula, H.; Hongisto, V.; Naatula, V.; Haapakangas, A.; Koskela, H. The effect of low ventilation rate with elevated bioeffluent concentration on work performance, perceived indoor air quality, and health symptoms. Indoor Air 2017, 27, 1141–1153. [Google Scholar] [CrossRef]
  20. Maurizio, G.; Fabio, F.; Michele, F.; Fabio, M.; Aurora, D.A.; Stefania, F.; Isabella, G.T.; Giulia, L.; Cristina, M.; Paolo, M.R.; et al. Topographic electroencephalogram changes associated with psychomotor vigilance task performance after sleep deprivation. Sleep Med. 2014, 15, 1132–1139. [Google Scholar] [CrossRef] [PubMed]
  21. Hedge, A.; Erickson, W.A. A study of indoor environment and sick building syndrome complaints in air conditioned offices: Benchmarks for facility performance. Int. J. Facil. Manag. 1997, 1, 185–192. [Google Scholar]
  22. Hoddes, E.; Dement, W.; Zarcone, V. The development and use of the Stanford sleepiness scale. Psychophysiology 1972, 9, 150. [Google Scholar]
  23. Kim, H.; Hong, T.; Kim, J.; Yeon, S. A psychophysiological effect of indoor thermal condition on college students’ learning performance through EEG measurement. Build. Environ. 2020, 184, 107223. [Google Scholar] [CrossRef]
  24. Kim, H.; Jung, D.; Choi, H.; Hong, T. Advanced prediction model for individual thermal comfort considering blood glucose and salivary cortisol. Build. Environ. 2022, 224, 109551. [Google Scholar] [CrossRef]
  25. Deng, M.; Wang, X.; Menassa, C.C. Measurement and prediction of work engagement under different indoor lighting conditions using physiological sensing. Build. Environ. 2021, 203, 108098. [Google Scholar] [CrossRef]
  26. Komuro, N.; Hashiguchi, T.; Hirai, K.; Ichikawa, M. Development of wireless sensor nodes to monitor working environment and human mental conditions. In IT Convergence and Security; Lecture Notes in Electrical Engineering; Springer: Singapore, 2021; Volume 712, pp. 149–157. [Google Scholar]
  27. Komuro, N.; Hashiguchi, T.; Hirai, K.; Ichikawa, M. Predicting Individual Emotion from Perceptionbased Non-contact Sensor Big Data. Sci. Rep. 2021, 11, 2317. [Google Scholar] [CrossRef]
  28. Kurebayashi, I.; Maeda, K.; Komuro, N.; Hirai, K.; Sekiya, H.; Ichikawa, M. Mental-state estimation model with time-series environmental data regarding cognitive function. Internet Things 2023, 22, 100730. [Google Scholar] [CrossRef]
  29. Abe, K.; Iwata, S. NEC s Emotion Analysis Solution Supports Work Style Reform and Health Management. NEC Tech. J. 2019, 14, 44–48. [Google Scholar]
  30. James, A.R. A Circumplex Model of Affect. J. Personal. Soc. Psychol. 1980, 39, 1161–1178. [Google Scholar]
  31. Jonathan, P.; James, A.R.; Bradley, S.P. The circumplex model of affect: An integrative approach to affective neuroscience, cognitive development, and psychopathology. Dev. Psychopathol. 2005, 17, 715–734. [Google Scholar] [CrossRef] [PubMed]
  32. Lundberg, S.M.; Lee, S. A unified approach to interpreting model predictions. Adv. Neural Inf. Process. Syst. 2017, 30, 4768–4777. [Google Scholar]
  33. Hoenen, M.; Wolf, O.T.; Pause, B.M. The impact of stress on odor perception. Perception 2017, 46, 366–376. [Google Scholar] [CrossRef]
  34. Ayata, D.; Yaslan, Y.; Kamasak, M.E. Emotion based music recommendation system using wearable physiological sensors. IEEE Trans. Consum. Electron. 2018, 64, 196–203. [Google Scholar] [CrossRef]
  35. Bombail, V. Perception and emotions: On the relationships between stress and olfaction. Appl. Anim. Behav. Sci. 2018, 212, 98–108. [Google Scholar] [CrossRef]
  36. Frescura, A.; Lee, P.J. Emotions and physiological responses elicited by neighbours sounds in wooden residential buildings. Build. Environ. 2022, 210, 108729. [Google Scholar] [CrossRef]
  37. Mir, M.; Nasirzadeh, F.; Bereznicki, H.; Enticott, P.; Lee, S. Investigating the effects of different levels and types of construction noise on emotions using EEG data. Build. Environ. 2022, 225, 109619. [Google Scholar] [CrossRef]
  38. Brosch, T. Affect and emotions as drivers of climate change perception and action: A review. Curr. Opin. Behav. Sci. 2021, 42, 15–21. [Google Scholar] [CrossRef]
  39. Scheneider, C.R.; Zaval, L.; Markowitz, E.M. Positive emotions and climate change. Curr. Opin. Behav. Sci. 2021, 42, 114–120. [Google Scholar] [CrossRef]
  40. Shin, Y.B.; Woo, S.H.; Kim, D.H.; Kim, J.J.; Park, J.Y. The effect on emotions and brain activity by the direct/indirect lighting in the residential environment. Neurosci. Lett. 2015, 584, 28–32. [Google Scholar] [CrossRef] [PubMed]
  41. Choi, H.; Kim, H.; Hong, T.; An, J. Examining the indirect effects of indoor environmental quality on task performance: The mediating roles of physiological response and emotion. Build. Environ. 2023, 236, 110298. [Google Scholar] [CrossRef]
Figure 1. Overview of of the sytem proposed in this work. (Source: authors’ own study.)
Figure 1. Overview of of the sytem proposed in this work. (Source: authors’ own study.)
Bdcc 09 00243 g001
Figure 2. Floor plan of the experimental house, which includes a bedroom, a working room, and a living room. (Source: authors’ own study.)
Figure 2. Floor plan of the experimental house, which includes a bedroom, a working room, and a living room. (Source: authors’ own study.)
Bdcc 09 00243 g002
Figure 3. SHAP value distributions for Random Forest models across participants: (a) ID1; (b) ID2; (c) ID3; (d) ID4; (e) ID5; (f) ID6. Each point represents an environmental factor (red = higher values, blue = lower values). Positive SHAP values indicate stronger contributions toward unpleasant emotional states, while negative values indicate contributions toward pleasant states. CO2 concentration and temperature consistently show strong influences across participants.
Figure 3. SHAP value distributions for Random Forest models across participants: (a) ID1; (b) ID2; (c) ID3; (d) ID4; (e) ID5; (f) ID6. Each point represents an environmental factor (red = higher values, blue = lower values). Positive SHAP values indicate stronger contributions toward unpleasant emotional states, while negative values indicate contributions toward pleasant states. CO2 concentration and temperature consistently show strong influences across participants.
Bdcc 09 00243 g003aBdcc 09 00243 g003b
Figure 4. SHAP value distributions for GBDT models across participants: (a) ID1; (b) ID2; (c) ID3; (d) ID4; (e) ID5; (f) ID6. Interpretation is analogous to Figure 3. While CO2 concentration and temperature remain key factors, humidity displays stronger inter-individual variation compared to Random Forest results.
Figure 4. SHAP value distributions for GBDT models across participants: (a) ID1; (b) ID2; (c) ID3; (d) ID4; (e) ID5; (f) ID6. Interpretation is analogous to Figure 3. While CO2 concentration and temperature remain key factors, humidity displays stronger inter-individual variation compared to Random Forest results.
Bdcc 09 00243 g004aBdcc 09 00243 g004b
Figure 5. Relationship between CO2 concentration and SHAP values for (af) participants ID1–ID6 using Random Forest. Stable emotional states are observed at 400–600 ppm (low SHAP values), while higher concentrations (≥600 ppm) correlate with increased SHAP values, suggesting negative emotional impact.
Figure 5. Relationship between CO2 concentration and SHAP values for (af) participants ID1–ID6 using Random Forest. Stable emotional states are observed at 400–600 ppm (low SHAP values), while higher concentrations (≥600 ppm) correlate with increased SHAP values, suggesting negative emotional impact.
Bdcc 09 00243 g005
Figure 6. Relationship between CO2 concentration and SHAP values for (af) participants ID1–ID6 using GBDT. Similar to the RF model, CO2 causes stable and negative emotional states.
Figure 6. Relationship between CO2 concentration and SHAP values for (af) participants ID1–ID6 using GBDT. Similar to the RF model, CO2 causes stable and negative emotional states.
Bdcc 09 00243 g006
Figure 7. Relationship between temperature and SHAP values for (af) participants ID1–ID6 using Random Forest. SHAP values are stable near zero between 23 and 26 °C, indicating optimal comfort. Beyond 26 °C, SHAP values increase, reflecting a tendency toward discomfort.
Figure 7. Relationship between temperature and SHAP values for (af) participants ID1–ID6 using Random Forest. SHAP values are stable near zero between 23 and 26 °C, indicating optimal comfort. Beyond 26 °C, SHAP values increase, reflecting a tendency toward discomfort.
Bdcc 09 00243 g007
Figure 8. Relationship between temperature and SHAP values for (af) participants ID1–ID6 using GBDT. Similar to the RF model, temperature causes comfort and uncomfortable states.
Figure 8. Relationship between temperature and SHAP values for (af) participants ID1–ID6 using GBDT. Similar to the RF model, temperature causes comfort and uncomfortable states.
Bdcc 09 00243 g008aBdcc 09 00243 g008b
Figure 9. Relationship between humidity and SHAP values for (af) participants ID1–ID6 using Random Forest. Unlike CO2 and temperature, no consistent trend is observed. Individual differences dominate, with some participants sensitive to low humidity (dry conditions) and others to high humidity.
Figure 9. Relationship between humidity and SHAP values for (af) participants ID1–ID6 using Random Forest. Unlike CO2 and temperature, no consistent trend is observed. Individual differences dominate, with some participants sensitive to low humidity (dry conditions) and others to high humidity.
Bdcc 09 00243 g009aBdcc 09 00243 g009b
Figure 10. Relationship between humidity and SHAP values for (af) participants ID1–ID6 using Random Forest. Similar to RF, individual differences dominate.
Figure 10. Relationship between humidity and SHAP values for (af) participants ID1–ID6 using Random Forest. Similar to RF, individual differences dominate.
Bdcc 09 00243 g010aBdcc 09 00243 g010b
Table 1. Specifications of the environmental sensors used in the experiment.
Table 1. Specifications of the environmental sensors used in the experiment.
Environmental DataSensor (Model, Principle)
Temperature (°C)DHT11—Measures temperature
Humidity (%)via resistance change and humidity via capacitance change.
CO2 concentration (ppm)MH-Z19C—Detects CO2 by infrared absorption at 4.26 μ m.
Illuminance (lux)TSL25721—Photodiodes sensitive to visible + IR and IR only.
Visible light (output, 0–1023)
Infrared light rays (output, 0–1023)
Full light rays (visible light + infrared rays)
Intensity of blue light (output, 0–1023)CDS393-B—CdS cell with blue filter; resistance varies with light intensity.
Sound level (output, 0–1023)DFR0034—Vibration of electrodes alters capacitance.
Smell (output, 0–1023)TGS2450—Conductivity change in response to sulfur compounds.
Ultrasonic distance (cm)HC-SR04—Distance measured by ultrasonic reflection time.
Barometric pressure (hPa)BME280—Thin-film resistance change due to pressure deformation.
Table 2. RF’s Parameter settings.
Table 2. RF’s Parameter settings.
ParameterValue
random_state42
max_depth10
n_estimator30
Table 3. GBDT’s parameter settings.
Table 3. GBDT’s parameter settings.
ParameterValue
num_boost_round100
early_stopping_rounds10
Table 4. Concrete dataset of the environmental feature data we used for model construction.
Table 4. Concrete dataset of the environmental feature data we used for model construction.
TimeTemp.Hum.Light (Inf.)Light (Full)Light (Vis.)Ill.BlueSoumdSmellUltra.BarometricCO2Emotion
09:18:2023.46183319236719841431017171.91010432HAPPY
09:18:3023.46283322239733847471016173.161010432HAPPY
09:18:4023.46183322239733847481017174.081010432HAPPY
09:18:5023.36184323239729852591017172.751010432HAPPY
09:19:0023.46184323239729848441018173.21010431HAPPY
Table 5. Detailed performance of the Random Forest model. This table presents a comprehensive analysis of the Random Forest model’s classification performance on both the training and test datasets. Results are disaggregated for each participant (ID1–ID6) and include key metrics such as precision, recall, F1 score, and accuracy. The table also specifies the number of samples for each emotion class (Pleasure, Unpleasantness) to provide context for the performance metrics. While a high average training accuracy of 95.2% indicates strong learning, the robust test accuracy of 87.8% confirms the model’s generalization capability.
Table 5. Detailed performance of the Random Forest model. This table presents a comprehensive analysis of the Random Forest model’s classification performance on both the training and test datasets. Results are disaggregated for each participant (ID1–ID6) and include key metrics such as precision, recall, F1 score, and accuracy. The table also specifies the number of samples for each emotion class (Pleasure, Unpleasantness) to provide context for the performance metrics. While a high average training accuracy of 95.2% indicates strong learning, the robust test accuracy of 87.8% confirms the model’s generalization capability.
TrainTest
ID1 PrecisionRecallF1 ScoreNumberID1 PrecisionRecallF1 ScoreNumber
Pleasure0.940.800.871353Pleasure0.840.690.76296
Unpleasure0.920.980.953187Unpleasure0.900.950.92844
Macro avg0.930.890.914540Macro avg0.870.820.841140
Accuracy 0.93 Accuracy 0.88
ID2 PrecisionRecallF1 ScoreNumberID2 PrecisionRecallF1 ScoreNumber
Pleasure0.940.990.973507Pleasure0.870.940.90853
Unpleasure0.970.810.881138Unpleasure0.790.610.69312
Macro avg0.960.900.934645Macro avg0.830.770.791165
Accuracy 0.95 Accuracy 0.85
ID3 PrecisionRecallF1 ScoreNumberID3 PrecisionRecallF1 ScoreNumber
Pleasure0.970.950.961972Pleasure0.930.860.90493
Unpleasure0.970.980.983548Unpleasure0.930.960.95887
Macro avg0.970.970.975520Macro avg0.930.910.921380
Accuracy 0.97 Accuracy 0.93
ID4 PrecisionRecallF1 ScoreNumberID4 PrecisionRecallF1 ScoreNumber
Pleasure0.940.890.921150Pleasure0.750.680.72256
Unpleasure0.960.980.973230Unpleasure0.910.930.92844
Macro avg0.950.940.944380Macro avg0.830.810.821100
Accuracy 0.96 Accuracy 0.87
ID5 PrecisionRecallF1 ScoreNumberID5 PrecisionRecallF1 ScoreNumber
Pleasure0.960.950.962496Pleasure0.920.920.92685
Unpleasure0.930.930.931564Unpleasure0.830.840.83335
Macro avg0.940.940.944060Macro avg0.880.880.881020
Accuracy 0.95 Accuracy 0.89
ID6 PrecisionRecallF1 ScoreNumberID6 PrecisionRecallF1 ScoreNumber
Pleasure0.950.970.962528Pleasure0.860.910.88655
Unpleasure0.940.920.931552Unpleasure0.820.740.77370
Macro avg0.950.940.944080Macro avg0.840.820.831025
Accuracy 0.95 Accuracy 0.85
Table 6. Detailed performance of the GBDT model. This table provides a comprehensive overview of the GBDT model’s performance, mirroring the structure of RF. The results, shown for each participant (ID1–ID6), are based on the same key metrics (precision, recall, F1 score, and accuracy) and sample counts. The GBDT model achieved a high average training accuracy of 95.3% and a strong test accuracy of 86.5%, demonstrating its effectiveness in distinguishing between pleasant and unpleasant emotional states.
Table 6. Detailed performance of the GBDT model. This table provides a comprehensive overview of the GBDT model’s performance, mirroring the structure of RF. The results, shown for each participant (ID1–ID6), are based on the same key metrics (precision, recall, F1 score, and accuracy) and sample counts. The GBDT model achieved a high average training accuracy of 95.3% and a strong test accuracy of 86.5%, demonstrating its effectiveness in distinguishing between pleasant and unpleasant emotional states.
TrainTest
ID1 PrecisionRecallF1 ScoreNumberID1 precisionRecallF1 ScoreNumber
Pleasure0.910.830.871288Pleasure0.750.650.70361
Unpleasure0.940.970.953252Unpleasure0.850.900.87779
Macro avg0.920.900.914540Macro avg0.800.770.781140
Accuracy 0.93 Accuracy 0.82
ID2 precisionRecallF1 ScoreNumberID2 PrecisionRecallF1 ScoreNumber
Pleasure0.960.990.983488Pleasure0.920.950.93872
Unpleasure0.970.890.931157Unpleasure0.840.740.79293
Macro avg0.960.940.954645Macro avg0.880.850.861165
Accuracy 0.96 Accuracy 0.90
ID3 PrecisionRecallF1 ScoreNumberID3 PrecisionRecallF1 ScoreNumber
Pleasure0.970.970.971998Pleasure0.870.850.86467
Unpleasure0.980.980.983522Unpleasure0.920.940.93913
Macro avg0.980.980.985520Macro avg0.900.890.891380
Accuracy 0.98 Accuracy 0.91
ID4 PrecisionRecallF1 ScoreNumberID4 PrecisionRecallF1 ScoreNumber
Pleasure0.940.840.891069Pleasure0.820.600.69337
Unpleasure0.950.980.973311Unpleasure0.840.940.89763
Macro avg0.940.910.934380Macro avg0.830.770.791100
Accuracy 0.95 Accuracy 0.84
ID5 PrecisionRecallF1 ScoreNumberID5 PrecisionRecallF1 ScoreNumber
Pleasure0.960.960.962613Pleasure0.850.880.87568
Unpleasure0.920.930.931447Unpleasure0.850.800.82452
Macro avg0.940.940.944060Macro avg0.850.840.851020
Accuracy 0.95 Accuracy 0.85
ID6 PrecisionRecallF1 ScoreNumberID6 PrecisionRecallF1 ScoreNumber
Pleasure0.960.960.962508Pleasure0.910.900.90675
Unpleasure0.930.930.931572Unpleasure0.810.830.82350
Macro avg0.940.950.954080Macro avg0.860.860.861025
Accuracy 0.95 Accuracy 0.87
Table 7. Frequency of top-four SHAP-ranked features for RF and GBDT. Counts represent how many times each environmental factor appeared in the top four across IDs 1–6. CO2 and temperature were consistently dominant, while humidity and barometric pressure had a moderate influence and light and odor exhibited weaker or individual-dependent effects.
Table 7. Frequency of top-four SHAP-ranked features for RF and GBDT. Counts represent how many times each environmental factor appeared in the top four across IDs 1–6. CO2 and temperature were consistently dominant, while humidity and barometric pressure had a moderate influence and light and odor exhibited weaker or individual-dependent effects.
Random ForestGBDT
ParameterCountParameterCount
CO2 Concentration6 CO2 Concentration
Temperature
6
Temperature5
Barometric Pressure4Humidity5
Humidity3 Barometric Pressure
Blue Light
2
Odor2
Blue Light
Full Ray
Infrared Ray
Ultrasonic Distance
1  Odor
Infrared Ray
Ultrasonic Distance
1
 Visible Ray
Sound level
Illuminance
0Visible Ray
Full Ray
Sound level
Illuminance
0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Toriyama, Y.; Isogami, T.; Komuro, N. Interpretable Emotion Estimation in Indoor Remote Work Environments via Environmental Sensor Data. Big Data Cogn. Comput. 2025, 9, 243. https://doi.org/10.3390/bdcc9100243

AMA Style

Toriyama Y, Isogami T, Komuro N. Interpretable Emotion Estimation in Indoor Remote Work Environments via Environmental Sensor Data. Big Data and Cognitive Computing. 2025; 9(10):243. https://doi.org/10.3390/bdcc9100243

Chicago/Turabian Style

Toriyama, Yuma, Tsumugi Isogami, and Nobuyoshi Komuro. 2025. "Interpretable Emotion Estimation in Indoor Remote Work Environments via Environmental Sensor Data" Big Data and Cognitive Computing 9, no. 10: 243. https://doi.org/10.3390/bdcc9100243

APA Style

Toriyama, Y., Isogami, T., & Komuro, N. (2025). Interpretable Emotion Estimation in Indoor Remote Work Environments via Environmental Sensor Data. Big Data and Cognitive Computing, 9(10), 243. https://doi.org/10.3390/bdcc9100243

Article Metrics

Back to TopTop