Skip Content
You are currently on the new version of our website. Access the old version .
  • Article
  • Open Access

25 January 2022

Assessing the Visualization-Based Decision Support System for Environmental Impact Assessments

,
,
,
,
and
1
College of Human Ecology, Cornell University, Ithaca, NY 14853, USA
2
Gaia 3D Incorporation, Seoul 08611, Korea
3
College of Engineering, Ewha Womans University, Seoul 03760, Korea
*
Author to whom correspondence should be addressed.

Abstract

Even though environmental impact assessments (EIAs) have been an important tool for environmental decision-making, most EIAs are published as a mix of text and tabular data that is not easily accessible to or understandable for the public. In this paper, we present a decision support system (DSS) that supports the decision-making of stakeholders in the EIA stage. The system was designed to improve the public’s understanding of stakeholders before and after a construction project by providing visualization of key environmental elements. We recruited 107 participants to test the usability of the system and examined the impacts of individual differences between the participants on their perceptions of the system, including their environmental expertise and computer self-efficacy. The results showed that the proposed system had high usability, especially for users with high computational efficacy and environment expertise. The system could thus help to improve the communication between the public and experts during public hearings and enhance the environmental literacy of the public.

1. Introduction

An environmental impact assessment (EIA) is a decision-making tool that helps to evaluate the likely positive and negative environmental impacts of a proposed project or development while also considering interrelated socio-economic, cultural, and health effects [1,2]. An EIA is designed to review, predict, and assess the potential environmental impacts of a project before its plan is finalized so that strategies can be developed to prevent or minimize possible adverse effects on the environment [2,3,4]. Although the EIA process may vary around the world, it is important to determine whether the predicted impacts occurred and whether the proposed mitigation measures were implemented. In particular, the early stages of an EIA in the process of public hearing/comments could lead to more engagement between the affected community and the developers [2,5].
In addition, in the upper level of the EIA process, SEA (strategic environmental assessment), the usability of visualization presumably could help SEA practitioners and planners in terms of more environmentally friendly development, a smoother planning process, greater plan transparency, and the like [6]. The visualization should bring public involvement of the EIA and SEA to speed up plans, policies, and projects by reducing the environmental impacts and improving the environmental benefits of plans, acting as an instrument of knowledge brokerage, and improve public participation in both levels [7].
There are some limitations that can hinder the communication between stakeholders in the process of environment-related decision-making [8]. In an EIA as well as SEA, experts often use only text and tables with numerical figures to present their findings to the public. Many stakeholders, such as the public, developers, administrators, and engineers, are involved in the decision-making process, and the role of the public and their participation is especially important. However, because the general public lacks professional knowledge regarding the environment, it is difficult for them to fully understand the details of projects under consideration. This lack of understanding could lead to serious miscommunication between stakeholders, inefficient decision-making, and unintended project outcomes. Previous studies only demonstrated modelled using a Bayesian network (BN) [9] and ANP (analytical network process) [10], but this lacks limited access to the internet (w/o installation of software). Furthermore, our study used real-time movement of water, airflow, and oil spill so that a modification of inputs can allow users to test alternatives in different ways, and this is a part of a wholistic model with visualization for a decision-making process.
To resolve this problem, a tool that can allow the public to easily understand the post-development associated with the EIA process is needed. Using this tool during the public review stage of a project, concerned citizens would be better able to understand the potential environmental consequences of a construction project. In this respect, the use of suitable data visualization elements such as maps, charts, and simulations would be useful because they can enhance the delivery of information for environmental topics [4]. It is also important for the tool to be accessible, which would encourage more widespread use and promote effective decision-making. Data visualization can also be important when conveying information about health impacts. For example, it has been found that how an impact assessment on the environmental and health impacts of coal mining is implemented and presented affects the outcomes of the project [11].
The main purpose of this paper is to design a system that provides effective data visualizations for environmental information that can be used to improve the communication between stakeholders. We thus propose a decision support system (DSS) that visualizes the environmental conditions before and after a proposed construction project to help EIA stakeholders communicate freely and make reasonable decisions. System usability testing that examined user satisfaction with the system was conducted, with the results analyzed depending on the individual differences of the respondents, in particular their computer self-efficacy and environmental expertise.
In the remainder of the paper, we first review past studies on EIAs and DSSs. We then describe the design and development of the system, before presenting the findings of the system usability testing. A discussion of these results and suggestions for future research then conclude the paper.

3. User Analysis

3.1. System Design

The system prototype was geared toward construction and civil engineering development projects. The system was designed to offer four simulations associated with a construction project: (1) hydrological water flow, (2) oil spills, (3) air quality due to microscopic or particulate matter, and (4) airflow. The water and oil simulations predicted the likelihood of floods before, during, and after the construction process, while the wind simulation was developed to visualize the airflow before and after construction. In addition, wind simulation has some application to be used in the context of air pollution. The algorithm process in this study did not compare quantitative measures with data in the field, but we believed that the methods should be quite similar and support in preparation of the EIA report.
The proposed system offers the advantages of openness and accessibility. Previous systems have mostly operated in stand-alone mode [4,24,25,26]. On the other hand, the proposed system is implemented as a pure web service that does not use non-standardized technologies such as Active X, applets, and Flash. Stand-alone and client server modes are commonly employed because simulations and 3D rendering require high computational power [26]. The proposed system attempts to overcome this limitation using a pure web service.

3.1.1. Algorithms and Models

The simulations were designed using the cellular automata simulation method [27]. For the simulation of the hydrological water flow, interactive terrain modeling using hydraulic erosion was used. Interactive terrain modeling was built upon different physics-based erosion and deposition algorithms [28]. Each frame of a pixel represents a water column, which has the following information: (1) the layers of the material it consists of, (2) the depth of those layers, (3) the tilt angle, and (4) the height of the water.
Using pixel-level information, the model automatically calculates the water flow, the sediment change, and how it spreads across the terrain. The height of a water column represents the acceleration of the water flow. The height of the columns indicates the sum of the height of different terrain layers and water layers.
Based on the height differences between water columns, the acceleration, flow, and velocity of the water can be calculated, which can then be used to determine the changes in the sediment by pixel. By analyzing the interaction between the water columns and the layers, tilt angle, and regolith thickness, the algorithm predicts the interactions of the following water columns. The algorithm predicts the change across the entire terrain, not just the interactions between the water columns.
For the wind flow, we visualized the results of the simulation data in raster format rather than the data itself (Figure 2). The streamline was calculated based on wind field data.
Figure 2. Raster data, the determination of wind flow markers, and the visualization of wind data.

3.1.2. Hydrological Water Flow System

The simulation presents the risk of flooding before and after a construction project (Figure 3). Given that the development area will change as the development progresses, the system presents the predicted changes to the river, particularly the damage occurring near the river due to flooding.
Figure 3. Hydrological water flow simulation.

3.1.3. Interactive Water Simulation

This simulation visualizes the hydrological water flow and simulates changes depending on user input, including embankment removal or construction (Figure 4). The orange and red shapes represent the height of the flood gate. In the interactive water simulation, users can freely adjust the placement and height of the flood gate. At the same time, they can set the area for water contaminant.
Figure 4. Interactive water simulation.

3.1.4. Wind Simulation

This simulation demonstrates the changes in the wind paths before and after the construction of a new town (Figure 5). When the town is constructed, the wind speeds increase. The white arrows represent the flow of the wind, showing both the direction and speed.
Figure 5. Wind simulation.

3.1.5. Oil and Wastewater Simulation

This simulation visualizes the flow of oil and wastewater by adjusting the geographical features in real-time (Figure 6). The yellow elements represent a mix of wastewater and oil that has not spread into the environment, while the red elements represent the same pollutants that have been emitted into the surroundings. The blue area depicts the flow of oil and wastewater over the surface.
Figure 6. Flow of oil and wastewater based on the adjustment of geographical features.

3.2. Participants and Measurements

We recruited 107 participants (55 female, 52 male) for this study online and on campus at Cornell University. We attempted to recruit a similar proportion of participants in each age group because we believed that age might be highly correlated with computer self-efficacy and perception of system usability. Thus, 27.41% of the participants were in their 20 s, 33.33% were in their 30 s, 9.63% were in their 40 s, and 29.63% were in their 50 s.

3.2.1. Computer Self-Efficacy Survey

Computer self-efficacy is a self-rated measure of specific computer-related knowledge and skills [29]. In the present survey, the computer self-efficacy survey asked the participants to imagine that they are given a new software package for some aspect of work. The type of software package does not matter, but the package is designed to make a job easier. Participants rate their confidence in their ability to use the new software package from 1 to 10. The full survey is provided in Appendix A.

3.2.2. System Usability Scale (SUS)

System usability surveys are considered a quick, reliable, and valid method for assessing the learnability and usability of a system [30]. Standardized questions are widely used to evaluate the usability of a system. An advantage of a system usability scale (SUS) is that it can be used with a diverse range of contexts, technologies, and systems. In addition, it demonstrates the general and overall quality of a system and its appropriateness for a specific context [30]. The SUS in the present study used a 5-point Likert scale [31] to assess the level of agreement with 10 statements, resulting in a score from 0 to 100 points, with higher scores demonstrating higher subjective usability. The full survey is provided in Appendix B.

3.3. Procedures

Participants first provided their demographic information, including gender, age, race, whether they were working in an environment-related field, and their computer self-efficacy [32]. They then explored the system, with the different simulations presented in random order. The participants used a MacBook Pro laptop with a mouse when exploring the system. The participants were given a brief explanation of each simulation. After the explanation, they were asked to click buttons and use the system without any instructions for each simulation. After five minutes of exploration, they answered the SUS [30].

4. Results

A linear mixed-model analysis was conducted to test the effects of the different simulations on the SUS scores. Several variables, such as gender, age, and environmental expertise, were included.

4.1. Impact of System Simulations on SUS Scores

The results showed that there were no significant differences between the four different simulations in the SUS scores (Table 2).
Table 2. Mean SUS scores for the four simulations on SUS (standard deviation in parentheses).

4.2. Impacts of Individual Differences on SUS Scores

4.2.1. Environmental Expertise

The results showed that people working in the environmental field scored higher (+0.631) on the SUS than did those not working in the field, and this difference was statistically significant (p < 0.05).

4.2.2. Computer Self-Efficacy

People who had higher computer self-efficacy had higher SUS scores than those with low self-efficacy (+0.102), and this difference was also statistically significant (p < 0.001)

4.2.3. Interaction Effect

Computer self-efficacy showed a significant interaction effect with environmental expertise (Table 3). Among participants who had higher computer self-efficacy, those who were also environmental experts scored 0.173 higher on the SUS than participants who were not experts (p < 0.001).
Table 3. Results of linear mixed models for computer self-efficacy and environmental expertise.

5. Discussion

Since their introduction as part of the NEPA [3], EIAs have been instituted in more than 100 countries, serving not only as a major instrument for environmental management [33] but also as an effective tool for sustainable development that takes into account environmental, social, and economic factors. However, due to the complexity of EIAs in terms of the elements they are expected to cover (Table 1), stakeholders often find it difficult to understand EIA reports, thus, they cannot participate effectively in the establishment of development plans and related decisions.
Furthermore, SEA aims to integrate environmental considerations into plan-making, and so to promote sustainable development and environmental protection. Global visualization of EIA and SEA processes generally entails the preparation of an environmental report that identifies the likely significant impacts of the levels in a plan/programme (SEA) as well as projects (EIA), proposal of measures to avoid or mitigate any significant negative impacts, consultation on the report with the public and government environmental bodies, taking the environmental report and consultation findings, and monitoring of the actual impacts of the plan/programme/projects helping with visualization [7].
In particular, for a potential construction project, local residents are a large stakeholder group, but they are typically not organized and have a lower social status and limited financial resources. Therefore, though EIAs play a positive role in the public decision-making process [34], meaningful public participation in the EIA process is lacking, leading to an imbalance between relevant stakeholders [35].
The importance of involving the public in the design process for visual tools has been clearly outlined by the International Association for Public Participation (IAP2), who identify it as a core value of public participation: “public participation seeks input from participants in designing how they participate” [36,37]. This is particularly important because effective images evoke powerful emotions, driving decisions and actions. This key feature of visual tools can be used to support an effective, meaningful, and participatory dialogue, contributing to enhanced knowledge brokerage [38]. The main purpose of the DSS proposed in the present study for EIAs was thus to help stakeholders communicate more effectively and enhance social sustainability. The results showed that respondents with a higher computer self-efficacy tended to rate the system as more usable than those who had lower computer self-efficacy. This suggests that engineers and environment administrative officers should benefit from the DDS in the decision-making process because they are assumed to have higher computer self-efficacy [34].
Visualization has also been advocated as an instrument of change towards more sustainable patterns of behavior and development [39] and as a tool to support the achievement of Sustainable Development Goals.
In addition, environmental experts were more satisfied with this system than were members of the general public with no environmental expertise. This indicates that the DDS could be further improved by servicing the general public who have no environmental background. However, this system could be employed in other countries with other stakeholders because there are no barriers in terms of language or cultural/policy differences. In addition, it could be useful in the context of public hearings.

6. Limitations and Future Studies

This study has some limitations. First, the system was designed in a 2D format that could be extended to a 3D form in the future. Thus, to design a system that helps the general public better understand the EIA process, future studies could apply 3D modeling techniques. Second, even though EIAs are employed in a variety of countries, differences exist between them. Third, a system that visualizes other elements that are not covered by the six types of environment and their associated items in a typical EIA is needed (Table 1). Our study shows that this system would help to increase environmental literacy. Moreover, in this study we only demonstrated the issues such as waterflow, air movement, and oil spill as examples, but in the future all of the 21 items (Table 1) will be implemented to help the decision-making process; therefore, items like landscape, noise/vibration, and sunlight should be further studied.

7. Conclusions

In this paper, we designed and presented a DSS that assists decision-making in the EIA process for a variety of stakeholders. The system visualizes for the user the environmental conditions before and after a construction project. To better understand the perceived usability for different stakeholders, we explored the impact of individual differences, including computer self-efficacy and environmental expertise, on SUS scores. The results obtained from 107 participants exploring the system and answering a survey showed that the participants with higher computer self-efficacy and environmental experience scored more highly on the SUS. We expect that our proposed system could help improve the communication between the public and experts during public hearings.

Author Contributions

Conceptualization, S.L.; methodology, S.-y.L., H.K.; software, S.S.; formal analysis, S.-y.L. and S.-Y.Y.; writing—original draft preparation, S.-y.L.; writing—review and edit-ing, S.L.; visualization, M.-K.K.; funding acquisition, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Development of ICT-based EIA Technology, Ministry of Environment (MOE-2020002990006 and 2020002990005) and National Research Foundation of Korea (NRF) by the Ministry of Science, Technology and Education (NRF-2021R1A2C1011213).

Institutional Review Board Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Computer Self-Efficiency Survey

I could complete the job using the software package if…
  • There was no one around to tell me what to do as I go.
  • I had never used a package like it before.
  • I had only the software manuals for reference.
  • I had seen someone else using it before trying it myself.
  • I could call someone for help if I got stuck.
  • Someone else had helped me get started.
  • I had a lot of time to complete the job for which the software was provided.
  • I had just the built-in help facility for assistance.
  • Someone showed me how to do it first.
  • I had used similar packages before this one to do the same job.

Appendix B. System Usability Scale Survey Items

  • I think that I would like to use this system frequently.
  • I found the system unnecessarily complex.
  • I thought the system was easy to use.
  • I think that I would need the support of a technical person to be able to use this system.
  • I found the various functions in this system were well integrated.
  • I thought there was too much inconsistency in this system.
  • I would imagine that most people would learn to use this system very quickly.
  • I found the system very cumbersome to use.
  • I felt very confident using the system.
  • I needed to learn a lot of things before I could get going with this system.

References

  1. Makhdoum, M.F. Degradation model: A quantitative EIA instrument, acting as a Decision Support System (DSS) for environmental management. Environ. Manag. 2002, 30, 151–156. [Google Scholar] [CrossRef]
  2. Glasson, J.; Therivel, R. Introduction to Environmental Impact Assessment; Routledge: London, UK, 2019. [Google Scholar]
  3. United States National Environmental Policy Act of 1969. Available online: https://www.epa.gov/laws-regulations/summary-national-environmental-policy-act (accessed on 20 January 2022).
  4. Li, W.; Zhu, J.; Zhang, Y.; Cao, Y.; Hu, Y.; Fu, L.; Huang, P.; Xie, Y.; Yin, L.; Xu, B. A Fusion Visualization Method for Disaster Information Based on Self-Explanatory Symbols and Photorealistic Scene Cooperation. ISPRS Int. J. Geo-Inform. 2019, 8, 104. [Google Scholar] [CrossRef]
  5. Song, Y.-I.; Glasson, J. A new paradigm for Environmental Assessment (EA) in Korea. Environ. Impact Assess. Rev. 2010, 30, 90–99. [Google Scholar] [CrossRef]
  6. Partidario, M.R.; Monteiro, M.B. Strategic environmental assessment effectiveness in Portugal. Impact Assess. Proj. Apprais. 2019, 37, 247–265. [Google Scholar] [CrossRef]
  7. Therivel, R.; Gonzalez, A. Is SEA worth it? Short-term costs v. long-term benefits of strategic environmental assessment. Environ. Impact Assess. Rev. 2020, 83, 106411. [Google Scholar] [CrossRef]
  8. Kim, E.; Song, W.; Lee, D. A multi-scale metrics approach to forest fragmentation for Strategic Environmental Impact Assessment. Environ. Impact Assess. Rev. 2013, 42, 31–38. [Google Scholar] [CrossRef]
  9. Gill, L.; Kumar, V.K.; Lange, E.; Lerner, D.; Morgan, E.; Romano, D.; Shaw, E. An interactive visual decision support tool for sustainable urban river corridor management. In Proceedings of the International Environmental Modelling and Software Society (iEMSs) 2010 International Congress on Environmental Modelling and Software, Ottawa, ON, Canada, 5–8 July 2010. [Google Scholar]
  10. Isaacs, J.; Falconer, R.; Blackwood, D. A unique approach to visualising sustainability in the built environment. In Proceedings of the 2008 International Conference Visualisation, Washington, DC, USA, 9–11 July 2008; IEEE Xplore: Manhattan, NY, USA, 2008. [Google Scholar] [CrossRef]
  11. Cortes-Ramirez, J.; Sly, P.D.; Ng, J.; Jagals, P. Using human epidemiological analyses to support the assessment of the impacts of coal mining on health. Rev. Environ. Health 2019, 34, 391–401. [Google Scholar] [CrossRef]
  12. Caldwell, L.K. Environment: A New Focus for Public Policy? Public Admin. Rev. 1963, 23, 132–139. [Google Scholar] [CrossRef]
  13. Eom, S.B.; Lee, S.M.; Kim, E.B.; Somarajan, C. A survey of decision support system applications (1988–1994). J. Oper. Res. Soc. 1998, 49, 109–120. [Google Scholar] [CrossRef]
  14. Power, D.J. Decision Support Systems: Concepts and Resources for Managers; Greenwood/Quorum: Westport, CT, USA, 2002. [Google Scholar]
  15. Gunaratnam, M.; Alva-Argáez, A.; Kokossis, A.; Kim, J.-K.; Smith, R. Automated Design of Total Water Systems. Ind. Eng. Chem. Res. 2005, 44, 588–599. [Google Scholar] [CrossRef]
  16. Rodgers, P.A.; Huxor, A.P.; Caldwell, N.H.M. Design Support Using Distributed Web-Based AI Tools. Res. Eng. Des. 1999, 11, 31–44. [Google Scholar] [CrossRef]
  17. Bang, H.; Virós Martin, A.; Prat, A.; Selva, D. Daphne: An intelligent assistant for architecting earth observing satellite systems. In Proceedings of the 2018 AIAA Information Systems-AIAA Infotech@ Aerospace 2018, Kissimmee, FL, USA, 8–12 January 2018; p. 1366. [Google Scholar]
  18. Varma, V.K.; Ferguson, I.; Wild, I. Decision support system for the sustainable forest management. For. Ecol. Manag. 2000, 128, 49–55. [Google Scholar] [CrossRef]
  19. Sugumaran, R.; Meyer, J.C.; Davis, J. A Web-based environmental decision support system (WEDSS) for environmental planning and watershed management. J. Geogr. Syst. 2004, 6, 307–322. [Google Scholar] [CrossRef]
  20. Booty, W.; Lam, D.; Wong, I.; Siconolfi, P. Design and implementation of an environmental decision support system. Environ. Model. Softw. 2001, 16, 453–458. [Google Scholar] [CrossRef]
  21. Choi, S.-W.; Lee, E.-B.; Kim, J.-H. The Engineering Machine-Learning Automation Platform (EMAP): A Big-Data-Driven AI Tool for Contractors’ Sustainable Management Solutions for Plant Projects. Sustainability 2021, 13, 10384. [Google Scholar] [CrossRef]
  22. Relvas, H.; Miranda, A.I. An urban air quality modeling system to support decision-making: Design and implementation. Air Qual. Atmos. Health 2018, 11, 815–824. [Google Scholar] [CrossRef]
  23. Vogt, A.; Babel, F.; Hock, P.; Baumann, M.; Seufert, T. Immersive virtual reality or auditory text first? Effects of adequate sequencing and prompting on learning outcome. Br. J. Educ. Technol. 2021, 52, 2058–2076. [Google Scholar] [CrossRef]
  24. Li, W.; Zhu, J.; Fu, L.; Zhu, Q.; Xie, Y.; Hu, Y. An augmented representation method of debris flow scenes to improve public perception. Int. J. Geogr. Inf. Sci. 2020, 35, 1521–1544. [Google Scholar] [CrossRef]
  25. Gontier, M. Scale issues in the assessment of ecological impacts using a GIS-based habitat model—A case study for the Stockholm region. Environ. Impact Assess. Rev. 2007, 27, 440–459. [Google Scholar] [CrossRef]
  26. Narayanan, S.; McIlraith, S. Analysis and simulation of web services. Comput. Netw. 2003, 42, 675–693. [Google Scholar] [CrossRef]
  27. Chopard, B.; Droz, M. Cellular Automata; Springer: Berlin, Germany, 1998; Volume 1. [Google Scholar]
  28. Št’ava, O.; Beneš, B.; Brisbin, M.; Křivánek, J. Interactive terrain modeling using hydraulic erosion. In Proceedings of the 2008 ACM SIGGRAPH/Eurographics Symposium on Computer Animation 2008, Dublin, Ireland, 7–9 July 2008; pp. 201–210. [Google Scholar]
  29. Murphy, C.A.; Coover, D.; Owen, S.V. Development and Validation of the Computer Self-Efficacy Scale. Educ. Psychol. Meas. 1989, 49, 893–899. [Google Scholar] [CrossRef]
  30. Brooke, J. SUS: A ’Quick and Dirty’ Usability Scale. Usability Eval. Ind. 1996, 189, 4–7. [Google Scholar]
  31. Lewis, J.R. The System Usability Scale: Past, Present, and Future. Int. J. Hum.-Comput. Interact. 2018, 34, 577–590. [Google Scholar] [CrossRef]
  32. Compeau, D.R.; Higgins, C.A. Computer Self-Efficacy: Development of a Measure and Initial Test. MIS Q. 1995, 19, 189–211. [Google Scholar] [CrossRef]
  33. The World Bank. Participation Sourcebook; ESD: Washington, DC, USA, 1996. [Google Scholar]
  34. Glucker, A.N.; Driessen, P.; Kolhoff, A.; Runhaar, H.A. Public participation in environmental impact assessment: Why, who and how? Environ. Impact Assess. Rev. 2013, 43, 104–111. [Google Scholar] [CrossRef]
  35. Hartley, N.; Wood, C. Public participation in environmental impact assessment—Implementing the Aarhus Convention. Environ. Impact Assess. Rev. 2005, 25, 319–340. [Google Scholar] [CrossRef]
  36. Sheppard, S.R.J. Guidance for crystal ball gazers: Developing a code of ethics for landscape visualization. Landsc. Urban Plan. 2001, 54, 183–199. [Google Scholar] [CrossRef]
  37. IAP2. Student Workbook—Planning for Effective Public Participation; International Association for Public Participation: Montreal, QC, Canada, 2006. [Google Scholar]
  38. Partidário, M.R.; Sheate, W.R. Knowledge brokerage—Potential for increased capacities and shared power in impact assessment. Environ. Impact Assess. Rev. 2013, 39, 26–36. [Google Scholar] [CrossRef]
  39. Partidärio, M.R. A strategic advocacy role in sea for sustainability. J. Environ. Assess. Policy Manag. 2015, 17, 1550015. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.