Next Article in Journal
Olive Oils Classification via Laser-Induced Breakdown Spectroscopy
Next Article in Special Issue
Skeleton-Based Dynamic Hand Gesture Recognition Using an Enhanced Network with One-Shot Learning
Previous Article in Journal
Optimal Design of a Plenum Fan with Three-Dimensional Blades
Previous Article in Special Issue
Semi-Immersive Virtual Reality as a Tool to Improve Cognitive and Social Abilities in Preschool Children
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exploring Visual Perceptions of Spatial Information for Wayfinding in Virtual Reality Environments

1
School of Architecture, Soongsil University, Seoul 06978, Korea
2
School of Architecture, Hanyang University, Seoul 04763, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(10), 3461; https://doi.org/10.3390/app10103461
Submission received: 31 March 2020 / Revised: 11 May 2020 / Accepted: 15 May 2020 / Published: 17 May 2020
(This article belongs to the Special Issue Augmented Reality, Virtual Reality & Semantic 3D Reconstruction)

Abstract

:
Human cognitive processes in wayfinding may differ depending on the time taken to accept visual information in environments. This study investigated users’ wayfinding processes using eye-tracking experiments, simulating a complex cultural space to analyze human visual movements in the perception and the cognitive processes through visual perception responses. The experiment set-up consisted of several paths in COEX Mall, Seoul—from the entrance of the shopping mall Starfield to the Star Hall Library to the COEX Exhibition Hall—using visual stimuli created by virtual reality (four stimuli and a total of 60 seconds stimulation time). The participants in the environment were 24 undergraduate or graduate students, with an average age of 24.8 years. Participants’ visual perception processes were analyzed in terms of the clarity and the recognition of spatial information and the activation of gaze fixation on spatial information. That is, the analysis of the visual perception process was performed by extracting “conscious gaze perspective” data comprising more than 50 consecutive 200 ms continuous gaze fixations; “visual understanding perspective” data were also extracted for more than 300 ms of continuous gaze fixation. The results show that the methods for analyzing the gaze data may vary in terms of processing, analysis, and scope of the data depending on the purpose of the virtual reality experiments. Further, they demonstrate the importance of what purpose statements are given to the subject during the experiment and the possibility of a technical approach being used for the interpretation of spatial information.

1. Introduction

Consumers have shifted from simply buying products to the consumption of improved quality in terms of culture and value [1,2,3]. At the same time, companies have adapted to locate consumers among diverse cultures, and large complex spaces combining commercial and cultural areas, such as library and music hall, have been created. However, as spaces have expanded to contain various content and scale, the distance and the complexity of visitor circulation in these indoor environments has increased. For this reason, consumers must recognize spatial information when traversing paths using signs and maps, even in a complex cultural space. Although designers have developed and provided various forms of signs for visitors to recognize spatial information, there are limitations in understanding how visitors perceive this information and use it for wayfinding.
Data for spatial cognitive processes may be collected through questionnaires and interviews. However, to collect the sensory physiological signal data of the direct visual perception process, it is necessary to use parallel scientific experimental methods as a supplement. Questionnaires and interviews are limited because they provide subjective opinions about the spatial experience, and survey participants can consciously change answers. The scientific measurement method using physiological signals can compensate for such limitations and directly extract any sensory signals generated unconsciously [4,5].
Human visual information can be measured by using eye-tracking equipment [6,7]. To collect information about the direct visual perception process, we experimented with a virtual reality (VR)-based head mounted display (HMD) system. VR-based eye-tracking equipment allows participants to experience spaces even if they are not in the real world [8]. The experience of the virtual space as if it is real using an HMD has the advantage of increasing the immersion in space, and it is possible to quantify the user’s gaze data. There are various constraints in a real environment. There are many variables in extracting data because the real environment cannot be regulated constantly as a consequence of population congestion or external environmental factors. However, in the laboratory using a VR-based HMD, it is possible to control the variable parts while realizing the commitment to space.
Although the diffusion of VR and the expansion of the market are expected to continue, the design and the evaluation of the space through VR can be considered an initial stage. In this study, we contributed to the field by suggesting a research method that involves experiments to apply the VR environment to spatial design using visual perception. VR equipment was used to convert human visual information into the cognition of sensation, and visual searches were monitored using eye-tracking experiments with VR equipment. In addition, data analysis was proposed to track the unconscious spatial search according to the time taken for cognitive processing to occur. This is significant in proposing a method for utilizing the developing VR technology according to human sensory information in architectural space planning.

2. Literature Review

2.1. Theory of Visual Recognition Using the Eye-Tracking Mechanism

Many studies have attempted to explain human visual recognition, which occurs via the eyes in space [9,10,11]. Early research on eye tracking was technically limited to simple eye observation and eye endoscopy but has since evolved into interdisciplinary areas, including psychophysics, cognitive neuroscience, brain science, and computer science [12,13]. Such studies have tried to identify the nature of visual perception, attempting to define the gaze in terms of its focus with respect to identity, meaning, or expectation. Studying the gaze, researchers can identify what people’s interests are; the gaze can be described as the act of focusing on something more clearly in the mind than several conceivable things considered simultaneously [14]. Julesz and Schumer noted that, when the gaze is grasped by sensation as an internally invisible mechanism manifesting as imagination, expectations, or thinking, the number of perceptions that can be immediately included in the present cognitive realm becomes small. That is, visual perception can be determined depending on what humans think [15]. Some concerns have been raised as to whether the essential mechanism of visual perception concerns the “where” of eye movement or visual gazes in relation to spatial location. In a study of physiological optics, a wandering exploration of the eye was identified and observed as the human gaze or visual attention. It was remarked that observing the field of view is the only way humans can see each individual part of a space as clearly as possible. Consciousness can voluntarily control the gaze, thus humans can pay attention to their surrounding objects without distraction [16]. The human eyes provide visual evidence of the world around them, and visual attention not only requires human cognitive function but is also related to preconceived factors. The features of visual attention have been developed into an important theory and inform the visual search of feature integration theory (FIT) [17,18], which maintains that the eye tracks things in space by encoding simple and useful properties, such as position and color, orientation, size, and stereo distance.
The modeling of computer applications for eye tracking in relation to visual attention is related to a bottom-up schema of dichotomous visual interest or a function-oriented description [19,20,21]. That is, when considering image stimulation, an interest in the cognitive stage considers a specific area of the image. This area can be recognized parafoveally in the sense of initially requesting detailed inspection through foveal vision. It should be emphasized that a complete model of visual attention requires a high level of visual and cognitive functioning. Human interest in space cannot be explained simply by considering visual features, but it is required to analyze the eye-tracking mechanism along with the cognitive factors of particular interest.

2.2. Visual Attention and Cognitive Processing Time

Visual perception responses can be measured by analyzing the reactions from the acquisition of visual information to the brain and the perception of visual stimuli. The cognitive stages are different depending on the time at which the subjects receive the visual information. Time associated with visual attention and the gaze toward objects and scenes can be interpreted as a cognitive process [22,23,24,25,26]. Related studies have ascertained the meaning of the relationship between cognitive processes and time, identified the temporal difference between cognition and exploration when reading, and examined the difference in gaze fixation for the perception process of vision in image scenes or in spatial information pertaining to complex objects or in-focus stimuli over time [27,28,29,30,31]. The time required for the human brain to perceive a visual stimulus and send commands to the body is 0.1 s [32,33,34,35]. In this time, it is not possible to judge the nature of the object by paying attention to it because this is the early stage of human perception. The minimum response time is 0.2 s before the acquired visual information is activated in the brain, and this visual gaze time applies to both objects and scenes [36]. The minimum response time from the visual stimulus to the reaction through the brain is 0.2 s, and the time it takes for the acquired visual information to move the mind is 0.3 s [37,38,39]. The visual understanding of an object is developed by the data obtained in a fixed view in about 0.3 s. The analysis of the data according to the time of eye movement is necessary for the gaze analysis because of differences in human perception and cognitive processing. The relationship between cognitive processing and time has been scientifically studied and verified by many scholars, as described above. Depending on the object and the purpose of each cognitive level, there may be a difference in the time analysis. In this study, based on the hypothesis of a difference in visual movement over 0.2 s and 0.3 s, which was verified in the previous study, these intervals were used as standards for the analysis of the experimental data.

2.3. The Potential of Virtual Reality for Eye-Tracking Experiments

The use of VR in cognitive psychology research has increased with the proliferation of VR tools [40]. The VR tool has many advantages over the existing experimental methodology, as it has the potential to create effective stimulus and response protocols by controlling the experimental environment [41]. Many studies have demonstrated the effectiveness of VR for experiments, and the potential scope of research can be expanded depending on the possibilities for immersion in and controllability of the VR environment [42]. In the VR environment, safe dynamic tests for brain-computer interfaces are possible, and a controlled experimental design is also conceivable in the form of an evoked environment. To obtain examples of cognitive experiments, the usefulness of the object situation or location in the experimental environment was investigated in a user-gaze comparative study, and the advantages of exploration in 3D were evaluated as increasing cognitive response speeds by 50% or more when compared to searching in a physical space, as VR can quickly induce an omnidirectional gaze during spatial exploration [43,44,45,46]. In addition, VR experiments create a useful paradigm because they allow for experimental control in cognitive psychology research. Psychological cognition experiments can discover new methodologies using the new research tools provided by VR and extending beyond traditional methods. As for the nature of cognitive psychological phenomena, approaches can vary and include specific perception, memory, problem solving, and mental image attention. However, the immersion and the visual fidelity of the VR experimental tool do not necessarily elicit realistic psychological responses from users [47,48,49,50,51,52]. It is necessary to supplement the experimental limitations by utilizing the advantages of immersive and controllable VR environments and data according to the search and the fixation of the user’s cognitive process. In VR environmental experiments, researchers must supplement the cognitive process consciously for the experiment’s participants through the steps of the questionnaire. By recognizing spatial information and applying the experimental method of intelligibility, the cognitive stage can be identified through VR environmental verification to provide guidelines not only for the spatial environment but also for future spaces [53,54,55,56]. In addition, the immersion and the visual fidelity in the virtual space were grasped through an analysis of the visual data, and the interview data of participants’ cognitive processes were applied to verify the analysis of the visual data and the data analysis depending on the level of their recognition. This is our experimental configuration and data analysis method that verifies both existing technologies and psychological approaches to virtual reality as distinct from the approaches of related studies.

3. Materials and Methods

3.1. Visual Stimuli

The complex space connected to the subway station was constructed as an experimental setting. As visual stimuli in COEX Mall, Seoul, we included paths commencing at the entrance of Starfield and passing through the Star Hall Library to the COEX Exhibition Hall. The experiment stimuli were presented to participants as a way to recognize the paths in the immersive VR space. The experimental stages were as follows. First, to create the visual stimuli, a 360-degree camera was installed on the path from the entrance to the crossroads at the Star Hall Library. Several images were recorded, and visual stimuli were selected in consideration of the congestion, the location of the signs, and the visibility. During the preliminary investigation, it was found that the sign, the path, and the surrounding spaces were used to identify the destination at the crossroads of the experimental space. To obtain spatial information, the destination name of the sign and the direction of the arrow were found to be confusing factors in selecting the path of the space.
Second, through a pretest consisting of two participants, the field of view and its visibility to the participants were confirmed when the captured 360-degree images were applied to the VR-based HMD device (SMI-HTC vive). Third, by analyzing the precautions of the experimental steps presented to the participants of the pretest, a pre-explanation was added to improve the participants’ understanding of the experiment. Fourth, in the experiment (with a total of 23 participants—ten males and 13 females), participants each wore an HMD headset and observed visual stimuli while responding to an interview conducted by the experimenter.
A total of four visual stimuli were presented in an order and under a time limit for the experiment. The perception processes of clarity and cognition according to the area of interest in this study were analyzed using one visual stimulus. In the visual stimulus shown in Figure 1a, the participant observed the environment for 30 s at the entrance of the experimental space. The experimenter explained to the participants the purpose of the experiment and the VR environment and asked the participant to recognize the visual information of the sign first. The next stimulus, shown in Figure 1b,c, was an image that allowed the first sign to be observed closely and could be seen as the participant moved toward the crossroads. Here, the participants could observe the guide signs at the front for 30 to 60 s, allowing them to understand and adapt to the next experimental stimulus. The next stimulus in Figure 1d was found at the first crossroad to the exhibition hall; participants were given the opportunity to select the next route to the exhibition hall after observing the spatial information on the signboard for one minute and recognizing the space. Finally, while observing the last visual stimulus in Figure 1e, the participants chose an appropriate path among the numbered crossroads and provided answers to the spatial perception items in the questionnaire interview. The data extracted from the last visual stimulus were analyzed as areas of interest (AOIs) for clarity and recognition.

3.2. Experiment Procedure and Participants

The VR experiment was conducted over two days from 10 January 2019 to 11 January 2019, and the participants were 24 undergraduate or graduate students. All participants had a visual acuity or corrected visual acuity score of 0.5 or higher, with an average age of 24.8 years (standard deviation: ±1.89).
A preliminary experiment was conducted to verify the experimental setting. In the preliminary experiment, participants were asked to follow simple instructions after focus adjustment, which allowed them to adapt to the immersive VR environment before entering the main experiment (see Figure 2). The respondents observed the visual stimuli freely for 60 s, and as they had in the preliminary experiments, they explored the virtual space by freely turning their heads and bodies within the range allowed by the HMD equipment. The experimenter ended the preliminary experiment after monitoring the participants’ behavior in real time. After the completion of the preliminary experiment, participants entered the main experiment, and the time required for each participant to complete it was different; the participants took a minimum of four minutes and a maximum of five minutes and 40 s. According to the validity rate of data collection, the gaze data of 11 final selected participants were selected and analyzed.
The initial data for visual stimulus were reviewed to exclude participants whose validity rate was less than 85%. This was to increase the accuracy and the reliability of the experimental data by examining the validity rate in the process of recording the participants’ gaze data in real time. Data from 11 out of 24 participants (45.83%) were valid and were selected. The percentage of valid initial data was 69.2%, but this rate was increased to 92.9% by the inclusion of 11 participants with validity rates of higher than 85%. The extracted data were analyzed according to the validation of the effective rate.
The visual stimuli used in the experiment were photographic images taken across 360 degrees. The participants performed in the experiment by observing a fixed virtual space image. AOI is an analysis method that can observe gaze fixation within the designated area to which the participants’ gazes were directed. For the setting of AOIs and the extraction of data, the SMI BeGaze 3.6 program was used. The experimenter can designate a specific region of interest within the entire spatial image to extract gaze movements generated within that region. The change of gaze can be analyzed using the quantitative data of gaze movements, which is used to obtain information on wayfinding by mapping participants’ cognitive processes of visual perception.
The area AOIs were construed as “path”, “sign”, and “background”. “Sign” was used as a guide to obtain spatial information (see Figure 3), thus participants could receive information on the direction of the space. The AOIs used in the research analysis were identified using the data obtained by the participants looking for visual information and fixing their gazes accordingly in the VR images in adherence to the experiment’s purpose statement. This allowed the researchers to analyze the research results to reveal characteristics of spatial perception according to the intelligibility and the recognition of the “sign” within the research site.

4. Results

4.1. Average Gaze Data for Understanding Spatial Information

Time-series analysis was performed to analyze the experimental data through the recognition of spatial information. To analyze the visual information process, the time was divided into six sequences of ten-second intervals. To extract the gaze information for the AOI set (“path”, “sign”, and “background”) of visual stimuli, the number of gaze fixation data retrieved from each section was identified along with the AOIs that participants observed and the information they acquired over time.
In the interview conducted at the final stage of the experiment, participants answered questions about the clarity and the recognition of the crossroads selection process using a five-point scale (“very difficult”, “difficult”, “normal”, “easy”, and “very easy”). According to their answers, they were divided into two groups. Participants who selected “very difficult”, “difficult”, or “normal” were grouped as “difficult”, and those who chose “easy” or “very easy” were classified as “easy”. Dividing the group according to the level of spatial information recognition allowed the researchers to analyze the correlation between each group’s changing characteristics in the cognitive understanding of spatial information and the movement of gaze information according to time-series analysis.
To analyze the spatial information, the movement of the gaze was studied over time to ascertain the average fixation number (see Figure 4) and the ratio of each AOI (see Figure 5). The data for the gaze increased and decreased over time in relation to the two groups (“easy” and “difficult”), which were divided according to their propensity for spatial understanding. Overall, using the averages and ratios extracted from the AOIs, the movement by which the gaze increases and decreases over time could be analyzed by applying the correlation between the understanding of space and the cognitive process.
Figure 4 shows the gaze fixation data in the AOIs for the two groups according to the scale of their understandings within the space. We abbreviated the “easy” group’s fixation AOI sign as E-FAS, the “difficult” group’s fixation AOI sign as D-FAS, the “easy” group’s fixation AOI path as E-FAP, and the “difficult” group’s fixation AOI path as D-FAP. After the start of the experiment, the gaze fixation of AOI for the “path” of the D-FAP group was 3.5 times higher than that of the E-FAS group in the experimental Time Sequence 1 (ten seconds). From the beginning of the gaze, the ratio of the D-FAP gaze gradually decreased throughout the entire experimental period, but the E-FAP gradually increased (see Figure 4a). The AOI gaze fixation for “sign” showed initially a high average gaze fixation rate in both the E-FAS and the D-FAS groups, which then gradually decreased (see Figure 4b).
Figure 5 shows the gaze fixation ratio for AOIs over the experimental period for the “easy” and the “difficult” groups. In the “easy” group, the gaze concentration ratio for the “sign” (55%) during the first section of the experiment (0 ~ 10 s) was approximately three times higher than that for the “path” (25%) and the “background” (19%). In the third section (20 ~ 30 s), which comprised the middle of the experiment, the “easy” group focused on the “path” rather than the “sign” in the order of “path” (52%), “background” (30%), and then “sign” (18%), unlike at the start of the experiment. After this, their attention was focused on the “path” rather than the other AOI areas, which shows that the gaze was fixed in accordance with the purpose of the experiment (see Figure 5a). During the first section of the experiment (0 ~ 10 s), the gaze fixation in the “difficult” group occurred in the order of “path” (40%), “sign” (36%), and “background” (23%), and the difference was less pronounced. In the second section of the experiment (10 ~ 20 s), the gaze fixation was higher for “path” (35%) and “background” (39%) than for “sign” (26%). In the middle of the experiment, a high gaze fixation ratio was shown for “path’” (50%). In the final section of the experiment, gaze fixation was also observed in the AOIs rather than the direction for wayfinding in the order of “path” (54%), “sign” (33%), and “background” (12%) (see Figure 5b).
In the interview during the final stage of the experiment, the groups were divided according to the difficulty experienced when perceiving space, and the extracted gaze fixation data were analyzed. As a result, differences in the capacity to understand the spatial information appeared according to the ratio of gaze fixation. The cognitive processing of information can be analyzed according to the time intervals of the experiment. In the experiment, the “easy” group’s understanding of spatial information was obtained by perceiving the “sign”, and their gaze was fixed on the wayfinding of the space for the purpose of the experiment. However, in the “difficult” group, the gaze fixation data appeared scattered in spaces other than the perception of the “sign” and the wayfinding.

4.2. Analysis of Conscious Gaze and the Visual Understanding of Data

To analyze the data of perceptual process, a time range of one visual movement was considered. The VR gaze tracking equipment recorded participant gaze data at 250 Hz per second, and the gaze time for each datum was 0.004 s—that is, one visual movement per 0.004 s was recorded and extracted at 250 Hz.
As seen in previous studies, the minimum time to consciously grasp an object in front of the eyes is 200 ms; it is understood that eyes perceive an object and consciously observe it when the gaze is fixed in one place. Therefore, to ensure a fixed gaze time of 0.2 s or more, at least 50 of the gaze data obtained in this experiment had to be continuous. However, less than 50 of the non-contiguous fixation data were also included in the gaze data recorded at 250 Hz. Using the above results, the raw gaze fixation data from the AOIs were analyzed by dividing the total experiment time (60 s) into six ten-second sections (see Figure 6). All fixation data were recorded regardless of dwelling time. The visual perception process was analyzed by extracting 50 or more 200 ms continuous gaze fixation data for “conscious attention” and 300 ms continuous gaze fixation data for “visual understanding” (see Figure 7).
The data for the participants’ gaze movement that passed the validity rate test indicates the value of gaze fixation in the AOI areas. The AOIs were classified into FAP, FAS, and FAB. Each AOI region allowed for the extraction of gaze data from the FAS to understand the spatial information as well as gaze data from the FAP to understand the direction of wayfinding within a target space. FAB accounted for the remaining areas beyond FAP and FAS and was interpreted as searching gaze. In Time Sequence 1 (0 ~ 10 s) shown in Figure 6, the FAS value for gaze fixation was the highest, thus a conscious gaze fixation in the cognition of spatial information appeared. In Time Sequence 3 (20 ~ 30 s), the value of the FAS was the lowest, indicating that the search by the cognitive gaze for the “sign” was over at the beginning of the experiment. As the FAP value increased in Time Sequence 3, it continued until the end of the experiment. When analyzing the participants’ perception of information about the space, it is understood participants used their gazes to search for 10 to 20 s at the beginning of the experiment, and then, after 30 s, with a greater understanding of the space, they proceeded to observe the direction of the destination. In the gaze movement analysis, there was a difference in the visual fixation data of the AOIs according to the time sequence (see Figure 6a).
A frequency of 200 ms when 50 or more gazes were fixed in succession was extracted from the data and analyzed as “conscious gaze” data. Figure 6c shows the analysis of more than 75 gaze fixation data at 300 ms as a “cognitive gaze” beyond consciousness (see Figure 6b). The 200 ms FAS (200-FAS) appears to have had a higher frequency than the other AOIs in Time Sequence 1 (0 ~ 10 s). Time Sequence 2 (10 ~ 20 s) revealed a decrease in raw data after the commencement of the experiment; however, it is understood that the “sign” area was perceived more consciously than the other AOI areas. The 300 ms FAP (300-FAP) increased in frequency from Time Sequence 2 (10 ~ 20 s) to Time Sequence 6 (50 ~ 60 s). It is conjectured that the spatial information was recognized in the “sign” during the initial Time Sequence 1.

4.3. Relative Proportions of the Conscious Gaze and Visual Understanding

In the relative ratio analysis, each change in participant gaze movement was noted during the perception and cognitive processing of the virtual space. The raw data, the visual perception movement of the conscious gaze at 200 ms and the visual perception at 300 ms were analyzed using visual steps as the gaze movements according to the relative ratio values of each AOI (see Figure 7). Eye-movement data, for which the visual perception and cognitive stages were relative, were analyzed in terms of the stimulus and position of the visual perception in each time sequence. Analyzing the rate at which the highest level of perception was recorded for each time step shows conscious gaze for 200-FAS to be relatively high and increasing in contrast to that of the 200-FAP and the 200-FAB. Examining the detailed time sequence, the 200-FAS shows a lower rate than the 200-FAP in Steps 3 and 5, but the 300-FAS also exhibits a low rate in Steps 2 and 5. By the gaze fixation data analysis (see Figure 6), it was confirmed that the visual attention related to the stimuli increased with the FAP. By analyzing the ratio, it is possible to define the meaning of visual attention by considering where the gaze dwelled.

4.4. Analysis of the Perception and Understanding of Spatial Information According to Cognitive Differences

The data extracted from the visual exploration of the “easy” and the “difficult” groups were divided into the perception and the understanding of spatial information through the analysis of AOIs (see Figure 8). The stages of visual perception and exploration of space were investigated by analyzing the “easy” group’s perceptions (see Figure 8a) and understandings (see Figure 8c) as well as the “difficult” group’s perceptions (see Figure 8b) and understandings (see Figure 8d). The perception processes of the “easy” group were analyzed to be high for the “sign” (E-200-FAS), which was consistently searched for throughout the areas of the AOIs. However, the perception processes of the “difficult” group were analyzed to be either frequent or infrequent in terms of the gaze rate for the “sign” (D-200-FAS) relative to the areas of AOIs and depending on the time sequence. In the process of understanding the space at 300 ms, the E-FAS group continued to seek to comprehend the space, as the gaze was observed in the “sign” area at this time point. However, the “difficult” group was observed to stay focused on the “path”, except for viewing the “sign” in the initial time step. Revealing the conscious difficulty response to spatial information, differences were witnessed between the E-FAS group and the D-FAP group in the visual scan path pattern throughout the experimental stages.

5. Discussion and Conclusion

Visual perception in the process of spatial experience was measured using various research methods. The visual perception process can be divided into conscious perception and unconscious perception, and the gaze movement can be classified according to James [14] “what you see” or Von Helmhotlz [16] “where you see”. This research focused on “how” and “what” is seen as the gaze changes over time, focusing on the levels of perception and cognition throughout the VR experiment.
Extracting the meaning of gaze fixation data
The processes of perception and cognition in information understanding were defined according to the time of the gaze fixation, and analysis was conducted accordingly. In the study of eye tracking involving statistical analysis according to the number of eye fixations and the differences in frequency, it cannot be established that people actually perceived the objects because they saw something. This study sought to ascertain whether the length of time at which a gaze is fixed can be regarded as the actual seeing of an object and to grasp the data that are meaningful to the process of visual perception. In the study of eye tracking, 0.1 s is interpreted as temporary eye fixation at a low level of spatial perception. Data above 0.3 s are analyzed as indicating visual understanding at the cognition stage. The interpretation of meaning may vary depending on how long the conscious and the unconscious gaze dwells rather than interpreting all data on gaze fixation as “seeing”.
When comparing the average frequency of the gaze fixation data according to the ratio of the total AOIs, the “difficult” group exhibited greater gaze fixation than the “easy” group, and this trend continued over time (see Figure 4). However, examining the gaze fixation according to the ratio among the AOIs, the “easy” group focused on the “sign” at the beginning of the experiment, and the ratio of gaze fixation increased after the middle of the experiment in other areas (see Figure 5). That is, according to the degrees of perception indicated by the groups, it is possible to understand the process of directional gazing and the comprehension of spatial information through the ratio analysis of the gaze data, which remains unapparent in the frequency of the raw data for the gaze. Data interpretation and extraction methods according to the research purpose are emphasized to reveal information that is unknown within the number of gaze scan paths and fixation points.
Processing eye-tracking data according to time series
A time-series analysis of data or a research method of data processing according to frequency and ratio is required to understand the stages at which the levels of consciousness and unconsciousness change in the processes of visual perception. In particular, in the study of spatial information, it is important to understand the flow of visual attention and information acquisition over time to identify the machinations of perception and cognition. When a cognitive process for wayfinding occurs as a result given information about a large and complex space, it is necessary to extract the data about when, what, and how to understand and make decisions. In this study, the gaze movement was tracked according to a time series, revealing the processes of perception and cognition using eye tracking in the VR environment.
This study sought to identify when and where to see and understand according to the purpose of wayfinding in response to spatial information. As a result, the characteristics concerning whether or not the spatial information was understood appeared in the classification of the participants, and the value of the gaze data according to the time flow was significantly different accordingly. In this study, signs affecting the processes of perception and cognition in the research of the nature of visual perception were validated, thus the thoughts and the conscious focus of participants appeared as gaze movements.
At 200 ms, defined as the recognition level of the gaze, the ratios of AOIs were evenly distributed over time for the “easy” group; however, a large deviation in the field of the “sign” was revealed for the “difficult” group (see Figure 8). It was interpreted that the “easy” group searched for “where to look” by determining the hierarchy of the gaze recognition area, whereas the “difficult” group was unable to locate this hierarchy of recognition. At 300 ms, defined as the understanding level of the gaze, it was shown that the “easy” group maintained intentional consciousness to continuously understand throughout the experiment, directing their gazes toward the “sign” for the entire experiment. However, the gazes of the “difficult” group were interpreted as searching the surrounding area rather than understanding the “sign” from the middle of the experiment. Accordingly, the tracking of eye movements is an invisible psychological mechanism for spatial information, which may be interpreted and verified.
Enhancing immersion using a VR eye-tracking experiment
There are many limitations to such experimentation with the consumer’s reaction to the spatial environment. However, an immersive experiment setting was created through the effectiveness of the VR technology, thus it was possible to quickly record the participants’ reactions to the spatial environment. These strengths of the VR experiment, such as immersion and control, will not apply to all studies. In this study, research on the control of the VR experiment was conducted in advance and found a limit to the participants’ immersion in the VR experiment over time. Thus, to enhance participants’ immersion for the purposes of the experiment, visual stimuli were presented for exploring the space, and cognitive psychological phenomena were examined through interviews. In the resulting analysis, the participants were divided into groups according to the degree of their cognition responses, and the data were classified and analyzed accordingly.
In conclusion, it was identified that visual information correlates with human perception. The data of the gaze can be interpreted as indicating a cognitive stage of the exploration and understanding of visual perception and may be judged as a useful mechanism for psychological decision making. Because of the attributes of the gaze data, researcher interpretation is essential in extracting meaning and interpreting “what”, “where”, and “how much” a person looks and comprehends. Depending on what purpose statement is given over the course of the VR experiment, the extraction and the analysis of the gaze data obtained may vary. Even in an experimental environment implemented with an efficient VR technique, quantitative data extraction and interpretation must be conducted in combination with a qualitative research method. Most of the subjects in this study were in their early twenties, since the study conducted VR experiments as a verification of gaze perception and cognition. It would be desirable to conduct another experiment with a mixed aged group, as data collected according to age may be expected to reveal interesting results concerning sensory perception and cognition.

Author Contributions

Funding acquisition, J.Y.K.; Investigation, J.Y.K.; Methodology, M.J.K.; Supervision, M.J.K.; Writing—original draft, J.Y.K.; Writing—review & editing, M.J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea Grant funded by the Korean Government (Grant no.2017R1A2B2007276).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sheth, J.N.; Mittal, B.; Newman, B.I. Consumer Behavior and Beyond; The Dryden Press: New York, NY, USA, 1999. [Google Scholar]
  2. Sojka, J.; Tansuhaj, P.S. Cross-cultural consumer research: A twenty-year review. In North American- Advances in Consumer Research 22; UT: Association for Consumer Research: Minneapolis, MN, USA, 1995; pp. 461–474. [Google Scholar]
  3. Venkatesh, A. Ethnoconsumerism: A new paradigm to study cultural and cross-cultural consumer behavior. In Marketing in a Multicultural World; SAGE Publications: Southend oaks, CA, USA, 1995; pp. 26–67. [Google Scholar]
  4. Kim, J.; Kim, J. Method Extracting Observation Data by Spatial Factor for Analysis of Selective Attention of Vision. Korean J. Sci. Emot. Sensib. 2015, 18, 3–14. [Google Scholar] [CrossRef]
  5. Kim, J. An Analyzed the Area of Interest based on the Visiting Intention and Existence of People in Cafe Space. J. Korean Inst. Inter. Des. 2016, 25, 130–139. [Google Scholar] [CrossRef] [Green Version]
  6. Bozomitu, R.G.; Pasarica, A.; Tarniceriu, D.; Rotariu, C. Development of an Eye Tracking-Based Human-Computer Interface for Real-Time Applications. Sensors 2019, 19. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Kim, J.; Kim, J. The Initial Value of Extraction for Visual Perception in VR(Virtual Reality) Eye-tracking Experiments. J. Korean Inst. Inter. Des. 2019, 28, 84–94. [Google Scholar] [CrossRef]
  8. Bodden, V. Virtual-Reality Headsets; Checkerboard Library, an imprint of Abdo Publishing: Minneapolis, MN, USA, 2018; p. 32. [Google Scholar]
  9. Madary, M. Visual Phenomenology; MIT Press: Cambridge, MA, USA, 2017. [Google Scholar]
  10. Palmer, S.E. Vision Science: Photons to Phenomenology; MIT Press: Cambridge, MA, USA, 1999. [Google Scholar]
  11. Siegel, S. How does visual phenomenology constrain object-seeing? Australas. J. Philos. 2006, 84, 429–441. [Google Scholar] [CrossRef]
  12. Duchowski, A.T. Eye tracking methodology. Theory Pract. 2007, 328, 2–3. [Google Scholar]
  13. Zeki, S. A Vision of the Brain; Blackwell scientific publications: Hoboken, NJ, USA, 1993. [Google Scholar]
  14. James, W.; Burkhardt, F.; Bowers, F.; Skrupskelis, I.K. The Principles of Psychology; Macmillan London: London, UK, 1890; Volume 1. [Google Scholar]
  15. James, W. Principles of Psychology; Harvard University Press (trad it. Principi di psicologia, Principato, Milano, 1983): Cambridge, MA, USA, 1890. [Google Scholar]
  16. Von Helmholtz, H. Treatise on Physiological Optics: Translated from the 3rd German Ed; Optical Society of America: Washington, DC, USA, 1925. [Google Scholar]
  17. Posner, M.I.; Snyder, C.R.; Davidson, B.J. Attention and the detection of signals. J. Exp. Psychol. Gen. 1980, 109, 160. [Google Scholar] [CrossRef]
  18. Sparks, D.L.; Mays, L.E. Signal transformations required for the generation of saccadic eye movements. Annu. Rev. Neurosci. 1990, 13, 309–336. [Google Scholar] [CrossRef]
  19. Bundesen, C. A theory of visual attention. Psychol. Rev. 1990, 97, 523. [Google Scholar] [CrossRef]
  20. Connor, C.E.; Egeth, H.E.; Yantis, S. Visual attention: Bottom-up versus top-down. Curr. Biol. 2004, 14, R850–R852. [Google Scholar] [CrossRef] [Green Version]
  21. Itti, L. Models of Bottom-up and Top-down Visual Attention; California Institute of Technology: Pasadena, CA, USA, 2000. [Google Scholar]
  22. Bryant, J.S. Feature detection process in speech perception. J. Exp. Psychol. Hum. Percept. Perform. 1978, 4, 610–620. [Google Scholar] [CrossRef] [PubMed]
  23. Fazio, R.H.; Herr, P.M.; Olney, T.J. Attitude accessibility following a self-perception process. J. Personal. Soc. Psychol. 1984, 47, 277–286. [Google Scholar] [CrossRef]
  24. Schag, K.; Rauch-Schmidt, M.; Wernz, F.; Zipfel, S.; Batra, A.; Giel, K.E. Transdiagnostic Investigation of Impulsivity in Alcohol Use Disorder and Binge Eating Disorder with Eye-Tracking Methodology-A Pilot Study. Front. Psychiatry 2019, 10, 724. [Google Scholar] [CrossRef] [PubMed]
  25. Allan, L.G. The perception of time. Percept. Psychophys. 1979, 26, 340–354. [Google Scholar] [CrossRef] [Green Version]
  26. Aubry, F.; Guillaume, N.; Mogicato, G.; Bergeret, L.; Celsis, P. Stimulus complexity and prospective timing: Clues for a parallel process model of time perception. Acta Psychol. 2008, 128, 63–74. [Google Scholar] [CrossRef]
  27. Gibson, W.C.; Herron, W.G. Psychotherapists’ religious beliefs and their perception of the psychotherapy process. Psychol. Rep. 1990, 66, 3–9. [Google Scholar] [CrossRef]
  28. Rayner, K. Eye movements and attention in reading, scene perception, and visual search. Q. J. Exp. Psychol. 2009, 62, 1457–1506. [Google Scholar] [CrossRef]
  29. Rayner, K. Eye movements in reading and information processing: 20 years of research. Psychol. Bull. 1998, 124, 372. [Google Scholar] [CrossRef]
  30. Chua, H.F.; Boland, J.E.; Nisbett, R.E. Cultural variation in eye movements during scene perception. Proc. Natl. Acad. Sci. USA 2005, 102, 12629–12633. [Google Scholar] [CrossRef] [Green Version]
  31. Rayner, K.; Pollatsek, A. Eye movements and scene perception. Can. J. Psychol. Rev. Can. Psychol. 1992, 46, 342. [Google Scholar] [CrossRef]
  32. Broadbent, D.E. Perception and Communication; Elsevier: Amsterdam, The Netherlands, 2013. [Google Scholar]
  33. Blamey, P.J.; Cowan, R.S.; Alcantara, J.I.; Whitford, L.A.; Clark, G.M. Speech perception using combinations of auditory, visual, and tactile information. J. Rehabil. Res. Dev. 1989, 26, 15–24. [Google Scholar] [PubMed]
  34. Lee, H.J.; Lee, J.; Kim, C.J.; Kim, G.J.; Kim, E.S.; Whang, M. Brain process for perception of the “out of the body” tactile illusion for virtual object interaction. Sensors 2015, 15, 7913–7932. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Julesz, B.; Schumer, R.A. Early visual perception. Annu. Rev. Psychol. 1981, 32, 575–627. [Google Scholar] [CrossRef] [PubMed]
  36. Abe, M.; Ueki, T.; Saeki, Y.; Tateoka, S.; Tomita, M. When and how of assistance in nursing. Report 1. Perception of the psychological process of the patient. Kangogaku Zasshi 1977, 41, 241–246. [Google Scholar]
  37. Brown, J. On time perception in visual movement fields. Psychol. Forsch. 1931, 14, 233–248. [Google Scholar] [CrossRef]
  38. Solso, R.L.; MacLin, M.K.; MacLin, O.H. Cognitive Psychology; Pearson Education New Zealand: Auckland, New Zealand, 2005. [Google Scholar]
  39. Davidoff, J. Differences in Visual Perception: The Individual Eye; Elsevier: Amsterdam, The Netherlands, 2012. [Google Scholar]
  40. Gaggioli, A. Using virtual reality in experimental psychology. In Towards Cyberpsychology; IOS Press: Amsterdam, The Netehrlands, 2001; pp. 157–174. [Google Scholar]
  41. Wilson, C.J.; Soranzo, A. The use of virtual reality in psychology: A case study in visual perception. Comput. Math. Methods Med. 2015, 1–5. [Google Scholar] [CrossRef] [Green Version]
  42. Bayliss, J.D.; Ballard, D.H. Recognizing evoked potentials in a virtual environment. In Advances in Neural Information Processing Systems; MIT Press: Vancouver, BC, Canada, 2000; pp. 3–9. [Google Scholar]
  43. Biocca, F.; Owen, C.; Tang, A.; Bohil, C. Attention issues in spatial information systems: Directing mobile users’ visual attention using augmented reality. J. Manag. Inf. Syst. 2007, 23, 163–184. [Google Scholar] [CrossRef]
  44. Checcucci, E.; Amparore, D.; Pecoraro, A.; Peretti, D.; Aimar, R.; De Cillis, S.; Piramide, F.; Volpi, G.; Piazzolla, P.; Manfrin, D.; et al. 3D mixed reality holograms for preoperative surgical planning of nephron-sparing surgery: Evaluation of surgeons’ perception. Minerva Urol. Nefrol. 2019. [Google Scholar] [CrossRef]
  45. Giroux, M.; Barra, J.; Barraud, P.A.; Graff, C.; Guerraz, M. From Embodiment of a Point-Light Display in Virtual Reality to Perception of One’s Own Movements. Neuroscience 2019, 416, 30–40. [Google Scholar] [CrossRef]
  46. Sapkaroski, D.; Mundy, M.; Dimmock, M.R. Virtual reality versus conventional clinical role-play for radiographic positioning training: A students’ perception study. Radiography 2020, 26, 57–62. [Google Scholar] [CrossRef] [Green Version]
  47. Akizuki, H.; Uno, A.; Arai, K.; Morioka, S.; Ohyama, S.; Nishiike, S.; Tamura, K.; Takeda, N. Effects of immersion in virtual reality on postural control. Neurosci. Lett. 2005, 379, 23–26. [Google Scholar] [CrossRef] [PubMed]
  48. Bowman, D.A.; McMahan, R.P. Virtual reality: How much immersion is enough? Computer 2007, 40, 36–43. [Google Scholar] [CrossRef]
  49. McMahan, R.P. Exploring the Effects of Higher-Fidelity Display and Interaction for Virtual Reality Games; Virginia Tech: Blacksburg, VA, USA, 2011. [Google Scholar]
  50. McMahan, R.P.; Bowman, D.A.; Zielinski, D.J.; Brady, R.B. Evaluating display fidelity and interaction fidelity in a virtual reality game. Ieee Trans. Vis. Comput. Graph. 2012, 18, 626–633. [Google Scholar] [CrossRef] [PubMed]
  51. McMahan, R.P.; Lai, C.; Pal, S.K. Interaction fidelity: The uncanny valley of virtual reality interactions. In International Conference on Virtual, Augmented and Mixed Reality; Springer: Berlin/Heidelberg, Germany, 2016; pp. 59–70. [Google Scholar]
  52. Plante, T.G.; Cage, C.; Clements, S.; Stover, A. Psychological benefits of exercise paired with virtual reality: Outdoor exercise energizes whereas indoor virtual exercise relaxes. Int. J. Stress Manag. 2006, 13, 108. [Google Scholar] [CrossRef]
  53. Vatavu, R.D.; Pentiuc, Ş.G.; Chaillou, C.; Grisoni, L.; Degrande, S. Visual recognition of hand postures for interacting with virtual environments. Adv. Electr. Comput. Eng. 2006, 6, 55–58. [Google Scholar]
  54. Wallet, G.; Sauzéon, H.; Pala, P.A.; Larrue, F.; Zheng, X.; N’Kaoua, B. Virtual/real transfer of spatial knowledge: Benefit from visual fidelity provided in a virtual environment and impact of active navigation. Cyberpsychol. Behav. Soc. Netw. 2011, 14, 417–423. [Google Scholar] [CrossRef]
  55. Hahm, J.; Lee, K.; Lim, S.L.; Kim, S.Y.; Kim, H.T.; Lee, J.H. Effects of active navigation on object recognition in virtual environments. Cyberpsychol. Behav. 2006, 10, 305–308. [Google Scholar] [CrossRef]
  56. Conroy-Dalton, R. Is spatial intelligibility critical to the design of largescale virtual environments? Int. J. Des. Comput. 2002, 4-19. [Google Scholar]
Figure 1. Visual stimuli for experimental procedures with a map of the experimental site. The stimulus images in the diagram are flattened, 360-degree images to indicate the experiment’s setting. (a) As a focus adjustment step for extracting the gaze data of each subject, subjects performed the focus data test in a 360-degree space. The images (b) and (c) show the subjects’ exploration of the surrounding environment, while the researcher explained the visual stimuli to the subjects, including signs and crossroads in the indoor space, via a “presentation of purpose statement”. (d) As a step in extracting the gaze data, the participants viewed the signs and searched the space according to the purpose statement. (e) The researchers interviewed the participants about how they understood the space in the process of exploring it.
Figure 1. Visual stimuli for experimental procedures with a map of the experimental site. The stimulus images in the diagram are flattened, 360-degree images to indicate the experiment’s setting. (a) As a focus adjustment step for extracting the gaze data of each subject, subjects performed the focus data test in a 360-degree space. The images (b) and (c) show the subjects’ exploration of the surrounding environment, while the researcher explained the visual stimuli to the subjects, including signs and crossroads in the indoor space, via a “presentation of purpose statement”. (d) As a step in extracting the gaze data, the participants viewed the signs and searched the space according to the purpose statement. (e) The researchers interviewed the participants about how they understood the space in the process of exploring it.
Applsci 10 03461 g001
Figure 2. Experimental steps and visualization of the laboratory layout. (a) Layout of the virtual reality (VR) experiment. (b) The experimenter helps the participant wear the head mounted display (HMD) and adjusts its focus according to the nine points. (c) The experiment proceeds according to the visual stimulus stages. (d) After the VR experiment, the participant answer questions about the spatial information and their perceived consciousness.
Figure 2. Experimental steps and visualization of the laboratory layout. (a) Layout of the virtual reality (VR) experiment. (b) The experimenter helps the participant wear the head mounted display (HMD) and adjusts its focus according to the nine points. (c) The experiment proceeds according to the visual stimulus stages. (d) After the VR experiment, the participant answer questions about the spatial information and their perceived consciousness.
Applsci 10 03461 g002
Figure 3. Areas of interest (AOIs) positioned in visual stimuli. (a) By setting the regions of interest in visual stimuli, the participants’ gaze data that were created while searching for directions according to spatial information in the experiment were extracted. In the visual stimulus, the area excluding “sign” and “path” was set as the background. (b) A visual stimulus shown when a participant wore an HMD and explored the VR environment.
Figure 3. Areas of interest (AOIs) positioned in visual stimuli. (a) By setting the regions of interest in visual stimuli, the participants’ gaze data that were created while searching for directions according to spatial information in the experiment were extracted. In the visual stimulus, the area excluding “sign” and “path” was set as the background. (b) A visual stimulus shown when a participant wore an HMD and explored the VR environment.
Applsci 10 03461 g003
Figure 4. The average fixation periods for the AOIs according to the two groups. (a) Average fixation of AOI paths in the “easy” group (E-FAP) and the “difficult” group (D-FAP); average fixation gradually decreased along the “easy” group’s “path” but increased for the “difficult” group. (b) The average number of fixed appearances of the “sign” over time in E-FAS and D-FAS; the average fixation decreased similarly for E-FAS and D-FAS.
Figure 4. The average fixation periods for the AOIs according to the two groups. (a) Average fixation of AOI paths in the “easy” group (E-FAP) and the “difficult” group (D-FAP); average fixation gradually decreased along the “easy” group’s “path” but increased for the “difficult” group. (b) The average number of fixed appearances of the “sign” over time in E-FAS and D-FAS; the average fixation decreased similarly for E-FAS and D-FAS.
Applsci 10 03461 g004
Figure 5. The extraction of the proportions of each AOI (path, sign, and background) from the entire area according to two groups. (a) The “easy” group’s gaze fixation on the AOI path (E-FAP), sign (E-FAS), and background (E-FAB) according to the time interval. (b) The “difficult” group’s gaze fixation on the AOI path (D-FAP), sign (D-FAS), and background (D-FAB) according to the time interval.
Figure 5. The extraction of the proportions of each AOI (path, sign, and background) from the entire area according to two groups. (a) The “easy” group’s gaze fixation on the AOI path (E-FAP), sign (E-FAS), and background (E-FAB) according to the time interval. (b) The “difficult” group’s gaze fixation on the AOI path (D-FAP), sign (D-FAS), and background (D-FAB) according to the time interval.
Applsci 10 03461 g005
Figure 6. Fixation data analysis in the AOIs from the time series. (a) Participants’ data from the AOIs. (b) 200 ms of visual fixation frequency; (c) 300 ms of visual fixation frequency.
Figure 6. Fixation data analysis in the AOIs from the time series. (a) Participants’ data from the AOIs. (b) 200 ms of visual fixation frequency; (c) 300 ms of visual fixation frequency.
Applsci 10 03461 g006
Figure 7. Relative values for eye-movement fixation in AOIs. (a) Conscious attention at 200 ms. (b) Visual understanding at 300 ms.
Figure 7. Relative values for eye-movement fixation in AOIs. (a) Conscious attention at 200 ms. (b) Visual understanding at 300 ms.
Applsci 10 03461 g007
Figure 8. Analysis of visual perception and understanding in AOIs according to cognitive differences. (a) Visual data for 200 ms in the AOIs for the “easy” group. (b) Visual data at 200 ms in the AOIs for the “difficult” group. (c) Visual data at 300 ms in the AOIs for the “easy” group. (d) Visual data at 300 ms in the AOIs for the “difficult” group.
Figure 8. Analysis of visual perception and understanding in AOIs according to cognitive differences. (a) Visual data for 200 ms in the AOIs for the “easy” group. (b) Visual data at 200 ms in the AOIs for the “difficult” group. (c) Visual data at 300 ms in the AOIs for the “easy” group. (d) Visual data at 300 ms in the AOIs for the “difficult” group.
Applsci 10 03461 g008

Share and Cite

MDPI and ACS Style

Kim, J.Y.; Kim, M.J. Exploring Visual Perceptions of Spatial Information for Wayfinding in Virtual Reality Environments. Appl. Sci. 2020, 10, 3461. https://doi.org/10.3390/app10103461

AMA Style

Kim JY, Kim MJ. Exploring Visual Perceptions of Spatial Information for Wayfinding in Virtual Reality Environments. Applied Sciences. 2020; 10(10):3461. https://doi.org/10.3390/app10103461

Chicago/Turabian Style

Kim, Ju Yeon, and Mi Jeong Kim. 2020. "Exploring Visual Perceptions of Spatial Information for Wayfinding in Virtual Reality Environments" Applied Sciences 10, no. 10: 3461. https://doi.org/10.3390/app10103461

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop