Next Article in Journal
Optimal Design of Electrothermal Microactuators for Many Criteria by Means of an Immune Game Theory Multiobjective Algorithm
Previous Article in Journal
Woc-Bots: An Agent-Based Approach to Decision-Making
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Effects of Augmented Reality Interaction Techniques on Egocentric Distance Estimation Accuracy

by
Chiuhsiang Joe Lin
1,*,
Dino Caesaron
1,2 and
Bereket Haile Woldegiorgis
1
1
Department of Industrial Management, National Taiwan University of Science and Technology, Taipei City 10607, Taiwan
2
Department of Industrial Engineering, School of Industrial Engineering, Telkom University, Bandung 40257, Indonesia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(21), 4652; https://doi.org/10.3390/app9214652
Submission received: 27 September 2019 / Revised: 23 October 2019 / Accepted: 30 October 2019 / Published: 1 November 2019

Abstract

:
Recent developments in virtual environment applications allow users to interact with three-dimensional (3D) objects in virtual environments. As interaction with 3D objects in virtual environments becomes more established, it is important to investigate user performance with such interaction techniques within a specific task. This study investigated two interaction modes, direct and indirect, depending on how the users interacted with the 3D objects, by measuring the accuracy of egocentric distance estimation in a stereoscopic environment. Fourteen participants were recruited to perform an acquisition task with both direct pointing and indirect cursor techniques at three egocentric distances and three task difficulty levels. The accuracy of the egocentric distance estimation, throughput, and task completion time were analyzed for each interaction technique. The indirect cursor technique was found to be more accurate than the direct pointing one. On the other hand, a higher throughput was observed with the direct pointing technique than with the indirect cursor technique. However, there were no significant differences in task completion time between the two interaction techniques. The results also showed accuracy to be higher at the greatest distance (150 cm from the participant) than at the closer distances of 90 cm and 120 cm. Furthermore, the difficulty of the task also significantly affected the accuracy, with accuracy lower in the highest difficulty condition than in the medium and low difficulty conditions. The findings of this study contribute to the understanding of user-interaction techniques in a stereoscopic environment. Furthermore, developers of virtual environments may refer to these findings in designing effective user interactions, especially those in which performance relies on accuracy.

1. Introduction

Virtual environment (VE) applications are rapidly increasing. Since the early 1990s, such applications have been developed in diverse domains, such as surgery [1], safety training for cabin safety procedures [2], advanced manufacturing systems integrating robots and humans to execute tasks [3], and usability evaluations [4]. In addition to being applicable in more extensive areas, VEs have recently received attention from computing technology providers such as Microsoft, Google, NVIDIA, HTC, and Samsung—all of which are seeking even more exciting applications. Many recent efforts have focused on applications that allow users to observe and interact in three-dimensional (3D) environments rather than just viewing 3D modeled images. In the development of these applications, immersive virtual environments (IVEs) featuring head-mounted displays (HMDs) are mostly used. However, such systems usually generate inconveniences such as musculoskeletal system discomfort, nausea, and difficulty with concentrating because of vertigo caused by simulated sickness [5]. Stereoscopic 3D displays (S3Ds) have recently gained more popularity. An S3D system needs a relatively lower investment in the hardware, software, and imagery produced than those with HMDs do. VR displays present slightly different images separately to each eye of the viewer. The separation between the two images, which can create slightly different views of a scene for each eye is called parallax. This phenomenon allows the viewer to perceive depth in a scene using different information [6]. Stereoscopic objects may be displayed with zero, negative, or positive parallax, corresponding to at the surface of, in front of, or behind the screen display. One technology that utilizes S3D is the 3D projection display. Unlike HMD-based systems, projection-based displays allow multiple users to view images in VE at the same time. Another advantage of projection displays is the satisfying interaction of augmentation [7], which allows users to interact with virtual images using real objects in a seamless manner. Nevertheless, integrating physical objects may lead to obscuration issues [8].

1.1. Direct vs. Indirect Interaction Methods

Since many applications in VEs require selecting a target object among other objects, prior studies have applied an acquisition task as their basis of study. To acquire the object (target), the user performs an action to position the selection tool (e.g., a finger, cursor) over the target. A tracking system also needs to acquire all the necessary information from the user through his/her actions or other movements [9,10]. Lin and Woldegiorgis [11] investigated the performance of a direct pointing technique, wherein participants directly moved a hand to reach for a real/virtual target shown continuously in a VE with projection display. Similar studies by Bruder et al. [12] and Swan et al. [13] employed direct-reaching by hand as a reporting method. In Bruder, Steinicke, and Sturzlinger [12], a comparison of interaction techniques with direct 3D mid-air interaction and a direct-2D touch screen found no significant difference in error rates in a selection task performed in a stereoscopic multi-touch tabletop setup. In contrast, Swan, Singh, and Ellis [13] conducted a comparative study of direct-matching and blind reaching in estimating the position of real/virtual objects in the stereoscopic display, and they found that perceptual direct-matching was more accurate than blind reaching in estimating egocentric distance. Overall, distances have been underestimated by up to 4 cm in blind reaching, but the disparity was reduced to 2.6 cm in perceptual direct matching. In our study, these techniques (direct pointing, reaching, and direct matching) are considered direct interaction methods.
Unlike direct interaction methods, which utilize the human body, including gestures and gaze direction, to complete a task [14], indirect interaction methods allow the user to observe representations in a display. The user controls the representation or an icon to perform a specific task. The icon can represent a virtual hand, a virtual cursor, or anything imaginable that can be used as a virtual control [14]. The user directly controls an intermediary, such as a mouse, gamepad, stylus or other physical controls, to manipulate the icons. Previous studies have applied virtual representations to perform a task in IVEs. Poupyrev and Ichikawa [15] presented the possibility of an interaction metaphor of a virtual pointer and a virtual hand for object selection and repositioning tasks in IVE. In their study, the participants interacted with objects by ray casting pointing technique, wherein when both the virtual pointer and the virtual hand intersected with an object, the object could be manipulated (selected or repositioned). The comparison showed that the virtual pointer provided more accuracy in a selection of objects within the user’s reaching distance than did the virtual hand. Werkhoven and Groen [16] evaluated manipulation (positioning, grasping) performance with 3D controllers (i.e., virtual hand control and 3D cursor control) in a stereoscopic environment. The virtual hand control was set such that a participant manipulated a virtual object by contacting a virtual hand created by wearing a DataGlove on the right hand. The position and orientation of the glove were tracked in real time while the participant was manipulating the object. In the 3D cursor control setting, the participant controlled a SpaceMouse to move the 3D cursor (i.e., a virtual arrow) in VE. The results showed that in positioning tasks, virtual hand control was more accurate than the 3D cursor. A recent study by Deng et al. [17] asked participants to position a ball-shaped object in a spherical area in a virtual space using a handheld controller in a ray casting technique setup. This technique allows a virtual light ray emitted from the controller to move an object remotely in a VE. Based on our definition of how users interact with objects, we considered those settings in indirect interaction methods.

1.2. Related Work

As user interactions with virtual objects become more popular, it is crucial to determine an appropriate interaction technique and to design effective interactions with the 3D target in VEs. To design an effective and efficient user interface, it is also essential to understand the factors affecting the users’ performance. Moreover, the design of the interface also needs to encompass an understanding of human behaviors—particularly when people are interacting in 3D space. Researchers have explored another 3D space input technique [18,19,20] as an alternative to traditional 2D input devices (e.g., mouse, trackball, etc.) and touch screens [21]. Generally, the technique utilizes 3D target pointing hand movement as an input function for various devices that require free-hand or touchless computer interaction. The question is whether such hand pointing interactions have the same performance as traditional 2D input devices. Several studies have successfully investigated user performance in human–computer tasks in 2D environments [22,23] by utilizing direct-touch screens, as well as in 3D environments [17,24,25]. These are considered to be the same type of indirect interactions used in the present study. Nevertheless, to the best of our knowledge, comparative studies on how users interact with virtual targets considered direct and indirect interactions in our study are limited in number. Although previous studies have examined the performance of target acquisition in VE, the two techniques were considered separately (see Table 1 for an overview of studies on interaction techniques). This study, therefore, investigated the effects of direct interaction and indirect interaction techniques on the accuracy of egocentric distance estimation in a VE. We simultaneously employed direct pointing and a virtual cursor controlled by a gamepad for interactions in direct and indirect interactions.
Another important factor in a VE is space and the perception of it. In addition, spatial information such as distance and size are fundamentally crucial in both the real world and VEs; furthermore, its information should be provided accurately [26]. However, biased perception of distance in VEs often causes users to over- or underestimate. The space in VEs can be divided into two measured distances: egocentric and exocentric distance. Egocentric distance is a distance between an observer and the object; usually, the observer estimates a depth toward the object. Exocentric distance is a distance between two objects. Cutting [27] defined a metric of depth from observer to target into peripersonal space, a distance extending a little beyond the arm’s reach (about a radius of one meter) and extrapersonal space, which is all space beyond 1 m. Distance underestimation has been reported in the majority of studies of VEs [28,29,30,31]. A systematic review of previous studies by Lin and Woldegiorgis [32] found that the accuracy of distance in VEs was only about 80%, as compared to about 94% in the real world. Distance estimation in VEs can also suffer from worse inaccuracy, sometimes by as much as 50% of the intended distance [33]. Another study of egocentric distance perception in an IVE system by Willemsen et al. [34] showed that distance accuracy was degraded by 45% (underestimated true distance within the range of 5 to 15 m). The study hypothesized that the inaccuracies and cue conflicts involving stereo viewing conditions in HMDs resulted in inaccurate absolute scaling of the virtual world. However, the results indicated that compressed egocentric distance judgment was not caused by the unnatural stereo viewing conditions commonly found in IVE systems. The study suggested further investigation of other factors as a source of the compression; e.g., the ergonomics of wearing the VR headset or sense of presence. Regarding the inaccuracy of distance judgment, previous research provides a variety of possible causes. An extensive review by Renner et al. [35] identified four classes of factors that possibly cause underestimation: measurement methods, technical factors, compositional factors, and human factors. However, it is still not clear why actions in VE are smaller than intended. Recent studies on the perception of target sizes in VE also revealed a compression issue [36,37], as it is dominantly observed in egocentric distance. One of the causes reported was the variations of judgment techniques used to respond. To date, studies of distance estimations have mostly employed verbal or variations of walking (blind walking, triangulation walking) reporting techniques to specify the perceived distance. As VR applications are constantly becoming more interactive, users are enabled to physically interact (e.g., touching, manipulating, etc.) with 3D objects rather than just viewing 3D images. This advancement has been a focus of concern for scholars conducting research on interaction techniques to measure the accuracy of distance estimation. Table 1 shows a summary of studies on interaction techniques (direct and indirect interactions) with different experimental conditions and the extent to which this factor affects the accuracy of egocentric distance estimation. Generally, egocentric distance estimation can be evaluated in one of three planes (frontal, sagittal/lateral, and horizontal/transversal), depending on where the effective motion or movement is to be performed. Although the majority of studies have investigated the accuracy of egocentric distance in the frontal plane, as it is important for many applications in VEs [38], the results have consistently showed that egocentric distance judgments in VEs are not as accurate as real-world estimates.
In this experiment, the accuracy of distance judgment using direct pointing and indirect cursor techniques was investigated. Participants estimated distances by reaching for/pointing at the virtual target at three egocentric distance levels: 90, 120, and 150 cm from the participant. Various applications in augmented reality and virtual environments require reaching by hand or holding other objects to touch a virtual target. For example, many mixed reality-assisted medical procedures involve positioning a medical device at a depth indicated by a (virtual) marker. Distance judgment by reaching/pointing has been studied in previous studies [13,45,46,47]. Based on some of the results of previous studies [12,40,46], it was expected that distance estimation would be less accurate with the direct pointing technique than with the indirect cursor technique. In particular, a recent study compared the exocentric distance judgment and spatial perception between direct and indirect interactions and found that the accuracy in the frontal space significantly differed [46]. In the current study, the experiment was carried out to find whether such differences also exist for egocentric distance judgment. We expected a more accurate distance estimation for longer distances between the participant and the target [11]. We also expected that underestimations would be less severe with the direct pointing technique [13] than with the indirect cursor technique.

2. Methods

In this study, we investigated the effects of two interaction techniques, i.e., direct pointing and indirect cursor techniques, on egocentric distance estimation in a stereoscopic environment. We examined the two techniques by considering a tapping (pointing) task wherein virtual targets were projected at three different levels of egocentric distance along the frontal plane. The interaction techniques were developed on the basis of terms for interactions between users and targets in VE by Mine [14].

2.1. Direct Interaction with Virtual Objects

Direct interaction is an impression of direct involvement with an object rather than by communicating with an intermediary [48]. With this interaction, a more natural type of interaction is expected to be easy to achieve because of the utilization of the user’s hand or other body parts to interact with the objects [49]. Direct 3D interaction has been the focus of many works in VEs over the last few decades. However, based on the statistics showing that accuracy is less accurate in virtual scenes than in the real world, direct interaction with a physical object (e.g., hand, pointing stick, etc.) could also create confusion and lead to inaccuracy because of the need to “touch” intangible 3D objects [50]. Although direct interaction could lead to a number of errors, most results from similar studies agree that this interaction could improve the performance of manipulation of objects when the condition is coupled with visual and motor spaces (e.g., hand positions) to allow better interactions [51,52,53].
In our study, in the direct interaction condition, participants directly pointed (i.e., direct pointing technique) to the center of the target surface using a pointing stick (Figure 1a). This reporting method is similar to the method employed in Singh, Swan, Jones, and Ellis [45], Lubos et al. [54], and Bruder, Steinicke, and Stuerzlinger [39], who employed a physical arm to reach a target in the egocentric distance in a near-field VE. Since the distances considered in their experiment were within the arm’s reach, direct pointing by hand was sufficient. In our study, we used three pointing sticks of different lengths with reflective markers attached at their tips. The data of the reflective marker positions were recorded and tracked by a 6D motion capture system (Optitrack) at a frame rate of 120 frames per second. To confirm the pointing actions, participants had to click a Logitech spotlight wireless remote attached at the lower end of the stick.

2.2. Indirect Interaction with Virtual Objects

Alternatively, to select an object that is not located within the arm’s reach, the indirect interaction method offers the possibility of selecting distant objects but is limited by decoupling of the user’s motor and visual spaces. Indirect interaction also often requires more cognition between input and output [55]. For example, a user who controls an intermediary of a slider on the panel to adjust the intensity of light must put thought into controlling the intermediary slider, wait for a response, and then interpret the results. Different indirect interaction techniques have been proposed, one example being the Go-Go technique [56], which can provide users with the ability to interact with objects in extrapersonal space by nonlinear scaling of hand positions within arm’s reach.
In the present study, for the indirect interaction technique, participants used a gamepad to control a virtual cursor (i.e., hand cursor) to acquire the target in VE. Participants had to place the cursor at the center of the target surface (Figure 1b). Since the virtual cursor and the targets were all displayed virtually, there were no conflicts between the real (physical) and virtual objects. This was expected to reduce the confusion caused by virtually touching an intangible 3D object. However, it is not clear whether the reduction of visual conflicts can improve the overall estimation performance.

2.3. Experimental Design and Variables

The experimental task is illustrated in Figure 2. We used a standard ISO 9241-9 tapping (pointing) task setup [57], with stereoscopic targets projected (on a frontal plane to participants) at different egocentric distances in front of the projection screen, i.e., with different negative parallaxes. The parallax itself defined the position of the targets relative to a fixed screen display position while the participant was seated 210 cm from the screen; the targets could be perceived at distances of 90, 120, and 150 cm from the participant. Eight spherical targets, rendered in red, were arranged in a circle and displayed one by one in a defined order.
Similarly, the arrangement of experimental facilities and their setup is shown in Figure 3. The experimental space was 4 m × 3 m × 2.5 m and partitioned by black curtains to create an excellent stereoscopic environment with no unwanted light. The room was equipped with tables and an adjustable chair. The background of the stereoscopic environment was a uniform dark blue scene, which was projected onto a 130 cm wide × 100 cm high projection screen placed at a fixed position of 210 cm from the participant, and the projector was positioned below the participant’s eyes, under the table.

2.3.1. Independent Variables

This experiment had three independent variables: two interaction techniques (direct pointing and indirect cursor technique), three egocentric distances (target displayed at 90 cm, 120 cm, and 150 cm from the participant), and six indices of task difficulty (2.7, 3.5, 4.1, 4.5, 6.05, and 6.15 bits). Therefore, the experiment was designed as a 2 (interaction technique) × 3 (egocentric distance) × 6 (index of difficulty) within-subject design, and repeated-measures ANOVA was used as the analysis method. The levels of task difficulty were further classified into low (easy task), medium (moderate task), and high (difficult task) indices of difficulty. These difficulty levels respectively indicated an index of difficulty (ID) less than or equal to four, greater than four but less than or equal to six, and greater than six [58]. The ID was computed using Equation (1):
I D = l o g 2 ( A W + 1 )
where A is the amplitude of inter-target distance and W is the target width, as shown in Figure 2. A farther inter-target distance or smaller target results in increased difficulty.

2.3.2. Dependent Variables

The major dependent variable considered in this study was accuracy. Accuracy measures how close the estimation is to the reference value, which can be defined as a fraction of the actual egocentric distance. Accuracy was previously used by Armbrüster et al. [59], Dey et al. [60], and Lin and Woldegiorgis [11] for evaluating egocentric distance perception, using Equation (2):
A c c u r a c y = ( 1 | D e D a D a | )
where De is the observer’s perceived distance, and Da is the corresponding actual egocentric distance. A value of accuracy closer to one indicates that the egocentric distance estimation is more accurate.
Task completion time was also measured in seconds from the moment the participant judged the first target to the moment they completed the trial (with eight targets). Movement time was considered as the time taken by the participant to move the tip of the pointing stick or virtual hand cursor from one target to the next target. As displayed in Equation (3), the movement time (MT) required to point at a target is affected by the distance moved (A) and the width of the target W measured with respect to the direction of movement [61]:
M o v e m e n t   T i m e   ( M T ) = a + b * l o g 2 ( A   W + 1 )
where a and b are regression coefficients, while the logarithmic phrase is an index of difficulty (ID). The movement time was used to calculate the throughput of each trial. The throughput represents a rate of information transmission of responses [57,62] and can be calculated using the following equation:
T h r o u g h p u t = I n d e x   o f   D i f f i c u l t y   ( I D ) M o v e m e n t   T i m e   ( M T )

2.4. Experimental Settings and Stimuli

The VE, the 3D stereoscopic targets, and the experimental task were developed in Unity 3D (version 4.3.4f1). The VE was projected by a ViewSonic 3D projector and the task ran on a high-speed computer that supported stereoscopic vision (processor Intel Core-i7-7700 CPU @3.60GHz, with installed RAM 8 GB, and NVIDIA GeForce GTX 1060 6GB graphics); the latency of the virtual system could be minimized such that it should not affect the performance of interaction. To perceive the 3D stereoscopic targets at three different egocentric distances, participants wore NVIDIA 3D glasses integrated with a 3D vision emitter. The parallax was adjusted based on the required egocentric distances (90 cm, 120 cm, or 150 cm from the participant) and the interpupillary distance (IPD) of each participant. The interaction by direct pointing was performed using light wooden sticks of three different lengths, 80 cm, 110 cm, and 140 cm, with 0.6 cm reflective markers attached to their tips. The material of the pointing stick was carefully chosen to minimize the adverse effects of its weight on the participants’ pointing postures or performance. The data of the reflective marker positions were recorded and tracked by a 6D motion capture system (Optitrack) at a frame rate of 120 frames per second. To confirm the pointing actions, participants had to click a Logitech spotlight wireless remote attached at the lower end of the stick. The interaction by indirect cursor was performed using a gamepad, a Sun-Yes R-0011 dual analog. The gamepad analog enabled the movement of the cursor along the x-, y-, and z-axes within the VE. The left analog of the gamepad controlled a virtual cursor in two-degrees-of-freedom (DoF) of translation (up and down, left and right), including diagonal movement, while the right analog of the gamepad controlled the depth (forward and backward) of the virtual cursor. The size of the virtual cursor was scaled to approximately the width of each target to provide good visualization when touching the target. Four programmable buttons were located on the gamepad grip, but for this study, only the primary button “X” was activated (to confirm the pointing). Greater force on the gamepad analog increased the velocity of the virtual cursor. We optimized the sensitivity value of the gamepad between the exerted force and the cursor speed. A preliminary experiment was conducted to obtain an optimum sensitivity value to control the cursor to a comfortable condition for precise and fast movement. The average value was approximately 2 m/s, and we used this value for subsequent experiments. The starting points of the stick tips and virtual cursors were kept fixed at three different positions, P1, P2, and P3, marked on the table (as shown in Figure 3). Considering the projector as a reference of measurement, the relative positions of the initial positions of the tips and cursors were P1 (82, 0, 82), P2 (82, 0, 90), and P3 (82, 0, 120) for targets at egocentric distances of 90, 120, and 150 cm, respectively. For each trial, both the stick tip and the cursor position were restarted at the initial positions. The data streams from the marker position, virtual cursor, and virtual target were collected and processed by PC to analyze how close the estimates of the egocentric distance were to the reference of the theoretical positions. To minimize the possibility of differences in the perceived egocentric distances because of the head motions, the participants rested their chins on a fixed chinrest on the tabletop and sat on an adjustable chair to ensure consistency in height.

2.5. Procedure

Prior to testing, participants read and completed a consent form, which also described the purposes, tasks, and procedures of the study. After reading the instructions, participants received an equivalent verbal description while their IPDs were measured. The IPD values were used to determine the separation of the horizontal images (in calculating the parallax) to create the 3D stereoscopic target at the required egocentric distance. Greater parallax brought the virtual target closer to the participant’s eyes.
Each participant trained for an average of 3 min in each interaction to become familiar with the experimental setting and apparatus. During the training session, the participant sat on a chair at a designated distance, wore the 3D glasses, and was instructed to view the VE and to obtain a good image of the virtual targets. Participants were also advised to quit at any time if they felt discomfort associated with the VE.
When each trial began, the spherical virtual targets were sequentially projected briefly and disappeared (just after the participant clicked the target) one after the other, and the participants were instructed to point at a target as quickly as possible while maintaining accuracy. No visual feedback (e.g., color change) was given for both interaction techniques when the target was chosen, other than the next target appearing.
We randomly divided the participants equally into two groups to start the experiment either in the direct or indirect interaction conditions, which were separated by at least two days to minimize the learning effect and fatigue. Each group was further divided into three to counterbalance the order of the trials in the three egocentric distance conditions (90 cm, 120 cm, and 150 cm from the participant). Once the egocentric distance was allocated, the participant had to complete all six IDs, within the same depth condition, in a completely randomized order. Each participant had to complete all combinations of tasks as 2 (interaction techniques) × 3 (egocentric distances) × 6 (indices of difficulty). Therefore, every participant completed 144 trials (3 egocentric distances × 6 indices of difficulty × 8 targets) for each interaction. The total time for each participant to finish all combinations of tasks including the training session was approximately 90 min.

2.6. Participants

Fourteen graduate students from various departments at the National Taiwan University of Science and Technology were recruited by advertisement on social media. They were ten males and four females between the ages of 23 and 28 years old (Mean = 24.36, SD = 1.45). All the participants were right-handed and reported normal vision or corrected-to-normal with glasses or contact lenses, and their IPDs were between 6.5 cm and 7 cm (Mean = 6.57, SD = 0.18). Among all the participants, the majority had not had any experience in VR. Written informed consent was provided to each participant prior to the experiment. All participants gave their informed consent for inclusion before they participated in the experiment. We tested the maximum stereo vision of the experiment for each participant by viewing the nearest target distance (90 cm from the participant) before the experiment started. None of the participants failed this test. The participants received neither payment nor compensation with academic credit. The experiment was approved by the research ethics committee of National Taiwan University (NTU-REC No: 201209HS002).

3. Results

Repeated-measures ANOVA with three independent variables was performed on each dependent variable, i.e., accuracy of estimation, task completion time, and throughput. Post hoc tests were conducted using the Tukey HSD test (α = 0.05) when applicable. Degrees of freedom (DoF) were corrected using Greenhouse-Geisser correction when Mauchly’s test indicated that the assumption of sphericity had been violated.

3.1. Accuracy of Estimation

As shown in Figure 4, the mean accuracy of the indirect cursor technique was higher (mean = 0.95, SD = 0.01) than the mean accuracy of the direct cursor technique (mean = 0.90, SD = 0.01) at all tested egocentric distances and IDs. The ANOVA exposed significant differences in accuracy between the two techniques (F [1, 13] = 12.26, p < 0.05). We found a significant main effect of egocentric distance (F [1.340, 17.419] = 6.719, p < 0.05) on accuracy. Tukey post-hoc test revealed that participants significantly performed more accurate judgment when the targets were 150 cm (p < 0.05) than when they were 120 cm and 90 cm from the participant. We found a two-way interaction between technique and egocentric distance (F [2, 26] = 8.759, p < 0.05) on accuracy. The post-hoc test showed that participants using direct pointing were significantly (p < 0.05) more precise when the targets were displayed at a distance of 150 cm than when they were displayed at the other distances. For the indirect cursor technique, we found no significant difference of distance on accuracy. In the direct pointing technique, the overall egocentric distance estimations were 96.797 cm (SD = 2.569 cm), 127.492 cm (SD = 2.151 cm), and 156.867 cm (SD = 1.543 cm) at distances of 90 cm, 120 cm, and 150 cm from the participant, respectively. These results showed that the overall egocentric distances were overestimated. Similarly, in the indirect cursor technique, the corresponding overall estimations were 86.015 cm (SD = 1.169 cm), 115.612 cm (SD = 1.295 cm), and 143.281 cm (SD = 1.794 cm) at 90 cm, 120 cm, and 150 cm, respectively. Unlike those with the direct pointing technique, the overall egocentric distances were underestimated.
The effect of ID on egocentric distance estimation accuracy was also examined. The results showed that the main effect of ID on the accuracy of distance judgement was significant (F [1.35, 17.58] = 8.24, p < 0.05). The highest mean accuracy was 0.934 (SD = 0.007) at a low level of ID, followed by 0.927 (SD = 0.008) at a medium level of ID, and then 0.919 (SD = 0.009) at a high level of ID. However, the interaction between ID and interaction technique was not significant (F [1.35, 17.49] = 2.94, p > 0.05).

3.2. Task Completion Time and Throughput

Figure 5 shows the mean completion time as functions of egocentric distance and ID. The repeated-measures ANOVA results showed that the task completion time for egocentric distances differed significantly (F [1.709, 22.217] = 37.382, p < 0.001). Tukey post-hoc tests further showed that the completion time was significantly different (p < 0.05) for all three distances. The completion time significantly increased when targets were displayed at the farthest distance of 150 cm (Mean = 18.342 s, SD = 0.930 s) as compared to 120 cm (Mean = 15.318 s, SD = 0.831 s) and 90 cm (Mean = 13.836 s, SD = 0.751 s). The completion time increased with ID for both direct pointing and indirect cursor techniques. The ANOVA confirmed that ID significantly affected completion time (F [2, 26] = 25.882, p < 0.001). The post-hoc Tukey’s analysis classified the independent variables into two groups. The results showed that the completion time was significantly higher (p < 0.01) when targets were displayed at high ID (mean = 17.020 s, SD = 0.893 s) than when they were displayed at medium (mean = 15.848 s, SD = 0.852 s) and low IDs (mean = 14.628 s, SD = 0.650 s). The post-hoc test did not indicate a significant difference between the low and medium IDs (p > 0.05) in task completion time. We found no significant main effect of interaction technique (F [1, 13] = 1.775, p > 0.05) on task completion time, and no significant (p > 0.05) two-way interaction effects of egocentric distance and interaction technique or ID and interaction technique. Furthermore, linear regression analyses showed that the movement time data were in agreement with Fitts’ Law. The R2 and the fitting equation of movement time for the direct pointing technique were 0.96 (MT = 1.75 + 0.10*ID), 0.85 (MT = 1.88 + 0.10*ID), and 0.87 (MT = 2.33 + 0.16*ID) for the distances of 90, 120, and 150 cm, respectively. The corresponding indirect cursor technique R2 and the fitting equation of movement time were 0.83 (MT = 1.79 + 0.11*ID), 0.92 (MT = 1.75 + 0.27*ID), and 0.92 (MT = 2.02 + 0.29*ID) for the distances of 90, 120, and 150 cm, respectively.
The results for throughput are illustrated in Figure 6. We found a weak main effect of interaction technique (F [1, 13] = 5.257, p = 0.039) on throughput. The overall throughput computed for the direct pointing technique was higher (mean = 2.314 bps, SD = 0.136 bps) than that for the indirect cursor technique (mean = 2.190 bps, SD = 0.095 bps). The ANOVA revealed a significant main effect of egocentric distance (F [1.989, 25.859] = 34.876, p < 0.001) on throughput. The post-hoc tests indicated that the throughput was significantly lower for the targets at 150 cm (p < 0.05) than for those at 120 cm and 90 cm. Similarly, we found a significant main effect of ID (F [1.548, 20.126] = 116.952, p < 0.001) on throughput. Tukey post-hoc tests further indicated that the throughputs were significantly different (p < 0.001) for all three IDs. The average throughput was significantly higher at the highest ID (Mean = 2.805 bps, SD = 0.157 bps) than at the medium ID (Mean = 2.172 bps, SD = 0.110 bps) and lowest ID (Mean = 1.779 bps, SD = 0.082 bps). We also found a marginally significant difference between interaction technique and index of difficulty (F [1.987, 25.832] = 3.574, p = 0.043) on throughput; however, no significant difference was found between interaction technique and egocentric distance (F [1.795, 23.329] = 2.819, p > 0.05).

4. Discussion

This study investigated the effects of two interaction techniques, direct pointing and indirect cursor, when a stereoscopic target was displayed in a projection-based display. The results revealed evident effects of the interaction technique, the egocentric distance, and the index of difficulty on the accuracy of interaction performance. Estimation of egocentric distance was less accurate with the direct pointing technique (90%) than with the indirect cursor technique (95%). This finding supports our experimental hypothesis that egocentric distance estimation is more accurate with the indirect cursor technique than with the direct pointing one. A plausible explanation is the visual conflicts that may occur when a virtual object is selected with the direct pointing technique [63]. In particular, when a participant tries to point a pointing stick at a virtual object in a stereoscopic environment, either the virtual object or the pointing stick will appear blurred. On the other hand, with the indirect cursor, since both the virtual targets and the virtual cursor were displayed stereoscopically in this study, the visual conflicts might have been reduced [40]. Another possible explanation might be related to an occlusion cue that occurred in the indirect cursor condition. When the virtual cursor passed through the virtual target, the cursor would occlude the spherical virtual target. Thus, with this cue, the participant might be aware that the cursor was passing behind the target. Although no visual feedback (e.g., color change) was provided when the cursor touched the target, the presence of the hand cursor could be a visual cue leading to the higher accuracy than that of the direct pointing technique. In addition, since the size of the virtual cursor was scaled approximately to the width of each target, as the cursor moved toward the target, the cursor was comparable until it was almost the same size as the target, which could indicate that they were at the same depth. In other words, when the hand cursor appeared smaller than the target, the size difference indicated that the cursor was farther from the target. Therefore, it was possible that the participants’ judgment benefited from this relative size cue.
This study also found that underestimations consistently occurred at each egocentric distance in the indirect cursor condition. On average, participants underestimated target positions by 4, 5, and 7 cm when the targets were displayed at distances of 90 cm, 120 cm, and 150 cm, respectively. This phenomenon was observed in other studies. A comprehensive review by Lin and Woldegiorgis [32] on studies of egocentric distance perspective in VR mostly showed underestimations. One may expect that this result indicates compression of the intended size of the virtual space [34,59]. Some researchers have claimed that such compression might be ascribed to measurement methods and technical issues [35], the low quality of graphics [64], experience in VR [65], and distance cues [66]. However, no consensus has been reached on the causes. In contrast to the indirect cursor technique, in this study, egocentric distance estimation was overestimated with the direct pointing technique in all target distance conditions. On average, distances were overestimated by 6 to 8 cm for all tested target distances (90 cm, 120 cm, and 150 cm from the participant). These results are in line with previous studies showing that the accuracy of egocentric distance estimations in stereoscopic displays up to a distance of 1.5 m was less precise and overestimated [11,13,67]. Using a technique similar to the direct pointing technique used in this study, Swan, Singh, and Ellis [13], who utilized direct matching and reaching for depth judgments, also observed a slightly overestimated judgment of 0.5 to 1.9 cm over reaching distances of up to 50 cm in a mixed environment of VR and real targets displayed by an nVisor ST60 head-mounted display (HMD) that allowed a 60° diagonal field of view (FOV). Likewise, using a widescreen display system and a triangulated pointing task, Bruder, Sanz, Olivier, and Lecuyer [67] found that overestimations of up to 50% occurred when a target was projected at a distance of 1 m from the observer. In a recent study by Lin and Woldegiorgis [11], vision-based and memory-based pointing methods were used to estimate the egocentric distances of 3D stereoscopic and real targets. Their results showed that egocentric distances were overestimated by 10 to 11 cm at distances of 100 cm and 150 cm in the stereoscopic environment, and their finding was consistent with their previous study [32]. Although earlier researchers based their conclusions on the response methods separately, their approaches can also be categorized as direct interaction methods. The overall estimations at the three tested egocentric distances showed underestimation with the indirect cursor technique and overestimation with the direct pointing technique. This contrast could be attributed to the difference in the methods of distance judgment, which has been interpreted as two representations of visual space with respect to the utilization of interaction techniques [68,69], namely, a cognitive/perception representation for the indirect cursor technique and a sensorimotor representation for the direct pointing technique. The activities in direct pointing involve vision-for-action of direct reaching to a reference target, in which the participant’s attention (perception) is focused on the target to be selected and the hand is moved closer to the goal through a motor (action) control. On the other hand, the indirect cursor technique involves only vision-for-perception or matching of a target object (i.e., virtual cursor) until it perceptually matches the reference target, with less or without any motor control (i.e., direct reaching to the target). These two different representations of visual space—a cognitive representation driving perception and a sensorimotor representation leading to action—could have influenced the estimation. At present, however, it is difficult to conclude that visual perception differs from participants’ responses or actions with respect to direct and indirect interaction techniques, as only a few related comparative studies are available. Further study with a greater focus on this variable is needed to clarify its effect on egocentric distance estimations. This study provides important information on how interaction techniques (direct pointing and indirect cursor) affect the accuracy of egocentric distance estimation for interaction with objects displayed stereoscopically at different parallaxes. It is also evident from the present study that interaction with the direct pointing technique could mitigate the underestimation problems commonly reported in VE studies.
In this experiment, we found that the accuracy of egocentric distance estimation with the direct pointing technique increased as the egocentric distance increased (as shown in Figure 4). This result was also supported by statistical evidence that egocentric distance estimation with the direct pointing technique was significantly more accurate (p < 0.05) at the farthest distance of 150 cm than at the other distances of 90 cm and 120 cm. The result is consistent with previous studies [11,35], which reported that the accuracy of stereoscopic target selection was lower for targets closer to the eyes. This finding can be explained by the occurrence of accommodation-converge mismatch [70]: 3D objects displayed closer to a participant generate higher accommodation-converge mismatch than do objects displayed farther away. This factor reduces the accuracy of target selection in stereoscopic environments.
The throughput performance was significantly higher with direct pointing than with the indirect cursor technique. This difference probably resulted from the target selection times of the direct pointing technique, which were faster than those of the indirect cursor technique. With the direct pointing technique, simplicity and efficiency of arm movement are easy to achieve. In addition, participants informally reported that they felt the direct pointing technique was easier to perform than the indirect cursor technique. In our study, a common pattern was found in movement time. The results indicated that movement time linearly increased with the target’s ID. However, an unanticipated finding was that throughput increased with ID, contrary to the intuitive expectation that a higher ID would result in lower throughput. Such results have been reported in some studies; for instance, Lin, Abreham, and Woldegiorgis [41] showed that throughput was higher at the highest ID than at the lowest ID. The explanation for this unexpected result could be that the rate of increment between IDs was not proportional to the rate of increment observed in the corresponding MT, where the slope of ID increased more sharply. Further work is required to comprehensively explain this issue.

5. Conclusions

In this study, the effects of interaction techniques on the interaction performance of accuracy in performing a pointing/selection task within a stereoscopic environment were investigated. The influences of indices of difficulty and different egocentric distances on the accuracy of interaction performance were also examined. The results showed that accuracy in the selection of stereoscopic targets was significantly affected by the interaction technique. Moreover, we also found substantial effects of the distance at which a target was projected at a different level of difficulty within the negative parallax of the stereoscopic display.
The indirect cursor technique was more accurate than the direct pointing technique. However, the results also showed that the throughput of the direct cursor technique was higher than that of the indirect cursor technique. This may have been caused by the ease of use and the naturalness of the direct cursor technique allowing faster target selection times, although the means of the task completion times between the direct pointing and indirect cursor techniques were not significantly different. Direct pointing could have interesting implications in VR. The results here suggest that the direct pointing technique may alleviate underestimation of distances in VE, despite the slight overestimation. Consequently, developers of VEs in which user task performance depends on accuracy (for instance, surgical training and simulations) should consider the relative advantages of the direct and indirect interaction techniques.
In addition, the experiment revealed that targets displayed at 150 cm from the participant provided the most accurate distance judgment of all three egocentric distances. This result implies that the accuracy of target selection in a stereoscopic environment improves as egocentric distance increases. This result could also be used to enhance the understanding of the effective space during interaction with stereoscopic targets and to propose proper interaction distances with stereoscopic targets and displays. However, future research might be needed to determine how much distance judgment improves with changes in distance by considering targets at farther distances from participants.
The present study examined task performance in the frontal plane under three egocentric distance conditions. However, performance in the lateral and transversal planes could be considered in future studies. In addition, the virtual target was displayed in a projection-based stereoscopic environment, and a particular tracking system and input device were employed. Future studies may use different virtual environment systems, such as head-mounted and wide-angle displays, to understand the generalizability of the results. In addition, a simple 3D pointing/selection task is commonly used to determine the effects of interaction techniques on egocentric distance accuracy. However, taking the growing interest in advanced forms of 3D virtual or augmented reality interactions into account, it may be necessary to consider a more complex task to show the performance of the interaction techniques. Therefore, further experimental investigations with a more complex task are recommended to clarify the performance of the interaction techniques. The results reported in this study provide useful information on how accuracy and throughput vary with different interaction (direct and indirect) techniques, levels of egocentric distances, and indices of difficulty in a stereoscopic display. Future studies on interaction in stereoscopic environments might be required to investigate user behaviors and kinematics by evaluating usability and analyzing the movements when different interaction techniques are employed.

Author Contributions

Conceptualization, C.J.L. and B.H.W.; methodology, D.C. and B.H.W.; software, D.C. and B.H.W.; data collection, D.C.; statistical analysis, D.C.; writing—original draft preparation, D.C.; writing—review and editing, C.J.L. and B.H.W.; supervision, C.J.L. and B.H.W.

Funding

This research was supported by the Ministry of Science and Technology of Taiwan (MOST 107-2218-E-011-019-MY3 and MOST 107-2221-E-011 -102 -MY3).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huber, T.; Paschold, M.; Hansen, C.; Wunderling, T.; Lang, H.; Kneist, W. New dimensions in surgical training: Immersive virtual reality laparoscopic simulation exhilarates surgical staff. Surg. Endosc. 2017, 31, 4472–4477. [Google Scholar] [CrossRef] [PubMed]
  2. Buttussi, F.; Chittaro, L. Effects of Different Types of Virtual Reality Display on Presence and Learning in a Safety Training Scenario. IEEE Trans. Vis. Comput. Graph. 2018, 24, 1063–1076. [Google Scholar] [CrossRef] [PubMed]
  3. Matsas, E.; Vosniakos, G.-C. Design of a virtual reality training system for human–robot collaboration in manufacturing tasks. Int. J. Interact. Des. Manuf. 2017, 11, 139–153. [Google Scholar] [CrossRef]
  4. Sun, H.-M.; Li, S.-P.; Zhu, Y.-Q.; Hsiao, B. The effect of user’s perceived presence and promotion focus on usability for interacting in virtual environments. Appl. Ergon. 2015, 50, 126–132. [Google Scholar] [CrossRef]
  5. Sharples, S.; Cobb, S.; Moody, A.; Wilson, J.R. Virtual reality induced symptoms and effects (VRISE): Comparison of head mounted display (HMD), desktop and projection display systems. Displays 2008, 29, 58–69. [Google Scholar] [CrossRef]
  6. McIntire, J.P.; Havig, P.R.; Geiselman, E.E. Stereoscopic 3D displays and human performance: A comprehensive review. Displays 2014, 35, 18–26. [Google Scholar] [CrossRef]
  7. Zhou, F.; Duh, H.; Billinghurst, M. Trends in Augmented Reality Tracking, Interaction and Display: A Review of Ten Years of ISMAR. In Proceedings of the 7th IEEE/ACM International Symposium on Mixed and Augmented Reality, Cambridge, UK, 15–18 September 2008; pp. 193–202. [Google Scholar]
  8. Fuhrmann, A.; Hesina, G.; Faure, F.; Gervautz, M. Occlusion in collaborative augmented environments. Comput. Gr. 1999, 23, 809–819. [Google Scholar] [CrossRef] [Green Version]
  9. Erazo, O.; Pino, J.A. Predicting Task Execution Time on Natural User Interfaces based on Touchless Hand Gestures. In Proceedings of the 20th International Conference on Intelligent User Interfaces, Atlanta, GA, USA, 29 March–1 April 2015; pp. 97–109. [Google Scholar]
  10. Lin, C.J.; Woldegiorgis, B.H. Kinematic analysis of direct pointing in projection-based stereoscopic environments. Hum. Mov. Sci. 2018, 57, 21–31. [Google Scholar] [CrossRef]
  11. Lin, C.J.; Woldegiorgis, B.H. Egocentric distance perception and performance of direct pointing in stereoscopic displays. Appl. Ergon. 2017, 64, 66–74. [Google Scholar] [CrossRef]
  12. Bruder, G.; Steinicke, F.; Sturzlinger, W. To touch or not to touch?: Comparing 2D touch and 3D mid-air interaction on stereoscopic tabletop surfaces. In Proceedings of the 1st symposium on Spatial user interaction, Los Angeles, CA, USA, 20–21 July 2013; pp. 9–16. [Google Scholar]
  13. Swan, J.E.; Singh, G.; Ellis, S.R. Matching and Reaching Depth Judgments with Real and Augmented Reality Targets. IEEE Trans. Vis. Comput. Graph. 2015, 21, 1289–1298. [Google Scholar] [CrossRef]
  14. Mine, M.R. Virtual Environment Interaction Techniques; University of North Carolina at Chapel Hill: Chapel Hill, CA, USA, 1995. [Google Scholar]
  15. Poupyrev, I.; Ichikawa, T. Manipulating Objects in Virtual Worlds: Categorization and Empirical Evaluation of Interaction Techniques. J. Vis. Lang. Comput. 1999, 10, 19–35. [Google Scholar] [CrossRef]
  16. Werkhoven, P.J.; Groen, J. Manipulation Performance in Interactive Virtual Environments. Hum. Fact. 1998, 40, 432–442. [Google Scholar] [CrossRef]
  17. Deng, C.-L.; Geng, P.; Hu, Y.-F.; Kuai, S.-G. Beyond Fitts’s Law: A Three-Phase Model Predicts Movement Time to Position an Object in an Immersive 3D Virtual Environment. Hum. Fact. 2019, 61, 879–894. [Google Scholar] [CrossRef] [PubMed]
  18. Nguyen, A.; Banic, A. 3DTouch: A wearable 3D input device for 3D applications. In Proceedings of the IEEE Virtual Reality (VR), Arles, France, 23–27 March 2015; pp. 55–61. [Google Scholar]
  19. Jang, Y.; Noh, S.; Chang, H.J.; Kim, T.; Woo, W. 3D Finger CAPE: Clicking Action and Position Estimation under Self-Occlusions in Egocentric Viewpoint. IEEE Trans. Vis. Comput. Graph. 2015, 21, 501–510. [Google Scholar] [CrossRef]
  20. Cha, Y.; Myung, R. Extended Fitts’ law for 3D pointing tasks using 3D target arrangements. Int. J. Ind. Ergon. 2013, 43, 350–355. [Google Scholar] [CrossRef]
  21. Chen, J.; Or, C. Assessing the use of immersive virtual reality, mouse and touchscreen in pointing and dragging-and-dropping tasks among young, middle-aged and older adults. Appl. Ergon. 2017, 65, 437–448. [Google Scholar] [CrossRef]
  22. Bi, X.; Li, Y.; Zhai, S. FFitts law: Modeling finger touch with fitts’ law. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Paris, France, 27 April–2 May 2013; pp. 1363–1372. [Google Scholar]
  23. MacKenzie, I.S.; Teather, R.J. FittsTilt: The application of Fitts’ law to tilt-based interaction. In Proceedings of the 7th Nordic Conference on Human-Computer Interaction: Making Sense Through Design, Copenhagen, Denmark, 14–17 October 2012; pp. 568–577. [Google Scholar]
  24. Lin, C.J.; Widyaningrum, R. Eye Pointing in Stereoscopic Displays. J. Eye Mov. Res. 2016, 9, 1–14. [Google Scholar] [CrossRef]
  25. Murata, A.; Iwase, H. Extending Fitts’ law to a three-dimensional pointing task. Hum. Mov. Sci. 2001, 20, 791–805. [Google Scholar] [CrossRef]
  26. Dey, A.; Jarvis, G.; Sandor, C.; Reitmayr, G. Tablet versus phone: Depth perception in handheld augmented reality. In Proceedings of the IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Atlanta, GA, USA, 5–8 November 2012; pp. 187–196. [Google Scholar]
  27. Cutting, J.E. Reconceiving perceptual space. In Looking into Pictures: An Interdisciplinary Approach to Pictorial Space; MIT Press: Cambridge, MA, USA, 2003; pp. 215–238. [Google Scholar]
  28. Naceri, A.; Chellali, R.; Dionnet, F.; Toma, S. Depth Perception within Virtual Environments: A Comparative Study Between Wide Screen Stereoscopic Displays and Head Mounted Devices. In Proceedings of the Computation World: Future Computing, Service Computation, Cognitive, Adaptive, Content, Patterns, Washington, DC, USA, 15–20 November 2009; pp. 460–466. [Google Scholar]
  29. Grechkin, T.Y.; Nguyen, T.D.; Plumert, J.M.; Cremer, J.F.; Kearney, J.K. How does presentation method and measurement protocol affect distance estimation in real and virtual environments? ACM Trans. Appl. Percept. 2010, 7, 1–18. [Google Scholar] [CrossRef]
  30. Iosa, M.; Fusco, A.; Morone, G.; Paolucci, S. Walking there: Environmental influence on walking-distance estimation. Behav. Brain Res. 2012, 226, 124–132. [Google Scholar] [CrossRef]
  31. Knapp, J.M.; Loomis, J.M. Limited field of view of head-mounted displays is not the cause of distance underestimation in virtual environments. Presence: Teleoper. Virtual Environ. 2004, 13, 572–577. [Google Scholar] [CrossRef]
  32. Lin, C.J.; Woldegiorgis, B.H. Interaction and visual performance in stereoscopic displays: A review. J. Soc. Inf. Disp. 2015, 23, 319–332. [Google Scholar] [CrossRef]
  33. Waller, D.; Richardson, A.R. Correcting distance estimates by interacting with immersive virtual environments: Effects of task and available sensory information. J. Exp. Psychol. Appl. 2008, 14, 61–72. [Google Scholar] [CrossRef] [PubMed]
  34. Willemsen, P.; Gooch, A.A.; Thompson, W.B.; Creem-Regehr, S.H. Effects of Stereo Viewing Conditions on Distance Perception in Virtual Environments. Presence 2008, 17, 91–101. [Google Scholar] [CrossRef]
  35. Renner, R.S.; Velichkovsky, B.M.; Helmert, J.R. The perception of egocentric distances in virtual environments—A review. ACM Comput. Surv. 2013, 46, 1–40. [Google Scholar] [CrossRef]
  36. Kelly, J.W.; Hammel, W.; Sjolund, L.A.; Siegel, Z.D. Frontal extents in virtual environments are not immune to underperception. Atten. Percept. Psychophys. 2015, 77, 1848–1853. [Google Scholar] [CrossRef] [Green Version]
  37. Stefanucci, J.K.; Creem-Regehr, S.H.; Thompson, W.B.; Lessard, D.A.; Geuss, M.N. Evaluating the accuracy of size perception on screen-based displays: Displayed objects appear smaller than real objects. J. Exp. Psychol. Appl. 2015, 21, 215–223. [Google Scholar] [CrossRef]
  38. Wartenberg, C.; Wiborg, P. Precision of Exocentric Distance Judgments in Desktop and Cube Presentation. Presence 2003, 12, 196–206. [Google Scholar] [CrossRef]
  39. Bruder, G.; Steinicke, F.; Stuerzlinger, W. Touching the Void Revisited: Analyses of Touch Behavior on and above Tabletop Surfaces. In Proceedings of the 14th IFIP Conference on Human-Computer Interaction, Cape Town, South Africa, 2–6 September 2013; pp. 278–296. [Google Scholar]
  40. Bruder, G.; Steinicke, F.; Sturzlinger, W. Effects of visual conflicts on 3D selection task performance in stereoscopic display environments. In Proceedings of the IEEE Symposium on 3D User Interfaces (3DUI), Orlando, FL, USA, 16–17 March 2013; pp. 115–118. [Google Scholar]
  41. Lin, C.J.; Abreham, B.T.; Woldegiorgis, B.H. Effects of displays on a direct reaching task: A comparative study of head mounted display and stereoscopic widescreen display. Int. J. Ind. Ergon. 2019, 72, 372–379. [Google Scholar] [CrossRef]
  42. Lin, C.J.; Ho, S.-H.; Chen, Y.-J. An investigation of pointing postures in a 3D stereoscopic environment. Appl. Ergon. 2015, 48, 154–163. [Google Scholar] [CrossRef]
  43. Napieralski, P.E.; Altenhoff, B.M.; Bertrand, J.W.; Long, L.O.; Babu, S.V.; Pagano, C.C.; Kern, J.; Davis, T.A. Near-field distance perception in real and virtual environments using both verbal and action responses. ACM Trans. Appl. Percept. 2011, 8, 1–19. [Google Scholar] [CrossRef]
  44. Poupyrev, I.; Weghorst, S.; Fels, S. Non-isomorphic 3D rotational techniques. In Proceedings of the SIGCHI conference on Human Factors in Computing Systems, The Hague, The Netherlands, 1–6 April 2000; pp. 540–547. [Google Scholar]
  45. Singh, G.; Swan, J.E.; Jones, J.A.; Ellis, S.R. Depth judgments by reaching and matching in near-field augmented reality. In Proceedings of the IEEE Virtual Reality Workshops (VRW), Costa Mesa, CA, USA, 4–8 March 2012; pp. 165–166. [Google Scholar]
  46. Lin, C.J.; Caesaron, D.; Woldegiorgis, B.H. The accuracy of the frontal extent in stereoscopic environments: A comparison of direct selection and virtual cursor techniques. PLoS ONE 2019, 14, e0222751. [Google Scholar] [CrossRef] [PubMed]
  47. Singh, G.; Swan, E.J., II; Jones, J.A.; Ellis, S. Depth judgment measures and occluding surfaces in near-field augmented reality. In Proceedings of the 7th Symposium on Applied Perception in Graphics and Visualization, Los Angeles, CA, USA, 23–24 July 2010; pp. 149–156. [Google Scholar]
  48. Hutchins, E.L.; Hollan, J.D.; Norman, D.A. Direct manipulation interfaces. Hum.-Comput. Interact. 1985, 1, 311–338. [Google Scholar] [CrossRef]
  49. Steinicke, F.; Benko, H.; Kruger, A.; Keefe, D.; Riviere, J.-B.d.l.; Anderson, K.; Hakkila, J.; Arhippainen, L.; Pakanen, M. The 3rd dimension of CHI (3DCHI): Touching and designing 3D user interfaces. In Proceedings of the CHI ’12 Extended Abstracts on Human Factors in Computing Systems, Austin, TX, USA, 5–10 May 2012; pp. 2695–2698. [Google Scholar]
  50. Chan, L.-W.; Kao, H.-S.; Chen, M.Y.; Lee, M.-S.; Hsu, J.; Hung, Y.-P. Touching the void: Direct-touch interaction for intangible displays. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Atlanta, GA, USA, 10–15 April 2010; pp. 2625–2634. [Google Scholar]
  51. Lemmerman, D.K.; LaViola, J.J., Jr. Effects of Interaction-Display Offset on User Performance in Surround Screen Virtual Environments. In Proceedings of the 2007 IEEE Virtual Reality Conference, Charlotte, NC, USA, 10–14 March 2007; pp. 303–304. [Google Scholar]
  52. Wang, Y.; MacKenzie, C. Effects of orientation disparity between haptic and graphic displays of objects in virtual environments. In Proceedings of the IFIP Conference on Human-Computer Interaction, Edinburgh, Scotland, 30 August–3 September 1999; pp. 391–398. [Google Scholar]
  53. Mine, M.R.; Frederick, P.; Brooks, J.; Sequin, C.H. Moving objects in space: Exploiting proprioception in virtual-environment interaction. In Proceedings of the 24th annual conference on Computer graphics and interactive techniques, San Jose, CA, USA, 20–22 July 1997; pp. 19–26. [Google Scholar]
  54. Lubos, P.; Bruder, G.; Steinicke, F. Analysis of direct selection in head-mounted display environments. In Proceedings of the IEEE Symposium on 3D User Interfaces (3DUI), Minneapolis, MN, USA, 29–30 March 2014; pp. 11–18. [Google Scholar]
  55. Jerald, J. The VR Book: Human-Centered Design for Virtual Reality; Association for Computing Machinery and Morgan Claypool: New York, USA, 2016; p. 635. [Google Scholar]
  56. Poupyrev, I.; Billinghurst, M.; Weghorst, S.; Ichikawa, T. The go-go interaction technique: Non-linear mapping for direct manipulation in VR. In Proceedings of the 9th annual ACM symposium on User interface software and technology, Seattle, WA, USA, 6–8 November 1996; pp. 79–80. [Google Scholar]
  57. ISO. Ergonomic Requirements for Office Work with Visual Display Terminals (VDTs)-Part 9-Requirements for Non-Keyboard Input Devices., 1st ed.; International Organization for Standardisation: Geneva, Switzerland, 2000; p. 57. [Google Scholar]
  58. Soukoreff, R.W.; MacKenzie, I.S. Towards a standard for pointing device evaluation, perspectives on 27 years of Fitts’ law research in HCI. Int. J. Hum.-Comput. Stud. 2004, 61, 751–789. [Google Scholar] [CrossRef]
  59. Armbrüster, C.; Wolter, M.; Kuhlen, T.; Spijkers, W.; Fimm, B. Depth Perception in Virtual Reality: Distance Estimations in Peri- and Extrapersonal Space. CyberPsychology Behav. 2008, 11, 9–15. [Google Scholar] [CrossRef] [PubMed]
  60. Dey, A.; Cunningham, A.; Sandor, C. Evaluating depth perception of photorealistic mixed reality visualizations for occluded objects in outdoor environments. In Proceedings of the IEEE Symposium on 3D User Interfaces (3DUI), Waltham, MA, USA, 20–21 March 2010; pp. 127–128. [Google Scholar]
  61. Fitts, P.M. The information capacity of the human motor system in controlling the amplitude of movement. J. Exp. Psychol. Gen. 1992, 121, 262–269. [Google Scholar] [CrossRef] [PubMed]
  62. Burno, R.A.; Wu, B.; Doherty, R.; Colett, H.; Elnaggar, R. Applying Fitts’ Law to Gesture Based Computer Interactions. Procedia Manuf. 2015, 3, 4342–4349. [Google Scholar] [CrossRef]
  63. Thompson, W.; Fleming, R.; Creem-Regehr, S.; Stefanucci, J.K. Visual Perception from a Computer Graphics Perspective; A. K. Peters: Natick, MA, USA, 2011; p. 540. [Google Scholar]
  64. Kunz, B.R.; Wouters, L.; Smith, D.; Thompson, W.B.; Creem-Regehr, S.H. Revisiting the effect of quality of graphics on distance judgments in virtual environments: A comparison of verbal reports and blind walking. Atten. Percept. Psychophys. 2009, 71, 1284–1293. [Google Scholar] [CrossRef] [Green Version]
  65. Richardson, A.R.; Waller, D. The effect of feedback training on distance estimation in virtual environments. Appl. Cogn. Psychol. 2005, 19, 1089–1108. [Google Scholar] [CrossRef] [Green Version]
  66. Jones, J.A.; Swan, J.E.; Bolas, M. Peripheral Stimulation and its Effect on Perceived Spatial Scale in Virtual Environments. IEEE Trans. Vis. Comput. Graph. 2013, 19, 701–710. [Google Scholar] [CrossRef]
  67. Bruder, G.; Sanz, F.A.; Olivier, A.H.; Lecuyer, A. Distance estimation in large immersive projection systems, revisited. In Proceedings of the IEEE Virtual Reality (VR), Arles, France, 23–27 March 2015; pp. 27–32. [Google Scholar]
  68. Bridgeman, B.; Gemmer, A.; Forsman, T.; Huemer, V. Processing spatial information in the sensorimotor branch of the visual system. Vis. Res. 2000, 40, 3539–3552. [Google Scholar] [CrossRef]
  69. Parks, T.E. Visual-illusion distance paradoxes: A resolution. Atten. Percept. Psychophys. 2012, 74, 1568–1569. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  70. Hoffman, D.M.; Girshick, A.R.; Akeley, K.; Banks, M.S. Vergence–accommodation conflicts hinder visual performance and cause visual fatigue. J. Vis. 2008, 8, 33. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Illustration of the interaction techniques. (a) Direct pointing technique: directly pointing at the stereoscopic target with a stick, and (b) indirect cursor technique: controlling the gamepad to place the virtual hand cursor on the stereoscopic target.
Figure 1. Illustration of the interaction techniques. (a) Direct pointing technique: directly pointing at the stereoscopic target with a stick, and (b) indirect cursor technique: controlling the gamepad to place the virtual hand cursor on the stereoscopic target.
Applsci 09 04652 g001
Figure 2. Illustration of the experimental task. (a) Eight virtual spherical targets appeared one at a time in sequence, with target width (W) and amplitude of inter-target distances (A), and (b) the sequence of pointing trials for the first four targets.
Figure 2. Illustration of the experimental task. (a) Eight virtual spherical targets appeared one at a time in sequence, with target width (W) and amplitude of inter-target distances (A), and (b) the sequence of pointing trials for the first four targets.
Applsci 09 04652 g002
Figure 3. Illustration of (a) the experimental setup and (b) a participant during the experiment in the direct interaction condition. A chinrest was used to keep the participant’s head fixed, and the 3D projector was placed under the right-hand table. The origin for reference of measurement was in line with the center of the projector and at table height (75 cm above the floor).
Figure 3. Illustration of (a) the experimental setup and (b) a participant during the experiment in the direct interaction condition. A chinrest was used to keep the participant’s head fixed, and the 3D projector was placed under the right-hand table. The origin for reference of measurement was in line with the center of the projector and at table height (75 cm above the floor).
Applsci 09 04652 g003
Figure 4. Results of accuracy measurement for direct pointing and indirect cursor interaction techniques with respect to egocentric distance (A) and index of difficulty (B). The error bar represents standard error of mean.
Figure 4. Results of accuracy measurement for direct pointing and indirect cursor interaction techniques with respect to egocentric distance (A) and index of difficulty (B). The error bar represents standard error of mean.
Applsci 09 04652 g004
Figure 5. Results of task completion time for direct pointing and indirect cursor interaction techniques with respect to egocentric distance (A) and index of difficulty (B). The error bar represents standard error of mean.
Figure 5. Results of task completion time for direct pointing and indirect cursor interaction techniques with respect to egocentric distance (A) and index of difficulty (B). The error bar represents standard error of mean.
Applsci 09 04652 g005
Figure 6. Results of throughput for direct pointing and indirect cursor interaction techniques with respect to egocentric distance (A) and index of difficulty (B). The error bar represents standard error of mean.
Figure 6. Results of throughput for direct pointing and indirect cursor interaction techniques with respect to egocentric distance (A) and index of difficulty (B). The error bar represents standard error of mean.
Applsci 09 04652 g006
Table 1. Summary of studies on interaction techniques (direct or indirect interaction). Interaction techniques influence the accuracy of egocentric distance estimation in different displays and under various experimental conditions.
Table 1. Summary of studies on interaction techniques (direct or indirect interaction). Interaction techniques influence the accuracy of egocentric distance estimation in different displays and under various experimental conditions.
Author/s, year [number in reference list]Display 1Motion plane (lateral, frontal, transversal)Interaction TechniqueExperimental ConditionsResults (findings) 2
Bruder et al. 2013. [39]Stereoscopic tabletopTransversal planeDirect mid-air selection and direct-2D touch screenDirect mid-air selection vs. direct-2D touch screen.Direct-2D touch screen was more accurate than 3D mid-air touching.
Bruder et al. 2013. [40]Stereoscopic tabletopTransversal planeDirect mid-air selection with the tip of the user’s index finger.Direct input with the user’s fingertip vs. offset-based input with a virtual offset cursor.Direct input with the user’s fingertip was more accurate than offset-based input with the virtual offset cursor.
Lin, et al. 2019. [41]Stereoscopic projection screen and HMDFrontal planeDirect mid-air selection using a pointing stickStereoscopic vs. immersive environments.The immersive environment was less accurate than the stereoscopic environment.
Overestimation was found in distance estimates.
Lin and Woldegiorgis 2017. [11]Stereoscopic projection screen and real worldFrontal planeDirect mid-air selection using a pointing stick Stereoscopic vs. real environments.
 
Pointing by vision vs. memory.
The stereoscopic environment was less accurate than the real world.
Not significant.
Overestimation was found in distance estimates.
Lin, et al. 2015. [42]Stereoscopic projection screenAll three planesDirect pointing with hand- and gaze-directed. Direct pointing with hand-directed vs. direct pointing with gaze-directed.Gaze-directed pointing was less accurate than hand-directed.
Hand-directed pointing is suggested for tapping and tracking tasks.
Napieralski et al. 2011. [43]HMD and real worldFrontal planeDirect reaching using a stylus and verbal responseDirect reaching vs. verbal responses.
Immersive virtual environment vs. real world.
Direct reaching tended to be more accurate and consistent than verbal responses.
Underestimation was found more in IVE as compared to the real world.
Poupyrev et al. 2000. [44]Desktop monitorFrontal planeIndirect-input with 6DOF controller Absolute mapping control vs. relative mapping control.Not significant.
Werkhoven and Groen 1998. [16]HMDFrontal planeDirect input with virtual hand control and indirect input with 3D cursor controlVirtual hand control vs. 3D cursor control.
 
Monoscopic virtual environment vs. stereoscopic virtual environment
3D cursor control was less accurate than virtual hand control in the positioning task.
The stereoscopic condition was more accurate than the monoscopic condition.
1 HMD, head-mounted display. 2 Not significant indicates an insignificant difference of accuracies made in the described experimental conditions.

Share and Cite

MDPI and ACS Style

Lin, C.J.; Caesaron, D.; Woldegiorgis, B.H. The Effects of Augmented Reality Interaction Techniques on Egocentric Distance Estimation Accuracy. Appl. Sci. 2019, 9, 4652. https://doi.org/10.3390/app9214652

AMA Style

Lin CJ, Caesaron D, Woldegiorgis BH. The Effects of Augmented Reality Interaction Techniques on Egocentric Distance Estimation Accuracy. Applied Sciences. 2019; 9(21):4652. https://doi.org/10.3390/app9214652

Chicago/Turabian Style

Lin, Chiuhsiang Joe, Dino Caesaron, and Bereket Haile Woldegiorgis. 2019. "The Effects of Augmented Reality Interaction Techniques on Egocentric Distance Estimation Accuracy" Applied Sciences 9, no. 21: 4652. https://doi.org/10.3390/app9214652

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop