Next Article in Journal
DOF Decoupling Task Graph Model: Reducing the Complexity of Touch-Based Active Sensing
Previous Article in Journal
Robotized Inspection of Vertical Structures of a Solar Power Plant Using NDT Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image-Based Navigation for the SnowEater Robot Using a Low-Resolution USB Camera

1
Department of Mechanical Systems Engineering, Yamagata University, Yamagata Prefecture, Yonezawa 992-8510, Jo-nan 4-3-16, Japan
2
Graduate School of Science and Engineering, University of Toyama, Toyama 930-8555, Gofuku 3190, Japan
*
Author to whom correspondence should be addressed.
Robotics 2015, 4(2), 120-140; https://doi.org/10.3390/robotics4020120
Submission received: 18 December 2014 / Revised: 27 March 2015 / Accepted: 30 March 2015 / Published: 8 April 2015

Abstract

:
This paper reports on a navigation method for the snow-removal robot called SnowEater. The robot is designed to work autonomously within small areas (around 30 m2 or less) following line segment paths. The line segment paths are laid out so as much snow as possible can be cleared from an area. Navigation is accomplished by using an onboard low-resolution USB camera and a small marker located in the area to be cleared. Low-resolution cameras allow only limited localization and present significant errors. However, these errors can be overcome by using an efficient navigation algorithm to exploit the merits of these cameras. For stable robust autonomous snow removal using this limited information, the most reliable data are selected and the travel paths are controlled. The navigation paths are a set of radially arranged line segments emanating from a marker placed in the environment area to be cleared, in a place where it is not covered by snow. With this method, by using a low-resolution camera (640 × 480 pixels) and a small marker (100 × 100 mm), the robot covered the testing area following line segments. For a reference angle of 4.5° between line paths, the average results are: 4° for motion on hard floor and 4.8° for motion on compacted snow. The main contribution of this study is the design of a path-following control algorithm capable of absorbing the errors generated by a low-cost camera.

Graphical Abstract

1. Introduction

Robot technology is focused on increasing the quality of life by the creation of new machines and methods. This paper reports on the development of a snow-removal robot called SnowEater. The concept for this robot is the creation of a safe, slow, small, light, low-powered and inexpensive autonomous snow-removal machine for home use. In essence, SnowEater can be compared with autonomous vacuum cleaner robots [1], but instead of working inside homes, it is designed to operate on house walkways, and around front doors or garages. A number of attempts to automate commercial snow-blower machines have being made [2,3], but such attempts have been delayed out of concerns over safety.
The basic SnowEater model is derived from the heavy snow-removal robot named Yukitaro [4], which was equipped with a high-resolution laser rangefinder for the navigation system. In contrast to the Yukitaro robot, which weights 400 kg, the SnowEater robot is planned to weigh 50 kg or less. In 2010, a snow-intake system was fitted to SnowEater [5] that enables snow removal using low auger and traveling speeds while following a line path. In 2012, a navigation system based on a low-cost camera was introduced [6]. The control system was designed using linear feedback to ensure paths were followed accurately. However, due to the reduced localization accuracy, the system did not prove to be reliable. Among the many challenges in the realization of autonomous snow-removal robots, is the navigation and localization issue, which currently remains unsolved.
An early work on autonomous motion on snow is presented in [7]; in which four cameras and a scan laser were used for simultaneous localization and mapping (SLAM); and long routes in polar environments were successfully traversed. The mobility of several small tracked vehicles moving in natural and deep-snow environments is discussed in [8,9]. The terramechanics theory for motion on snow presented in [10] can be applied to improve the performance of robots moving on snow.
The basic motion models presented in [11,12,13,14,15,16,17,18] can be used for motion control. In addition, Gonzalez et al. [19] presented a model for off-road conditions. Also, the motion of a tracked mobile robot (TMR) on snow is notably affected by slip disturbance. In order to ensure the performance of path tracking or path following control against disturbances, robust controllers [20,21] and controllers based on advanced sensors [22] exist. However, advanced feedback compensation based on precise motion modeling is not necessary for this application because strict tracking is not required in our snow-removal robot.
One of the goals of this study is to develop a simple controller without the need for precise motion modeling. Another goal is to use a low-cost vision-based navigation system that does not require a large budget or an elaborate setup in the working environment. In contrast to other existing navigation strategies that use advanced sensors, we use only one low-cost USB camera and one marker for motion controlling.
This paper presents an effective method of utilizing a simple directional controller based on a camera-marker combination. In addition, a path-following method to enhance the reliability of navigation is proposed. Although the directional controller itself does not provide asymptotic stability of the path to follow, the simplicity of the system is a significant merit.
The rest of this paper is organized as follows. The task overview and prototype robot are presented in Section 2. The control law, motion and mathematical model are shown in Section 3. The experimental results are shown in Section 4 and our conclusions are given in Section 5.

2. Task Overview and Prototype Robots

The SnowEater robot prototype was developed in our laboratory. The prototype is a tracked mobile robot (TMR). Its weight currently is 26 kg, and the size is 540 mm × 740 mm × 420 mm. The main body is made of aluminum and has a flat shape for stability on irregular terrain. The intake system consists of steel and aluminum conveying screws that collect and compact snow at a low rotation speed. Two 10 W DC motors are used for the tracks, and two 15 W DC motors are used for the screws.
The front screw collects snow when the robot is moving with slow linear motion. The two internal screws compact the snow into blocks for easy transportation. Once the snow is compacted, the blocks are dropped out from the rear of the robot. In addition, the intake system compacts the snow in front of the robot reducing the influence of the track sinkage. Since the robot requires a linear motion to collect snow [5], the path to follow consists of line segments that cover the working area. In our plan, another robot carries the snow blocks to a storage location. Figure 1 shows the prototype SnowEater robot, and Figure 2 shows the line arrangement of the path-following method.
Figure 1. SnowEater robot prototype.
Figure 1. SnowEater robot prototype.
Robotics 04 00120 g001
Figure 2. Line paths that cover the snow-removal area.
Figure 2. Line paths that cover the snow-removal area.
Robotics 04 00120 g002

3. Image-Based Navigation System

One of the objectives of the SnowEater project is to use a low-resolution camera as the sole navigation sensor. In our strategy, a square marker is placed in the center of the working area, so as to be visible from any direction, and a camera is mounted on the robot. Robot camera-marker position/orientation is obtained by using the ARToolKit library [23]. This library uses the marker four corners position in the camera image and the information given a priori (camera calibration file and marker size) to calculate the transformation matrix between the camera and marker coordinate systems. Later the transformation matrix is used to estimate the translation components [24]. However, with low-resolution cameras, the accuracy and reliability of the measurement vary significantly, depending on the camera-marker distance. This is one of the problems associated with localization when using only vision. Gonzales et al. [25] discusses this issue and provides an innovative solution using two cameras. A solution for the localization problem using stereoscopic vision is given in [26,27], while [28] presents a solution using rotational stereo cameras. In [29], a trajectory-tracking method using an inexpensive camera without direct position measurement is presented.
Some studies have utilized different sensors such as a laser scanner to map outdoor environments [30]. Our study places a high priority on keeping the system as a whole inexpensive and simple; hence the challenge is to exploit robot performance using a low-resolution camera.
In this section, the term “recognized marker position” refers to the marker blob in the captured image, and the term “localization” means identification of the robot location with respect to the marker coordinate system.

3.1. Localization Performance Evaluation

To evaluate localization performance, outdoor and indoor experiments were carried out using a low-cost USB camera (Sony, PlayStation Eye 640 × 480 pixels) mounted on a small version of the SnowEater robot. Results with different cameras can be seen in Appendix A. The dimensions and weight of the small version are 420 mm × 310 mm × 190 mm and 2.28 kg, respectively. Because only a path-following strategy with visual feedback is considered, both robots (SnowEater and the small version) behave similarly. Figure 3 shows the small version of the SnowEater robot.
Figure 3. Monochromatic square marker and small version of the SnowEater robot.
Figure 3. Monochromatic square marker and small version of the SnowEater robot.
Robotics 04 00120 g003
The ARToolKit library uses monochromatic square markers to calculate the camera-marker distance and orientation, and the system can be used under poor lighting conditions [31]. Our objective is to exploit the library localization merits and performance with low-resolution cameras. Localization control calculations and track commands are completed every 300 ms. Following the results presented in [6] and Appendix A, a monochromatic square marker measuring 100 × 100 mm provides enough accuracy to be used in our application; hence the following experiments are done with this marker.
The camera is mounted on the robot facing toward the marker. The robot was placed above the X-axis of the marker-based coordinate system at different xn positions for 1 min. The outdoor experiment was carried out in snow environments; with air and snow temperatures of −5 °C and −1 °C, respectively. Figure 4 shows the experimental setup.
Figure 4. Experimental setup and different camera-marker position during the outdoor experiments.
Figure 4. Experimental setup and different camera-marker position during the outdoor experiments.
Robotics 04 00120 g004
Figure 5 shows the results for the robot position.
Figure 5. Localization position vs. actual robot-marker distance.
Figure 5. Localization position vs. actual robot-marker distance.
Robotics 04 00120 g005
The results show that the robot x-coordinate is more reliable compared to the y-coordinate. The average values are stable within 800 mm. However, to use the average value the robot must remain in a static state. For motion-state applications, the x-value (camera-marker distance) is more reliable because its variance is smaller than the y-value.
Figure 6 shows the orientation angle experimental results. Since the robot was oriented toward the marker, the angle is 0°.
Figure 6. Orientation angle vs. robot-marker distance.
Figure 6. Orientation angle vs. robot-marker distance.
Robotics 04 00120 g006
The orientation angle variance increases with the camera-marker distance, and its application is limited to the marker proximity (x < 300 mm) when using a 100 × 100 mm marker.
A simple way to obtain the camera-marker direction is to count the number of pixels between the center of the marker blob and the image center in the camera image. These pixel counts are converted into degrees by a simple pixel-degrees relation. This relation is found by finding the number of pixel counts related to a fixed angle. Figure 7 and Figure 8 show the results.
Figure 7. Recognized marker position in the camera image.
Figure 7. Recognized marker position in the camera image.
Robotics 04 00120 g007
Figure 8. Recognized marker position vs. robot-marker distance.
Figure 8. Recognized marker position vs. robot-marker distance.
Robotics 04 00120 g008
Figure 8 shows the results for the recognized marker position and, as can be seen, there is not a large variance, even when x = 1400 mm. Therefore, the data are reliable and can be used during motion.
In contrast to the orientation results shown in Figure 6, as the camera-marker distance becomes larger, the recognized marker position error becomes smaller. This characteristic is very useful for our research because the relative angle between the marker direction and robot forward direction can be obtained directly from the image.
In summary, from the controlling perspective, the recognized marker position is reliable when the camera-marker distance is large. When the camera-marker distance is small, localization data can be used. The x-coordinate of the camera-marker distance can be obtained more accurate than the y-coordinate. Based on these properties, a navigation method for distant and vicinity regions can be created. Section 3.2 describes a novel navigation algorithm based on these properties.

3.2. Motion Model

Although the SnowEater is a TMR, a differential-drive robot model [11,12,13,14,15,16,17,18] is used in the motion algorithm.
Figure 9. Motion model of the tracked robot.
Figure 9. Motion model of the tracked robot.
Robotics 04 00120 g009
Using the marker coordinate system, the robot coordinates are defined as shown in Figure 9. The longitudinal and angular velocities of the robot, v and φ ˙ , are related to the right and left track velocities v r and v l , respectively, by:
v = v r + v l 2
φ ˙ = v r v l L
where L represents the distance between the left and right tracks. Among these values, φ ˙ is used as the control input signal for path following. Velocity v is selected in relation to the snow-processing mechanism of the SnowEater robot [5].
Using the robot orientation angle φ , the robot velocity in the Cartesian coordinates is expressed as
y ˙ =   v s i n   φ x ˙ = v   c o s   φ
By differentiating y = r   s i n   θ , x = r   c o s   θ and using Equation (3), the velocity in polar coordinates is expressed as:
r ˙ = v   c o s ( θ φ ) θ ˙ = v r s i n ( θ φ )
where r is the distance from the marker. The relative angle θ φ   ( = α ) corresponds to the recognized marker position, and it is used in the motion controller described below.

3.3. Recursive Back-Forward Motion

A recursive back-forward motion was created to cover the working area. Figure 10 shows the travel route of the robot. The robot trajectory consists of two motions: Motion 1 follows a linear path, while Motion 2 is used to connect two consecutive linear paths through a curved motion in the vicinity region.
Figure 10. Recursive back-forward algorithm.
Figure 10. Recursive back-forward algorithm.
Robotics 04 00120 g010
Point A1 is the initial point of the task. The robot moves in the line segment from A1 to B1 using a straight motion (Motion 1). Point B1 is in the limit of the working area in the distant region. After reaching B1, the robot approaches point C1 in the vicinity region using a straight motion (Motion 1). Later, the robot travels to point A2 using a curve motion (Motion 2). Then, the robot embarks upon a new linear path.
Motion 1 control is executed using only the recognized marker position, while Motion 2 uses the distance from the marker in addition to the recognized marker position. With this combined motion, the low-resolution limitation is overcome.
The camera-marker direction is obtained from the information in the image using a geometric relation. Figure 11 show this relation, where α i m a g e , θ i m a g e , φ i m a g e are the distance in the camera image corresponding to α, θ and φ.
Figure 11. Orientation angle of the robot in the marker coordinate system and in the camera image.
Figure 11. Orientation angle of the robot in the marker coordinate system and in the camera image.
Robotics 04 00120 g011
Point C is the image center, which is on a straight heading in front of the robot. Since point P (the projective point of the robot onto the Y axis) is not recognized in the image, the orientation angle φ is not detected directly in the camera image. Hence the angles θ and φ are unknown, even in terms of pixels ( θ i m a g e or φ i m a g e ) .
However, the relative angle α between the marker location and robot direction is available in terms of pixels ( θ i m a g e φ i m a g e = α i m a g e ). In the following discussion, α i m a g e and α are not distinguished.

3.4. Control in Distant Region (Motion 1)

Assuming the track speeds v r and v l can be manipulated, the feedback controller is expressed using the angular velocity ( φ ˙ ) as the input signal. Throughout the travel, accurate localization can be expected at points A1, C1, A2, C2,…, because of the camera-marker distance.
Motion 1 navigation is executed by a simple direction controller using only the relative angle α in the feedback law.
φ ˙   : = K α = K ( θ φ )
where K is positive feedback gain. A linear path can be executed when the marker is set to the image center ( α = 0 ). Note this control does not provide asymptotic stability of the target path itself, because the tracking error from the target path cannot be included in control Equation (5) without accurate localization. The path-following error depends on the initial robot position. For this reason, the initial positioning is executed in the vicinity region. Motion 2 control is described later.
Considering the snow-removal task, this method is a reasonable solution because strict tracking is not required. Moreover, the certainty of returning to the marker position is a good advantage.
The basic property of the direction control used in Motion 1 is described below. The control purpose is to converge the relative angle ( θ φ ) to zero. This convergence can be checked by a Lyapunov function [32].
V 1 = 1 2 ( θ φ ) 2
By differentiating Equation (6) along the solution of Equations (4) and (5),
V ˙ 1 = ( θ φ ) ( θ ˙ φ ˙ ) = v r ( θ φ ) s i n ( θ φ ) K ( θ φ ) 2 < ( v r K ) ( θ φ ) 2 = 2 ( v r K ) V 1
A sufficient condition for the negative definiteness of V ˙ 1 is:
v r < K
When the distance r becomes small, the condition is not satisfied. One solution for this issue is to change the velocity as the distance gets smaller to v = r . Under this condition, Equation (7) becomes:
V ˙ 1 ( 1 K ) ( θ φ ) 2 2 ( 1 K ) V 1
If there exist ε > 0 such that 2 ( 1 K ) ε , then:
V 1 < V 1 ( 0 ) e ε t
Therefore, for any δ > 0 , there exists a time T such that for all t > T ,   θ φ < δ . The reachability of the vehicle to the marker can be understood directly from Equation (4) because the robot is oriented to the marker. The distance r decreases when | θ φ | π 2 . In appendix C, the behavior of this controller and the conventional PI controller using the available signals is shown.

3.5. Control for Switching the Target Line (Motion 2)

In Motion 2, the robot changes the target path from the current target line (path i) to the next line (path i + 1). During this phase, the robot trajectory is a curve convergent to the next straight target line (path i + 1). To achieve this, a modification is made to feedback Equation (5).
φ ˙   : = K ( α + α r e f ) = K ( θ φ + α r e f )
where α r e f is added to change the robot direction. The value of α r e f must be designed to move the robot onto the next target line. We define the value of α r e f as a function of the camera-marker distance ( α r e f = f ( r ) ) as described below.
Considering a polar coordinate system oriented to the next target line (path i + 1), the curved path is defined to satisfy the relation r = A θ , where A is a constant. This path smoothly converges to the origin of the O-YX coordinate system. Relying on the good accuracy of the camera localization within the vicinity region, the initial position is assumed to be known. Figure 12 shows this path.
Figure 12. Curved path to reach the marker.
Figure 12. Curved path to reach the marker.
Robotics 04 00120 g012
The A value is selected to create the curved path. Because the initial position is known, A can be determined. Next, α r e f is related to θ by:
α r e f + θ φ = 0
Since the robot motion follows Equation (3), the angle φ can be expressed as:
tan φ = y ˙ x ˙
if the motion satisfies the relation r = A θ
y ˙ = r ˙   s i n   θ + r   c o s   θ θ ˙ = A θ ˙ y r + r x r θ ˙
x ˙ = r ˙ cos θ r sin θ θ ˙ = A θ ˙ x r r y r θ ˙
Then, α r e f is defined by substituting φ = tan 1 y ˙ x ˙ and θ = r A into Equation (12)
α r e f = r A + tan 1 A y + r x A x r y
The parameter A has to be selected to satisfy |   α r e f | < v i s i o n   r a n g e 2 , because the marker must remain within the camera vision range. In this particular case the vision range is limited and r x (because the y value is small compared to x). This approximation makes the implementation much easier. Then Equation (16) can be approximated as:
α r e f = x A + tan 1 x A
Additionally, the minimum marker recognition distance (B) must be considered. The coordinate system is shifted from O-YX to O’-Y’X’; hence Equation (17) becomes:
α r e f = x B A + tan 1 x B A
In this method, the asymptotic stability of the path itself is not provided. The stability of the direction control used in Motion 2 is checked by using a Lyapunov-like function:
V 2 = 1 2 e 2 ,   e = ( θ φ + α r e f )
By differentiating along the solution of Equation (4) and the feedback equation:
V ˙ 2 = e ( θ ˙ φ ˙ + α ˙ r e f ) = e ( θ ˙ + α ˙ r e f K e ) = K e 2 + e ξ ξ = θ ˙ + α ˙ r e f = v r sin ( θ φ ) + α ˙ r e f  
Since V ˙ 2 < 0 for | e | > s u p | ξ | K , the error e will settle in the region {   e   |   s u p | ξ | K < | e |   } . Hence the expectable accuracy is less than s u p | ξ | K . Note that the error can be suppressed by selecting the correct parameter K . The velocity v must be small when the robot is closer to the marker to avoid the growth of ξ .

4. Experimental Results

Experimental verification was carried out using the small version of the SnowEater robot. The track motors of the small version are Maxon RE25 20 W motors. Each track speed is PI feedback controlled by using 500 ppr rotary encoders (Copal, RE30E) with 10 ms of sampling period. The control signals are sent through a USB-serial connection between an external PC and the robot microcontrollers (dsPIC30F4012). The PC has an Intel Celeron CPU B830 (1.8 GHz) processor with 4 GB of RAM. The interface is made via Microsoft Visual Studio 2010. Figure 13 shows the system diagram. The track response to different robot angular velocity commands ( φ ˙ ) can be seen in Appendix B.
Figure 13. Control system diagram for the SnowEater robot.
Figure 13. Control system diagram for the SnowEater robot.
Robotics 04 00120 g013

4.1. Motion 2

In this experiment, Motion 2 navigation is tested. The control objective is to move the robot to the target line. The experiment is carried out using two different conditions: (1) a hard floor, and (2) a slippery floor of 6 mm polystyrene beads. Figure 14 shows the experimental setups.
Figure 14. Motion 2 experimental setup on a hard floor, and on a slippery floor of polystyrene beads.
Figure 14. Motion 2 experimental setup on a hard floor, and on a slippery floor of polystyrene beads.
Robotics 04 00120 g014
The robot position results are obtained using a high-resolution camera mounted on the top of the robot. The robot and marker locations are generated by processing the captured image. The initial position was xo = 1200 mm, yo = 100 mm. Figure 15 and Figure 16 show the experimental results.
Figure 15. Robot position throughout the experiment.
Figure 15. Robot position throughout the experiment.
Robotics 04 00120 g015
Figure 16. α r e f experimental results when Motion 2 is executed.
Figure 16. α r e f experimental results when Motion 2 is executed.
Robotics 04 00120 g016
Figure 15 shows a comparison of the Motion 2 experiment on a hard floor and on polystyrene beads repeated five times.
As Figure 16 shows, direction control for the angle α is accomplished.
Because the feedback does not provide asymptotic stability of the path to follow, deviations occur in each experiment. If the final position is not accurate, tracking of the path generated by Motion 1 cannot be achieved. If the final error is outside the allowable range, a new curve using the Motion 2 control needs to be generated again. Due to the marker proximity, tracking of the new path is more accurate.

4.2. Motion 1 in Outdoor Conditions

To confirm the applicability in snow environments of this path-following strategy, outdoor experiments were carried out. These experiments were conducted at different times of the day under different lighting conditions. Figure 17 shows the setup of the test area.
Figure 17. Test area setup.
Figure 17. Test area setup.
Robotics 04 00120 g017
The test area measured 3000 mm × 1500 mm. The snow on the ground was lightly compacted, the terrain was irregular, and conditions were slippery. For the first experiment, the snow temperature was −1.0 °C and the air temperature is −2.4 °C with good lighting conditions. Figure 18 shows the experimental results for Motion 1 in outdoor conditions. In both experiments, the robot was oriented toward the marker and the path to follow was 3000 mm long. The color lines in the figures highlight the path followed.
Figure 18. Motion 1 experimental results on snow in outdoor conditions.
Figure 18. Motion 1 experimental results on snow in outdoor conditions.
Robotics 04 00120 g018
As can be seen in Figure 18, the robot follows a straight-line path while using Motion 1.

4.3. Motion 2 in Outdoor Conditions

Figure 19 shows the results for the Motion 2 experiment conducted under the same experimental conditions as for the Motion 1 experiment. The color lines highlight the previous path (brown), the path followed (red), and the next path (white).
In the left-hand photograph in Figure 19, the next path is in front of the marker. In the right-hand photograph, the next path has an angle with respect to the marker.
Figure 19. Motion 2 on snow.
Figure 19. Motion 2 on snow.
Robotics 04 00120 g019

4.4. Recursive Back-Forward Motion

Figure 20 shows the results for the recursive back-forward motion in indoor conditions on a hard floor and on snow. The Motion 2 region is 1200 mm. The test travelled distance is 2000 mm.
Figure 20. Recursive back-forward motion experimental results in indoor conditions.
Figure 20. Recursive back-forward motion experimental results in indoor conditions.
Robotics 04 00120 g020aRobotics 04 00120 g020b
Table 1. Recursive back-forward motion experimental results in indoor conditions.
Table 1. Recursive back-forward motion experimental results in indoor conditions.
On Hard FloorOn Snow
Reference angle between paths4.5°4.5°
Mean angle between paths4.0°4.8°
Max. angle between paths4.6 °6.9 °
Min. angle between paths3 °1.6 °
Angle between paths deviation 0.49°1.54°
The robot covered the testing area using the recursive back-forward motion. For a reference angle of 4.5° between line paths, the average results are: 4° for motion on hard floor and 4.8° for motion on compacted snow. The number of paths in the experiment on snow is different because the test area on snow is smaller than the test area on hard floor. Although the motion performance on snow is reduced (compared to the motion on a hard floor), the robot returned to the marker and covered the area. Table 1 shows the results of Figure 20.
In snow removal and cleaning applications, full area coverage rather than strict path-following is required. Therefore, this method can be applied for such tasks.
Figure 21 shows the recursive back-forward motion experiment on snow in outdoor conditions. The snow temperature was −1.0 °C and the air temperature was −1.3 °C.
Figure 21. Recursive back-forward motion experimental results in outdoor conditions.
Figure 21. Recursive back-forward motion experimental results in outdoor conditions.
Robotics 04 00120 g021
Figure 21 results confirm the applicability of the method under outdoor snow conditions. Due to poor lighting, the longest distance traveled was 1500 mm. For longer routes, a bigger marker is required and to improve the library performance in poor lighting conditions, the luminous markers shown in [31] can be used.

5. Conclusions

In this paper, we presented a new approach for a snow-removal robot utilizing a path-following strategy based on a low-cost camera. Using only a simple direction controller, an area with radially arranged line segments was swept. The advantages of the proposed controller are its simplicity and reliability.
The required localization values, as measured by the camera, are the position and orientation of the robot. These values are measured in the marker vicinity in a stationary condition. During motion or in the marker distant region, only the position of the marker blob in the captured image is used.
With 100 mm × 100 mm square monochromatic marker, a low-cost USB camera (640 × 480 pixels), and the ARToolKit library, the robot followed the radially arranged line paths. For a reference angle of 4.5° between line paths, the average results are: 4° for motion on hard floor and 4.8° for motion on compacted snow. With good lighting conditions, 3000 mm long paths were traveled. We believe with this method our intended goal area can be covered.
Although the asymptotic stability of the path is not provided, our method presents a simple and convenient solution for SnowEater motion in small areas. The results showed that the robot can cover all the area using just one landmark. Finally, because the algorithm grants area coverage, it can be applied not only to snow-removal but also to other tasks such as cleaning.
In the future, the method will be evaluated by using the SnowEater prototype in outdoor snow environments. Also, the use of natural passive markers (e.g., houses and trees) will be considered.

Appendix A

The camera, marker size and the pattern (inside the marker) selection was done after the following indoor experiments. In these experiments, three different parameters were considered: the camera, the marker size and the pattern inside the marker.
In the first experiment, three different cameras (Buffalo USB camera, SONY PlayStation Eye and Logicool HD Pro webcam) were tested. The cameras were set to 30 fps and 640 × 480 pixels. The marker size was 100 × 100 mm and the pattern inside the marker was the same. Figure A1 shows the results.
In the second experiment, the pattern inside the marker was changed. The camera was Sony PlayStation Eye set to 30 fps and 640 × 480 pixels. The size for each marker was 100 × 100 mm. Figure A2 shows these results.
In the third experiment, the marker size was changed. The camera was Sony PlayStation Eye set to 30 fps and 640 × 480 pixels. The pattern inside each marker was the same. Figure A3 shows these results.
Figure A1. Localization and recognized marker position errors for different cameras.
Figure A1. Localization and recognized marker position errors for different cameras.
Robotics 04 00120 g022
Figure A2. Localization and recognized marker position errors for different patterns.
Figure A2. Localization and recognized marker position errors for different patterns.
Robotics 04 00120 g023
Figure A3. Localization and recognized marker position errors for different marker size.
Figure A3. Localization and recognized marker position errors for different marker size.
Robotics 04 00120 g024
These are the conclusions from these experiments: The camera hardware has not direct influence in the localization results when set to the same resolution and frame rate. The pattern inside the marker has not relevant influence in the localization results. The marker size is the most relevant factor in the localization accuracy. A larger marker will have a large accurate region but this region is in the vicinity of the marker and proportional to the marker size.

Appendix B

An experiment to check the track response to different robot angular velocity commands ( φ ˙ ) was done. The floor was hard and dry. In the test, the robot rotated 2 revolutions (2π rad) for a given angular velocity, thus the test duration was shorter for higher rotation ratio. Figure B1 shows these results.
Figure B1. Track response to different angular velocity commands.
Figure B1. Track response to different angular velocity commands.
Robotics 04 00120 g025

Appendix C

An experiment to compare our controller (using the recognized marker position signal) and the conventional path following linear feedback [12] with PI controller (using position/orientation signals from ARToolKit) was done. The control equation of the latter is:
φ ˙ = u + u d t ,   u = K 1 γ K 2 l
where γ and l represents the relative angle and distance from the target trajectory, respectively, and K 1 , K 2 are constant feedback gains.
The marker size was 100 × 100 mm and the camera in use is Sony PlayStation Eye. The floor was hard and dry. Figure C1 show the results.
Figure C1. Direction controller (recognized marker position) vs. PI controller (inaccurate localization signals).
Figure C1. Direction controller (recognized marker position) vs. PI controller (inaccurate localization signals).
Robotics 04 00120 g026
These results show the behavior of both controllers. In the PI01, PI03 and PI05 experiments the robot lost the marker from the camera vision range; hence the robot stopped.
The PI controller works only in the marker vicinity where the localization signals are accurate. Once the camera-marker distance increases the controller does not work properly due to the poor accuracy of the data. In contrast the direction controller remains stable even if the camera-marker distance increases.

Author Contributions

These authors contributed equally to this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Irobot home page. Available online: http://www.irobot.com/For-the-Home/Vacuum-Cleaning/Roomba (accessed on 14 November 2014).
  2. Sato, T.; Makino, T.; Naruse, K.; Yokoi, H.; Kakazu, Y. Development of the Autonomous Snowplow. In Proceedings of ANNIE 2003, St. Louis, MI, USA, 2–5 November 2003; pp. 877–882.
  3. Suzuki, S.; Kumagami, H.; Haniu, H.; Miyakoshi, K.; Tsunemoto, H. Experimental study on an Autonomous Snowplow with a Vision Sensor. Bull. Univ. Electro-Commun. 2003, 15, 2–30. [Google Scholar]
  4. Industrial Research Institute of Niigata Prefecture. “Yuki-Taro” the autonomous snowplow; Report on the Prototype Robot Exhibition EXPO 2005 AICHI; Industrial Research Institute of Niigata Prefecture: Niigata, Japan, March 2006; ( In Japanese). [Google Scholar]
  5. Mitobe, K.; Rivas, E.; Hasegawa, K.; Kasamatsu, R.; Nakajima, S.; Ono, H. Development of the Snow Dragging Mechanism for an Autonomous Snow Eater Robot. In Proceedings of the 2010 IEEE/SICE International Symposium on System Integration, Sendai, Japan, 22–23 December 2010; pp. 73–77.
  6. Rivas, E.; Mitobe, K. Development of a Navigation System for the SnowEater Robot. In Proceedings of the 2012 IEEE/SICE International Symposium on System Integration, Fukuoka, Japan, 16–18 December 2012; pp. 378–383.
  7. Moorehead, S.; Simmons, R.; Apostolopoulos, D.; Whittaker, W. Autonomous Navigation Field Results of a Planetary Analog Robot in America. In Proceedings of the International Symposium on Artificial Intelligence, Robotics and Automation in Space, Noordwijk, The Netherlands, 1–3 June 1999.
  8. Lever, J.; Denton, D.; Phetteplace, G.; Wood, S.; Shoop, S. Mobility of a lightweight tracked robot over deep snow. J. Terramech. 2006, 43, 527–551. [Google Scholar] [CrossRef]
  9. Lever, J.; shoop, S.; Bernhard, R. Design of lightweight robots for over-snow mobility. J. Terramech. 2009, 46, 67–74. [Google Scholar] [CrossRef]
  10. Wong, J.Y. Theory of Ground Vehicles, 4th ed.; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2008; Chapter 6, Steering of tracked vehicles; pp. 419–449. [Google Scholar]
  11. Dudek, G.; Jenkin, M. Fundamental Problems. In Computational Principles of Mobile Robotics, 2nd ed.; Cambridge University Press: New York, NY, USA, 2010; pp. 19–25. [Google Scholar]
  12. Dudek, G.; Jenkin, M. Mobile Robot Hardware. In Computational Principles of Mobile Robotics, 2nd ed.; Cambridge University Press: New York, NY, USA, 2010; pp. 36–43. [Google Scholar]
  13. Dudek, G.; Jenkin, M. Visual Sensors and Algorithms. In Computational Principles of Mobile Robotics, 2nd ed.; Cambridge University Press: New York, NY, USA, 2010; pp. 123–149. [Google Scholar]
  14. Dudek, G.; Jenkin, M. Pose Maintenance and Localization. In Computational Principles of Mobile Robotics, 2nd ed.; Cambridge University Press: New York, NY, USA, 2010; pp. 240–243. [Google Scholar]
  15. Canudas, C.; Siciliano, B.; Bastin, G. Modelling and structural properties. In Theory of Robot Control, 1st ed.; Springer: London, 1997; p. 278. [Google Scholar]
  16. Canudas, C.; Siciliano, B.; Bastin, G. Nonlinear feedback control. In Theory of Robot Control, 1st ed.; Springer: London, 1997; pp. 331–341. [Google Scholar]
  17. Morin, P.; Samson, C. Motion control of wheeled mobile robots. In Springer Handbook of Robotics; Siciliano, K., Ed.; Springer-Verlag: Berlin Heidelberg, Germany, 2008; pp. 799–826. [Google Scholar]
  18. Campion, G.; Chung, W. Wheeled Robots. In Springer Handbook of Robotics; Siciliano, K., Ed.; Springer-Verlag: Berlin Heidelberg, Germany, 2008; pp. 391–410. [Google Scholar]
  19. Gonzalez, R.; Rodriguez, F.; Guzman, J. Modelling Tracked Robots in Planar Off-Road Conditions. In Autonomous Tracked Robots in Planar Off-Road Conditions; Springer International Publishing: Cham, Switzerland, 2014; pp. 11–33. [Google Scholar]
  20. Normey-Rico, J.; Alcala, I.; Gomez-Ortega, J.; Camacho, E. Mobile robot path tracking using robust PID controller. Control Eng. Pract. 2001, 9, 1209–1214. [Google Scholar] [CrossRef]
  21. Coelho, P.; Nunes, U. Path-following Control of Mobile Robots in Presence of Uncertainties. IEEE Trans. Robot. 2005, 21, 252–261. [Google Scholar] [CrossRef]
  22. Low, C.B.; Wang, D.W. GPS-Based Path Following Control for a Car-Like Wheeled Mobile Robot With Skidding and Slipping. IEEE Trans. Control Syst. Technol. 2008, 16, 340–347. [Google Scholar] [CrossRef]
  23. ARToolKit home page. Available online: http://www.hitl.washington.edu/artoolkit/ (accessed on 14 November 2014).
  24. Kato, H.; Bullinghurst, M. Marker tracking and HMD calibration for a video-based augmented reality conferencing system. In Proceedings of the 2nd IEEE and ACM International Workshop, San Francisco, CA, USA, 20–21 October1999; pp. 85–94.
  25. Gonzalez, R.; Rodriguez, F.; Guzman, J.; Pradalier, C.; Siegward, S. Control of off-road mobile robots using visual odometry and slip compensation. Adv. Robot. 2013, 27, 893–906. [Google Scholar] [CrossRef]
  26. Mirkhani, M.; Forsati, R.; Shahri, A.M.; Moayedikia, A. A novel efficient algorithm for mobile robot localization. Robot. Auton. Syst. 2013, 61, 920–931. [Google Scholar] [CrossRef]
  27. Savino, S. An algorithm for robot motion detection by means of a stereoscopic vision system. Adv. Robot. 2013, 27, 981–991. [Google Scholar] [CrossRef]
  28. Yoshida, T.; Shiozawa, H.; Fukao, T.; Yokokohji, Y. Dense 3D Reconstruction Using a Rotational Stereo Camera and a Correspondence with Epipolar Transfer. J. Robot. Soc. Jpn. 2013, 31, 1019–1027. (In Japanese) [Google Scholar] [CrossRef]
  29. Wang, K.; Liu, Y.; Li, L. Visual Servoing Trajectory Tracking of Nonholonomic Mobile Robots Without Direct Position Measurement. IEEE Trans. Robot. 2014, 30, 1026–1035. [Google Scholar] [CrossRef]
  30. Shiozawa, H.; Yoshida, T.; Fukao, T.; Yokokohji, Y. 3D Reconstruction using Airbone Velodyne Laser Scanner. J. Robot. Soc. Jpn. 2013, 31, 992–1000. (In Japanese) [Google Scholar] [CrossRef]
  31. Ishida, M.; Shimonomura, K. Marker Based Camera Pose Estimation for Underwater Robots. In Proceedings of the 2012 IEEE/SICE International Symposium on System Integration, Fukuoka, Japan, 16–18 December 2012; pp. 629–634.
  32. Haddad, W.M.; Chellaboina, V. Stability Theory for Nonlinear Dynamical Systems. In Nonlinear Dynamical Systems and Control; Princenton University Press: New Jersey, USA, 2008; pp. 135–182. [Google Scholar]

Share and Cite

MDPI and ACS Style

Rivas, E.; Komagome, K.; Mitobe, K.; Capi, G. Image-Based Navigation for the SnowEater Robot Using a Low-Resolution USB Camera. Robotics 2015, 4, 120-140. https://doi.org/10.3390/robotics4020120

AMA Style

Rivas E, Komagome K, Mitobe K, Capi G. Image-Based Navigation for the SnowEater Robot Using a Low-Resolution USB Camera. Robotics. 2015; 4(2):120-140. https://doi.org/10.3390/robotics4020120

Chicago/Turabian Style

Rivas, Ernesto, Koutarou Komagome, Kazuhisa Mitobe, and Genci Capi. 2015. "Image-Based Navigation for the SnowEater Robot Using a Low-Resolution USB Camera" Robotics 4, no. 2: 120-140. https://doi.org/10.3390/robotics4020120

APA Style

Rivas, E., Komagome, K., Mitobe, K., & Capi, G. (2015). Image-Based Navigation for the SnowEater Robot Using a Low-Resolution USB Camera. Robotics, 4(2), 120-140. https://doi.org/10.3390/robotics4020120

Article Metrics

Back to TopTop