You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

19 June 2017

Localization Framework for Real-Time UAV Autonomous Landing: An On-Ground Deployed Visual Approach

,
,
,
and
1
College of Mechatronics and Automation, National University of Defense Technology, Changsha 410073, China
2
Naval Academy of Armament, Beijing 100161, China
3
Institute of Technical Aspects of Multimodal Systems(TAMS), Department of Computer Science, University of Hamburg, 22527 Hamburg, Germany
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue UAV or Drones for Remote Sensing Applications

Abstract

One of the greatest challenges for fixed-wing unmanned aircraft vehicles (UAVs) is safe landing. Hereafter, an on-ground deployed visual approach is developed in this paper. This approach is definitely suitable for landing within the global navigation satellite system (GNSS)-denied environments. As for applications, the deployed guidance system makes full use of the ground computing resource and feedbacks the aircraft’s real-time localization to its on-board autopilot. Under such circumstances, a separate long baseline stereo architecture is proposed to possess an extendable baseline and wide-angle field of view (FOV) against the traditional fixed baseline schemes. Furthermore, accuracy evaluation of the new type of architecture is conducted by theoretical modeling and computational analysis. Dataset-driven experimental results demonstrate the feasibility and effectiveness of the developed approach.

1. Introduction

Over the past few decades, the application of unmanned aircraft has increased enormously in both civil and military scenarios. Although aerial robots have successfully been implemented in several applications, there are still new research directions related to them. Floreano [1] and Kumar et al. [2] outlined the opportunities and challenges of this developing field, from the model design to high-level perception capability. All of these issues are concentrating on improving the degree of autonomy, which supports that UAVs continue to be used in novel and surprising ways. No matter whether fixed-wing or rotor-way platforms, a standard fully-unmanned autonomous system (UAS) involves performs takeoffs, waypoint flight and landings. Among them, the landing maneuver is the most delicate and critical phase of UAV flights. Two technical reports [3] argued that nearly 70% of mishaps of Pioneer UAVs were encountered during the landing process caused by human factors. Therefore, a proper assist system is needed to enhance the reliability of the landing task. Generally, two main capabilities of the system are required. The first one is localization and navigation of UAVs, and the second one is generating the appropriate guidance command to guide UAVs for a safe landing.
For manned aircraft, the traditional landing system uses a radio beam directed upward from the ground [4,5]. By measuring the angular deviation from the beam through onboard equipment, the pilot knows the perpendicular displacement of the aircraft in the vertical channel. For the azimuth information, additional equipment is required. However, due to the size, weight and power (SWaP) constraints, it is impossible to equip these instruments in UAV. Thanks to the GNSS technology, we have seen many successful practical applications of autonomous UAVs in outdoor environments such as transportation, aerial photography and intelligent farming. Unfortunately, in some circumstances, such as urban or low altitude operations, the GNSS receiver antenna is prone to lose line-of-sight with satellites, making GNSS unable to deliver high quality position information [6]. Therefore, autonomous landing in an unknown or global navigation satellite system (GNSS)-denied environment is still an open problem.
The visual-based approach is an obvious way to achieve the autonomous landing by estimating flight speed and distance to the landing area, in a moment-to-moment fashion. Generally, two types of visual methods can be considered. The first category is the vision-based onboard system, which has been widely studied. The other is to guide the aircraft using a ground-based camera system. Once the aircraft is detected by the camera during the landing process, its characteristics, such as type, location, heading and velocity, can be derived by the guidance system. Based on this information, the UAV could align itself carefully towards the landing area and adapt its velocity and acceleration to achieve safe landing. In summary, two key elements of the landing problem are detecting the UAV and its motion, calculating the location of the UAV relative to the landing field.
To achieve better performance in GNSS-denied environments, some other types of sensors, such as laser range finders, millimeter wavelength radar, have been explored for UAV autonomous landing. Swiss company RUAG (Bern, Switzerland) solved the landing task by the OPATS (object position and tracking sensor) [7]. Figure 1a presents this laser-based automatic landing system, the infrared laser beam of which is echoed back from a passive and optionally heated retro reflector on the aircraft. This system could measure the position of approaching aircraft around 4000 m. Moreover, the Sierra Nevada Corporation provides an alternative to the laser-based method. They developed the UAS common automatic recovery system (UCARS) [8] based on millimeter wavelength ground radar for MQ-8B Fire Scout autonomous landing, as shown in Figure 1b. Benefiting from the short bandwidth, UCARS provides precision approach (within 2.5 cm) in adverse weather condition. While those solutions are effective, they require the use of radar or laser emissions, which can be undesirable in a tactical situation. Furthermore, the limited payload of a small UAV constrains the onboard modules.
Figure 1. (a) The object position and tracking sensor (OPTAS) system is mounted on the tripod. Adapted from [7]; (b) UAS common automatic recovery system (UCARS) for MQ-8B autonomous landing. Adapted from [8].
Motivated by these mentioned challenges, we propose and develop a novel on-ground deployment of the visual landing system. In this paper, we mainly focus on the localization and navigation issue and try to improve the navigation accuracy and robustness. The essential contributions of this work are as follows: (1) an extendable baseline and wide-angle field of view (FOV) vision guidance system is developed by using a physically-separated and informationally-connected deployment of the two PTUs on both sides of the runway; (2) localization error and its transferring mechanism in practical situations are unveiled with both theoretical and computational analyses. In particular, the developed approach is experimentally validated with fair accuracy and better performance in timeliness, as well as practicality against the previous works.
The remainder of this paper is organized as follows. Section 2 briefly reviews the related works. In Section 3, the architecture of the on-ground deployed stereo system is proposed and designed. Section 4 conducts the accuracy evaluation, and its transferring mechanism is conducted through theoretical and computational analysis. Dataset-driven validation is followed in Section 5. Finally, concluding remarks are presented in Section 6.

3. System Architecture and Deployment

In this section, we introduce the theoretical model for the ground-to-air visual system. We first recap the traditional stereo vision model, which has a limited baseline, restraining the detection distance. To enlarge the system working boundary, we setup the camera and other sensor modules on the two separated PTUs and then calculate the target according to the image information and rotation angle from PTU. Each vision unit works independently and transfers the results of image processing and PTU status to the navigation computer, which calculates the estimated relative position of the UAV. The architecture of the ground stereo vision system is shown in Figure 3.

3.1. Fundamental Principles of Ground-Based Stereo Systems

The standard camera model is a pin-hole camera model. The coordinate of the target M is ( x , y , z ) , and its position on the image plane is ( u , v ) . The camera focus is f; then, the relationship of the coordinate between the 3D world and 2D image plane can be calculated by:
λ u v f = x y z
where λ is the scale factor.
Although the above model is simple, it could be helpful to estimate the theoretical camera lens according to the expected distance and resolution or to measure the target size roughly based on the pixel length on the image plane. Let the width and height of the target be W and H; the distance between the camera and target be L; the target projection on image plane be w and h; the relationship between them is:
f = w L W f = h L H
We define the coordinates of the left and right navigation module as shown in Figure 4b. When the optical axes of these two cameras are parallel, we could calculate the target in 3D space by:
x y z = b d u l u r f
where b is the baseline and d = u l - u r is the pixel disparity, as shown in Figure 4a. Even though some calibration methods could manage the axes’ nonparallel situation, it is still difficult to calculate the system correctly, as the baseline is large.
Figure 4. (a) Theoretical stereo vision model; (b) theoretical optical model. The target M is projected at the center of the optical axis.

3.2. Separated Long Baseline Deployment

In order to detect the target at long distance, a large baseline, more than 5 m, is required. Benefiting the camera assembled on the PTU separately, we could switch the baseline freely according to the expected detection distance and target size.
In this paper, we assumed that the world coordinate system ( X , Y , Z ) is located on the origin of the left vision unit, the rotation center of the PTU. For the sake of simplicity, the camera is installed on the PTU in the way that the axes of the camera frame are parallel to those of the PTU frame. The origins of these two frames are close. Therefore, it can be assumed that the camera frame coincides with the body frame. Figure 4b reveals the theoretical model for visual measurement. After installing the right camera system on the X-axis, the left and right optical center can be expressed as O l and O r , respectively. Then, the baseline of the optical system is O l O r , whose distance is D. Considering the center of mass of the UAV as a point M, O l M and O r M illustrate the connections between the each optical center and the UAV. In addition, ϕ l , ϕ r , ψ l , ψ r denote the tilt and pan angle on both sides. Therefore, we define ϕ l = 0 , ϕ r = 0 , ψ l = 0 and ψ r = 0 , as the PTU is set to the initial state, i.e., the optical axis parallel to the runway; the measurement of the counterclockwise direction is positive.
Since the point M does not coincide with the principle point, which is the center of the image plane, the pixel deviation compensation in the longitudinal and horizontal direction should be considered. As shown in Figure 5, we calculate pixel deviation compensation on the left side by:
ψ c l = arctan ( u - u 0 ) d u f ϕ c l = arctan ( v - v 0 ) cos ψ c l d v f
where the optical point is o ( u o , v o ) , d u and d v are the pixel length of the u- and v-axis in image plane and f is the focus. The current PTU rotation angle can be directly obtained through the serial ports during the experiments. Let ϕ p l and ψ p l be the left pan and tilt angle separately. Then, the total pan and tilt angle on the left side can be detailed as:
ϕ l = ϕ c l + ϕ p l ψ l = ψ c r + ψ p r
Figure 5. The geometry of one PTUwith respect to the optical center and the image plane.
For the other side, we could also calculate the angle in the same way.
The world coordinates of point M is ( x M , y M , z M ) R 3 . Point N is the vertical projection of point M on the X O Y plane, and N A is perpendicular to the X-axis. If we define N A = h , the following navigation parameters can be obtained:
x M = h tan ψ l = D tan ψ l tan ψ l - tan ψ r y M = h = D tan ψ l - tan ψ r z M = h tan ϕ l cos ψ l = D tan ϕ l cos ψ l ( tan ψ l - tan ψ r )
Furthermore, errors in the internal and external camera calibration parameters marginally affect some of the estimates: the x-position and z-position, in particular.

4. Accuracy Evaluation of the On-Ground Stereo System

4.1. Theoretical Modeling

We are now in the position to analyze the error related to the PTU rotation angle. The discussion was first presented in our previous works [15]. According to Equation (6), the partial derivatives of each equation with respect to the pan angle and the tilt angle are denoted in the following way,
x M ψ l = D tan ψ r cos 2 ψ l ( tan ψ l - tan ψ r ) 2 x M ψ r = D tan ψ l cos 2 ψ r ( tan ψ l - tan ψ r ) 2
y M ψ l = D cos 2 ψ l ( tan ψ l - tan ψ r ) 2 y M ψ r = D cos 2 ψ r ( tan ψ l - tan ψ r ) 2
z M ϕ l = D cos ψ l cos 2 ϕ l ( tan ψ l - tan ψ r ) z M ψ l = D tan ϕ l ( cos ψ l + sin ψ l tan ψ r ) cos 2 ψ r ( tan ψ l - tan ψ r ) 2 z M ψ r = D tan ϕ l cos ψ l cos 2 ψ r ( tan ψ l - tan ψ r ) 2
To analyze the influence of the error from the angle, we define the gradient of the world coordinate as:
x M ( ψ l , ψ r ) : = x M ψ l ( ψ l , ψ r ) , x M ψ r ( ψ l , ψ r )
y M ( ψ l , ψ r ) : = y M ψ l ( ψ l , ψ r ) , y M ψ r ( ψ l , ψ r )
z M ( ψ l , ψ r ) : = z M ψ l ( ψ l , ψ r ) , z M ψ r ( ψ l , ψ r )
In this case, simulation is needed to evaluate the behavior of our visual system. Figure 6a–c is the vector field distribution of x M ( ψ l , ψ r ) , y M ( ψ l , ψ r ) and z M ( ψ l , ψ r ) , which give us an intuitive result under different types of errors. The length of each vector describes the strength at a specific point; the direction along the vector points to the direction of the fastest error increase. However, only when y M 0 (the aircraft is in front of two cameras), the area ψ l - ψ r > 0 has the physics meaning. Figure 7a shows that x M has a significant variation when ψ l is approximate to ψ r , namely the optical axes are nearly parallel. Further, y M and z M have similar variations. Considering the general working status of the ground-based system, we mainly focus on the second quadrant of the aforementioned vector fields as shown in Figure 7a–c. In these areas, there are slight variations that theoretically demonstrate the feasibility of the system.
Figure 6. Vector field distribution of (a) x M ( ψ l , ψ r ) , (b) y M ( ψ l , ψ r ) and (c) z M ( ψ l , ψ r ) .
Figure 7. (ac) Vector field distribution of x M ( ψ l , ψ r ) , y M ( ψ l , ψ r ) , and z M ( ψ l , ψ r ) in the second quadrant; (df) gradient of X, Y, Z with θ l and θ r in the second quadrant.

4.2. Computational Analysis

In theory, O l M and O r M should intersect perfectly at one point all of the time, as shown in Figure 4b. Due to the inevitable errors from PTU rotation and tracking algorithms, we estimate the intersecting point by combing the vertical line of two different planes in space.
(1) We set ( x o l , y o l , z o l ) = ( 0 , 0 , 0 ) , and ( x o r , y o r , z o r ) = ( D , 0 , 0 ) is the optical center of each camera. Assuming that a l 0 , b l 0 , c l 0 and a r 0 , b r 0 , c r 0 , we obtain the parametric equations of lines O l M and O r M :
x - x o l a l = y - y o l b l = z - z o l c l = t l , x - x o r a r = y - y o r b r = z - z o r c r = t r ,
a l = cos ϕ l sin ψ l b l = cos ϕ l cos ψ l c l = sin ϕ l a r = cos ϕ r sin ψ r b r = cos ϕ r cos ψ r c r = sin ϕ r
where t l , t r are the parameters for the line O l M and O r M separately. Therefore, any point ( x , y , z ) on each line is usually written parametrically as a function of t l and t r :
x l = a l t l + x o l y l = b l t l + y o l z l = c l t l + z o l x r = a r l r + x o r y r = b r t r + y o r z r = c r t r + z o r
(2) In our situation, O l M and O r M are skew lines, such that these two lines are no parallel and do not intersect in 3D. Generally, the shortest distance between the two skew lines lies along the line that is perpendicular to both of them. By defining the intersection points of the shortest segment line for each line by ( x l p , y l p , z l p ) and ( x r p , y r p , z r p ) , we get the parametric equations:
x l p = a l t l + x o l y l p = b l t l + y o l z l p = c l t l + z o l x r p = a r l r + x o r y r p = b r t r + y o r z r p = c r t r + z o r
(3) Knowing the position of the intersection points on each line, the distance is calculated by the square Euclidean norm:
J = ( x l p , y l p , z l p ) - ( x r p , y r p , z r p ) 2 2
(4) By deriving the function J, we achieved the minimum distance when J t l = 0 and J t r = 0 . Then, the above functions derive the following equation:
a l 2 + b l 2 + c l 2 - ( a l a r + b l b r + c l c r ) - ( a l a r + b l b r + c l c r ) a l 2 + b l 2 + c l 2 t l t r = ( x o l - x o r ) - a l a r + ( y o l - y o r ) - b l b r + ( z o l - z o r ) - c l c r .
We could define the matrix on the left side as:
H = a l 2 + b l 2 + c l 2 - ( a l a r + b l b r + c l c r ) - ( a l a r + b l b r + c l c r ) a l 2 + b l 2 + c l 2
Considering that there is a uniqueness vertical line, so det H 0 , and the position of the target point Min the world coordinate is:
x M y M z M = w x l p y l p z l p + ( 1 - w ) x r p y r p z r p , w [ 0 , 1 ] .
where w is weight, and the other parameters are:
x l p = a l D a l ( a l 2 + b l 2 + c l 2 ) - a r ( a l a r + b l b r + c l c r ) ( a l b r - b l a r ) 2 + ( b l c r - c l b r ) 2 + ( a l c r - c l a r ) 2 y l p = b l D a l ( a l 2 + b l 2 + c l 2 ) - a r ( a l a r + b l b r + c l c r ) ( a l b r - b l a r ) 2 + ( b l c r - c l b r ) 2 + ( a l c r - c l a r ) 2 z l p = c l D a l ( a l 2 + b l 2 + c l 2 ) - a r ( a l a r + b l b r + c l c r ) ( a l b r - b l a r ) 2 + ( b l c r - c l b r ) 2 + ( a l c r - c l a r ) 2
and:
x r p = D a r a l ( a l a r + b l b r + c l c r ) - a r ( a l 2 + b l 2 + c l 2 ) ( a l b r - b l a r ) 2 + ( b l c r - c l b r ) 2 + ( a l c r - c l a r ) 2 + 1 y r p = b r D a l ( a l a r + b l b r + c l c r ) - a r ( a l 2 + b l 2 + c l 2 ) ( a l b r - b l a r ) 2 + ( b l c r - c l b r ) 2 + ( a l c r - c l a r ) 2 z r p = c r D a l ( a l a r + b l b r + c l c r ) - a r ( a l 2 + b l 2 + c l 2 ) ( a l b r - b l a r ) 2 + ( b l c r - c l b r ) 2 + ( a l c r - c l a r ) 2
The angle between the UAV landing trajectory and the runway area is usually between 3° and 7°. By considering 1 mrad normal distributed disturbance (the accuracy of the PTU is 0.006 ° ), Figure 8 illustrates measurement errors of x M , y M and z M in the case of different points ( x , y ) S , where S = { ( x , y ) | - 50 x 50 , 20 y 1000 } .
Figure 8. The baseline is 15 m, and the focus is 100 mm. The measurement errors in the X, Z and Y axis with 2000 m.
Obviously, the errors at a considerable distance are notable, but their incidence declines while the aircraft is close to the runway. When the UAV is only 100 m to the landing area, the error of altitude is about 0.02 m, which is dependable for the landing task, as shown in Figure 9. Figure 10 shows that the navigation could be improved at the same distance with a large baseline configuration (20 m).
Figure 9. The baseline is 15 m, and the focus is 100 mm. The measurement errors in the X, Z and Y axis with 280 m.
Figure 10. The baseline is 20 m, and the focus is 100 mm. The measurement errors in the X, Z and Y axis with 2000 m.
Furthermore, the errors are much smaller as the UAV lands aligned with the center line of the runway. Table 1 and Table 2 illustrate that the error varies non-linearly and decreases significantly as the target approaches the touch down point. The smaller the disturbance of the system is, the better the accuracy in each axis will be.
Table 1. Errors when the center extraction of the aircraft with 1-pixel disturbance, f = 100 mm and D = 10 m.
Table 2. Errors when the center extraction of the aircraft with 5-pixel disturbance, f = 100 mm and D = 10 m.
Different from the traditional binocular vision system, the optical axes of each vision unit are not parallel during the operation, and there is an initial offset between the camera optical center and the rotation axes of PTU. Therefore, the traditional checkerboard pattern calibration method is not sufficient and convenient to obtain the stereo system parameters for our large baseline system. To solve the calibration issue, we firstly chose the intrinsic camera model, which includes the principal point displacement, optical distortions, skew angle, etc. Each camera should be calibrated separately by the classical black-white chessboard method with the help of the calibration module from OpenCV. Secondly, we setup the setting points with the help of the differential Global Positioning System (DGPS) module and calibrate the system based on the PTU rotation angle, coordinates and the ground-truth position of the setting points.

4.3. Rotation Compensation

According to the above discussion, the 3D location estimation depends largely on the precision of the target center position in the image plane. Our previous work [16,17,20] introduced saliency-inspired and the Chan–Vese model method to track the UAV during the landing progress. Both of these approaches predict and extract the center of the UAV position without considering the PTU rotation. However, in the practical situation, the PTU might jump suddenly due to the disturbance of the control signal and the unexpected maneuver of the UAV. We define the target center position as ( x t , y t ) in the image frame, which can be predicted iteratively by:
x t = f x t - 1 - f ψ x t - 1 ψ - ϕ y t - 1 + f
y t = f y t - 1 - f ϕ x t - 1 ψ - ϕ y t - 1 + f
where ψ and ϕ are the PTU rotation angles and f is the camera focal length. The precision of the bounding box prediction (BBP) could be improved by the PTU rotation compensation.

4.4. Localization Framework

In the ensemble configuration, we separate the vehicle guidance and control into an inner loop and an outer loop, because it is a much simpler and well-tested design approach. As the inner loop controller already exists in the autopilot, we developed an efficient and robust outer navigation loop, which manages the visual information with the on-board sensors. Figure 11 presents the separated long baseline stereo localization frame.
Figure 11. PTU-based UAV localization framework.

5. Experiments and Discussion

5.1. Experiments’ Setup

For visible light camera, we selected DFK 23G445, which was developed by Imaging Source GmbH. The sensor of this camera is the Sony ICX445AQA equipped with the GigE interface, which has high data transfer rates, typically up to 1000 Mbit/s. This camera has an image resolution of 1280 × 960 with the RGB32 color model and a maximum frame rate of 30 fps. The lens of the vision system we adopted is 100 mm, and the baseline is 10 m. To extend the field of view, we adopted precision PTU to actuate the camera. PTU-D300E is a high performance product from FLIR. Its pan/tilt speeds up to 50°/s with the position resolution of 0.006°. Moreover, it is a user-programmable product integrating Ethernet and RS-232 interface. The real-time command interface supports advanced applications such as video tracking. We set up the camera on the top bracketing, and the assembled individual vision system is illustrated in Figure 3.
This experimental test-bed is a customized fixed-wing aircraft, which is a gasoline-powered radio-controlled model aircraft. The on-board autopilot allowed for the aircraft to perform simple commanded maneuvers. Our autopilot module is iFLY-F1A, and the navigation module is iFLY-G2 [21], which is a small six-DOF (degree of freedom) navigation system. This module supports real-time 3D information including attitude angle, angular rate, position, speed, acceleration, true air speed and calibrated air speed. F1A is connected with G2 through the RS-232 serial port. Table 3 lists the other technical specifications of the UAV platform. Communication is crucial in the landing navigation framework, because the relative localization is broadcast through the radio. The navigation data are sent using an advanced radio modem that transmits and receives on the 900-MHz band. The XTend RF modems support up to 22 km outdoor communication with the interface data rates from 10 bps–23,000 bps, which is sufficient to transfer GNSS data and predicted position from the ground station to the on-board navigation modem.
Table 3. The technical specifications of Pioneer.
The landing procedure was divided into four sections: (1) the UAV takeoff from the runway; (2) cruise near the landing area in a large range to test the control system; (3) cruise near the landing area in a small range; and after the UAV is locked by the visual system and received the visual references, the UAV control system was using vision-based localization data, and the GPS data was only recorded as the benchmark; (4) safely landing back on the runway.

5.2. Results and Discussion

Based on the results of the simulation, eight sets of experimental results are conduced to establish the feasibility of the proposed approach. The results are shown in Table 4. Considering the real-time capability and the precision of the target detection, we modified the original discriminative scale space tracker (DSST) [22], which additionally calculates a one-dimensional discriminative scale filter to evaluate the target size. In the realistic application, it is a very critical requirement that the lateral deviation error from the middle line of the runway and the lateral acceleration of the vehicle be perfectly eliminated to minimize the damage of the vehicle. Figure 12 illustrates the approach results. The left image shows the landing waypoints projecting on a satellite map where A is the locking point of the ground landing and B is the desired touch down point on the runway. In addition, the three 3D landing curves represent the calculated results from the Chan–Vese, DDST and DDST with BBP methods. To compare with the ground truth, recording during the landing process by DGPS, the location errors of each axis are lists on the right side. In X and Z-axis, the location error decreases while the vehicle approaches the landing area. The error in Y-axis has larger error compared with X and Z-axis, and the disturbance is significant.
Table 4. Eight experiment results in different weather condition.
Figure 12. Final approach of the fixed-wing UAV.
As the theoretical and simulation result discussed, the localization errors in each axis are large when the UAV is far way from the ground visual system. To illustrate the result more clearly, we compared the localization results with DGPS at separate intervals, which are shown in Table 5. Previously, the average errors of each axis at a large distance (more than 400 m) are large, especially in the depth dimension. The comparison of the average frame rate of different kinds of tracking algorithms is listed in Table 6. The DSST results have the best real-time performance, which reaches 21.345 fps, and has better accuracy compared with the mean shift method, which has similar process speed. We also calculated the errors in O c ( i O , c , j O , c , k O , c ) coordinates by implementing those tracking methods with the identical landing imaging streaming and PTU status. The results are shown at separate interval in Table 7, Table 8 and Table 9. In the accuracy measurement, the DSST with BBP calculates the 3D position more precisely at the cost of a slower frame rate.
Table 5. Localization error in each axis at separated intervals with the DSST and bounding box prediction (BBP) algorithms.
Table 6. Average frame rate of the tracking algorithms.
Table 7. Errors (m) in the i O , c -axis with tracking algorithms at separate intervals (m).
Table 8. Errors (m) in the j O , c -axis with tracking algorithms at separate intervals (m).
Table 9. Errors (m) in the k O , c -axis with tracking algorithms at separate intervals (m).

6. Concluding Remarks

This paper presents a complete localization framework of the ground-based stereo guidance system. This system could be used to pilot the UAV for landing autonomously and safely in the GNSS-denied scenario. Compared with the onboard solutions and other state-of-the-art ground-based approaches, this ground-based system profited enormously from the computation capacity and flexible configuration with the baseline and sensors. The separate deployed configurations did not improve the detection distance, which was discussed in our previous works [15]; however, they enhance the maximum range for depth estimation. Although the system has some pitfalls, such as the low accuracy at a long distance in the depth axis and not supporting the attitude measurement, this low-cost system could be arranged quickly for any proposed environment. Additional future work will focus on estimate errors over time and investigate methods to improve inevitable error propagation through the inclusion of additional sensors, such as GNSS and on-board sensors.

Acknowledgments

The work is supported by Major Application Basic Research Project of NUDT with Granted No. ZDYYJCYJ20140601. The authors would like to thank Dianle Zhou and Zhiwei Zhong for their contribution on the experimental prototype development. Tengxiang Li, Dou Hu, Shao Hua and Yexun Xi made a great contribution on the practical experiment configuration.

Author Contributions

W.K. proposed and designed the ground-based navigation system architecture. W.K. and T.H. implemented the vision algorithms to test the capability of the system. D.Z., L.S., and J.Z. participated in the development of the guidance system and conceive the experiments. All authors read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

References

  1. Floreano, D.; Wood, R.J. Science, technology and the future of small autonomous drones. Nature 2015, 521, 460–466. [Google Scholar] [CrossRef] [PubMed]
  2. Kushleyev, A.; Mellinger, D.; Powers, C.; Kumar, V. Towards a swarm of agile micro quadrotors. Auton. Robot. 2013, 35, 287–300. [Google Scholar] [CrossRef]
  3. Manning, S.D.; Rash, C.E.; LeDuc, P.A.; Noback, R.K.; McKeon, J. The Role of Human Causal Factors in US Army Unmanned Aerial Vehicle Accidents; Army Aeromedical Research Lab: Fort Rucker, AL, USA, 2004. [Google Scholar]
  4. McLean, D. Automatic Flight Control Systems; Prentice Hall: Englewood Cliffs, NJ, USA, 1990; p. 606. [Google Scholar]
  5. Stevens, B.L.; Lewis, F.L. Aircraft Control and Simulation; John Wiley & Sons: Hoboken, NJ, USA, 2003. [Google Scholar]
  6. Farrell, J.; Barth, M.; Galijan, R.; Sinko, J. GPS/INS Based Lateral and Longitudinal Control Demonstration: Final Report; California Partners for Advanced Transit and Highways (PATH): Richmond, CA, USA, 1998. [Google Scholar]
  7. RUAG. Available online: https://www.ruag.com/de/node/307 (accessed on 8 June 2017).
  8. Northrop Grumman. MQ-8B Fire Scout Vertical Takeoff and Landing Tactical Unmanned Aerial Vehicle System; Northrop Grumman: Falls Church, VA, USA.
  9. Wang, W.; Song, G.; Nonami, K.; Hirata, M.; Miyazawa, O. Autonomous Control for Micro-Flying Robot and Small Wireless Helicopter X.R.B. In Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 9–15 October 2006; pp. 2906–2911. [Google Scholar]
  10. Pebrianti, D.; Kendoul, F.; Azrad, S.; Wang, W.; Nonami, K. Autonomous hovering and landing of a quad-rotor micro aerial vehicle by means of on ground stereo vision system. J. Syst. Des. Dyn. 2010, 4, 269–284. [Google Scholar] [CrossRef]
  11. Martinez, C.; Campoy, P.; Mondragon, I.; Olivares-Mendez, M.A. Trinocular ground system to control UAVs. In Proceedings of the 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, St. Louis, MO, USA, 10–15 October 2009; pp. 3361–3367. [Google Scholar]
  12. Guan, B.; Sun, X.; Shang, Y.; Zhang, X.; Hofer, M. Multi-camera networks for motion parameter estimation of an aircraft. Int. J. Adv. Robot. Syst. 2017, 14. [Google Scholar] [CrossRef]
  13. Kim, E.; Choi, D. A UWB positioning network enabling unmanned aircraft systems auto land. Aerosp. Sci. Technol. 2016, 58, 418–426. [Google Scholar] [CrossRef]
  14. Kong, W.; Zhang, D.; Wang, X.; Xian, Z.; Zhang, J. Autonomous landing of an UAV with a ground-based actuated infrared stereo vision system. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 2963–2970. [Google Scholar]
  15. Kong, W.; Zhou, D.; Zhang, Y.; Zhang, D.; Wang, X.; Zhao, B.; Yan, C.; Shen, L.; Zhang, J. A ground-based optical system for autonomous landing of a fixed wing UAV. In Proceedings of the 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, Chicago, IL, USA, 14–18 September 2014; pp. 4797–4804. [Google Scholar]
  16. Tang, D.; Hu, T.; Shen, L.; Zhang, D.; Kong, W.; Low, K.H. Ground Stereo Vision-based Navigation for Autonomous Take-off and Landing of UAVs: A Chan-Vese Model Approach. Int. J. Adv. Robot. Syst. 2016, 13. [Google Scholar] [CrossRef]
  17. Ma, Z.; Hu, T.; Shen, L. Stereo vision guiding for the autonomous landing of fixed-wing UAVs: A saliency-inspired approach. Int. J. Adv. Robot. Syst. 2016, 13, 43. [Google Scholar] [CrossRef]
  18. Kong, W.; Zhou, D.; Zhang, D.; Zhang, J. Vision-based autonomous landing system for unmanned aerial vehicle: A survey. In Proceedings of the 2014 International Conference on Multisensor Fusion and Information Integration for Intelligent Systems (MFI), Beijing, China, 28–29 September 2014; pp. 1–8. [Google Scholar]
  19. Gautam, A.; Sujit, P.; Saripalli, S. A survey of autonomous landing techniques for UAVs. In Proceedings of the 2014 International Conference on Unmanned Aircraft Systems (ICUAS), Orlando, FL, USA, 27–30 May 2014; pp. 1210–1218. [Google Scholar]
  20. Hu, T.; Zhao, B.; Tang, D.; Zhang, D.; Kong, W.; Shen, L. ROS-based ground stereo vision detection: Implementation and experiments. Robot. Biomim. 2016, 3, 14. [Google Scholar] [CrossRef] [PubMed]
  21. Beijing Universal Pioneering Technology Co., Ltd. iFLYUAS; Beijing Universal Pioneering Technology Co., Ltd.: Beijing, China, 2016. [Google Scholar]
  22. Danelljan, M.; Häger, G.; Khan, F.; Felsberg, M. Accurate scale estimation for robust visual tracking. In British Machine Vision Conference, Nottingham, 1–5 September 2014; BMVA Press: Durham, UK, 2014. [Google Scholar]
  23. Oza, N.C. Online bagging and boosting. In Proceedings of the 2005 IEEE International Conference on Systems, Man and Cybernetics, Waikoloa, HI, USA, 10–12 October 2005; Volume 3, pp. 2340–2345. [Google Scholar]
  24. Bradski, G.R. Real time face and object tracking as a component of a perceptual user interface. In Proceedings of the IEEE Workshop on Applications of Computer Vision (WACV’98), Princeton, NJ, USA, 19–21 October 1998; pp. 214–219. [Google Scholar]
  25. Chan, T.F.; Sandberg, B.Y.; Vese, L.A. Active contours without edges for vector-valued images. J. Vis. Commun. Image Represent. 2000, 11, 130–141. [Google Scholar] [CrossRef]
  26. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.