Next Article in Journal
Efficient FPGA-Based Architecture of the Overlap-Add Method for Short-Time Fourier Analysis/Synthesis
Next Article in Special Issue
An Efficient and Provably Secure Certificateless Blind Signature Scheme for Flying Ad-Hoc Network Based on Multi-Access Edge Computing
Previous Article in Journal
SEE Sensitivity Evaluation for Commercial 16 nm SRAM-FPGA
Previous Article in Special Issue
Design and Analysis of Refined Inspection of Field Conditions of Oilfield Pumping Wells Based on Rotorcraft UAV Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Accurate Landing of Unmanned Aerial Vehicles Using Ground Pattern Recognition

1
Faculty of Applied Engineering, Electronics-ICT, IDLab University of Antwerp, 2000 Antwerp, Belgium
2
Departament of Computer Engineering (DISCA), Universitat Politècnica de València, 46022 Valencia, Spain
3
Faculty of Electrical and Computer Engineering, Rzeszow University of Technology, 35-959 Rzeszow, Poland
4
imec Antwerp, 2000 Antwerpen, Belgium
*
Author to whom correspondence should be addressed.
Electronics 2019, 8(12), 1532; https://doi.org/10.3390/electronics8121532
Submission received: 6 November 2019 / Revised: 29 November 2019 / Accepted: 3 December 2019 / Published: 12 December 2019

Abstract

:
Over the last few years, several researchers have been developing protocols and applications in order to autonomously land unmanned aerial vehicles (UAVs). However, most of the proposed protocols rely on expensive equipment or do not satisfy the high precision needs of some UAV applications such as package retrieval and delivery or the compact landing of UAV swarms. Therefore, in this work, a solution for high precision landing based on the use of ArUco markers is presented. In the proposed solution, a UAV equipped with a low-cost camera is able to detect ArUco markers sized 56 × 56 cm from an altitude of up to 30 m. Once the marker is detected, the UAV changes its flight behavior in order to land on the exact position where the marker is located. The proposal was evaluated and validated using both the ArduSim simulation platform and real UAV flights. The results show an average offset of only 11 cm from the target position, which vastly improves the landing accuracy compared to the traditional GPS-based landing, which typically deviates from the intended target by 1 to 3 m.

1. Introduction

Recently there has been a growing interest in unmanned aerial vehicles (UAVs). Their applications are diverse, ranging from surveillance, inspection and monitoring, to precision agriculture and package retrieval/delivery. Landing a UAV is acknowledged as the last and most critical stage of navigation [1]. According to statistics, the number of accidents associated to the UAV landing process represent 80% of the hazard cases [2]. Therefore, improved landing techniques are being intensely explored. Furthermore, some of the applications mentioned above, such as package retrieval, require a high level of accuracy so as to make sure the UAV lands exactly on the desired target.
Previous proposals heavily rely on the Global Positioning System (GPS) and inertial navigation sensors (INS) as the main positioning approaches [3]. However, altitude data provided by the GPS is typically inaccurate and needs to be compensated with a close-range sensor, such as a barometric pressure sensor or a radar altimeter [4]. Despite such compensation, these methods still remain inaccurate, especially in the horizontal plane, resulting in a landing position that typically deviates from the intended one by 1 to 3 m. Furthermore, the GPS cannot be used indoors. For these reasons, GPS and INS systems are mostly used for long range, outdoor flights having low accuracy requirements [3].
Taking the aforementioned issues into consideration, the aim of this work is to develop a novel vision-based landing system that is able to make a UAV land in a very specific place with high precision. This challenge is addressed by developing a solution that combines the use of a camera and ArUco markers [5,6]. This way, the relative offset of the UAV towards the target landing position is calculated using the ArUco library [7] (based on OpenCV). After computing its relative offset, the UAV adjusts its position so as to move towards the center of the marker and start descending, performing additional adjustments dynamically.
The rest of this work is structured as follows: in the next section some related works on UAV landing strategies are presented. In Section 3, our proposed solution is introduced, detailing the methodology followed in order to track the target for landing and how to perform the necessary calculations to adjust the UAV position. Then, Section 4 provides some technical details about the UAV used to deploy our proposal, highlighting the main issues and restrictions to be taken into consideration in the design of our novel landing algorithm. Section 5 describes how the different experiments were made. The main results are then presented in Section 6, where our solution is compared to a GPS-based approach, with an appropriate discussion. Finally, Section 7 concludes this work and refers to future works.

2. Related Work

Recently, different UAV landing approaches have been studied. These studies can be categorized based on the different types of landing platforms adopted. According to Reference [8], the vast amount of landing platforms belong to one of these three categories:
  • Category I–fixed platforms:
    Fixed platforms include all platforms that are stationary. This type of platforms are the easiest to land on, since the target remains stationary.
  • Category II–moving platforms:
    Moving platforms are defined by the ability to move with two degrees (surge, sway) of freedom. In this case, the UAV needs to track the platform first and then land on it.
  • Category III–Landing on a ship:
    The ultimate landing platform is on a moving ship. In this case the ship has six degrees of freedom (heave, sway, surge, yaw, roll, pitch) and, therefore, developing a safe landing approach becomes a troublesome task.
It is also possible to categorize the different landing approaches based on the sensor(s) used in the process. Many different sensors can be used, such as sonar, infrared, LIDAR, cameras or a combination of these. Below, some approaches that rely on computer vision will be discussed.
Chen et al. [9] succeeded in landing a real UAV on an object moving with a speed of 1 m/s (category II). A camera was used to track the position of the landing platform (xy-coordinates) and a LIDAR sensor provided detailed information about the altitude. This research work introduced a robust method to track and land on a moving object. However, the use of a LIDAR sensor discourages the solution, as it tends to be too expensive when scaling to a high number of UAVs.
Nowak et al. [10] proposed a system in which a UAV could land both at night, as well as during the day. The idea is simple yet robust and elegant: a beacon is placed on the ground. The light emitted from the beacon is then captured by a camera (without a infrared filter) and the drone moves (in the xy-plane) such that the beacon is in the center of the picture. Once centerd, the height is estimated based on the image area occupied by the beacon and the drone’s altitude is decreased in order for it to land safely.
Shaker et al. [11] suggested another approach: reinforcement learning. In this approach, the UAV agent learns and adapts its behaviour when required. Usually, reinforcement learning takes a lot of time. To accelerate this process, a technique called Least-Squares Policy Iteration (LSPI) is used. With this method, a simulated UAV (AR100) was able to achieve a smooth landing trajectory swiftly.
In the work of Lange et al. [12] an approach for landing and position control, similar to our work, was developed. Their approach was also based on OpenCV and on recognizing a landing pattern. However, their landing pattern was not built with the use of ArUco markers. In fact, the landing pattern used, with a diameter of 45 cm, was only detected from a distance of 70 cm. Therefore, this strategy cannot be used in an outdoor environment where the flight altitude is typically much higher. However, in this approach, the UAV does not need to see the entire marker, which is an advantage of this scheme.
A system that can land on and track a slow moving vehicle (180 cm/s) was developed by Araar et al. [13]. Indoor experiments show that the UAV used was able to successfully land on the target landing platform (which also consists of multiple ArUco markers) from a height of approximately 80 cm.
More recently, Patruno et al. [14] presented a solution for the landing of UAVs on a human-made landing target. Their target was similar to traditional heliplatforms but with specific aspect ratios, so that it can be detected from long distances. The geometric properties of the H-shaped marks adopted are used to estimate the pose with high accuracy, achieving an average RMSE value of only 0.0137 m in pose and 1.04 in orientation.
Baca et al. [15] were able to detect a moving car at 15 km/h, predict its future movement and attach to it. To achieve it they equipped the UAV with onboard sensors and a computer, which detects the car using a monocular camera and predicts the car future movement using a nonlinear motion model. While following the car, the UAV lands on its roof and it attaches itself using magnetic legs.
De Souza et al. [16] developed a autonomous landing system based on Artificial Neural Network (ANN) supervised by Fuzzy Mamdani Logic. Their method introduced low computational complexity while maintaining the characteristics and intelligence of the fuzzy logic controller. They validated their solution using both simulation and real tests for static and dynamic landing spots.
Fraczek et al. [17] presented an embedded video system that allows the UAV to automatically detect safe landing sites. Their solution was implemented on a heterogeneous Zynq SoC device from Xilinx and a Jetson GPU. Differently from the previous works, this work does not rely on a human made marker. Through the use of machine learning and computer vision techniques, the UAV classifies the terrain into three classes. The proposed solution was tested on 100 test images and classified the different terrains correctly in 80% of the cases. Furthermore, in these tests the performance between the Zynq SoC device and the Jetson GPU was compared.
Our work differs from the former ones as we want the UAV to detect the landing area when it is high above the ground (height > 20 m) to compensate for possibly high GPS error values, while using cheap sensors (only a Raspberry Pi camera is needed) and yet still achieving very low errors in terms of landing accuracy (<20 cm). To that extent, only one other source (the Ardupilot community [18]) was found that is attempting to achieve results similar to ours. Their method follows the same strategy as the work of Nowak et al. [10], as they also make use of an IR-beacon. According to the ArduPilot authors, their proposed method is able to land a UAV from an altitude of 15 m, reliably under all lighting conditions and with an maximum offset of only 30 cm. While the results of this approach are impressive, ours still outperforms it in terms of both accuracy, altitude and price.

3. Proposed Solution

The aim of this work is to make a UAV land on a specific location. First the UAV has to make a coarse approach to the landing zone. As stated before, the UAV will typically fail to hover above its exact target location, being usually within 1 to 3 m away from the intended landing position. Once the UAV is close to the target location our protocol is activated. The first step deals with finding the marker. The ArUco marker library [5,6] (based on OpenCV) provides a function which takes the camera feed and returns information about the marker(s). ArUco markers resemble the well-known QR-codes. They carry less information than the latter ones (only an id), which makes them easier to detect. A typical ArUco marker consists of a black border and a 6x6 square of black and white smaller squares. There are different types of configurations (e.g., 3 × 3, 4 × 4, 7 × 7) known as dictionaries. A marker from a dictionary with less squares is of course easier to detect but only a small number of ids can be provided. In this work, dictionary “DICT_6X6_250” is used. As the name suggests, it provides 250 different ids, which is more than enough for our purposes. The second part consists of descending the UAV while trying to keep it centered over the marker.
In order to detect a marker two conditions must be met: (i) the marker must be fully inside the picture and (ii) each square must be uniquely identified (black or white). In this application, it is possible that the two conditions are not simultaneously met in some cases. For instance, when the drone is at a low altitude (i.e., 0.5 m) the marker is too big to fit inside the field of view of the camera; in addition, the shadow of the drone may “corrupt” the image. On the other hand, when the drone is flying at a higher altitude (e.g., 12 m) the image may be too small to be detected. Therefore, a strategy was developed that combines markers of different sizes (see Figure 1), so that the drone can find its target from a higher altitude. If the UAV is able to detect a smaller marker, it will switch to it, adjusting its course accordingly. Figure 2 shows a real scenario where the UAV is able to see two markers but chooses to move towards the smaller marker. The center of this marker is indicated by the red spot.
Once the marker is detected, the drone has to move towards the center of the marker and descend from there. Due to the effect of wind and to the inherent instability of the UAV itself, the drone will also move in the horizontal plane while descending. This unwanted movement should be compensated in order to land the drone more precisely. To achieve this behavior, the strategy described in Algorithm 1 is proposed, which works as follows: In line 3, the UAV searches for an ArUco marker. If no marker is detected, the flight mode of the UAV is changed to loiter. If this is the case for 30 consecutive seconds, the mission is aborted and the UAV will land using GPS only. Otherwise, from the potential list of detected markers, the marker with the highest ID (i.e., the smallest marker) is selected (line 11). With the use of the ArUco library, the location of the marker with respect to the drone is estimated. If the altitude of the UAV is greater than z 2 (see empirical values in Table 1), α is set to 20 degrees; otherwise, it is set to 10 degrees. These values are based on: the detection distance of the markers, size of the UAV, size of the markers and a margin which is optimized empirically. In line 20 it is checked if the marker is within the virtual border (explained later). If so, the UAV descends; otherwise, it moves horizontally towards the target position. This algorithm will be executed continuously as long as the altitude of the UAV is greater than z 1 . From the moment the UAV’s altitude drops below z 1 (very near to ground), the control will be handed over to the flight controller, which will land the UAV in a safe manner and disarm the engines.
Algorithm 1 Static vision-based landing strategy.
1:
Start timer 30 s
2:
while altitude > z 1 do
3:
 IDs, detected ← SearchMarker()
4:
if ¬ detected then
5:
  Loiter()
6:
  if timer exceeded then
7:
   AbortLanding()
8:
  end if
9:
else
10:
  reset timer()
11:
  ID ← highest detected ID
12:
  Get P ( x , y , z ) i d
13:
  if z > z 2 then
14:
    α = 20
15:
  else
16:
    α = 10
17:
  end if
18:
   β x = | arctan x / z |
19:
   β y = | arctan y / z |
20:
  if β x > α or β y > α then
21:
    Move(x,y)
22:
  else
23:
   Descend()
24:
  end if
25:
end if
26:
end while
27:
Descend and disarm UAV
In the description above, a virtual border is mentioned. This border defines an area which should enclose the marker (illustrated in Figure 3). This virtual border is created in order to distinguish the two cases of descending or moving horizontally. The size a of this square is defined as:
a = 2 × tan α × h
where h refers to the relative altitude of the UAV.
The main advantage of defining the area in this way is that it will decrease as the drone lowers its altitude. Therefore, the drone will be more centerd above its target position when it flies at low altitude. However, when flying at higher altitude, the drone should descend whenever possible to avoid excessive landing times. For this reason, α is increased to 20 if the UAV is flying above 13 m ( z 2 ).
There are multiple ways to move the UAV towards the target point. In ArduSim, a UAV can be moved for example, by overriding the remote control with function “channelOverride”. With the use of this function the UAV can be moved at a constant speed, first along the roll axis (left/right) and afterwards along the pitch (forward/backward) axis, until the target point is met.
This algorithm experiences difficulties tracking a marker, because it will quickly change between marker IDs whenever one of the markers is not visible for a short amount of time. At first, a small timer was used to eliminate this problem. However, this approach was not satisfactory and so an extension of this algorithm is proposed as Algorithm 2. This second version uses the same general ideas but introduces improvements such as hysteresis that make adjustments more dynamic, achieving a higher effectiveness than the former one. In particular, for this second version, the altitude (activationLevel[i]) is saved whenever a new marker is detected. Contrary to the former version, the algorithm will only switch between markers when its current altitude is lower than half of the activationLevel[i] value. With this modification, the typical glitching behaviour (of detecting and not detecting a marker) is bypassed. Furthermore, if the UAV is not able to detect a marker, it will switch to a recovery mode. This means that it will increase its altitude by one meter, thereby increasing the chances of finding the marker again. Finally, the horizontal speed of the UAV depends on the horizontal distance to the marker. The pitch and roll values are set to v 1 (see empirical values in Table 2) if the distance between marker and UAV is greater then 1 m (see Table 1); otherwise, it is set to v 2 .
Algorithm 2 Adaptive vision-based landing strategy.
1:
Start timer 30 s
2:
while altitude > z 1 do
3:
  IDs ← Search
4:
  if ¬ detected then
5:
   Recover
6:
   if timer exceeded then
7:
    Abort
8:
   end if
9:
else
10:
  reset timer
11:
  for all IDs do
12:
   if First time detected then
13:
    activationLevel[i] = altitude
14:
   end if
15:
   if altitude < activationLevel[i]/2 then
16:
    id ← i
17:
   end if
18:
  end for
19:
  Get P ( x , y , z ) i d
20:
  if z > z 2 then
21:
    α = 20
22:
  else
23:
    α = 10
24:
  end if
25:
   β x = | arctan x / z |
26:
   β y = | arctan y / z |
27:
  if β x > α or β y > α then
28:
   Move(x,y)
29:
  else
30:
   Descend(speed)
31:
  end if
32:
end if
33:
end while
34:
Descend and disarm UAV
Since the area captured by the camera is large when flying at high altitudes, there is less risk of missing the marker. Therefore, the UAV‘s descending speed can be varied with respect to its altitude in the following Equation (1):
d e s c e n d i n g s p e e d = m i n 60 % , a l t i t u d e × 2 % i f a l t i t u d e > 6 m e t e r s v 3 o t h e r w i s e .

4. UAV Specification

Due to the vast amount and diversity of UAV models available, it is worth mentioning the actual characteristics of the UAV used for our experiments. The UAV adopted belongs to the Vertical Take-Off and Landing (VTOL) category, more commonly known as a multirotor UAV. In the experiments described in this work, a hexacopter model is used (see Figure 4), being equipped with a remote control operating in the 2.4 GHz band, a telemetry channel in the 433 MHz band, a GPS receiver, a Pixhawk flight controller and a Raspberry Pi with external camera (see Figure 5). The Raspberry Pi creates an ad-hoc WiFi connection in the 5 GHz band, which is used to communicate with a ground station or with other UAVs. Below we detail the purpose and the connections (as shown in Figure 6) between these different devices.

4.1. 2.4 GHz Remote Control

The FrSky X8r receiver provides communication between the UAV and the remote control. This device makes it possible to fly the UAV manually. It receives signals from the remote control and passes them to the flight controller. Furthermore, it sends basic information about the UAV to the remote control for example, flight mode, so that the pilot is informed about the UAV state. It accomplishes these tasks by using the entire 2.4 GHz ISM band. Therefore, it becomes nearly impossible to receive/send any WiFi signal in this band (2.4 GHz), as shown in Reference [19]. Hence, our ad-hoc WiFi connections rely on the 5 GHz band instead. This 5 GHz WiFi connection is used to connect an UAV or multiple UAVs to the ArduSim ground station.

4.2. 433 MHz Telemetry

The telemetry channel operates at a lower frequency (433 MHz). Its purpose is sending UAV information from the flight controller to a ground station (typically a smartphone), including data such as heading, tilt, speed, battery lifetime, flight mode, altitude, and so forth. It is also able to receive instructions to follow a specific mission, that is, return to home or to perform an emergency landing. However, these set of instructions are very limited since the 433 MHz telemetry is downlink focused. For this reason we have chosen to use the above mentioned 5 GHz WiFi connection to control the UAV from ArduSim. The 433 MHz telemetry will only be used in case of emergency.

4.3. Flight Controller

The flight controller, in this case the Pixhawk 4, is a device that receives information from sensors and that processes this information in order to control the UAV at a low-level, while a Raspberry Pi is in charge of application level control. In particular, information from different sensors such as GPS, barometer, magnetometer, accelerometer and gyroscope are combined in order to provide an accurate representation of the UAV state. This information is then used in order to stabilize the UAV and make it controllable. Some of this information can also be sent via a serial link towards the Raspberry Pi. We opted to use the Pixhawk flight controller (containing an implementation of the Ardupilot firmware [18]) because it is an open source solution and thus new protocols are easily implemented and deployed on real UAVs. For the communication between the Raspberry Pi and the Pixhawk, the open source MAVLink protocol [20] is used. It is a lightweight messaging protocol for communicating with most open-source flight controllers, as is the case of the Pixhawk.

4.4. Raspberry Pi

The Raspberry Pi 3 Model B+ serves three purposes in our custom UAV. First, it runs the open source ArduSim [21] simulation platform, which controls the UAV at a high-level. This program is capable of coordinating an autonomous flight. Second, it also runs a Python application that processes the camera information (see Figure 5) using the ArUco marker library and sends the resulting information to ArduSim via a TCP connection. Finally, it is used to setup an ad-hoc network in the 5 GHz band. This network can be used to communicate with other UAVs or, in our case, with a ground station (laptop). As an alternative, the Raspberry Pi can be equipped with a 4G LTE dongle so as to operate the UAV from a remote location. However, this option is currently not supported by ArduSim.

4.5. General Components

In addition to the elements referred above, there are mandatory components for each UAV such as:
  • Electronic Speed Controller (ESC): provides power to motors and controls their speed individually with the use of a Pulse Width Modulation (PWM) signal. The signal needed to vary the motor speed is provided by the flight controller.
  • Brushless DC motors to rotate the propellers and thus create thrust.
  • Li-Po battery and power module used to deliver and transmit power to the flight controller, ESCs and all other electrical components.
  • Safety switch which has to be manually pressed by the pilot to ensure no unintended takeoff takes place.
  • An optional buzzer to provide feedback about the current state of the UAV.

5. Experimental Settings

Once we finished the implementation of our solution in ArduSim, we performed validation experiments with the multicopter described in Section 4 in order to assess the effectiveness of our proposed solutions. In the first set of experiments, the UAV was instructed to fly up to an altitude of 20 m, to move toward a specific GPS location and to land automatically (by giving the flight controller full control) once that position was reached. During these experiments, the landing time was recorded, as well as the actual landing position. Those experiments, which do not use our protocol, are used as reference.
In the second set of experiments, the landing accuracy of Algorithm 1 was evaluated. Again, the UAV took off until an altitude of 20 m was reached and then flew towards the target GPS location. The largest available marker ( 56 × 56 cm) was placed at that location and the UAV used this marker as the initial reference point for landing. When the UAV was able to detect the smaller marker ( 18 × 18 cm), it used that marker as the reference point instead. For these experiments, the descending speed was defined by lowering the throttle by 10% and the roll and pitch values were set to a value of 5%. After each experiment, the landing time and the distance between the marker and the actual landing position were recorded. We define the landing time as the time interval from the moment when the UAV detected the largest marker until the time when the landing procedure was finished.
In the last set of experiments, the landing time of Algorithm 2 was measured. The markers used were the same. However, in this optimized version, the descending, roll and pitch values were dynamic, as described in Section 3.

6. Results

Without the use of our approach, the UAV was able to land consistently within a time span ranging from 27 to 30 s. Nonetheless, this rapid landing comes at a price. As shown in Figure 7 and Figure 8, the actual landing position varies substantially, ranging from a maximum error of 1.44 m to a minimum error of 0.51 m; the mean value for our experiments was of 0.85 m.
Notice that these errors are smaller than expected (1–3 m). This is most likely due to the small travelled distance between the takeoff and landing locations. In fact, in these experiments, the UAV flew for only 14 m and the total flight time was of about 52 s. Longer flights will introduce higher errors, as reported in the literature [22].
As shown in Figure 7, the landing position accuracy increased substantially when Algorithm 1 was adopted. In particular, experiments showed that the error ranged from only 3 to 18 cm, with a mean value of 11 cm (see Figure 8). Overall, this means that the proposed landing approach is able to reduce the landing error by about 96%. However, in three out of ten of the experiments performed, the UAV moved away from the marker due to the effect of wind. Since at these moments the altitude was already quite low, the UAV could no longer detect the marker, causing the mission to be stopped after 30 s. Furthermore, the average landing time was increased to 162 s. This is due to the fact that, during the transition from one marker to another, the algorithm experienced problems at detecting the smaller marker during some time periods, as illustrated in Figure 9a (from second 22 to 35).
Besides this malfunction situations, the UAV showed a smooth landing trajectory (see Figure 10).
Notice how the UAV makes more aggressive adjustments in the X axis when the altitude drops below 13 m; this is due to the fact that parameter α becomes smaller, restricting the error range. If the malfunction cases are removed, the average descending speed was of 0.3 m/s, which could be considered too conservative. Furthermore, Figure 10 shows that most of the adjustments are made when the UAV is close to the ground (constant altitude). This can be better observed in Figure 11, where the center of the camera frame and the actual location of the marker in regard to both axes is plotted ( b e t a x , b e t a y ). It can be seen that the drone only moves when the b e t a x or b e t a y angle exceeds the value of α . The range of estimated values captured is shown in Figure 12; we can observe that there is a higher variability in the X axis due to wind compensation requirements along that direction during the experiments, something that occurs to a much lower extent for the Y axis.
The malfunctions of Algorithm 1 are solved in the second version. As shown in Figure 9b, Algorithm 2 does not have any issues when switching between markers. Therefore, the landing time is decreased significantly to an average of only 55 s. Furthermore, as shown in Figure 10, the UAV is descending faster, which also contributes to reduce the landing time. The decrease in landing time did not have any affect on the accuracy of the application. The new recovery mode was also found to be beneficial to ensure that the UAV landed on the marker every time.
To compare the proposed solution to those mentioned in Section 2, Table 3 summarizes the results obtained, highlighting the main differences between them.
Finally, an illustrative video (https://youtu.be/NPNi5YC9AeI) has been made available to show how the proposed solution performs in real environments.

7. Conclusions and Future Work

Achieving accurate landing of multirotor UAVs remains a challenging issue, as GPS-based landing procedures are associated with errors of a few meters even under ideal satellite reception conditions, performing worse in many cases. In addition, GPS-assisted landing is not an option for indoor operations. To address this issue, in this work a vision-based landing solution that relies on ArUco markers is presented. These markers allow the UAV to detect the exact landing position from a high altitude (30 m), paving the way for sophisticated applications including automated package retrieval or the landing of large UAV swarms in a very restricted area, among others.
Experimental results using a real UAV have validated the proposed approach, showing that accurate landing (mean error of 0.11m) can be achieved while introducing an additional (but small) time overhead in the landing procedure compared to the standard landing command.
As future work, it is planned to improve the overall efficiency of the protocol. This can be done by improving the flight behaviour. In addition to decreasing the landing time even further, the accuracy can be increased by integrating the UAV‘s state (acceleration, velocity, etc.) in the control loop. Finally, the algorithm can be further optimized so that the UAV is able to land under less favourable weather conditions.

Author Contributions

C.T.C. conceived the idea; J.W. designed and implemented the protocol with the help of F.F. and T.K.; J.W. analyzed the data and wrote the paper; C.T.C., J.-C.C., F.F., T.K., J.M.M.-B. and P.M. reviewed the paper.

Funding

This work was funded by the “Ministerio de Ciencia, Innovación y Universidades, Programa Estatal de Investigación, Desarrollo e Innovación Orientada a los Retos de la Sociedad, Proyectos I+D+I 2018”, Spain, under Grant RTI2018-096384-B-I00.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pan, X.; Ma, D.Q.; Jin, L.L.; Jiang, Z.S. Vision-based approach angle and height estimation for UAV landing. In Proceedings of the 1st International Congress on Image and Signal Processing, CISP 2008, Sanya, China, 27–30 May 2008; Volume 3, pp. 801–805. [Google Scholar] [CrossRef]
  2. Tang, D.; Li, F.; Shen, N.; Guo, S. UAV attitude and position estimation for vision-based landing. In Proceedings of the 2011 International Conference on Electronic and Mechanical Engineering and Information Technology, EMEIT 2011, Harbin, China, 12–14 August 2011; Volume 9, pp. 4446–4450. [Google Scholar] [CrossRef]
  3. Gautam, A.; Sujit, P.B.; Saripalli, S. A survey of autonomous landing techniques for UAVs. In Proceedings of the 2014 International Conference on Unmanned Aircraft Systems, ICUAS 2014, Orlando, FL, USA, 27–30 May 2014; pp. 1210–1218. [Google Scholar] [CrossRef]
  4. PX4 Dev Team. Holybro Pixhawk 4 · PX4 v1.9.0 User Guide. 2019. Available online: https://docs.px4.io/v1.9.0/en/flight_controller/pixhawk4.html (accessed on 23 November 2019).
  5. Garrido-Jurado, S.; Muñoz-Salinas, R.; Madrid-Cuevas, F.J.; Medina-Carnicer, R. Generation of fiducial marker dictionaries using Mixed Integer Linear Programming. Pattern Recognit. 2016, 51, 481–491. [Google Scholar] [CrossRef]
  6. Romero-Ramirez, F.J.; Muñoz-Salinas, R.; Medina-Carnicer, R. Speeded up detection of squared fiducial markers. Image Vis. Comput. 2018, 76, 38–47. [Google Scholar] [CrossRef]
  7. Romero-Ramirez, F.J.; Muñoz-Salinas, R.; Medina-Carnicer, R. ArUco: Augmented reality library based on OpenCV. Available online: https://sourceforge.net/projects/aruco/ (accessed on 7 June 2019).
  8. Jin, S.; Zhang, J.; Shen, L.; Li, T. On-board vision autonomous landing techniques for quadrotor: A survey. In Proceedings of the Chinese Control Conference, Chengdu, China, 27–29 July 2016; pp. 10284–10289. [Google Scholar] [CrossRef]
  9. Chen, X.; Phang, S.K.; Shan, M.; Chen, B.M. System integration of a vision-guided UAV for autonomous landing on moving platform. In Proceedings of the IEEE International Conference on Control and Automation, Kathmandu, Nepal, 1–3 June 2016; pp. 761–766. [Google Scholar] [CrossRef]
  10. Nowak, E.; Gupta, K.; Najjaran, H. Development of a Plug-and-Play Infrared Landing System for Multirotor Unmanned Aerial Vehicles. In Proceedings of the 2017 14th Conference on Computer and Robot Vision, Edmonton, AB, Canada, 16–19 May 2017; pp. 256–260. [Google Scholar] [CrossRef]
  11. Shaker, M.; Smith, M.N.R.; Yue, S.; Duckett, T. Vision-Based Landing of a Simulated Unmanned Aerial Vehicle with Fast Reinforcement Learning. In Proceedings of the 2010 International Conference on Emerging Security Technologies, Canterbury, UK, 6–7 September 2010; pp. 183–188. [Google Scholar] [CrossRef] [Green Version]
  12. Lange, S.; Sünderhauf, N.; Protzel, P. Autonomous Landing for a Multirotor UAV Using Vision. In Proceedings of the SIMPAR 2008 International Conference on Simulation, Modeling and Programming for Autonomous Robots, Venice, Italy, 3–7 November 2008; pp. 482–491. [Google Scholar]
  13. Araar, O.; Aouf, N.; Vitanov, I. Vision Based Autonomous Landing of Multirotor UAV on Moving Platform. J. Intell. Robot. Syst. 2016. [Google Scholar] [CrossRef]
  14. Patruno, C.; Nitti, M.; Petitti, A.; Stella, E.; D’Orazio, T. A Vision-Based Approach for Unmanned Aerial Vehicle Landing. J. Intell. Robot. Syst. 2019, 95, 645–664. [Google Scholar] [CrossRef]
  15. Baca, T.; Stepan, P.; Spurny, V.; Hert, D.; Penicka, R.; Saska, M.; Thomas, J.; Loianno, G.; Kumar, V. Autonomous landing on a moving vehicle with an unmanned aerial vehicle. J. Field Robot. 2019, 36, 874–891. [Google Scholar] [CrossRef]
  16. De Souza, J.P.C.; Marcato, A.L.M.; de Aguiar, E.P.; Jucá, M.A.; Teixeira, A.M. Autonomous Landing of UAV Based on Artificial Neural Network Supervised by Fuzzy Logic. J. Control. Autom. Electr. Syst. 2019, 30, 522–531. [Google Scholar] [CrossRef]
  17. Fraczek, P.; Mora, A.; Kryjak, T. Embedded Vision System for Automated Drone Landing Site Detection. In Computer Vision and Graphics; Chmielewski, L.J., Kozera, R., Orłowski, A., Wojciechowski, K., Bruckstein, A.M., Petkov, N., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 397–409. [Google Scholar]
  18. Team, A.D. SITL Simulator (Software in the Loop). 2016. Available online: http://ardupilot.org/dev/docs/sitl-simulator-software-in-the-loop.html (accessed on 10 July 2019).
  19. Fabra, F.; Calafate, C.T.; Cano, J.C.; Manzoni, P. On the impact of inter-UAV communications interference in the 2.4 GHz band. In Proceedings of the 2017 13th International Wireless Communications and Mobile Computing Conference, Valencia, Spain, 26–30 June 2017; pp. 945–950. [Google Scholar] [CrossRef]
  20. Meier, L.; QGroundControl. MAVLink Micro Air Vehicle Communication Protocol. Available online: http://qgroundcontrol.org/mavlink/start (accessed on 30 January 2019).
  21. Fabra, F.; Calafate, C.T.; Cano, J.C.; Manzoni, P. ArduSim: Accurate and real-time multicopter simulation. Simul. Model. Pract. Theory 2018, 87, 170–190. [Google Scholar] [CrossRef]
  22. Careem, M.A.A.; Gomez, J.; Saha, D.; Dutta, A. HiPER-V: A High Precision Radio Frequency Vehicle for Aerial Measurements. In Proceedings of the 2019 16th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON), Boston, MA, USA, 10–13 June 2019; pp. 1–6. [Google Scholar] [CrossRef]
Figure 1. Two examples of ArUco markers of different sizes.
Figure 1. Two examples of ArUco markers of different sizes.
Electronics 08 01532 g001
Figure 2. Image retrieved by the Unmanned Aerial Vehicle (UAV) camera after processing using OpenCV/ArUco libraries.
Figure 2. Image retrieved by the Unmanned Aerial Vehicle (UAV) camera after processing using OpenCV/ArUco libraries.
Electronics 08 01532 g002
Figure 3. Visual representation of the virtual border.
Figure 3. Visual representation of the virtual border.
Electronics 08 01532 g003
Figure 4. Hexacopter used in our experiments.
Figure 4. Hexacopter used in our experiments.
Electronics 08 01532 g004
Figure 5. Raspberry Pi camera attached to the UAV.
Figure 5. Raspberry Pi camera attached to the UAV.
Electronics 08 01532 g005
Figure 6. Wiring scheme of our UAV.
Figure 6. Wiring scheme of our UAV.
Electronics 08 01532 g006
Figure 7. UAV landing position comparison.
Figure 7. UAV landing position comparison.
Electronics 08 01532 g007
Figure 8. Landing offset Global Positioning System (GPS) vs visual based approached.
Figure 8. Landing offset Global Positioning System (GPS) vs visual based approached.
Electronics 08 01532 g008
Figure 9. Number of consecutive dropped camera frames.
Figure 9. Number of consecutive dropped camera frames.
Electronics 08 01532 g009
Figure 10. Algortihms: static and dynamic approach.
Figure 10. Algortihms: static and dynamic approach.
Electronics 08 01532 g010
Figure 11. b e t a x and b e t a y angle variations vs. flight time.
Figure 11. b e t a x and b e t a y angle variations vs. flight time.
Electronics 08 01532 g011
Figure 12. Estimated X and Y variations associated to UAV positions during landing.
Figure 12. Estimated X and Y variations associated to UAV positions during landing.
Electronics 08 01532 g012
Table 1. Parameter values adopted regarding Algorithms 1 and 2.
Table 1. Parameter values adopted regarding Algorithms 1 and 2.
Altitude threshold z 1 0.30 m
Altitude threshold z 2 13 m
Virtual border angle α 10 , 20
Table 2. Speed values adopted regarding Algorithm 2.
Table 2. Speed values adopted regarding Algorithm 2.
speed v 1 15%
speed v 2 5%
speed v 3 10%
Table 3. Comparative table of the different schemes.
Table 3. Comparative table of the different schemes.
SourceAccuracy [m]Landing Speed [m/s]Maximum Altitude [m]Moving TargetOutdoor
Ours/dynamic0.110.330NoYes
Ours/static0.110.120NoYes
GPS-based1-30.6NoYes
Chen et al. [9]n/a0.232.5YesNo
Araar et al. [13]0.130.060.8YesNo
Patruno et al. [14]0.01n/an/aNoYes

Share and Cite

MDPI and ACS Style

Wubben, J.; Fabra, F.; Calafate, C.T.; Krzeszowski, T.; Marquez-Barja, J.M.; Cano, J.-C.; Manzoni, P. Accurate Landing of Unmanned Aerial Vehicles Using Ground Pattern Recognition. Electronics 2019, 8, 1532. https://doi.org/10.3390/electronics8121532

AMA Style

Wubben J, Fabra F, Calafate CT, Krzeszowski T, Marquez-Barja JM, Cano J-C, Manzoni P. Accurate Landing of Unmanned Aerial Vehicles Using Ground Pattern Recognition. Electronics. 2019; 8(12):1532. https://doi.org/10.3390/electronics8121532

Chicago/Turabian Style

Wubben, Jamie, Francisco Fabra, Carlos T. Calafate, Tomasz Krzeszowski, Johann M. Marquez-Barja, Juan-Carlos Cano, and Pietro Manzoni. 2019. "Accurate Landing of Unmanned Aerial Vehicles Using Ground Pattern Recognition" Electronics 8, no. 12: 1532. https://doi.org/10.3390/electronics8121532

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop