Next Article in Journal
Adaptive Impact-Time-Control Cooperative Guidance Law for UAVs Under Time-Varying Velocity Based on Reinforcement Learning
Previous Article in Journal
Distributed Event-Triggered-Based Adaptive Formation Tracking Control for Multi-UAV Systems Under Fixed and Switched Topologies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Precision Landing on a Moving Platform Based on Drone Vision Using YOLO Algorithm

1
Graduate School of Engineering, Chiba University, 1-33 Yayoi-cho, Inage-ku, Chiba 263-8522, Japan
2
Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET), Nanjing University of Information Science & Technology, Nanjing 210044, China
*
Author to whom correspondence should be addressed.
Drones 2025, 9(4), 261; https://doi.org/10.3390/drones9040261
Submission received: 12 February 2025 / Revised: 26 March 2025 / Accepted: 27 March 2025 / Published: 29 March 2025

Abstract

:
High-precision landing is a key technical problem that Unmanned Aerial Vehicles (UAVs) will encounter in all application fields, especially for the landing of moving targets. This paper focuses on developing a landing system designed to achieve real-time precise navigation by integrating the Global Navigation Satellite System (GNSS) with the quadcopter’s vision data. To overcome the challenge of the flight altitude being too high to detect the landing target, this paper first detects large-volume targets, followed by the precise identification of smaller targets, achieving enhanced recognition accuracy and speed through an improved YOLOv8 OBB algorithm. To maintain the UAV’s safety and stability throughout the landing process, this paper applies a position control approach using a reference model-based sliding mode controller (RMSMC). The quadcopter’s position is then controlled by the RMSMC throughout the entire landing procedure. The reference value of each state is determined by the reference model, which improves the stability and safety of the entire position control system. During the final experiment, the results demonstrate that the enhanced YOLOv8 OBB identification model increases the mAP0.5:0.95 index for landing target point detection by 2.22 percentage points compared to the original YOLOv8 OBB model, running at 53 FPS on Nvidia AGX. Through multiple actual flights, the proposed landing system consistently achieves an average position error of just 0.07 m.

1. Introduction

With the advancement in quadcopter technology, quadcopters have quickly become a popular flying platform across various application fields. Their simple design, adaptability in control, and ability to perform vertical flight maneuvers have made them widely adopted. They are not limited only to traditional military reconnaissance and geographic mapping, but also play an important role in logistics and transportation, environmental monitoring, agricultural management, and emergency rescue [1,2,3]. However, due to the limitations of a quadcopter’s batteries, quadcopters cannot sustain prolonged flight and must land frequently. Ensuring a safe and stable landing in various conditions is thus a critical challenge in enhancing the autonomy of quadcopter missions. Particularly in dynamic environments, autonomously identifying and accurately locking onto a landing point without human intervention presents a significant challenge for quadcopter systems. Therefore, intelligent recognition systems and precise autonomous guidance strategies are crucial. They not only ensure the safety of quadcopter landings but also directly impact the success rate of the mission [4,5].
In the early days, many scientists used traditional guidance strategies based only on GNSS [6,7]. However, GNSS signals are often blocked or interfered with in complex or restricted environments, leading to significant positioning errors. As a result, GNSS alone cannot meet the requirements for the high-precision fixed-point landings of quadcopters. With the advancement in image recognition technology, vision-based and image processing techniques have gradually emerged, offering a more accurate and reliable solution. This approach improves the precision of quadcopter landings while minimizing dependence on external positioning systems, fostering the further development of autonomous quadcopter mission capabilities. Landing based on drone vision is usually affected by two factors: target recognition accuracy and landing control stability.
For target recognition, while traditional edge detection algorithms can identify the outlines of landing targets in simple environments, they have significant limitations in more complex environments. Edge detection algorithms are highly susceptible to changes in lighting, noise interference, and background complexity, which can lead to reduced target recognition accuracy [8,9,10]. Additionally, in our previous experiments, during high-altitude flight, the landing target appears too small in the field of view, making it challenging for the edge detection algorithm to identify it accurately [11]. In such cases, traditional edge recognition algorithms struggle to effectively identify larger secondary targets, such as dynamic objects like vehicles, and are unable to perform pre-tracking and positioning to assist with landing. Mainstream approaches are now categorized into one-stage and two-stage detection methods as a response to the rapid development of object detection technologies and the challenges in target recognition. First, the two-stage methods, for example, Faster R-CNN [12], offer high accuracy but are slow and computationally intensive. In contrast, one-stage methods, including well-known approaches like YOLO and SSD, significantly improve detection speed by integrating target positioning and classification tasks, making them particularly suited for real-time scenarios [13,14,15,16]. With advancements in algorithm and architecture optimization, YOLO’s accuracy has been significantly improved, even surpassing two-stage methods in some scenarios, making it the preferred solution for complex target detection tasks. Lee et al. developed a deep learning-driven solution for automatic landing area positioning and obstacle detection for quadcopters. However, this approach is susceptible to false detections due to perspective changes and the presence of obstacles [17]. Leijian Yu et al. integrated the improved SqueezeNet architecture into YOLO to detect landing landmarks and make it robust to different lighting conditions. However, it was not tested in dynamic and complex environments [18]. BY Suprapto et al. used Mean-Shift and Tiny YOLO VOC to detect falling targets, but the Mean-Shift method had a poor detection effect on moving objects and, although Tiny YOLO VOC had high detection speed and accuracy, its detection performance would decrease at higher altitudes [19]. Ying Xu et al. developed an advanced A-YOLOX algorithm, incorporating an attention mechanism to enhance the autonomous landing capabilities of quadcopters, particularly improving the recognition of small and mid-sized objects as well as adapting to varying lighting conditions. However, the article mentioned that the control algorithm still needs to be further optimized to enhance the safety of landing [20]. Tilemahos Mitroudas et al. compared the YOLOv7 and YOLOv8 algorithms in a quadcopter safe landing experiment and found that the performance improvement of YOLOv8 made detection and data more reliable [21]. Kanny Krizzy D. Serrano et al. compared four generations of algorithms from YOLOv5 to YOLOv8 to ensure the safe landing of quadcopters, and found that YOLOv8 achieved the highest average accuracy, outperforming other YOLO architecture models [22]. During the dynamic landing of UAVs, factors such as complex background environment, changes in lighting conditions, and local blur of targets will affect the accuracy of target recognition, and the difficulty in detecting small targets at high altitudes further exacerbates this challenge. Therefore, improving the performance of the recognition algorithm, enhancing the ability to extract key features, enabling UAVs to accurately focus on the target area under complex background conditions and improving the compatible detection capabilities of multi-scale targets are of great significance to improving the success rate in the stable landing of UAVs.
In conclusion, once the target is identified, it is essential to implement a controller that guarantees the stability throughout the landing procedure [20]. Lin et al. proposed a visual servo-based landing controller (PBVS) for landing on a mobile platform. The controller has low computational complexity and strong robustness, but it relies on the camera’s field of view and has high requirements for parameter adjustment [23]. Wu et al. proposed a RL-PID controller that can automatically adjust the PID parameters during the landing process, but its landing accuracy is greatly affected by environmental changes [24]. Alireza Mohammadi et al. employed the extended Kalman filter for position estimation using UAV vision, while utilizing an MPC controller to achieve dynamic landing. But, the calculation complexity was high and it was more dependent on the model [25]. In addition to the methods mentioned, various controllers like PID control [26,27], backstepping control [23,28], and adaptive control [29,30] have also been used for quadcopter landing. Among them, the sliding mode controller excels in dynamic landing due to its strong robustness and adaptability to uncertainty and disturbances. It can effectively handle environmental changes and uncertainties in system models [31,32,33], and therefore can help improve landing accuracy and stability during landing to a certain extent. Therefore, this paper selects a sliding mode controller to enhance the stability of the UAV in dealing with disturbances during landing.
Based on the previous discussion, to achieve the precise dynamic landing of quadcopters, the landing target detection system in this paper is built upon the advanced YOLOv8 OBB algorithm. Additionally, RMSMC is utilized during the landing process to ensure both stability and accuracy in landing control. The primary contributions of this paper include the following:
  • To ensure precise positioning of the landing point, we enhanced the YOLOv8 OBB algorithm and deployed the optimized model on the quadcopter. By combining pixel error conversion with attitude angle compensation, we were able to obtain accurate distance errors, enabling the quadcopter to precisely identify the landing point. The accuracy and feasibility of this method were validated through experiments.
  • The performance and accuracy of the landing system were verified through flight tests. The results show that the improved YOLOv8 OBB algorithm improves mAP@0.5 by 3.1% and mAP@0.5:0.95 by 2.61% on the VisDrone/DroneVehicle datasets, and mAP@0.5 by 0.5% and mAP@0.5:0.95 by 2.1% on private datasets, compared with the latest version of YOLOv8 OBB. In addition, when the proposed controller is deployed, the maximum position tracking error is kept within 0.23 m and the maximum landing error is kept within 0.12 m. The experimental video can be found at https://doi.org/10.6084/m9.figshare.28600868.v1 (accessed on 26 March 2025).
The structure of this paper is as follows: Section 2 and Section 3 provide an overview of the system architecture, including the design of the vision algorithm and controller. Section 4 describes the experimental setup and presents the results. Finally, Section 5 concludes the paper and outlines future work.

2. Visual Identification System Design

The quadcopter used in this system was independently developed by our team, as shown in Figure 1.
Figure 2 illustrates the system structure. First, quadcopters are used to collect the necessary experimental data. Then, the landing target recognition model is trained on an external computer before being deployed to the onboard Nvidia AGX platform. The overall system workflow is as follows: the image captured by the camera in real time is transmitted to the Nvidia AGX for recognition processing. Specifically, Nvidia AGX calculates the pixel error, converts it into the actual distance error based on the recognition results, and feeds this information back to the flight control system. Finally, the quadcopter adjusts its position using position and altitude control to ensure that the landing point remains centered beneath the quadcopter. The bottom layer of the system consists of the camera driver and quadcopter sensors. The framework layer includes open-source libraries such as Robot Operating System (ROS), PyTorch, and OpenCV. The application layer handles the flight control and dynamic landing of the quadcopter.

2.1. YOLOv8 OBB Algorithm

To improve landing adaptability, this paper adopts the YOLOv8 OBB algorithm, which incorporates the target’s rotation angle as part of the landing recognition process. YOLOv8 OBB integrates an oriented bounding box (OBB) mechanism into the YOLOv8 framework to enhance detection of objects with strong directional features. To enhance the input data, some advanced augmentation methods, including cropping accompanied by certain rotation operations, are applied, achieving a 3.1% accuracy improvement on the COCO dataset. During training, the model stabilizes learning by refining the data augmentation strategy in the final epochs, enabling efficient training from scratch without requiring pre-trained models. Structurally, YOLOv8 and YOLOv8 OBB share the same backbone (EfficientNet) and neck, which combines a feature pyramid network (FPN) with a spatial attention module (SAM).
YOLOv8 OBB builds upon YOLOv8 by modifying the output layer to include an oriented bounding box detection head, enhancing rotated object detection accuracy and speeding up convergence, albeit with slightly higher computational demands. To optimize performance and efficiency, it uses a 1 × 1 convolution in the output layer for dimensionality reduction and employs a 5 × 5 convolution in subsequent layers to control parameter expansion, ensuring high-precision detection is preserved.

2.2. Improved YOLOv8 OBB Algorithm

The accuracy of landing target recognition is crucial to ensure the safety and precision of the landing process. Factors such as complex backgrounds and lighting changes can significantly affect recognition performance, and the difficulty in detecting small targets at high flight altitudes further increases the challenge. To enhance recognition accuracy, this paper improves the baseline YOLOv8 OBB model by integrating three key components: the Convolutional Block Attention Module (CBAM) [34], Dyhead [35], and EIoU loss.
First, a CBAM is incorporated following the last layer of the backbone to capture spatial information while preserving precise positional data. This integration improves the model’s focus on critical areas within complex scenes, enabling accurate landings across diverse environments. Second, the Dyhead module is employed to dynamically refine feature representations across multiple scales and directions, enhancing the receptive field and robustness to objects of varying orientations and sizes. This further improves recognition accuracy and model stability. Additionally, in the dynamic quadcopter landing task, where large vehicles are identified before smaller landing spots, the original CIoU loss is replaced with EIoU loss. This adjustment enhances accuracy, ensuring the quadcopter reliably identifies landing targets even in challenging environments. Together, these modifications significantly improve detection performance, particularly in dense and directionally complex scenes. The improved YOLOv8 OBB structure is illustrated in Figure 3.

2.3. Convolutional Block Attention Module

In the realm of object detection, the attention module is commonly employed to improve the model’s ability to focus on important features, boosting both selectivity and sensitivity. Common attention mechanisms include compressed excitation network (SENet) [36], coordinated attention (CA) [37], efficient channel attention (ECA) [38], and CBAM [34]. SENet improves the importance of features by compressing feature maps along spatial dimensions using global average pooling to generate global descriptors, followed by a scaling step that amplifies key features and reduces the influence of less relevant ones. ECA improves the attention of the channel by replacing SENet’s fully connected layer with a 1 × 1 convolution, eliminating dimensionality reduction and reducing parameters while maintaining efficiency and accuracy. CBAM extends SENet by incorporating spatial attention alongside channel attention, enabling more precise feature processing. In the dynamic landing task, the integration of CBAM can significantly improve the recognition accuracy of vehicles and landing targets by adaptively refining feature representations through channel and spatial attention mechanisms [39]. The channel attention module prioritizes the most informative features by reweighting feature maps, while the spatial attention module focuses on key spatial regions to improve the detection of small targets or partially occluded targets. Moreover, its lightweight design minimizes computational overhead, making CBAM particularly suitable for resource-constrained UAV systems that require real-time processing. In addition, it can enhance feature discrimination capabilities under different lighting conditions and complex backgrounds to ensure robust performance, which is critical for the safety and reliability of UAV dynamic landing operations.
To improve recognition accuracy, CBAM optimizes convolutional neural networks for better recognition accuracy by leveraging both channel and spatial attention features. It first employs a channel attention module, using global average and maximum pooling to generate importance weights that emphasize critical features. Next, the spatial attention module processes the feature map using pooling followed by convolution operations, concentrating on key spatial regions.
In summary, CBAM is made up of two parts: the channel attention module (CAM) and the spatial attention module (SAM), with their outputs fused through element-wise multiplication to generate the final result. The overall structure is shown in Figure 4:
Figure 5 depicts the design of the channel attention module (CAM). The feature map F is first processed using both global average and maximum pooling operations. The pooled results are subsequently fed into a multi-layer perceptron (MLP) to generate the channel attention weights M C . These weights are then applied to the original feature map F to emphasize the most relevant feature channels, and the final output is obtained by applying a sigmoid activation.
Figure 6 illustrates the design of the SAM. The module takes as input the feature map produced by the CAM, from which two separate feature maps are generated using global maximum and average pooling operations to capture spatial importance. These pooled maps are then concatenated along the channel axis and passed through a convolutional layer to compute the spatial attention map M S . This attention map is then element-wise multiplied with the input feature map, producing the final output of the SAM.

2.4. Dynamic Head

In dynamic landing tasks, when the flight altitude is high, the landing target is usually smaller, and the detection difficulty increases accordingly. In recent years, Dyhead, as a novel dynamic detection head framework, has significantly improved the performance of small target detection and object detection in complex scenes [40]. Dyhead improves the capability of the detection head by incorporating several self-attention layers, which allows for better feature representation and enhanced focus on important regions within the input data, combining the scale, space, and task perception mechanisms of different feature layers, spatial positions, and task-related channels to achieve more efficient feature representation. More importantly, this optimization does not increase significant computational overhead. Its architecture is shown in Figure 7:
In Figure 7, from left to right, the first is π L , which is the scale-aware attention mechanism and can be expressed as follows:
π L ( F ) F = σ f 1 S 1 C s = 1 S c = 1 C F s , c F
In Equation (1), F R L × H × W × C refers to a tensor with four dimensions representing the feature pyramid. The first dimension, L, corresponds to the total number of levels, while H and W capture the height and width of the spatial structure, respectively. Finally, C denotes the quantity of channels within each feature map. For computational efficiency, the tensor is reshaped into a three-dimensional structure F R L × S × C , where S is the total spatial dimension of each layer, calculated as the product of height and width S = H × W . The operation f ( · ) performs a linear mapping, which is approximated using a 1 × 1 convolution layer. Additionally, the hard sigmoid activation σ ( x ) is defined as max ( 0 , min ( 1 , x + x + 1 2 ) ) and is applied to map the output values to the range [0, 1]. This simple yet efficient nonlinear activation function helps reduce the computational burden while improving performance by providing a smooth, bounded output.
The second one is π S , which is the spatial perception attention, and its equation can be written as follows:
π S ( F ) F = 1 L l = 1 L k = 1 K w l , k F ( l , p k + Δ p k , c ) Δ m k
Method π S involves two main steps: sparse attention learning and cross-layer feature aggregation at the same spatial position. First, the method performs sparse sampling at k positions, where each position is denoted as p k . A self-learned spatial offset, Δ p k , shifts the sampling position from p k to p k + Δ p k . This modification enables the model to concentrate more accurately on the discriminative regions within the feature map. Additionally, Δ m k is a scalar representing the importance of the feature at the adjusted position p k + Δ p k , and is learned through the training process. At the intermediate layer of F , the feature activations are used to generate both Δ p k and Δ m k .
By leveraging these learned spatial offsets and importance scalars, the method enables the model to selectively attend to the most relevant areas within the input feature maps. It can allocate varying significance to each spatial position, thereby improving the overall detection performance by focusing on critical areas of the feature map.
Finally, π C represents the attention mechanism tailored to the task at hand, and its equation can be expressed as follows:
π C ( F ) F = max α 1 ( F ) F c + β 1 ( F ) , α 2 ( F ) F c + β 2 ( F )
Equation (3) divides the feature map into two key terms to separately capture and adjust different contributions or interactions, F c represents the portion of the feature map linked to the c-th channel. The vector [ α 1 ; β 1 ; α 2 ; β 2 ] = θ * ( · ) represents an extended function designed to learn and regulate the activation threshold. The function θ * ( · ) is implemented in several stages. Initially, average pooling is applied across the L × S dimensions to reduce the feature’s size. The pooled features undergo two fully connected layers for learning intricate patterns, followed by normalization to standardize the outputs. Finally, the outputs are mapped to the interval [ 1 , 1 ] through the application of the sigmoid function, which enables the model to dynamically adjust the activation level of the features based on the learned threshold.

2.5. Optimization of Loss Function

In the YOLO task, an appropriate loss function is chosen to guide model convergence and ensure task completion. The Probabilistic IoU (ProbIoU) used in YOLOv8 OBB is an effective method for measuring the overlap between oriented bounding boxes (OBBs). Based on the Hellinger distance, ProbIoU assesses the similarity between two OBBs by calculating their covariance matrices and the differences in their center coordinates. Specifically, ProbIoU can be computed using the following equation:
D B = 1 4 ( a 1 + a 2 ) ( y 1 y 2 ) 2 + ( b 1 + b 2 ) ( x 1 x 2 ) 2 ( a 1 + a 2 ) ( b 1 + b 2 ) ( c 1 + c 2 ) 2 + 1 2 ( c 1 + c 2 ) ( x 1 x 2 ) ( y 1 y 2 ) ( a 1 + a 2 ) ( b 1 + b 2 ) ( c 1 + c 2 ) 2 + 1 2 ln ( a 1 + a 2 ) ( b 1 + b 2 ) ( c 1 + c 2 ) 2 4 ( a 1 b 1 c 1 2 ) ( a 2 b 2 c 2 2 )
In Equation (4), D B represents the Bhattacharyya distance, a metric that quantifies the geometric similarity between two oriented bounding boxes (OBBs). This distance is calculated by considering the covariance matrices of the two oriented bounding boxes (OBBs), represented by a 1 , b 1 , c 1 for the first OBB and a 2 , b 2 , c 2 for the second OBB. Additionally, the difference in their center coordinates, given by ( x 1 x 2 ) and ( y 1 y 2 ) , is also factored in. The covariance matrices describe the orientation, shape, and spread of each OBB, while the differences in center coordinates capture their relative positions. The resulting value of D B provides a measure of how closely the two OBBs align geometrically. This distance is then used to compute the Probabilistic IoU (ProbIoU), which assesses the overlap between the OBBs. Ultimately, the final result of the ProbIoU is determined by the following:
P r o b I o U = 1 1 e D B
While ProbIoU is effective in assessing the overlap between bounding boxes, the introduction of Complete IoU (CIoU) significantly improves the accuracy of the evaluation. CIoU extends ProbIoU by adding an aspect ratio consistency term, which is expressed in the following equation:
C I O U = P r o b I o U v α
Here, the aspect ratio consistency term v is defined as follows:
v = 4 π 2 tan 1 w 2 h 2 tan 1 w 1 h 1 2
The weight α is then used to adjust the influence of the aspect ratio on the final IoU value; this can be obtained from the following:
α = v v P r o b I o U + ( 1 + ϵ )
This enhancement allows CIoU to capture the geometric relationships between OBBs more comprehensively, thereby improving the detection performance.
However, while CIoU enhances accuracy, it primarily focuses on the aspect ratio and does not fully capture the relationship between the bounding box’s width, height, and its confidence. To address this limitation, we replace CIoU with the EIoU loss function in ProbIoU. EIoU incorporates not just the center distance, but also the differences in area and size of the bounding box, enhancing both the stability and precision of target box regression.
The EIoU loss function is defined as follows:
L E I O U = L d i s + L I o U + L a s p = 1 IoU + ρ ( b , b g t ) 2 1 C d 2 + ρ ( w , w g t ) 2 1 C w 2 + ρ ( h , h g t ) 2 1 C h 2
where L E I O U is the expression of the overall loss, which consists of three losses: L d i s , L I o U , and L a s p . Specifically, L d i s represents the loss function that quantifies the disparity in distance between the predicted center points and the corresponding ground truth center points of the bounding boxes. L I o U evaluates the degree of overlap between the predicted and actual bounding boxes by considering the ratio of their intersection to the total area covered by both. L a s p penalizes discrepancies in shape between the estimated and true bounding boxes by ensuring their aspect ratios are closely aligned. The integration of these three components enables a more thorough refinement of the bounding box’s alignment, proportions, and dimensions, thereby enhancing the precision and robustness of the detection process.

2.6. Pixel Error Converted to Actual Distance Error

During the process of mobile target recognition and landing, the visual system detects the moving landing target as a pixel error, which must be converted into a distance error before it can be provided to the flight control system for flight control. This conversion involves using parameters such as camera focal length, pinhole imaging model, imaging relationship, and camera–target distance to convert the pixel error in the image coordinate system into the true distance error in the world coordinate system. The relevant camera parameters used in this article are shown in Table 1.
The error conversion process uses camera parameters, including the focal length f c and pixel size ( p x , p y ). For the Y-axis, the pixel error ϵ camera is converted to the image width error R y in meters using the following:
R y = ϵ camera M y M y p y 10 3
Here, M y and M y are the pixel counts of the original and processed images, and p y is the physical pixel size.
To convert the image width error into the actual distance error in the geographic coordinate system, a proportional relationship between the original and processed images is used to scale the pixel size. Using the pinhole imaging model, we can obtain the following:
1 h c + 1 v = 1 f c
Here, h c (corrected distance) is obtained by subtracting the vehicle height from the radar-measured distance if the straight-line error exceeds 0.7 m. f c is the camera’s focal length, and v is the image height. Based on the principle of similar triangles, Equation (12) is established:
P u R y = h c v
where P u is the distance between the quadcopter and the landing target, and its ratio to the image error R y equals the ratio of h c to v.
According to Equations (11)–(12), we can obtain the final true distance error in the world coordinate system P u :
P u = R y h c f c R y
Since the camera used in this study is not equipped with a gimbal, the attitude angle generated during the quadcopter’s flight will impact the accuracy of the conversion in the above equation. Therefore, compensation is necessary to ensure that the pixel error can be accurately converted to the actual distance error, even when the quadcopter has an attitude angle. The conversion relationship is illustrated in Figure 8:
Figure 8. Compensation by attitude angle.
Figure 8. Compensation by attitude angle.
Drones 09 00261 g008
P c = P u cos ( θ ) P h = P u cos ( θ ) h c tan ( θ )
Here, P c is the compensated distance error, θ is the quadcopter’s current attitude angle, and P h is the auxiliary calculated distance. Using Equations (10)–(14), the actual distance error available to the controller is successfully obtained using the detected pixel error.

3. Control System Design

In a dynamic quadcopter visual landing system, identifying the landing target is crucial, but ensuring stable and reliable tracking during landing is equally important. To handle inevitable disturbances, the system incorporates sliding mode control, valued for its robustness against external interference. However, its application often generates large control outputs to counter deviations, potentially causing abrupt attitude changes during landing. To address this, a reference model is introduced to ensure smooth and safe state transitions, enhancing control safety without sacrificing performance. A Kalman filter is also employed to improve stability through accurate state estimation. Combining these, a sliding mode controller integrated with a reference model and Kalman filter is proposed, achieving safe and precise control during dynamic visual landing. The complete controller layout is shown in Figure 9.
Figure 9 depicts the control system architecture, which includes an outer loop for position control and an inner loop for attitude control. The outer loop generates the desired Roll and Pitch targets based on the quadcopter’s position, which are then fed into the inner control loop to compute required torque to achieve the desired attitude and motion, ensuring stable flight behavior. The establishment of the model can refer to our previous study [11].

3.1. Reference Model Design

The dynamics of the system are described by using the Newton–Euler formulation:
r ˙ = v m v ˙ = m g e z + R e z F ϕ ˙ = ω I ω ˙ = ω × I ω + M
where r and v are the position and velocity, m is the mass, and g is the gravitational acceleration with unit vector e z . F is the total thrust, ϕ and ω denote attitude and angular velocity, I is the inertia matrix, and M is the control torque.
Then, the three-axis virtual input quantities are introduced as defined in the following equations:
u x = F cos ϕ sin θ cos ψ + sin ϕ sin ψ
u y = F cos ϕ sin θ sin ψ sin ϕ cos ψ
u z = F cos ϕ cos θ
The position movement equations are simplified as in the following equations:
x ¨ = u x m
y ¨ = u y m
z ¨ = u z m g
Finally, the general form of the position equations can then be rewritten as follows:
X ˙ = A X + B u
Y = C X
where the input matrix is defined as B = [ 0 , 0 , 1 m ] T , the output matrix is C = [ 1 , 0 , 0 ] , and the A is given by the following:
A = 0 1 0 0 0 1 0 0 0
In the entire control process, ensuring the safety and stability of the quadcopter is crucial. To achieve this, the quadcopter must be able to track the target in a stable manner, with minimal abrupt changes in the target value. Therefore, the design of the transition process for the tracking state plays a key role. In light of this, this paper proposes the design of a reference model to facilitate the smooth and comprehensive state transitions required during the control process.
The reference model is designed with the key consideration that A r must ensure the model’s stability and ( A r , B r ) should be controllable.
X ˙ r = A r X r + B r r Y r = C r X r
Here, A r , B r , C r are the reference state matrix, reference input matrix, and reference output matrix, respectively, and X r denotes the reference state associated with X, where the target values of position and velocity are expressed as [ P o s m , V e l m ] T . The structure of the matrix A r is as follows:
A r = 0 1 k m 1 k m 2
This matrix is crucial as it governs the change in the reference state. By designing the parameters k m 1 and k m 2 , the tracking speed of the transition state can be adjusted. And B r = B ( C r ( A r ) 1 B ) 1 .

3.2. Design of Sliding Mode Controller

In order to enable the quadcopter to land stably and safely at the center of the moving target, we designed a sliding mode controller to help achieve this goal. Following the design of the Kalman filter and reference model, obtaining the reference transition state, and estimating the system states, the RMSMC can be designed. The difference between the quadcopter’s current state and the transition state given by the reference is defined as E x = X X r , and the time derivative of the error, E x ˙ , can be computed as follows:
E x ˙ = X ˙ X r ˙ E x ˙ = ( A A r ) X + A r E x + B u B r r
Then, the sliding surface can be designed as follows:
σ = S E x
where S R 3 × 1 is the weight vector that can be adjusted for each state error. Therefore, by differentiating Equation (27), we obtain the following:
σ ˙ = S ( ( A A r ) X + A r E x + B u B r r )
Once the sliding mode condition is met, the system reaches the sliding surface, meaning that, at this point, σ ˙ = σ = 0 . Assuming K c = A A r and K 0 = ( S B ) 1 S B r , the equivalent control input u e q is given by the following:
u e q = ( S B ) 1 S K c X ( S B ) 1 S A r E x + K 0 r
In addition, to minimize the chattering caused by the sliding mode, a smooth function is used in this paper to replace the traditional sign function:
f ( σ ) = σ σ +
Among them, is a parameter that can be adjusted artificially, which is related to the disturbance size. Therefore, the nonlinear output is defined as U s w = K s w f ( σ ) where K s w = m ( S B ) 1 , m > 0 is the nonlinear gain. The final controller output is shown in the following equation:
u = u e q + u s w
To confirm the stability of the designed RMSMC, the Lyapunov function can be defined as follows:
V = 1 2 σ 2
To find the time derivative of V, we proceed as follows:
V ˙ = σ σ ˙
By incorporating the derivative of the sliding mode surface, we obtain the following equation:
V ˙ = σ [ S ( K c X + A r E x + B u B r r ) ]
Simplify the expression by expanding and rearranging terms:
V ˙ = σ [ S K c X + S A r E x S K c X + S B r r S A r E x + S B K s w f ( σ ) S B r r ]
Simplify further to the following:
V ˙ = σ [ S B K s w f ( σ ) ]
Substitute K s w = m ( S B ) 1 ; we can obtain:
V ˙ = σ [ S B ( m ( S B ) 1 ) f ( σ ) ]
Simplify further to the following:
V ˙ = m σ f ( σ ) V ˙ = m σ 2 σ +
Since both m and are set to positive values and can be adjusted, it follows that V ˙ 0 . Therefore, by the LaSalle invariance principle, the sliding mode controller designed above is asymptotically stable.

4. Simulation and Experiment

This section first presents the evaluation of the improved YOLOv8 OBB model, followed by experimental verification of the pixel error conversion. Finally, it provides validation of the controller and verification through actual flight tests.

4.1. Indicators of the Evaluation

This paper evaluates the improved model using precision, recall, mAP@0.5, and mAP@0.5:0.95. Precision measures the proportion of correctly identified positive samples, while recall indicates the model’s ability to detect all actual positives. The formulas for both metrics are as follows:
P r e c i s i o n = TP TP + FP
R e c a l l = TP TP + FN
AP is the area under the precision–recall curve, and mAP is the average across categories. mAP@0.5 corresponds to an IoU threshold of 0.5, while mAP@0.5:0.95 averages over multiple IoU thresholds ( 0.5 I O U 0.95 ). These metrics, along with C and K, provide a full evaluation of the YOLO model. Precision and recall at IoU threshold K are P ( K ) and R ( K ) , as shown as follows:
Δ R ( k ) = R ( k ) R ( k 1 )
A P = k = 1 N P ( k ) · Δ R ( k )
m A P = 1 C k = 1 N P ( k ) · Δ R ( k )

4.2. Experimental Dataset

During the data collection process, the first-level landing target in this paper is a self-designed pattern of 50 cm × 50 cm attached to a 1 m × 1 m wooden board, and the other secondary guided landing target is a vehicle. During the data collection process, photos of the first-level, secondary, and mixed landing targets at different heights and angles were taken at different time periods such as daytime (strong sunlight, weak sunlight) and night. At the same time, photos of the wrong landing points without targets and wrong patterns were taken. The quadcopter took images of 1920 × 1080 pixels at high altitude. The collected images were flipped, denoised, and color converted to enhance the richness and robustness of the dataset. After collecting the data, the roLabelImg open source software (https://github.com/cgvict/roLabelImg (accessed on 29 Jun 2020); version 1.8.6.) was used to mark the targets in the image to complete the annotation of the datasets.

4.3. Process of Model Training

The model was trained in a Windows 11 environment, utilizing a GeForce RTX3090Ti GPU and an Intel(R) Core(TM) i7-14700K CPU. Python 3.8.18 was used for application development; the deep learning framework was Pytorch 2.2.1, CUDA 11.8. The model was trained with the following initial settings: an input image size of 640 × 480, a warmup period of 5 epochs, a weight decay of 0.0003, and a total of 200 training epochs.

4.4. Ablation Experiments

4.4.1. Attentional Mechanisms’ Experiment

In this experiment, we incorporated different attention mechanisms with different coordinate configurations after the initial convolutional layer and trained the model on the public VisDrone/DroneVehicle dataset. The results are summarized in the Table 2. As observed, under various attention mechanisms, the mAP @ 0.5 : 0.95 is 0.882 percentage points higher than the baseline network and 0.572 percentage points higher than CA. Although it is 0.035 percentage points lower than GAM, the parameter count is significantly lower than that of GAM. Therefore, CBAM demonstrates the most noticeable improvement overall.

4.4.2. Loss Function

This section presents an evaluation of the performance of several loss functions, including CIoU, DIoU, and EIoU, on the public VisDrone/DroneVehicle dataset. Table 3 displays the experimental results. As observed, the EIoU loss function outperforms the others, with a 0.198% improvement over the CIoU loss function and a 0.036% improvement over the DIoU loss function based on the mAP@0.5:0.95.

4.4.3. Detect Head

This section assesses the effectiveness of various detection heads, including LSCD, LSDECD, LADH, and Dyhead, on the public VisDrone/DroneVehicle dataset. Table 4 presents the experimental results. As observed, under different detection head mechanisms, the mAP@0.5:0.95 is 2.586% higher than the baseline network, 2.685% higher than LSCD, 4.741% higher than LADH, and 0.23 percentage points lower than LSDECD. However, the mAP@0.5 is 2.554% higher than LSDECD. Overall, Dyhead performs the best.

4.4.4. Model Improvement Results

Finally, we conducted ablation experiments on both public and private datasets. The first row represents the baseline YOLOv8 OBB model, and the second to fourth rows correspond to the progressively improved models, where “” indicates the activation of the introduced modules.
As shown in Table 5, on the public dataset, the improved model demonstrates a 3.1% increase in the mAP@0.5:0.95 and a 2.61% improvement in the mAP@0.5, compared to the baseline.
As shown in Table 6, on the private dataset, the improved model achieves a 0.5% higher mAP@0.5:0.95 and a 2.1% higher mAP@0.5 compared to the baseline model.
In conclusion, we evaluated the model on both the VisDrone/DroneVehicle and private datasets. This is shown in Figure 10, where Figure 10a,c represent the performance results of YOLOv8 OBB, and Figure 10b,d represent the performance results of our improved model. As observed in the figure, the improved model enhances vehicle recognition accuracy by 0.07 on the VisDrone/DroneVehicle dataset. On the private dataset, both vehicle and mark recognition accuracies are improved by 0.05. In summary, our improved model effectively enhances both recognition accuracy and precision.
To assess the robustness of the model, we conducted tests in various environmental conditions, as shown in Figure 11. These tests included scenarios with rain, strong sunlight reflection (which caused the small mark to appear white), and night-time street lighting. The results demonstrate that our model can accurately detect landing targets, exhibiting a degree of safety and stability across these challenging environments.

4.5. System Verification

Experiment Platform

This paper employs a quadcopter, the structure and key sensors of which are depicted in Figure 12. The flight control system and several sensors of the quad-rotor were developed within our laboratory. The quadcopter is equipped with a six-axis inertial measurement unit (IMU), a geomagnetic sensor (MAG), a GNSS, a flight data recorder (FDR), LEDs, an NVIDIA AGX, a radar, and various other modules. With an arm length of 0.762 m and a total weight of 5.7 kg, the quad-rotor is engineered for superior performance and versatility.

4.6. The Verification of Pixel Conversion Result

To assess the efficacy of the proposed pixel conversion method and attitude angle range error compensation, two experiments were conducted. The first experiment focused on validating the pixel conversion process. As shown in Figure 13, a landing identification target was placed beneath the camera while the quadcopter maintained a fixed altitude. The landing mark was shifted along the aircraft’s X-axis in 20 cm steps. The flight control system recorded the range error data to evaluate the precision of the proposed pixel conversion method.
Figure 14 displays the data of the conversion verification experiment. In Figure 14a,b, the solid red line represents the actual distance measured in meters; meanwhile, the blue dotted line represents the distance calculated after the conversion. According to the gap in data, it can be seen that the distance conversion performed by the visual system is both accurate and reliable.
It can also be found from Table 7, that the gap between our converted distance error and the given standard value is very small, proving that our proposed conversion method is effective and feasible.
The second experiment tests attitude angle compensation by tilting the quadcopter, holding it, and repeating the process while calculating distance error, without using a gimbal.
As the quadcopter tilts, compensations are applied to the front and rear X-axis (Pitch), Y-axis (Roll), and height based on the attitude angle information. Figure 15 shows the results of the height and position error compensation. Both distance and height are measured in meters. In Figure 15a, the quadcopter’s attitude angle is shown in degrees. Figure 15b shows height compensation, with actual, pre-compensation, and post-compensation heights represented by red, solid blue, and light blue dotted lines, respectively. Figure 15c and Figure 15d show the attitude angle compensation for X and Y axis distance errors, with similar line conventions. These results confirm the method’s effectiveness.
Table 8 describes the height and distance errors before and after compensation. Table 8 shows that the maximum error of height before compensation reaches 0.18 m, while after compensation it is only 0.028 m, and the error is reduced by 84%. When the Roll angle moves, the distance measurement in the Y direction is significantly influenced by the attitude change, and the maximum error of the Y direction before compensation reaches 0.2922 m, while after compensation it is only 0.049 m, and the error is reduced by 83.6%.

4.7. Dynamic Landing Tracking Experiment

The specific design process of the sliding mode controller follows the procedure outlined in [11]. The dynamic platform moves at an average speed of about 1 m/s and the actual flight experiment environment is shown in Figure 16:
During the flight, the drone visual recognition map is as shown in Figure 17:
This experiment demonstrated the quadcopter’s control performance using a sliding mode controller integrated with the YOLOv8 visual tracking system, as illustrated in Figure 18. For position tracking, Figure 18a shows the reference position in red, while the actual tracking position is depicted by the blue dashed line. The results show that the sliding mode controller accurately tracks the reference trajectory, with the largest tracking error observed being 0.356 meters. The speed tracking performance is shown in Figure 18b, where the red line is the reference speed; meanwhile, the blue dashed line shows the actual speed, with the maximum recorded tracking error being 0.1258 m/s. Errors in position and speed tracking are detailed in Figure 19. In terms of visual tracking, Figure 18c,d display the tracking deviations in the north and east components. The northward tracking error gradually converges to within 0.13 m after the initial phase, reducing to 0.08 m upon landing. Meanwhile, the eastward tracking error remains stable within 0.25 m, with a final landing error of 0.11 m. The maximum error magnification before landing is shown in Figure 20. Throughout the process, the landing target was consistently detected.
This experiment verifies the effectiveness of the sliding mode controller and YOLOv8 visual tracking system in the dynamic landing of the UAV. As summarized in Table 9, the maximum position tracking error is 0.356 m, and the maximum velocity tracking error is 0.1258 m/s, showing good trajectory tracking ability. In terms of visual tracking, the north and east position deviations converge to 0.08 m and 0.11 m, respectively, showing the tracking ability of the system in dynamic landing. The overall experimental results prove the stability and robustness of the method in dealing with complex scenes, and provide reliable support for the dynamic landing of the UAV. Overall, the sliding mode controller and the YOLOv8 visual system show good robustness, providing reliable support for the dynamic landing of the UAV, thus proving the effectiveness and reliability of the method in actual scenes.

5. Discussion

The improved dynamic landing system proposed in this paper has demonstrated effectiveness in key aspects such as target detection, pixel conversion, and dynamic landing control. Comprehensive experimental results show that, for the baseline YOLOv8 OBB model, challenges commonly encountered during the landing process—such as difficulties in identifying small targets and the presence of complex, high-background environments—were addressed by integrating the CBAM attention mechanism, the DyHead detection head, and the EIoU loss function. These enhancements led to a steady improvement in recognition accuracy across both public and private datasets, validating the effectiveness of each module in boosting model performance and target detection accuracy within the experimental setting.
Additionally, for landing guidance in the control process, experiments on pixel error conversion and attitude angle compensation revealed that the proposed conversion algorithm and compensation strategy not only significantly reduced measurement errors under fixed height and varying attitudes but also provided precise position information for subsequent dynamic control. Furthermore, dynamic landing experiments employing a sliding mode controller demonstrated that the system maintained low maximum position and velocity errors while tracking the reference position, ensuring the successful completion of the landing task.
In conclusion, this study enhanced the accuracy and stability of UAV target detection and visual tracking by optimizing the landing recognition algorithm and implementing a sliding mode controller based on a reference model. However, future work should further explore the adaptability and robustness of the proposed approach in larger-scale environments with dynamic and unpredictable conditions.

6. Conclusions

This paper designs and implements a dynamic quadcopter landing system with drone vision and sliding mode control. Landing target detection uses the improved YOLOv8 OBB algorithm. CA, Dyhead, and EIoU Loss are added to the original model to improve the accuracy of the detection. Then, the detected pixel error is converted to the actual position error. At the same time, the attitude compensation is corrected. Furthermore, in order to ensure the stability and rapidity of tracking, a position sliding mode controller based on the reference model is designed, verified, and applied to help the quadcopter track the landing target. Finally, the practicality of the proposed system is ultimately validated through real-world flight experiments. The results demonstrate that the improved YOLOv8 OBB demonstrates an improvement in the mAP0.5:0.95 index by 2.23 percentage points. The position tracking error is always below 0.2 m, and the final landing error is 0.11 m, which proves that the system in this paper can complete accurate dynamic landing.
However, it should be noted that, although the RMSMC has a resistance to disturbances, no additional artificially generated interference was added in this experiment. In future research, we will add additional interference factors, such as continuous wind interference or a swaying landing surface.

Author Contributions

Conceptualization, H.W. and W.W.; methodology, H.W.; software, T.W.; validation, H.W. and T.W.; formal analysis, H.W.; investigation, H.W.; resources, H.W.; writing—original draft preparation, H.W.; writing—review and editing, S.S. and W.W. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is based on results obtained from a project, JPNP22002, commissioned by the New Energy and Industrial Technology Development Organization (NEDO).

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Budiyono, A.; Higashino, S.I. A review of the latest innovations in uav technology. J. Instrumentation, Autom. Syst. 2023, 10, 7–16. [Google Scholar]
  2. Laghari, A.A.; Jumani, A.K.; Laghari, R.A.; Nawaz, H. Unmanned aerial vehicles: A review. Cogn. Robot. 2023, 3, 8–22. [Google Scholar] [CrossRef]
  3. Su, J.; Zhu, X.; Li, S.; Chen, W.H. AI meets UAVs: A survey on AI empowered UAV perception systems for precision agriculture. Neurocomputing 2023, 518, 242–270. [Google Scholar] [CrossRef]
  4. Mu, L.; Li, Q.; Wang, B.; Zhang, Y.; Feng, N.; Xue, X.; Sun, W. A Vision-Based Autonomous Landing Guidance Strategy for a Micro-UAV by the Modified Camera View. Drones 2023, 7, 400. [Google Scholar] [CrossRef]
  5. Pieczyński, D.; Ptak, B.; Kraft, M.; Piechocki, M.; Aszkowski, P. A fast, lightweight deep learning vision pipeline for autonomous UAV landing support with added robustness. Eng. Appl. Artif. Intell. 2024, 131, 107864. [Google Scholar] [CrossRef]
  6. Zeng, Q.; Jin, Y.; Yu, H.; You, X. A UAV localization system based on double UWB tags and IMU for landing platform. IEEE Sensors J. 2023, 23, 10100–10108. [Google Scholar] [CrossRef]
  7. Cui, Q.; Liu, M.; Huang, X.; Gao, M. Coarse-to-fine visual autonomous unmanned aerial vehicle landing on a moving platform. Biomim. Intell. Robot. 2023, 3, 100088. [Google Scholar]
  8. Xin, L.; Tang, Z.; Gai, W.; Liu, H. Vision-based autonomous landing for the UAV: A review. Aerospace 2022, 9, 634. [Google Scholar] [CrossRef]
  9. Khazetdinov, A.; Zakiev, A.; Tsoy, T.; Svinin, M.; Magid, E. Embedded ArUco: A novel approach for high precision UAV landing. In Proceedings of the 2021 International Siberian Conference on Control and Communications (SIBCON), Kazan, Russia, 13–15 May 2021; pp. 1–6. [Google Scholar]
  10. Morales, J.; Castelo, I.; Serra, R.; Lima, P.U.; Basiri, M. Vision-based autonomous following of a moving platform and landing for an unmanned aerial vehicle. Sensors 2023, 23, 829. [Google Scholar] [CrossRef]
  11. Wu, H.; Wang, W.; Wang, T.; Suzuki, S. Sliding Mode Control Approach for Vision-Based High-Precision Unmanned Aerial Vehicle Landing System Under Disturbances. Drones 2025, 9, 3. [Google Scholar] [CrossRef]
  12. Gavrilescu, R.; Zet, C.; Foșalău, C.; Skoczylas, M.; Cotovanu, D. Faster R-CNN: An approach to real-time object detection. In Proceedings of the 2018 International Conference and Exposition on Electrical And Power Engineering (EPE), Iasi, Romania, 18–19 October 2018; pp. 0165–0168. [Google Scholar]
  13. Redmon, J. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  14. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  15. Sohan, M.; Sai Ram, T.; Reddy, R.; Venkata, C. A review on yolov8 and its advancements. In Proceedings of the International Conference on Data Intelligence and Cognitive Informatics, Tirunelveli, India, 18–20 November 2024; pp. 529–545. [Google Scholar]
  16. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  17. Lee, M.F.R.; Nugroho, A.; Le, T.T.; Bastida, S.N. Landing area recognition using deep learning for unammaned aerial vehicles. In Proceedings of the 2020 International Conference on Advanced Robotics and Intelligent Systems (ARIS), Taipei, Taiwan, 19–21 August 2020; pp. 1–6. [Google Scholar]
  18. Yu, L.; Luo, C.; Yu, X.; Jiang, X.; Yang, E.; Luo, C.; Ren, P. Deep learning for vision-based micro aerial vehicle autonomous landing. Int. J. Micro Air Veh. 2018, 10, 171–185. [Google Scholar]
  19. Suprapto, B.Y.; Wahyudin, A.; Hikmarika, H.; Dwijayanti, S. The detection system of helipad for unmanned aerial vehicle landing using yolo algorithm. J. Ilm. Tek. Elektro Komput. Dan Inform. 2021, 7, 193–206. [Google Scholar]
  20. Xu, Y.; Zhong, D.; Zhou, J.; Jiang, Z.; Zhai, Y.; Ying, Z. A novel uav visual positioning algorithm based on a-yolox. Drones 2022, 6, 362. [Google Scholar] [CrossRef]
  21. Mitroudas, T.; Balaska, V.; Psomoulis, A.; Gasteratos, A. Multi-criteria Decision Making for Autonomous UAV Landing. In Proceedings of the 2023 IEEE International Conference on Imaging Systems and Techniques (IST), Copenhagen, Denmark, 17–19 October 2023; pp. 1–5. [Google Scholar]
  22. Serrano, K.K.D.; Bandala, A.A. YOLO-Based Terrain Classification for UAV Safe Landing Zone Detection. In Proceedings of the 2023 IEEE Region 10 Symposium (TENSYMP), Canberra, Australia, 6–8 September 2023; pp. 1–5. [Google Scholar]
  23. Lin, J.; Wang, Y.; Miao, Z.; Zhong, H.; Fierro, R. Low-complexity control for vision-based landing of quadrotor UAV on unknown moving platform. IEEE Trans. Ind. Inform. 2022, 18, 5348–5358. [Google Scholar] [CrossRef]
  24. Wu, L.; Wang, C.; Zhang, P.; Wei, C. Deep reinforcement learning with corrective feedback for autonomous uav landing on a mobile platform. Drones 2022, 6, 238. [Google Scholar] [CrossRef]
  25. Mohammadi, A.; Feng, Y.; Zhang, C.; Rawashdeh, S.; Baek, S. Vision-based autonomous landing using an MPC-controlled micro UAV on a moving platform. In Proceedings of the 2020 International Conference on Unmanned Aircraft Systems (ICUAS), Athens, Greece, 1–4 September 2020; pp. 771–780. [Google Scholar]
  26. Qassab, A.; Khan, M.U.; Irfanoglu, B. Autonomous landing of a quadrotor on a moving platform using motion capture system. Discov. Appl. Sci. 2024, 6, 304. [Google Scholar]
  27. Ghasemi, A.; Parivash, F.; Ebrahimian, S. Autonomous landing of a quadrotor on a moving platform using vision-based FOFPID control. Robotica 2022, 40, 1431–1449. [Google Scholar]
  28. Ghommam, J.; Saad, M. Autonomous landing of a quadrotor on a moving platform. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 1504–1519. [Google Scholar] [CrossRef]
  29. Sun, L.; Huang, Y.; Zheng, Z.; Zhu, B.; Jiang, J. Adaptive nonlinear relative motion control of quadrotors in autonomous shipboard landings. J. Frankl. Inst. 2020, 357, 13569–13592. [Google Scholar] [CrossRef]
  30. Xia, K.; Lee, S.; Son, H. Adaptive control for multi-rotor UAVs autonomous ship landing with mission planning. Aerosp. Sci. Technol. 2020, 96, 105549. [Google Scholar]
  31. Wang, Q.; Wang, W.; Suzuki, S.; Namiki, A.; Liu, H.; Li, Z. Design and implementation of UAV velocity controller based on reference model sliding mode control. Drones 2023, 7, 130. [Google Scholar] [CrossRef]
  32. Wang, Q.; Namiki, A.; Asignacion, A., Jr.; Li, Z.; Suzuki, S. Chattering Reduction of Sliding Mode Control for Quadrotor UAVs Based on Reinforcement Learning. Drones 2023, 7, 420. [Google Scholar] [CrossRef]
  33. Wang, Q.; Wang, W.; Suzuki, S. UAV trajectory tracking under wind disturbance based on novel antidisturbance sliding mode control. Aerosp. Sci. Technol. 2024, 149, 109138. [Google Scholar]
  34. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  35. Dai, X.; Chen, Y.; Xiao, B.; Chen, D.; Liu, M.; Yuan, L.; Zhang, L. Dynamic head: Unifying object detection heads with attentions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7373–7382. [Google Scholar]
  36. Jie, H.; Li, S.; Gang, S. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  37. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  38. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11534–11542. [Google Scholar]
  39. Wang, S.; Li, K.; Chen, J.; Zhang, T. Unmanned Aerial Vehicle Autonomous Visual Landing through Visual Attention-Based Deep Reinforcement Learning. In Proceedings of the 2023 42nd Chinese Control Conference (CCC), Tianjin, China, 24–26 July 2023; pp. 4143–4148. [Google Scholar]
  40. Zhang, S.; Zang, S.; Liu, S. Deep Learning-Based Ship Detection in Maritime Environments. In Proceedings of the 2024 7th International Conference on Computer Information Science and Artificial Intelligence, Shaoxing, China, 13–15 September 2024; pp. 413–418. [Google Scholar]
Figure 1. Quadcopter structure diagram: using the center of the quadcopter body (point A) as the origin (0, 0, 0), measured in meters, the camera is located at (0, 0, −0.27) of the quadcopter, and Nvidia AGX is located at (0, 0, −0.22).
Figure 1. Quadcopter structure diagram: using the center of the quadcopter body (point A) as the origin (0, 0, 0), measured in meters, the camera is located at (0, 0, −0.27) of the quadcopter, and Nvidia AGX is located at (0, 0, −0.22).
Drones 09 00261 g001
Figure 2. Control Systems.
Figure 2. Control Systems.
Drones 09 00261 g002
Figure 3. The structure of the improved YOLOv8 OBB.
Figure 3. The structure of the improved YOLOv8 OBB.
Drones 09 00261 g003
Figure 4. CBAM’s overall structure.
Figure 4. CBAM’s overall structure.
Drones 09 00261 g004
Figure 5. CAM’s structure.
Figure 5. CAM’s structure.
Drones 09 00261 g005
Figure 6. SAM’s structure.
Figure 6. SAM’s structure.
Drones 09 00261 g006
Figure 7. Dynamic head’s structure.
Figure 7. Dynamic head’s structure.
Drones 09 00261 g007
Figure 9. Control system.
Figure 9. Control system.
Drones 09 00261 g009
Figure 10. Experiment platform.
Figure 10. Experiment platform.
Drones 09 00261 g010
Figure 11. Different environmental model detection experiments. Among them (a) is a cloudy and rainy environment, (b) is a sunny environment, and (c) is an evening environment.
Figure 11. Different environmental model detection experiments. Among them (a) is a cloudy and rainy environment, (b) is a sunny environment, and (c) is an evening environment.
Drones 09 00261 g011
Figure 12. Experimental quadcopter’s platform.
Figure 12. Experimental quadcopter’s platform.
Drones 09 00261 g012
Figure 13. Pixel error conversion experiment.
Figure 13. Pixel error conversion experiment.
Drones 09 00261 g013
Figure 14. Fixed angle pixel error conversion.
Figure 14. Fixed angle pixel error conversion.
Drones 09 00261 g014
Figure 15. Height and position compensation experiment. Among them, (a) is the real-time attitude angle change in cloudy and rainy weather, (b) is the height change before and after compensation, (c) is the distance change in the X-axis direction before and after compensation, and (d) is the distance change in the Y-axis direction before and after compensation.
Figure 15. Height and position compensation experiment. Among them, (a) is the real-time attitude angle change in cloudy and rainy weather, (b) is the height change before and after compensation, (c) is the distance change in the X-axis direction before and after compensation, and (d) is the distance change in the Y-axis direction before and after compensation.
Drones 09 00261 g015
Figure 16. Experiment environment.
Figure 16. Experiment environment.
Drones 09 00261 g016
Figure 17. Experiment environment.
Figure 17. Experiment environment.
Drones 09 00261 g017
Figure 18. Tracking experiment result.
Figure 18. Tracking experiment result.
Drones 09 00261 g018
Figure 19. Tracking error.
Figure 19. Tracking error.
Drones 09 00261 g019
Figure 20. Position error magnified.
Figure 20. Position error magnified.
Drones 09 00261 g020
Table 1. Camera specifications.
Table 1. Camera specifications.
Camera ParametersValue
Optical Aperture (f-stop)f/2.25
Focal Length3.35 mm
Diagonal Field of View180°
Resolution2592 × 1944 px
Table 2. Performance comparison of YOLOv8 OBB and its variants with different attention modules.
Table 2. Performance comparison of YOLOv8 OBB and its variants with different attention modules.
MethodsParameters (M)GflopsmAP0.5 (%)mAP0.5:0.95 (%)Precision (%)Recall
Baseline (YOLOv8_obb)3.0838.4474.350.5174.930.70401
Baseline + CA3.098.45374.76850.86573.9460.705
Baseline + ShuffleAttention3.0838.44674.87251.10976.0040.699
Baseline + GAM3.5188.79475.751.42776.3380.712
Baseline + CBAM3.0928.4675.451.39276.8180.705
Table 3. Performance comparison of YOLOv8 OBB and its variants with different loss functions.
Table 3. Performance comparison of YOLOv8 OBB and its variants with different loss functions.
MethodsmAP0.5 (%)mAP0.5:0.95 (%)box_lossdfl_losscls_loss
Baseline (YOLOv8_obb)74.29550.5090.7421.5430.516
Baseline + CIoU75.70251.4030.7181.3770.519
Baseline + DIoU75.41951.5560.7561.5770.492
Baseline + EIoU75.84351.5920.7181.3740.498
Table 4. Performance comparison of YOLOv8 OBB and its variants with different detection heads.
Table 4. Performance comparison of YOLOv8 OBB and its variants with different detection heads.
MethodsParameters (M)GflopsmAP0.5 (%)mAP0.5:0.95 (%)Precision (%)Recall
Baseline (YOLOv8_obb)3.0838.44174.30150.51174.9310.704
Baseline + LSCD2.3816.72772.54250.42170.9540.689
Baseline + LSDECD2.5975.40674.76153.30873.9490.726
Baseline + LADH3.1288.08169.45148.33771.8570.699
Baseline + Dyhead3.3237.71677.31553.07876.5960.728
Table 5. Comparison of different configurations in VisDrone/DroneVehicle.
Table 5. Comparison of different configurations in VisDrone/DroneVehicle.
YOLOv8_obbCBAMDyheadEIoUParameters (M)GflopsmAP0.5 (%)mAP0.5:0.95 (%)
3.0838.44174.29150.501
3.0928.46275.401 (+1.11)51.391 (+0.89)
3.3327.71176.891 (+2.6)52.712 (+2.211)
3.3327.72177.394 (+3.103)53.114 (+2.613)
Table 6. Comparison of different configurations in MarkWithCar.
Table 6. Comparison of different configurations in MarkWithCar.
YOLOv8_obbCBAMDyheadEIoUParameters (M)GflopsmAP0.5 (%)mAP0.5:0.95 (%)
3.0828.44195.90069.301
3.0918.45296.301 (+0.401)69.902 (+0.601)
3.3317.72296.303 (+0.403)71.102 (+1.801)
3.3317.72296.403 (+0.503)71.403 (+2.102)
Table 7. Analysis of fixed-angle pixel error conversion.
Table 7. Analysis of fixed-angle pixel error conversion.
ParameterTrue ValueDistance Max Error
X Distance (m)[−0.4, 0.4]0.017
Y Distance (m)0.00.0039
Table 8. Comparison of height and distance before and after compensation.
Table 8. Comparison of height and distance before and after compensation.
ParameterWithout CompensationWith Compensation (Increase)True Value
Measuring Height (m)2.54 to 2.722.50 to 2.57 ( 95 % )2.54
Measuring X Distance (m)−0.84 to −0.77−0.82 to −0.78 ( 74 . 32 % )−0.8
Measuring Y Distance (m)0.93 to 1.271.18 to 1.28 ( 96 . 2 % )1.22
Table 9. Summary of quadcopter landing experiment results.
Table 9. Summary of quadcopter landing experiment results.
ParametersMetricMaximum ErrorFinal Error
Position TrackingReference vs. Actual Position0.36 m
Speed TrackingReference vs. Actual Speed0.13 m/s
Visual Tracking (N)Northward Deviation0.13 m0.08 m
Visual Tracking (E)Eastward Deviation0.25 m0.11 m
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, H.; Wang, W.; Wang, T.; Suzuki, S. High-Precision Landing on a Moving Platform Based on Drone Vision Using YOLO Algorithm. Drones 2025, 9, 261. https://doi.org/10.3390/drones9040261

AMA Style

Wu H, Wang W, Wang T, Suzuki S. High-Precision Landing on a Moving Platform Based on Drone Vision Using YOLO Algorithm. Drones. 2025; 9(4):261. https://doi.org/10.3390/drones9040261

Chicago/Turabian Style

Wu, Hao, Wei Wang, Tong Wang, and Satoshi Suzuki. 2025. "High-Precision Landing on a Moving Platform Based on Drone Vision Using YOLO Algorithm" Drones 9, no. 4: 261. https://doi.org/10.3390/drones9040261

APA Style

Wu, H., Wang, W., Wang, T., & Suzuki, S. (2025). High-Precision Landing on a Moving Platform Based on Drone Vision Using YOLO Algorithm. Drones, 9(4), 261. https://doi.org/10.3390/drones9040261

Article Metrics

Back to TopTop