Next Article in Journal
Conversational Agents: Goals, Technologies, Vision and Challenges
Previous Article in Journal
Optimizing Traffic Engineering for Resilient Services in NFV-Based Connected Autonomous Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Step CNN-Based Estimation of Aircraft Landing Gear Angles

1
Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2
University of the Chinese Academy of Sciences, Beijing 100049, China
3
School of Computer Science, Peking University, Beijing 100091, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(24), 8440; https://doi.org/10.3390/s21248440
Submission received: 25 October 2021 / Revised: 2 December 2021 / Accepted: 11 December 2021 / Published: 17 December 2021
(This article belongs to the Section Physical Sensors)

Abstract

:
This paper presents a method for measuring aircraft landing gear angles based on a monocular camera and the CAD aircraft model. Condition monitoring of the aircraft landing gear is a prerequisite for the safe landing of the aircraft. Traditional manual observation has an intense subjectivity. In recent years, target detection models dependent on deep learning and pose estimation methods relying on a single RGB image have made significant progress. Based on these advanced algorithms, this paper proposes a method for measuring the actual angles of landing gears in two-dimensional images. A single RGB image of an aircraft is inputted to the target detection module to obtain the key points of landing gears. The vector field network votes the key points of the fuselage after extraction and scale normalization of the pixels inside the aircraft prediction box. Knowing the pixel position of the key points and the constraints on the aircraft, the angle between the landing gear and fuselage plane can be calculated even without depth information. The vector field loss function is improved based on the distance between pixels and key points, and synthetic datasets of aircraft with different angle landing gears are created to verify the validity of the proposed algorithm. The experimental results show that the mean error of the proposed algorithm for the landing gears is less than 5 degrees on the light-varying dataset.

1. Introduction

The landing gear status monitoring of the aircraft aims to simultaneously detect the aircraft target to obtain landing gear angle information, which is necessary for the safe landing of the aircraft. According to flight data worldwide, more than 60% of all flight accidents occurred during the landing phase, in just a few minutes [1]. Therefore, international airports attach great importance to the tracking and monitoring of aircraft in the landing stage, and the status monitoring of landing gear is an essential aspect of aircraft attitude monitoring.
There are special inspection requirements for aircraft landing gear status monitoring. Given an image of an aircraft, the purpose of detection is to calculate the angle between the landing gears and the fuselage plain. When the measured angle is 90 degrees, the landing gears are fully down, and the plane is deemed safe to land. This goal can be divided into two aspects: to determine whether the aircraft has landing gear, and the other is to measure the actual angle between the landing gear and the fuselage plane.
The traditional landing gear observation method is to equip observers with telescopes to observe the aircraft that is about to land. The disadvantage of this method is that it is pretty dependent on the observer’s eyesight and experience judgment, and it is fatiguable and subjective when observing for a long time. Although the instrument landing system can judge the aircraft’s position, the tower still lacks direct observation to evaluate the state of the landing gears. Suppose a monitoring system can be designed to judge the status of aircraft landing gear intelligently by using aircraft images taken from the ground to give timely warning. In that case, it will save many human resources and improve the stability and objectivity of aircraft monitoring. It is of great significance to ensure the safe and stable operation of the airport.
The landing gear status monitoring of aircraft has unique application characteristics. First of all, the plane has an extensive range of motion in the flight process, and the aircraft’s speed is fast, so the scale of the aircraft varies widely in the field of the camera view. Secondly, during the aircraft’s flight, the landing gears and some key points of the plane are easily blocked due to the different angles of the camera view. Third, the cost of depth cameras and radar equipment that can obtain depth information is relatively high. The monocular camera cannot get absolute depth information of aircraft. Fourthly, pose estimation networks based on a single RGB image, such as PVNet, require the object to be a constant rigid body [2]. Since the landing gear is mobilizable, the aircraft cannot wholly be detected as a rigid body.
To solve these problems, this paper proposes a method to measure the landing gear angle of aircraft based on a monocular camera and a 3D model of aircraft without depth information. This paper estimates the key points using multi-step CNN networks, using aircraft images taken from the ground monocular camera. Then, angles of landing gears are calculated according to the constraints of the aircraft. The monocular camera is attached to a theodolite, which is used to track the plane. The aircraft is approximatively imaged at infinity relative to the camera with a telephoto fixed focal length camera. And image processing methods such as wavelet transform are used to keep the plane clear. In contrast to obtaining angles through aircraft angle sensors, the proposed method utilizes computer vision directly obtained by ground control systems. The proposed plan will work well if aircraft sensors fail to communicate with ground control systems, adding another layer of assurance for the safe landing. Additionally, the vision equipment is cheap and easy to assemble. The average landing gear angle error of this method is less than 8 degrees.
The main contributions of our work are summarized as follows:
  • Object extraction and normalization modules are added to adapt to the change in aircraft target scale. The target detection module extracts the area of the aircraft, and the normalization module normalizes the area size. Then, the normalized area is inputted into the subsequent vector field regression network, voting for the key points of the fuselage. The normalization module could effectively avoid the error of the vector field network, which is caused by dramatic changes in the input image. The target detection module adopts an efficient YOLO-V4 model.
  • The aircraft is divided into the landing gear and the fuselage to detect key points. The landing gears are movable, and the shape of the aircraft without the landing gear changes very little and can be regarded as a rigid body. Therefore, the aircraft as a whole can be divided into the landing gear and the two parts of the fuselage. Aircraft key points consist of landing gear key points and fuselage key points. The target detection module obtains the key points of the landing gears. The key points of the fuselage are acquired by a robust pixel-level voting network, which requires the model to be a rigid body. In addition, the distance-based coefficients are multiplied by the loss function to optimize the vector field;
  • To resolve the difficulty of obtaining depth information, we propose a method to directly calculate the absolute angle between the landing gears and fuselage plane by using key positions in 2D images according to the constraints of aircraft, omitting the step of regaining 3D spatial coordinates;
  • This article contributes a synthetic aircraft dataset of different camera views containing landing gears with different angles to verify the algorithm performance.

2. Related Works

Image enhancement is intended to be easy for machines to recognize [3]. It includes improving the contrast between the partial and whole image and improving the signal-to-noise ratio of the picture. Histogram equalization technology is used to transform the gray histogram of the original image from a relatively concentrated gray range to a uniform distribution in all gray areas [4,5,6]. Wavelet transform decomposes the image signal to enhance the low-frequency coefficient and appropriately weaken the high-frequency coefficient, and the enhanced image is obtained after reconstruction [7,8,9,10].
Target detection models and pose estimation methods based on a single RGB image have significantly been improved in recent years with the development of deep learning. Target detection based on the convolutional neural network is divided into the two-stage algorithm represented by R-CNN series with high detection accuracy [11,12,13], as well as the single-stage algorithm described by YOLO series with high precision and high speed [14,15,16]. Based on the original YOLO target detection architecture, the YOLO-V4 [17] algorithm adopts the best optimization strategy in the field of CNNs in recent years, which has varying levels of optimization for different aspects such as data processing, backbone network, network training, and so forth.
Attitude estimation based on deep learning can be divided into three parts. The RGB image fusion depth information obtained by depth camera or lidar has the highest accuracy, but it relies on expensive hardware and requires a large number of calculations [18,19,20]. The overall direct regressions based on deep learning, such as 6DVNet [21], SSD-6D [22], and so forth, lack explicit geometric constraints and have low accuracy. In addition, there are two-stage attitude estimations, such as Deepmanta [23], YOLO-6D [24], PVNet [2], and other methods. Knowing the target model, the PVNet regresses the vector field network, voting for the target key points with strong robustness. Then, the PNP algorithm is used to calculate the space pose of the target.
To represent the deformable targets and for attitude estimation, Wang et al. [25] proposed a rigid target category representation method using Normalized Object Coordinate Space (NOCS) to standardize the given target in space. NOCS only estimates the 6D posture and scale transformation of the rigid bold mark and does not have a unique representation of each part’s posture and joint connection. Li et al. [26] proposed a class-oriented representation of variable goals, ANCSH. This method divides the deformable target into different rigid parts, and the rigid parts are expressed by normalized part coordinate space.
The deformable target attitude estimation is mainly concentrated in certain areas. The primary processing method is to obtain the target instance-level information from the accurate CAD model of the deformable target. For example, Brachmann et al. [27] used the random forest to vote for the parameters of each point in the image. Pavlasek et al. expressed the attitude estimation problem of deformable objects as Markov random field. These methods are constrained by the need to model accurately. In addition, [28,29,30,31] proposed proposed ways that manipulate and interact with the target and then estimate the deformable target attitude. These methods can assess the attitude of unknown targets, but it takes a lot of time to manipulate and interact with the target, and this can only be applied to simple deformable targets. For example, Schmidt et al. [32,33] tracked deformable targets through probability inference, which requires the definition of a standard geometric structure. Daniele et al. [34] combined natural motion language information and computer vision for the attitude estimation of deformable targets, but this method requires natural motion language description as an additional mode.
In recent years, two special deformable objects, the hand and the human body, have received much attention from researchers due to their wide application. The present hand pose estimation methods are mainly divided into generation, discriminant, and mixed methods. The generation method primarily uses optimization algorithms to fit the predefined hand pose model into the given image [35,36]. The discriminant method estimates hand posture by learning the mapping relationship from the input data to the key points of the hand, such as coordinates or joint angles [37,38,39,40]. The hybrid method usually uses the optimization model in the generation method to optimize the estimation results of the discriminant method [41]. Human pose estimation can be divided into bottom-up and top-down research methods. The bottom-up approach, such as OpenPose, MultiPoseNet, and so forth, mainly predicts heat maps of different junction points and then inputs images to detect human joints and groups them according to human bodies [42,43,44]. The top-down approach sees the human body boundary frame first and predicts the position of the key nodes in the boundary frame to realize the human body pose estimations [45,46,47]. These attitude-estimation methods for particular deformable targets are only applicable to unique targets and are difficult to extend to other deformable targets.

3. Proposed Methods

Figure 1 shows how the proposed method works. The angle estimation consists of two stages. The first step is to normalize the aircraft image and to detect key points of the aircraft fuselage and landing gear using the vector field network and YOLO-V4. Then, according to the aircraft’s structure and the pixel positions of key points, an algorithm is designed to calculate the angle between the landing gears and the fuselage.
YOLO-V4 network was used for the target detection. CSPDarkNet53 was used for the main feature extraction network. Feature pyramid SPP, PAN, and other structures were used to extract features [17]. Mish activation functions were used to improve accuracy and generalization performance [48]. Bilinear interpolation [49] was used in normalization, and the output size was 608 by 608 pixels. After normalization, the unstable error caused by scale variation was eliminated, and the detection accuracy was improved. Resnet18 [50] was used to obtain classification mask and the regression vector field voting for key points based on the RANSC method [2].

3.1. Norm Module

Firstly, the target detection module regresses out the detection frame of landing gear tires to calculate the pixel coordinates of the tires’ center points. Then, we can obtain the normalized center points according to the normalized size and the length of the object box of the aircraft. The relationship proposed in [14] is as follows:
( x , y ) = ( ( x m a x , y m a x ) + ( x m i n , y m i n ) ) / 2 ,
X = ( x x m i n ) / ( ( x m a x x m i n ) ) × d ,
Y = ( y y m i n ) / ( ( y m a x y m i n ) ) × d ,
where ( x , y ) is the pixel position of the original image, ( X , Y ) is the pixel position after normalization, ( x m i n , y m i n ) and ( x m a x y m a x ) is the upper left and lower right corner point of the detection frame, and d is the width of the normalized image.
In our work, we adopt a normalization model to obtain the key points of the landing gear and the normalized aircraft object.

3.2. Vector Field Loss Function Optimization

The target extraction module obtains the key points of the landing gear. The fuselage is treated as a rigid body, and a pixel-level vector field regression network votes for the key points of the fuselage. OpenPose [42] utilizes the vector field network to model the correspondence between key points, and PoseCNN [51] and PVNet [2] utilize it as an intermediate quantity to replace the regression of key points.
To detect the fuselage key points, the vector field regression network has two prediction objects, mask at pixel level in the target area, and the key point vector field in the target zone. The loss function proposed in [52] is composed of two parts.
L = λ 1 L s e g + λ 2 L v e c ,
where λ 1 and λ 2 represent the weight of loss function. L s e g and L v e r represent the loss values of the two tasks of target semantic segmentation and the key point vector field regression, respectively.
In the vector field network, the binary cross entropy loss function L s e g is used to segment aircraft and background region at the pixel level. Aircraft pixels are represented by 1 and background pixels by 0. The L v e r use the smooth L 1 loss proposed in [52] for learning unit vectors.
d k , p = ( k x p x ) 2 + ( k y p y ) 2 ,
v ( p ) = ( k x p x d k , p , k y p y d k , p ) ,
L v e c = k K p P ( s m o o t h L 1 ( v x | ( p , k ) g x | ( p , k ) ) + s m o o t h L 1 ( v y | ( p , k ) g y | ( p , k ) ) ,
where k is the key points. p is the current pixel. d k , p is the distance from k to p. ( k x , k y ) and ( p x , p y ) are the image coordinates of the corresponding key point and the current pixel. K is all the key points, P is the set of all instance pixels belong to the target, v x | ( p , k ) , g x | ( p , k ) , v y | ( p , k ) , g y | ( p , k ) respectively represent the x-axis and y-axis components of the predicted values and the true values.
The vector field regression network has two prediction objects: the target area’s pixel-level mask and the key point vector field.
Depending on whether it belongs to the target pixel or not, the pixel belonging to the aircraft is represented by one, and the background is represented by zero. The classification loss, L s e g , is calculated by the cross-entropy loss function. For vector field regression mission, a Smooth L1 loss function is used to calculate the L v e c on behalf of the sum of the loss values of normalized vectors from any key point to all pixel positions of the target.
The final goal of the vector field regression network is to predict the coordinates of key points on the image. However, the loss function of Formula (6) loses the distance information from pixels to key points. Figure 2a shows that the pixels around the key point are closer to the key point features. The pixels around key points account for the main body of loss in the later period of training. Still, the number of pixels around key points only accounts for a small part of all pixels of aircraft targets, making it more challenging to optimize the main part of the hole pixels [53]. In this paper, an alterable vector field loss function is proposed to enhance the pixel learning of the principal amount as follows:
L v e c K = K × L v e c ,
w h e n x p , k < λ , K = x p , k / λ , w h e n , x p , k λ , K = 1 ,
where L v e r is the key point vector field loss function proposed in [2], K is the distance-based weight value, x p , k represents the distance between the current pixel point p and the corresponding key point k, and λ is a constant.
Figure 2b shows that the loss values within 5 pixels from the key point are weakened. The closer the pixel is to the key point, the smaller the value of K. The loss coefficient of the main part of the pixel is relatively strengthened.

3.3. CAL Module

As we have obtained the pixel position of the key points of the aircraft through the target detection module and vector field network, we can calculate landing gear angles without the depth information according to the constraint relations of the plane.
Geometry has a wide range of applications in the structure [54,55,56,57]. We use the right triangle relationship and length ratio to calculate the landing gear angles. The structure of the airplane is abstracted into lines of key points as shown in Figure 3a. According to the part of the plane where the landing gear is, the landing gear to be measured is divided into the nose landing gear and the rear landing gear. The nose landing gear angle is formed by the nose landing gear and the fuselage, and the rear landing gear angle is formed by the rear landing gear and the wing belonging to the fuselage plane. Steps for calculating angles are as follows:
  • Find the right triangle between the landing gear and fuselage shown in Figure 3b. Because the aircraft’s vertical tail is perpendicular to the fuselage plane, in Figure 3a, the line that is parallel to the aircraft’s vertical line is perpendicular to the fuselage plane. Considering the rotating direction of the landing gear, the vertical foot of the nose landing gear is on the fuselage line, and the vertical foot of the rear landing gear is on the wing belonging to the fuselage plane.
  • Given the aircraft model, the true length in the triangle is calculated from the ratio of the length of each side to the length of the fuselage, wing, and tail respectively.
  • Then the sines and cosines is used to calculate the angle θ between the landing gears and fuselage plane.
sine:
θ s = arcsin ( x 1 x ) 2 + ( y 1 y ) 2 R × D w d w ,
cosine:
θ c = arccos ( x 1 x ) 2 + ( y 1 y ) 2 R × D k d k ,
sine and cosine:
θ = arctan ( x 1 x ) 2 + ( y 1 y ) 2 × D w d w ( x 1 x ) 2 + ( y 1 y ) 2 × D k d k ,
θ z = sin 2 θ × θ s + cos 2 θ × θ c ,
combine:
θ = C o m b ( [ θ s , θ c , θ z ] ) ,
where ( x 1 , y 1 ) is the key point of the landing gear, ( x 2 , y 2 ) is the connection point between landing gear and fuselage, and ( x , y ) is the intersection point required. d w , D w is the pixel length of the vertical tail line in the image and the real length of the space. For the nose gear d k , D k is the pixel length of the fuselage line in the image and the absolute length of the space. For the rear landing gear, d k , D k is the pixel length of the wing line in the image and the real length of the space.
The combined method is to choose different calculation methods according to the lines’ length, based on the experience in Figure 3. Consider a combined method of sines and cosines: when the length of the fuselage or wing is greater than 340 pixels size, the cosine method is used as { θ = θ c } . When the measurement is less than 120 pixels, the sine method is used as { θ = θ s } . Otherwise, the sine and cosine method is used as { θ = θ z } .
The angles of both rear landing gears are the same. The average value will be taken if both rear landing gears are in the field of view. If one of them is blocked, only the visible landing gear will be measured.

3.4. Synthetic Aircraft Datasets

Existing aircraft datasets are generally remote-sensing images used for target detection and classification. In contrast, few aircraft datasets are taken from the ground to the sky, and there is no dataset specifically used to obtain the attitude of the aircraft landing gear. Tobin et al. [58] proved the broad application prospect of random field sample generation through ablation experiments. Shravastava et al. [59], and Bousmalis et al. [60] trained the network by using an adversarial generation network to generate data in the target detection task and 3D pose estimation task, respectively, improving the accuracy in small sample cases. Hinterstoisser et al. [61] rendered textured CAD models from multiple perspectives to generate a large number of composite samples for training networks. The aircraft dataset with different attitude landing gears was made using an aircraft CAD model to verify the angle estimation algorithm’s effectiveness on aircraft landing gears.
Figure 4 shows the RGB image of the plane, the corresponding mask image, and the pixel coordinate marks of the key points. A 3D model of rendered aircraft is loaded into CAD software. Then the landing gears are rotated at different angles, as well as the aircraft is operated by rotation and translation. Different camera views are elected, and the real images and corresponding mask images are projected. At the same time, the spatial coordinates of the key points are derived, and the pixel positions of the key points are obtained according to the projection relations.
Figure 5 shows the input data form of the vector field network model made by the dataset. The normalized aircraft image and the mask of the aircraft area are used to carry out semantic segmentation of aircraft. The vector field images are used to regress the pixel-level vectors of each key point.
The key points of the aircraft are selected as corner points, and points with unique characteristics [2] that can be easily identified. A total of eight points are chosen for each image, including the endpoint of the aircraft wing, the head of the aircraft, the tail of the plane, and the center point of the aircraft. Three landing gear tires’ key points and three joining points between the landing gear and the aircraft are added to calculate the angle. The total quantity of the key points ranges up to 14. The tires’ center can be regarded as the key point, so their detection box is centrosymmetric, reducing the error of angle floating. The advantage is that the tires are apparent, and their features are easy to identify. The aircraft target detection box is a rectangle for normalization purposes. In addition, the datasets use different sky background images to increase the generalization ability of the algorithm.

4. Experiments

In the scale normalization module, YOLO-V4 is trained to detect the aircraft and landing gears targets. Label smoothing is set to 0.005, and training techniques were used, such as data enhancement, cyclic cosine annealing learning rate schedule, and freezing training. Considering the shielding of landing gears, it is marked for the positions of the landing gears and whether they were visible in the picture. The YOLO-V4 model is trained for about 200 rounds. For the vector field regression network, the fuselage key points is extracted by PVNet. The Adam optimizer is used, and the initial learning rate was set to 0.001, using MULTISTEP learning rate adjustment strategy. Over 200 rounds of training are taken, using an NVIDIA 3080Ti graphics card.
This paper uses the average pixel error of key points, the average angle error of landing gear, and the accuracy curve to measure the algorithm’s performance comprehensively. The landing gear to be measured is divided into the nose landing gear and the two types of the rear landing gear. The method proposed in this paper was tested on the synthetic aircraft dataset presented in Section 3.4, including parallelism and perspective. Of the dataset, 90% was randomly selected as the training set and 10% as the test set.

4.1. Angle Measurement Results

Figure 6 and Table 1 show that when the positioning error of key points is not considered, the combined method and cosine method have better outputs. Table 1 shows that the combined process reduces the average error by 5.1 degrees of the nose landing gears, 2.5 degrees of the rear landing gears, and 1.5 degrees of the mean angle of the two types.

4.2. Normalized Module

Table 2 shows that the average error of the mean angle has been reduced obviously by 8.6 degrees with the normalization module, and the mean error of fuselage points is reduced by around seven pixels in size. When the error threshold is 10, the accuracy has significantly improved approximately 40% and 20% of the two types of landing gears, respectively. When the angle error of prediction increases, it becomes more difficult to correctly determine the landing gear type since the type is judged depending on the distance between the key points.

4.3. K Loss Function

Table 3 shows that the vector field loss function is efficient when enhancing the pixel learning of the principal part based on distance weighting. The key points pixels error of the fuselage is reduced by 1.2 pixels size, and the angle accuracy of the nose and rear landing gear increases by 0.7% and 1.1%, respectively.

4.4. Experiment on Different Datasets

Figure 7 shows how the algorithm works. An RGB image is inputted to detect the key points of the aircraft after normalization. Then the angle between the landing gear and the fuselage plane of the aircraft is calculated and displayed using the relationship between the key points of the aircraft.
Considering a civil airplane landing with a fixed attitude range, the angle between the vertical tail and the fuselage is close to 90 degrees during landing. The landing gear is down when the corner is 90 degrees. We subdivided the dataset into three categories. We used the data whose tail and fuselage wing angles ranged from 80 to 100 degrees as the first dataset and the landing gear of 90 degrees as the second dataset. The third kind of dataset is selected from the approximate parallel perspective of the total dataset. The projection is approximately parallel when the aircraft is far away, and the parallel projection data were also used for testing.
Figure 8 and Table 4 show that the data within 10 degrees of tail and fuselage wing angles are much better than those obtained using a random camera view. When the landing gear is perpendicular to the fuselage plane, the mean angle error is fewer than 8 degrees. The angle accuracy of the rear landing gear is much better than that of the nose landing gear since there are two rear landing gears to obtain the average.
This experiment also shows that the proposed algorithm satisfies real-time operation requirements (>16 fps) under the conditions of the NVIDIA 3080Ti graphics card.

4.5. Robustness

Table 5 summarizes the results on the datasets with simulated illumination variation and motion blur of the aircraft. The CAD software created about 5000 images that simulated changing sunlight conditions with different light intensities and sun positions. Then all photos were randomly resized to 0.8–1.2 times as extensive supplements, and half of the images were arbitrarily selected for motion blur in the overall dataset. The table report in the overall dataset shows the negative impact on the performances of the model trained on the total dataset.
However, we improved model performances by training on datasets dissemination in this part. Data enhancements were used to improve the robustness, including random resizing, random rotation, and overlap. Table 5 shows that the model trained on the light-varying dataset has a better effect on the nose landing gear accuracy compared with the all-meet dataset. All images were resized to 256 × 256 pixels in the low-resolution dataset, and results show that the rear landing gear accuracy is slightly 3% lower. The accuracy ratio of the nose landing gear decreases significantly. The average angle error is within 5.6 degrees.

4.6. Discussion

Figure 9 is a partial result tested on the light-varying dataset and the motion-blurry dataset. Input images are available in 608 × 608 pixels and 256 × 256 pixels. The left side of the output pictures is the angles between the vertical tail and the aircraft structure, including the airframe and the wing in RGB images. The measured landing gear angles are written at the top of the pictures.
To avoid the difficulty of angle prediction under some unique perspectives, we filter the input data of the CAL module according to the angles between the tail and the aircraft structure. Figure 9 shows the vertical tail angles to judge the accuracy of the model. When the difference between the angle and 90 degrees is greater than 10 degrees, we do not make predictions to ensure the high accuracy of the model. When the vertical tail angles are close to 90°, the algorithm is more effective. When the angle is close to 0 or 180 degrees, it becomes more challenging to determine landing gear angles correctly since they were calculated depending on the distance between the key points. Because of the fixed landing attitude of civil aviation, the vertical tail angles in the camera view will not fluctuate too much. When the steep tail angle is within 10 degrees, we can make a more accurate prediction. When the angle exceeds the range, we can also judge that the accuracy has decreased in a problematic perspective.

5. Conclusions

This paper proposes a multi-step CNN-based estimation of the angles between the landing gear and the fuselage plane on the monocular camera. First, the aircraft is divided into two parts to predict key points using a single RGB image. Then, the data are filtered into the CAL module to calculate the aircraft landing gear angles. Experiments have shown that the angle accuracy is improved with the target normalization module, the loss function with coefficient K, and the combined angle measurement method based on pixel coordinates of key points. In addition, we contributed a synthetic aircraft dataset containing landing gears with different angles. And we increased the robustness of illumination, making the model closer to the actual environment. We also considered wavelet transform and other methods to improve the image resolution and trained the model with high- and low-resolution images to compare the effect. The result obtained for the light-varying dataset of the proposed algorithm demonstrates that the average angle errors of the nose and rear landing gears are 6.7 and 4.5 degrees, respectively. The average angle error of total landing gears is 4.6 degrees, and the accuracy within 10 degrees is over 80%.
In future research, we will incorporate the CAL module into the end-to-end CNN, classifying rough angles and regressing pleasing results. Second, we will optimize the angle prediction under challenging perspectives. In addition, we are going to increase the variety of aircraft.

Author Contributions

Conceptualization, F.L. and Z.W.; methodology, F.L.; software, F.L.; validation, Z.L., Z.W., B.Z. and C.M.; formal analysis, F.L. and C.M.; investigation, F.L.; resources, Z.L.; data curation, F.L., Z.W. and C.M.; writing—original draft preparation, Z.W.; writing—review and editing, F.L., B.Z. and Z.L.; visualization, F.L.; supervision, Z.W., J.L. and C.M.; project administration, Z.W. and J.L.; funding acquisition, Z.L. and Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Bethune Medical Engineering and Instrumentation Center, grant number Y92342T.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Acknowledgments

We appreciate the help of Department of Image Processing of Institute of Optics, Fine Mechanics and Physics (CIOMP), Chinese Academy of Sciences in providing GPUs for us.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, H.L. Video Monitoring Method Based on Multi-Camera Collaboration for Aircraft Landing Process. Master’s Thesis, Harbin Institute of Technology, Harbin, China, 2019. (In Chinese). [Google Scholar]
  2. Peng, S.; Liu, Y.; Huang, Q.; Zhou, X.; Bao, H. PVNet: Pixel-wise Voting Network for 6DoF Object Pose Estimation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4556–4565. [Google Scholar]
  3. Murthy, K.; Shearn, M.; Smiley, B.D.; Chau, A.H.; Levine, J.; Robinson, M.D. Sky Sat-1: Very High-Resolution Imagery from a Small Satellite. In Proceedings of the SPIE Remote Sensing, Amsterdam, The Netherlands, 22–25 September 2014. [Google Scholar]
  4. Khodambashi, S.; Moghaddam, M. An Impulse Noise Fading Technique Based On Local Histogram Processing. In Proceedings of the 2009 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Ajman, United Arab Emirates, 14–17 December 2009; pp. 95–100. [Google Scholar]
  5. Kim, Y.T. Contrast enhancement using brightness preserving bi-histogram equalization. IEEE Trans. Consum. Electron. 1997, 43, 1–8. [Google Scholar]
  6. Chen, S.D.; Ramli, R. Minimum mean brightness error bi-histogram equalization in contrast enhancement. IEEE Trans. Consum. Electron. 2003, 49, 1310–1319. [Google Scholar] [CrossRef]
  7. Qian, G. Detecting Transformer Winding Deformation Fault Types Based on Continuous Wavelet Transform. In Proceedings of the IEEE International Conference on Mechatronics and Automation, Harbin, China, 7–10 August 2016; pp. 1886–1891. [Google Scholar]
  8. Best, S. A discussion on the significance of geometry in determining the resonant behavior of fractal and other non-Euclidean wire antennas. IEEE Antennas Propag. Mag. 2003, 45, 9–28. [Google Scholar] [CrossRef]
  9. Jiao, T.; Wang, M.; Zhang, J.; Wang, Y. Wheel image enhancement based on wavelet analysis and pseudo-color processing. Autom. Instrum. 2020, 1, 47–51. [Google Scholar]
  10. Wang, Z.; Wang, F.; Chi, G. A research on defect image enhancement based on partial differential equation of quantum mechanics. Int. J. Comput. Sci. Math. 2018, 9, 122–132. [Google Scholar] [CrossRef]
  11. Li, J.; Liang, X.; Shen, S.M.; Xu, T.; Yan, S. Scale-aware Fast R-CNN for Pedestrian Detection. IEEE Trans. Multimed. 2018, 20, 985–996. [Google Scholar] [CrossRef] [Green Version]
  12. Razavian, A.S.; Azizpour, H.; Sullivan, J.; Carlsson, S. CNN Features off-the-shelf: An Astounding Baseline for Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR Workshops 2014, Columbus, OH, USA, 23–28 June 2014; pp. 512–519. [Google Scholar]
  13. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  15. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  16. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018. Available online: http://timmurphy.org/2009/07/22/line-spacing-in-latex-documents/ (accessed on 8 April 2018).
  17. Bochkovskiy, A.; Wang, C.Y.; Liao, H. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  18. Ali, W.; Abdelkarim, S.; Zahran, M.; Zidan, M.; Sallab, A.E. YOLO3D: End-to-end real-time 3D Oriented Object Bounding Box Detection from LiDAR Point Cloud. In Proceedings of the ECCV 2018 Workshops, Munich, Germany, 8–14 September 2019; pp. 716–728. [Google Scholar]
  19. Beltran, J.; Guindel, C.; Moreno, F.M.; Cruzado, D.; Garcia, F.; Arturo, D. BirdNet: A 3D Object Detection Framework from LiDAR information. In Proceedings of the 21st International Conference on Intelligent Transportation Systems, ITSC 2018, Maui, HI, USA, 4–7 November 2018; pp. 3517–3523. [Google Scholar]
  20. Zhou, Y.; Sun, P.; Zhang, Y.; Anguelov, D.; Gao, J.; Ouyang, T.; Guo, J.; Ngiam, J.; Vasudevan, V. End-to-End Multi-View Fusion for 3D Object Detection in LiDAR Point Clouds. arXiv 2019, arXiv:1910.06528. [Google Scholar]
  21. Wu, D.; Zhuang, Z.; Xiang, C.; Zou, W.; Li, X. 6D-VNet: End-to-end 6DoF Vehicle Pose Estimation from Monocular RGB Images. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; pp. 1238–1247. [Google Scholar]
  22. Kehl, W.; Manhardt, F.; Tombari, F.; Ilic, S.; Navab, N. SSD-6D: Making RGB-Based 3D Detection and 6D Pose Estimation Great Again. In Proceedings of the IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, 22–29 October 2017; pp. 1530–1538. [Google Scholar]
  23. Chabot, F.; Chaouch, M.; Rabarisoa, J.; Teulière, C.; Chateau, T. Deep MANTA: A Coarse-to-Fine Many-Task Network for Joint 2D and 3D Vehicle Analysis from Monocular Image. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 1827–1836. [Google Scholar]
  24. Tekin, B.; Sinha, S.N.; Fua, P. Real-Time Seamless Single Shot 6D Object Pose Prediction. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 292–301. [Google Scholar]
  25. Wang, H.; Sridhar, S.; Huang, J.; Valentin, J.; Song, S.; Guibas, L.J. Normalized Object Coordinate Space for Category-Level 6D Object Pose and Size Estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 2637–2646. [Google Scholar]
  26. Lee, T.; Lee, B.; Kim, M.; Kweon, I. Category-Level Metric Scale Object Shape and Pose Estimation. IEEE Robot. Autom. Lett. 2021, 6, 8575–8582. [Google Scholar] [CrossRef]
  27. Brachmann, E.; Michel, F.; Krull, A.; Yang, M.Y.; Gumhold, S.; Rother, C. Uncertainty-Driven 6D Pose Estimation of Objects and Scenes from a Single RGB Image. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 3364–3372. [Google Scholar]
  28. Desingh, K.; Lu, S.; Opipari, A.; Jenkins, O. Factored Pose Estimation of Articulated Objects using Efficient Nonparametric Belief Propagation. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 7221–7227. [Google Scholar]
  29. Mao, Y.; Yang, S.; Chao, D.; Pan, Z.; Yang, R. Real-Time Simultaneous Pose and Shape Estimation for Articulated Objects Using a Single Depth Camera. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 1517–1532. [Google Scholar]
  30. Martín-Martín, R.; Höfer, S.; Brock, O. An Integrated Approach to Visual Perception of Articulated Objects. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 5091–5097. [Google Scholar]
  31. Kumar, S.; Dhiman, V.; Ganesh, M.R.; Corso, J.J. Spatiotemporal Articulated Models for Dynamic SLAM. arXiv 2016, arXiv:1604.03526. [Google Scholar]
  32. Schmidt, T.; Hertkorn, K.; Newcombe, R.; Marton, Z.C.; Fox, D. Depth-based tracking with physical constraints for robot manipulation. In Proceedings of the IEEE International Conference on Robotics and Automation, ICRA 2015, Seattle, WA, USA, 26–30 May 2015; pp. 119–126. [Google Scholar]
  33. Schmidt, T.; Newcombe, R.; Fox, D. Self-Supervised Visual Descriptor Learning for Dense Correspondence. IEEE Robot. Autom. Lett. 2017, 2, 420–427. [Google Scholar] [CrossRef]
  34. Daniele, A.; Howard, T.; Walter, M.A.; Howard, T.; Walter, M. A Multiview Approach to Learning Articulated Motion Models. In Robotics Research, Proceedings of the 18th International Symposium, ISRR 2017, Puerto Varas, Chile, 11–14 December 2017; Springer: Cham, Switzerland, 2019; pp. 371–386. [Google Scholar]
  35. Ranjan, A.; Hoffmann, D.T.; Tzionas, D.; Tang, S.; Romero, J.; Black, M.J. Learning Multi-Human Optical Flow. Int. J. Comput. Vis. 2020, 128, 873–890. [Google Scholar] [CrossRef] [Green Version]
  36. Tkach, A.; Tagliasacchi, A.; Remelli, E.; Pauly, M.; Fitzgibbon, A. Online generative model personalization for hand tracking. ACM Trans. Graph. 2017, 36, 1–11. [Google Scholar] [CrossRef]
  37. Chen, X.; Wang, G.; Guo, H.; Zhang, C. Pose guided structured region ensemble network for cascaded hand pose estimation. Neurocomputing 2020, 395, 138–149. [Google Scholar] [CrossRef] [Green Version]
  38. Ge, L.; Ren, Z.; Yuan, J. Point-to-Point Regression PointNet for 3D Hand Pose Estimation. In Proceedings of the 15th European Conference, Munich, Germany, 8–14 September 2018; pp. 489–505. [Google Scholar]
  39. Chang, J.Y.; Moon, G.; Lee, K.M. V2V-PoseNet: Voxel-to-Voxel Prediction Network for Accurate 3D Hand and Human Pose Estimation from a Single Depth Map. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5079–5088. [Google Scholar]
  40. Wan, C.; Probst, T.; Gool, L.V.; Yao, A. Dense 3D Regression for Hand Pose Estimation. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5147–5156. [Google Scholar]
  41. Li, W.; Lei, H.; Zhang, J.; Wang, X. 3D hand pose estimation based on label distribution learning. J. Comput. Appl. 2021, 41, 550–555. [Google Scholar]
  42. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 1302–1310. [Google Scholar]
  43. Newell, A.; Huang, Z.; Deng, J. Associative Embedding: End-to-End Learning for Joint Detection and Grouping. In Proceedings of the Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  44. Kocabas, M.; Karagoz, S.; Akbas, E. MultiPoseNet: Fast Multi-Person Pose Estimation using Pose Residual Network. In Proceedings of the ECCV 2018, 15th European Conference, Munich, Germany, 8–14 September 2018; pp. 437–453. [Google Scholar]
  45. Chen, Y.; Wang, Z.; Peng, Y.; Zhang, Z.; Jian, S. Cascaded Pyramid Network for Multi-person Pose Estimation. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7103–7112. [Google Scholar]
  46. Xiao, B.; Wu, H.; Wei, Y. Simple Baselines for Human Pose Estimation and Tracking. In Proceedings of the ECCV 2018, 15th European Conference, Munich, Germany, 8–14 September 2018; pp. 472–487. [Google Scholar]
  47. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep High-Resolution Representation Learning for Human Pose Estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 3364–3372. [Google Scholar]
  48. Misra, D. Mish: A Self Regularized Non-Monotonic Neural Activation Function. arXiv 2019, arXiv:1908.08681. [Google Scholar]
  49. Kim, H.; Park, S.; Wang, J.; Kim, Y.; Jeong, J. Advanced Bilinear Image Interpolation Based On Edge Features. In Proceedings of the First International Conference on Advances in Multimedia, Colmar, France, 20–25 July 2009; pp. 33–36. [Google Scholar]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  51. Xiang, Y.; Schmidt, T.; Narayanan, V.; Fox, D. PoseCNN: A Convolutional Neural Network for 6D Object Pose Estimation in Cluttered Scenes. In Proceedings of the Robotics: Science and Systems XIV, Pittsburgh, PA, USA, 26–30 June 2018. [Google Scholar]
  52. Girshick, R. Fast R-CNN. arXiv 2015, arXiv:1504.08083. [Google Scholar]
  53. Xu, L.; Fu, Q.; Tao, W.; Zhao, H. Monocular vehicle pose estimation based on 3D model. Opt. Precis. Eng. 2021, 29, 1346–1355. [Google Scholar] [CrossRef]
  54. Best, S.R. Operating band comparison of the perturbed Sierpinski and modified Parany Gasket antennas. IEEE Antennas Wirel. Propag. Lett. 2002, 1, 35–38. [Google Scholar] [CrossRef]
  55. Guariglia, E. Harmonic Sierpinski Gasket and Applications. Entropy 2018, 20, 714. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Krzysztofik, W.J. Fractal Geometry in Electromagnetics Applications- from Antenna to Metamaterials. Microw. Rev. 2013, 19, 3–14. [Google Scholar]
  57. Hohlfeld, R.G.; Cohen, N. Self-similarity and the geometric requirements for frequency independence in antennae. Fractals 1999, 7, 79–84. [Google Scholar] [CrossRef]
  58. Tobin, J.; Fong, R.; Ray, A.; Schneider, J.; Abbeel, P. Domain randomization for transferring deep neural networks from simulation to the real world. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017. [Google Scholar]
  59. Shrivastava, A.; Pfister, T.; Tuzel, O.; Susskind, J.; Webb, R. Learning from Simulated and Unsupervised Images through Adversarial Training. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2242–2251. [Google Scholar]
  60. Bousmalis, K.; Silberman, N.; Dohan, D.; Erhan, D.; Krishnan, D. Unsupervised Pixel-Level Domain Adaptation with Generative Adversarial Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 95–104. [Google Scholar]
  61. Tremblay, J.; Prakash, A.; Acuna, D.; Brophy, M.; Jampani, V.; Anil, C.; To, T.; Cameracci, E.; Boochoon, S.; Birchfield, S. Training Deep Networks with Synthetic Data: Bridging the Reality Gap by Domain Randomization. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 23–30. [Google Scholar]
Figure 1. Principle of the angle estimation. Key points are abbreviated to kpts. We use YOLO-V4 to get two types of detection boxes, including the landing gear tires and the aircraft. Then the norm module gets the normalized central points of landing gear tires. At the same time, the norm module extracts the overall image inside the aircraft detection frame and normalizes the image to the PVNet network. In the PVNet network, vector fields are regressed to vote for the fuselage key points. Finally, CAL module is used to compute the landing gear angles according to key points and the aircraft model.
Figure 1. Principle of the angle estimation. Key points are abbreviated to kpts. We use YOLO-V4 to get two types of detection boxes, including the landing gear tires and the aircraft. Then the norm module gets the normalized central points of landing gear tires. At the same time, the norm module extracts the overall image inside the aircraft detection frame and normalizes the image to the PVNet network. In the PVNet network, vector fields are regressed to vote for the fuselage key points. Finally, CAL module is used to compute the landing gear angles according to key points and the aircraft model.
Sensors 21 08440 g001
Figure 2. (a) is the visual loss value, the higher the brightness, the greater the loss value. (b) is the weight value K based on the distance between the pixel and key point when λ = 5. The x-coordinate is the distance and the y-coordinate is the value of K.
Figure 2. (a) is the visual loss value, the higher the brightness, the greater the loss value. (b) is the weight value K based on the distance between the pixel and key point when λ = 5. The x-coordinate is the distance and the y-coordinate is the value of K.
Sensors 21 08440 g002
Figure 3. (a) is the mask image of the aircraft. The solid line is the key points of the fuselage, wing, and tail, the red dot is the key point of the landing gear, the pink dot is the connection point of the landing gear and the fuselage, the green dotted line is the straight line of the connection point parallel to the fuselage and wing, and the red dotted line is the straight line of the key point of the landing gear parallel to the vertical tail. (b) is the triangle formed by the landing gear key points, connecting points, and dotted line intersection.
Figure 3. (a) is the mask image of the aircraft. The solid line is the key points of the fuselage, wing, and tail, the red dot is the key point of the landing gear, the pink dot is the connection point of the landing gear and the fuselage, the green dotted line is the straight line of the connection point parallel to the fuselage and wing, and the red dotted line is the straight line of the key point of the landing gear parallel to the vertical tail. (b) is the triangle formed by the landing gear key points, connecting points, and dotted line intersection.
Sensors 21 08440 g003
Figure 4. (a) is the original image in the dataset. (b) is the mask image, and (c) is the mask image displaying labeled key points. The dataset contains about 40,000 images.
Figure 4. (a) is the original image in the dataset. (b) is the mask image, and (c) is the mask image displaying labeled key points. The dataset contains about 40,000 images.
Sensors 21 08440 g004
Figure 5. The figures above show the input data form of the vector field network model made by the dataset. In the top line, the first picture from left to right is the original image. The second picture in the top line is the mask with the labeled information of the displayed key points. The rest of the pictures are pixel vector field labels belonging to corresponding key points.
Figure 5. The figures above show the input data form of the vector field network model made by the dataset. In the top line, the first picture from left to right is the original image. The second picture in the top line is the mask with the labeled information of the displayed key points. The rest of the pictures are pixel vector field labels belonging to corresponding key points.
Sensors 21 08440 g005
Figure 6. The input is the real pixel coordinates of key points, and the output is the curve of error threshold and the angle accuracy. (a) is the accuracy of the different methods about the rear landing gear, and (b) is the accuracy of the different methods about the nose landing gear.
Figure 6. The input is the real pixel coordinates of key points, and the output is the curve of error threshold and the angle accuracy. (a) is the accuracy of the different methods about the rear landing gear, and (b) is the accuracy of the different methods about the nose landing gear.
Sensors 21 08440 g006
Figure 7. The proposed method works on datasets; predicts key points in two parts, calculates angle with no depth information.
Figure 7. The proposed method works on datasets; predicts key points in two parts, calculates angle with no depth information.
Sensors 21 08440 g007
Figure 8. The accuracy of the total datasets and the down-undercart datasets.
Figure 8. The accuracy of the total datasets and the down-undercart datasets.
Sensors 21 08440 g008
Figure 9. Parts of output images. The predicted angles are magenta font at the top of the graph. The number after the letter forward is for the nose landing gear angle. The number after the letter left is for the left rear landing gear angle, and the number after the letter right is for the right rear landing gear angle when facing the front. The number after the letter airframe represents the angle between the vertical tail and the airframe. The number after the letter wing indicates the angle between the vertical tail and the wing.
Figure 9. Parts of output images. The predicted angles are magenta font at the top of the graph. The number after the letter forward is for the nose landing gear angle. The number after the letter left is for the left rear landing gear angle, and the number after the letter right is for the right rear landing gear angle when facing the front. The number after the letter airframe represents the angle between the vertical tail and the airframe. The number after the letter wing indicates the angle between the vertical tail and the wing.
Sensors 21 08440 g009
Table 1. The mean error of angles and accuracy rate with different measurement methods when threshold equals 10.
Table 1. The mean error of angles and accuracy rate with different measurement methods when threshold equals 10.
MethodThe Nose ErrorThe Rear ErrorMean Angle ErrorThe Nose AccuracyThe Rear Accuracy
Sine18.712.412.233.8%52.5%
Cosine15.810.211.871.2%77.5%
Sine and cosine16.611.211.645.8%62.0%
Combined method11.58.79.168.9%77.8%
Table 2. The mean error and accuracy rate table with or without normalization. F means false, while T indicates true.
Table 2. The mean error and accuracy rate table with or without normalization. F means false, while T indicates true.
NormalizeThe Nose AngleThe Rear AngleMean Angle ErrorKpt FuselageKpt Landing GearThe Nose AccuracyThe Rear Accuracy
F30.616.617.710.93.827.9%58.1%
T11.58.79.13.63.768.7%78.0%
Table 3. This is a mean error and accuracy rate table with different loss functions. F means false, while T means true.
Table 3. This is a mean error and accuracy rate table with different loss functions. F means false, while T means true.
K LossThe Nose AngleThe Rear AngleMean Angle ErrorKpt FuselageKpt Landing GearThe Nose AccuracyThe Rear Accuracy
F11.88.18.94.53.866.7%79.9%
T11.97.98.83.33.767.4%81.0%
Table 4. The mean error and accuracy rate for different datasets.
Table 4. The mean error and accuracy rate for different datasets.
DatasetsThe Nose AngleThe Rear AngleMean Angle ErrorKpt FuselageKpt Landing GearThe Nose AccuracyThe Rear Accuracy
Total18.3811.754.649.0%82.2%
Parallel11.58.79.13.63.768.9%77.8%
Undercart down12.08.510.22.83.070.4%77.1%
Within 10 degrees11.55.19.654.666.0%92.1%
All meet8.34.87.233.181.9%91.9%
Table 5. The mean error and accuracy rate on different datasets.
Table 5. The mean error and accuracy rate on different datasets.
DatasetsThe Nose AngleThe Rear AngleMean Angle ErrorKpt FuselageKpt Landing GearThe Nose AccuracyThe Rear Accuracy
Overall64.059.960.854.813.59.4%5.3%
Varying light6.74.54.63.73.384.8%91.4%
Low resolution9.15.25.61.91.464.9%88.6%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, F.; Wu, Z.; Li, J.; Lai, Z.; Zhao, B.; Min, C. A Multi-Step CNN-Based Estimation of Aircraft Landing Gear Angles. Sensors 2021, 21, 8440. https://doi.org/10.3390/s21248440

AMA Style

Li F, Wu Z, Li J, Lai Z, Zhao B, Min C. A Multi-Step CNN-Based Estimation of Aircraft Landing Gear Angles. Sensors. 2021; 21(24):8440. https://doi.org/10.3390/s21248440

Chicago/Turabian Style

Li, Fuyang, Zhiguo Wu, Jingyu Li, Zhitong Lai, Botong Zhao, and Chen Min. 2021. "A Multi-Step CNN-Based Estimation of Aircraft Landing Gear Angles" Sensors 21, no. 24: 8440. https://doi.org/10.3390/s21248440

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop