Next Article in Journal
Face Detection Ensemble with Methods Using Depth Information to Filter False Positives
Previous Article in Journal
Engineering Design of an Active–Passive Combined Thermal Control Technology for an Aerial Optoelectronic Platform
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-Time Docking Ring Detection Based on the Geometrical Shape for an On-Orbit Spacecraft

School of Electrical and Electronic Engineering, Shanghai Institute of Technology, Shanghai 201418, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(23), 5243; https://doi.org/10.3390/s19235243
Submission received: 28 October 2019 / Revised: 19 November 2019 / Accepted: 25 November 2019 / Published: 28 November 2019
(This article belongs to the Section Optical Sensors)

Abstract

:
Docking ring is a circular hatch of spacecraft that allows servicing spacecraft to dock in various space missions. The detection of the ring is greatly beneficial to automatic capture, rendezvous and docking. Based on its geometrical shape, we propose a real-time docking ring detection method for on-orbit spacecraft. Firstly, we extract arcs from the edge mask and classify them into four classes according to edge direction and convexity. By developing the arc selection strategy, we select a combination of arcs possibly belonging to the same ellipse, and then estimate its parameters via the least squares fitting technique. Candidate ellipses are validated according to the fitness of the estimation with the actual edge pixels. The experiments show that our method is superior to the state-of-the-art methods, and can be used in real time application. The method can also be extended to other applications.

1. Introduction

With the development of space exploration, the on-orbit service [1] of automatic orbital capture and repair mechanism for faulty spacecraft [2,3,4,5] has become an urgent need. In order to achieve the automatic on-orbit service, vision system is widely used to acquire the pose of the target spacecraft. A docking ring [6] is a circular hatch of spacecraft that allows servicing spacecraft to dock in various space missions. The accurate detection of docking ring can be greatly beneficial to the target’s visual tracking, recognition [7,8] and localization [9,10,11,12,13]. The geometrical shape of the docking ring is distinctive, since it is a regular circular part on the man-made spacecraft. Besides, a circle in the scene is projected into an ellipse in the image, the docking ring can be detected by finding the ellipse in the image. The ellipse parameters are usually used as input in the pose estimation procedure, so our goal is to get the ellipse parameters in the image. It is different from conventional object detection tasks [14] that try to locate the image coordinates of the object, and draw a bounding box around the object in the image. In the past few years, researchers have utilized edge contour methods to detect ellipses. Many approaches to generate the arcs are presented, such as linking short straight lines [15,16,17,18], splitting the edge contour [19,20,21,22], or connecting the edge pixels [23,24]. Instead of exhaustive search, arcs are grouped according to their relative position and constraints on the curvature [15,16,19,20,22,23,24], or ellipse fitting error [17,18,21]. In this work, the inner boundary of docking ring is the single target ellipse to be found. This is in contrast to ellipse detection approaches finding all ellipses in the image. In practice, we must deal with several major problems. Suffering from partial occlusion, poor illumination or complex background, the on-orbit spacecraft image may be of low quality, which causes the detection more challenging. Furthermore, docking ring detection algorithm need reach high accuracy and efficiency because it is mostly just a previous step for localization, which is based on the feature correspondence between the 3D model and the 2D image.
We introduce a detection method of docking ring based on the gradient direction of edge points. Arcs are extracted from the edge mask and classified into four quadrants with edge direction and convexity. By putting forwards arc selection strategy, we select a cluster of arcs possibly belonging to the same ellipse, and then estimate its parameters via the least squares fitting method. Candidate ellipses are validated according to the fitness between the estimations and actual edge pixels. The contributions of our work are summarized in the following:
(1)
Our method is a general way to detect the docking ring from on-orbit spacecraft images. Based on the geometric properties and reflection characteristics of the target, the method adapts to various types of spacecraft in the complex and changeable space environment.
(2)
We develop novel arc selection strategies according to the geometric properties of the ellipse, and achieve better performance than the state-of-art approaches.

2. The Proposed Method

We proposed an accurate and fast detection method for docking ring, which consists of three steps: arc extraction, ellipse parameters estimation, and validity verification. Firstly, the candidate arcs are extracted from the edge image by linking the edge points. Secondly, we designed arc selection strategies to group the arcs that possibly belong to the same ellipse, and then the least square method [25,26] was used to estimate elliptic parameters efficiently. Finally, the validity of the ellipses was verified by measuring the fitness between the arc and the candidate ellipse.

2.1. Arc Extraction

The docking ring has different reflection characteristics from the surrounding background, resulting image brightness changes sharply. Therefore, we extracted the edge pixels to determine the boundary of the docking ring. According to the basic geometric properties of the ellipse in mathematics, the gradient direction of the ellipse edge point changes continuously. Based on this feature, we proposed a method for extracting elliptic arcs from a spatially connected point sequence. The edge points with stable gradient direction changing in the edge image were extracted and connected into arc segments. The contour of the ellipse remained in these arc segments. This step was to identify candidate arcs in the image, whose key phases were elaborated as follows.
(1) Edge detection:
Our method was mainly based on the edge information of the image. We applied the Canny edge detector [27] to the input image to get the edge set. Here the edge point e i   =   ( x i , y i , θ i ) is defined by its image position ( x i , y i ) and gradient direction θ i .
(2) Arc segment detection:
The edge points classification is based on the grayscale difference between the target and the surrounding background. Considering the reflectance properties of the docking ring, the gray value of the region outside the contour is higher than that inside the contour, as shown in Figure 1a. According to the gradient direction of edge points, the classification of the edge point e i is expressed as follows:
D ( e i )   =   { I ,               i f   d x > 0 d y > 0 II ,           i f   d x < 0 d y > 0 III ,       i f   d x < 0 d y < 0 IV ,       i f   d x > 0 d y < 0
where the symbol ∧ represents the logical relation “and”, and d x , d y are the derivative of Sobel in the x and y direction respectively. According to the signs of d x and d y , we divided the edge points into four categories, namely the four quadrants shown in Figure 1b.
For edge points in the same quadrant, the 8-connected edge pixels are connected into arc segments. It means that each arc α i is a set of 8-connected edge points, and the gradient direction of its edge points is within a fixed range. Then, short arcs whose pixel number is less than the threshold T p were eliminated. We could acquire arc segments set { α i } of four quadrants.
(3) Arc segment classification:
Since the elliptic arcs located in different quadrants have different convexity, arc segments were further selected according to their convexity. The convexity was judged by the position of the point on the arc relative to the line by connecting the two endpoints, as shown in Figure 2. P represents one point on the arc segment α i , the point P L and P R are the endpoints of α i , and l i is the line by connecting the endpoints.
The convexity of arc α i is represented by (2), in which “+” represents that the arc is convex, and “−” represents a concave arc. Elliptic arcs lying in the first and second quadrants are convex arcs, and the arcs lying in the third and fourth quadrants are concave arcs.
C ( α i )   =   { + ,     ( P L P R   ×   P L P ) · ( 0 , 0 , 1 ) > 0 ,     ( P L P R   ×   P L P ) · ( 0 , 0 , 1 ) < 0
By considering both the edge gradient direction and the convexity, the classification of arc segments is represented in (3).
Q ( α i )   =   { I ,       if   D ( e i ) = I C ( α i ) = + II ,     if   D ( e i ) = II C ( α i ) = +   III ,   if   D ( e i ) = III C ( α i ) = IV ,   if   D ( e i ) = IV C ( α i ) =
This step breaks the spatially connectivity of the contour with gradient direction changing sharply, and removes some arcs that do not belong to the ellipse.

2.2. Ellipse Parameters Estimation

This step is to identify candidate ellipses in the image, including arc grouping and ellipse fitting.

2.2.1. Arc Selection Strategy

By classifying the arc segments, the contour of a complete ellipse will be divided into four parts and located in four quadrants. We recombined the arcs from the same ellipse so as to estimate the ellipse parameters. Suffering from some challenges, including image noise, uneven illumination, and partial occlusion, arcs in real images may be fractional absence or divided into several short arc segments. In order for accurate elliptic parameter estimation, sufficient contour information is needed. Taking the above factors into consideration, arc segments of three different quadrants are selected to compose a candidate ellipse ε i . The rules we made were robust to the possible arc fracture of the real image, and enough information was provided for the subsequent parameter estimation step. In order to effectively group three arcs that belong to the same ellipse, we developed arc selection strategies according to the geometric properties of an ellipse.
1. Arc selection for two arcs:
Firstly, the constraints for the combination of two arcs were established. Two candidate arcs α i and α j were considered belonging to the same ellipse if they satisfied the following three constraints:
(1)
Quadrant constraint. We selected three arcs of the four quadrants to estimate an ellipse in this paper. The possible quadrants of the three arcs were indicated as follows: (I, II, III), (II, III, IV), (III, IV, I), and (I, II, IV). It shows that the three arcs were situated in three adjacent quadrants. Therefore, the combination of two arcs only combined the arcs in adjacent quadrants, as described by (4).
(2)
Position constraint. Based on the quadrant constraint, if arc α i and arc α j belong to a same ellipse, the endpoints of the arcs are constrained by their relative position in the image, as shown in Figure 3a. Position constraint is described by (5).
(3)
Tangent constraint. Assume we drew a tangent line, the ellipse was completely on one side of the tangent line. We considered the endpoint of the arc as tangent point to draw the tangent, and the tangent direction was perpendicular to the gradient direction of the endpoint, as shown in Figure 3b. Then, two arcs for combination must satisfy the constraints in (6).
C q   =   { T ,   if   ( Q ( α i ) , Q ( α j ) ) { ( I , II ) , ( II , III ) , ( III , IV ) , ( IV , I ) } F ,   otherwise   ,
C p   =   { T ,       i f       ( Q ( α i ) , Q ( α j ) ) ( I , I I ) ( P i L . x > P j R . x )           T ,       i f       ( Q ( α i ) , Q ( α j ) ) ( I I , I I I ) ( P i L . y > P j L . y )     T ,       i f       ( Q ( α i ) , Q ( α j ) ) ( I I I , I V ) ( P i R . x < P j L . x ) T ,       i f       ( Q ( α i ) , Q ( α j ) ) ( I V , I ) ( P i R . y < P j R . y ) F ,       o t h e r w i s e ,
C t   =   { T ,       i f   ( Q ( α i ) , Q ( α j ) ) ( I , I I ) ( P j R . y < f i L ( P j R . x ) ) ( P i L . y < f j R ( P i L . x ) )         T ,       i f   ( Q ( α i ) , Q ( α j ) ) ( I I , I I I ) ( P j L . x > g i L ( P j L . y ) ) ( P i L . x > g j L ( P i L . y ) )     T ,       i f   ( Q ( α i ) , Q ( α j ) ) ( I I I , I V ) ( P j L . y > f i R ( P j L . x ) ) ( P i R . y > f j L ( P i R . x ) ) T ,       i f   ( Q ( α i ) , Q ( α j ) ) ( I V , I ) ( P j R . x < g i R ( P j R . y ) ) ( P i R . x < g j R ( P i R . y ) ) F ,       o t h e r w i s e ,
where y   =   f k L ( x ) and y   =   f k R ( x ) represent the tangent equations of the left endpoint P k L and the right endpoint P k R of the arc α k respectively, the inverse equations are expressed as x   =   g k L ( y ) and x   =   g k R ( y ) .
For arcs α i and α j that satisfy the above constraints, the center of the ellipse was calculated, as shown in Figure 4. We connected the endpoint P i L of arc α i and the middle point P j M of α j into a line. Then we created multiple parallel lines, and fitted a line l i m to the middle points of the parallel lines. Similarly, l j m was another fitted line. The intersection P c of two straight lines l i m and l j m was the estimation of the ellipse center.
2. Arc selection for three arcs:
Next, the rule of combining three arcs was made. Three candidate arcs would be grouped, if the following constraints were satisfied:
(1)
Any two of three arcs in the adjacent quadrant met the arc selection constraints for two arcs. There were two groups of arc pair in the adjacent quadrant, each group of arcs needed to meet the arc selection for two arcs.
(2)
Constraints on the center of the ellipse. Three arcs originated from the same ellipse had the same ellipse center. To avoid the effect of the image noise, the centers of the ellipses calculated by two groups of arcs were required to be located within a preset distance.

2.2.2. Parameter Estimation for Candidate Ellipse

When dealing with 5D ellipse parameters estimation, the voting mechanism of Hough transform is time consuming. So, we adopted the least square method to get the conic expression of the ellipse, and then acquired the 5D parameters of the ellipse through the conic expression. An ellipse is a special form of a conic:
A x 2 + B x 2 + C x y + D x + E y + F   =   0
By using the least square method, all points on the arcs could be fitted to the quadratic curve of (7), and then the coefficients (A, B, C, D, E, F) could be obtained.
The geometric parameters of the ellipse include the coordinates of the center ( x 0 , y 0 ) , the angle of ellipse rotation θ , the semi-major axis a and the semi-minor axis b , and then the standard equation of the ellipse can be expressed as
( ( x x 0 ) cos θ + ( y y 0 ) sin θ a ) 2 + ( ( x x 0 ) sin θ + ( y y 0 ) cos θ b ) 2   =   1
The parameters of the ellipse can be calculated by equating the corresponding coefficients in (7) and (8).
{   x 0   =   C E 2 B D 4 A B C 2                           y 0   =   C D 2 A E 4 A B C 2                           θ   =   1 2 a r c t a n C A B         a   =   2 | A δ + B δ C δ sin 2 θ | b   =   2 | A δ + B δ + C δ sin 2 θ |
where
δ   =   1 A x 0 2 + B x 0 y 0 + C y 0 2 F

2.3. Validity Verification

The ellipse set was further verified to reduce the false detection rate. In this paper, the validity of candidate ellipse was evaluated according to the fitness between the estimated ellipse and the points on the arcs.
(1) Distance definition:
In this paper, the distance was calculated between pixel P on the arc and fitting ellipse E , as shown in Figure 5. We connected the edge point P and the ellipse center O , then the line O P and the ellipse E were intersected at the point M . P M is defined as the distance from the edge point to the ellipse, see Figure 5b. The distance is described in (10).
d ( P )   =   d O P d O M
where,
d O M   =   a 2 cos 2 φ + b 2 sin 2 φ d O P   =   ( x x 0 ) 2 + ( y y 0 ) 2 φ   =   a r c t a n ( a b tan φ 0 ) φ 0   =   a r c t a n ( y y 0 x x 0 ) θ
(2) Evaluation of candidate ellipses:
The pixels on the arc of the ellipse should be located at the boundary of the estimated ellipse. Therefore, the distance between the edge point on the arc and the fitting ellipse was used to evaluate whether the candidate target meets the ellipse shape.
d ( P i )   =   { = 0 i f   P i   i s   o n   t h e   b o u n d a r y   o f   E i < 0 i f   P i   i s   o n   t h e   o u t s i d e   o f   E i         > 0 i f   P i   i s   o n   t h e   i n s i d e   o f   E i              
We could define the set N   =   { P i : | d ( P i ) | < T d } that contains the edge points that are close to the elliptic boundary. The score s [ 0 ,   1 ] describes how well the points of the three arcs composing E i fit the boundary of the estimated ellipse:
s   =   N N i + N j + N k
where N i , N j , and N k represent the number of pixels on the three arcs respectively. The candidate ellipse E i satisfying s > T s is a valid ellipse, and the ellipse E i is added to the ellipse set { E i } ; Otherwise, E i is discarded which is considered as a false detection.
Finally, the uniqueness detection was carried out in the ellipse set { E i } . If there were two or more ellipses with similar parameters, only the elliptic parameters with the highest score were valid.

3. Experiment

In order to verify the performance of the proposed algorithm, we conducted experiments both on real images and video sequences. All the experiments are executed via the C++ language on the PC equipped with an Intel(R) Core(TM) i7-7700HQ [email protected] GHz processor, 8.00 GB of RAM and Windows 10 as the operating system.

3.1. Test on Real Images

The test on real images was to evaluate the detection effect and processing time of the proposed algorithm for the docking ring. Meanwhile, it aimed at assessing its adaptability to images of various types of spacecraft captured in the complex and changeable space environment. We compared our method with Libuda’s method [16] and Fornaciari’s method [23] on real images. The codes of the compared methods are provided by their authors. The image set contained 31 images ranging in size from 275 × 183 to 1024 × 852. The real images used in this section, including various types of spacecraft in different poses, are available on the Internet.
The detection results of Libuda’s algorithm, Fornaciari’s algorithm and our algorithm are illustrated in Figure 6. The targets detected by each algorithm are marked with green ellipses. Our method can acquire a unique ellipse, which is the inner boundary of the docking ring, yet Libuda’s and Fornaciari’s algorithms have many cases of false positives and false negatives. Taking full advantage of the characteristics of the docking ring, our algorithm gets better performance than others.
The running time of Libuda’s algorithm, Fornaciari’s algorithm and ours in real images is shown in Figure 7. The average running time of our method was 27.38 ms and the maximum running time was 132.80 ms. Compared with the other two algorithms, our method took the least computational time, which could meet the demand of real-time application.
Libuda et al. presents a fast data driven four stage filtering process, which uses geometric features in each stage to synthesize ellipses from binary image data with the help of lines, arcs, and extended arcs. However, it suffers from a long computation time. The ellipse detector proposed by Fornaciari et al. assigns a bounding box for each arc, removes the straight edges and determines the convexity of the arc by comparing the areas of region under and over the arc. However, the method raises the detection speed at the cost of detection accuracy and robustness. The good performance of our method could be attributed to the reasonable arc selection strategy, parameter estimation, and ellipse verification. Arc selection strategy could accelerate the process of arc grouping significantly. The accuracy and efficiency of parameter estimation were guaranteed by the least square fitting technique. A reasonable ellipse validation strategy could ensure that the false detection results were filtered out. However, since our algorithm required three arcs in different quadrants in the parameter estimation of candidate ellipses, it was difficult to detect the ellipses with heavy occlusion or small ellipses that had few pixels on the boundary.
The above experiments proved the superiority of our algorithm over other algorithms. Then we conducted experiments to verify the performance of our algorithm against illumination, complex background, and occlusion. Figure 8 shows our detection results in the challenge images. Figure 8a–c illustrates the adaptability of our method to various illumination conditions. Figure 8d–i show the robustness of our algorithm against a cluttered background and image blur. Some failures of the method are indicated in Figure 8j–l. In Figure 8k, it is difficult to extract the edge points of the ellipse because of the low contrast. In Figure 8j,l, the reason of the failure lies in the heavy occlusion and image blur.

3.2. Video Sequence Test

In this section, we tested the performance of the proposed algorithm on two video sequences. The video sequences were derived from the Orbital Express project and the docking mission of the Space Shuttle Atlantis to the International Space Station. The videos are available at https://www.youtube.com/watch?v=YUnQCs77PoY and https://www.youtube.com/watch?v=7M1PeXWTSSQ.
The experiments on video sequences were used to test the performance of the proposed algorithm in practical space application and the effectiveness of dealing with real docking mission. In addition, the running time of each step of the algorithm was analyzed in detail.
(1) Orbital Express project autonomously docking video sequence.
The autonomously docking video sequence in the Orbital Express project [28] was used to test the effectiveness of the algorithm. The video sequence consisted of 144 images with a resolution of 480 × 360. The images were captured in the open space with no earth or other stars in the background, as shown in Figure 9a. Canny edge detection results are shown in Figure 9b. According to the gradient direction, edge points were divided into four quadrants as shown in Figure 9c–f. The similar pixel points connected in the eight adjacent areas were connected into an arc, and the arcs with a length less than the threshold are removed. The arcs shown in Figure 9g were obtained by judging the convexity. The four quadrants of arcs are represented in white, yellow, red, and green respectively. As can be seen from Figure 9g, after the steps, there were six arcs obtained for the complex image, including one arc in quadrant I, one arc in quadrant II, two arcs in quadrant III, and two arcs in quadrant IV. Finally, the detection result was obtained through ellipse fitting and validation, and the detection result in green is shown in Figure 9h.
Due to the low contrast, the edge extraction was incomplete, and the docking rings in some images were not detected. The right detection rate of our algorithm for this video sequence was 98.61%, and the average detection time was 6.96 milliseconds. Figure 10 illustrates some examples of results.
The size of each image in video sequence was consistent and the content of the image was similar, so the factors affecting the speed of the algorithm could be analyzed through the average running time of each step in the algorithm. Table 1 shows the average process time of each step of our algorithm. It could be seen that edge detection step took the most time, followed by arc detection and classification. After screening, the number of remaining arcs was small and the processing speed of subsequent steps was high.
(2) Video sequence of the Space Shuttle Atlantis docked with the International Space Station.
The video sequence of the Space Shuttle Atlantis docked to the International Space Station (ISS) [29] was selected to further test the effectiveness of our algorithm in a space application. The video sequence included 5692 frames of images with a resolution of 852 × 480. As shown in Figure 11a, the Space Shuttle itself occupied most of the background of the image. Figure 11 shows the intermediate results, among which Figure 11b was the edge detection result, Figure 11c–f were four quadrants of edge points divided according to the direction of gradient, Figure 11g was the extraction result of four quadrants of arc segments, and Figure 11h shows the green ellipse as the final detection result.
Our algorithm achieved a 100% right detection rate for the video sequence of the Space Shuttle Atlantis docked with the International Space Station. The average detection time was 41.79 ms. Figure 12 illustrates some examples of the detection results. The green ellipse represents the detection results of our algorithm.
Table 2 shows the average processing time of each step of our algorithm. Compared with the video sequence of the Orbital Express project autonomously docking, the image resolution of this sequence was larger and the background was more complex. There were more edge points to be dealt with and the process time was longer. The running time was related to the size of the image and the complexity of image content. These two factors mainly affected the efficiency of edge detection and arc extraction. In addition, the speed of arc classification mainly depended on the threshold of arc length and the number of arcs that changing continuously in the direction of gradient. After extracting the arcs, the arc selection strategy effectively reduced the number of arc combinations and improved the accuracy, thus reducing the subsequent time of ellipse fitting and verification.
To sum up, the two video sequences verified the accuracy and efficiency of our detection algorithm, and also proved that it could be used for actual space applications.

4. Conclusions

In this paper, we proposed a docking ring detection algorithm based on geometrical shape from on-orbit spacecraft images. By taking advantage of the continuous change of gradient direction of the target elliptic contour, this algorithm selects pixels with relatively stable gradient direction change from the edge image and connects them into an arc to avoid the edge points with abrupt gradient direction change in the arc. On this basis, a variety of arc selection strategies are adopted to effectively reduce the number of ellipse fitting in the arc combination process, thus improving the detection efficiency. Finally, the target ellipse obtained after arc grouping is further verified to reduce the false detection rate. The method is general for spacecraft with docking rings. Experimental verification proved that our algorithm had better performance compared to other algorithms and it could be extended to other applications. Future work should investigate detecting ellipse with heavy occlusion and small ellipse.

Author Contributions

Conceptualization, L.Z. and W.P.; Formal analysis, L.Z. and W.P.; Funding acquisition, X.M.; Investigation, L.Z. and W.P.; Methodology, L.Z. and W.P.; Software, L.Z. and W.P.; Visualization, L.Z. and W.P.; Writing—original draft, L.Z. and W.P.; Writing—review & editing, L.Z. and W.P.

Funding

This research was funded in part by the Shanghai Science and Technology Commission Research Project (Grant No. 18511103900 and No. 19511103402), and National Science and Technology Major Project of China (Grant No. 2019ZX04026001).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Flores-Abad, A.; Ma, O.; Pham, K.; Ulrich, S. A review of space robotics technologies for on-orbit servicing. Prog. Aerosp. Sci. 2014, 68, 1–26. [Google Scholar] [CrossRef]
  2. Huang, P.; Lu, Y.; Wang, M.; Meng, Z.; Zhang, Y.; Zhang, F. Postcapture Attitude Takeover Control of a Partially Failed Spacecraft with Parametric Uncertainties. IEEE Trans. Autom. Sci. Eng. 2019, 16, 919–930. [Google Scholar] [CrossRef]
  3. Li, X.; Tai, Y.; Zhang, L.; Li, H.; Li, L. Characterization of dynamic random process using optical vortex metrology. Appl. Phys. B 2014, 116, 901–909. [Google Scholar] [CrossRef]
  4. Li, X. Digital speckle correlation method based on phase vortices. Opt. Eng. 2012, 51, 077004. [Google Scholar] [CrossRef]
  5. Sun, Q.; Niu, Z.; Wang, W.; Li, H.; Luo, L.; Lin, X. An Adaptive Real-Time Detection Algorithm for Dim and Small Photoelectric GSO Debris. Sensors 2019, 19, 4026. [Google Scholar] [CrossRef]
  6. Finkbeiner, J.R.; Dunlap, P.H.; Steinetz, B.M.; Daniels, C.C. Review of seal designs on the Apollo spacecraft. J. Spacecr. Rocket. 2008, 45, 900–910. [Google Scholar] [CrossRef]
  7. Velasquez, A.F.; Luckett, J.; Napolitano, M.; Marani, G.; Evans, T.; Fravolini, M. Experimental evaluation of a machine vision based pose estimation system for autonomous capture of satellites with interface rings. In Proceedings of the AIAA Guidance, Navigation, and Control (GNC) Conference, Boston, MA, USA, 19–22 August 2013; p. 4758. [Google Scholar]
  8. Teutsch, C.; Berndt, D.; Trostmann, E.; Weber, M. Real-time detection of elliptic shapes for automated object recognition and object tracking. In Machine Vision Applications in Industrial Inspection XIV; International Society for Optics and Photonics: Bellingham, WA, USA, 2006; Volume 6070, p. 60700J. [Google Scholar]
  9. Ye, C.; Hong, S.; Tamjidi, A. 6-DOF Pose Estimation of a Robotic Navigation Aid by Tracking Visual and Geometric Features. IEEE Trans. Autom. Sci. Eng. 2015, 12, 1169–1180. [Google Scholar] [CrossRef]
  10. Kim, J.-U.; Kang, H.-B. A New 3D Object Pose Detection Method Using LIDAR Shape Set. Sensors 2018, 18, 882. [Google Scholar] [CrossRef]
  11. Kwon, Y.C.; Jang, J.W.; Hwang, Y.; Choi, O. Multi-Cue-Based Circle Detection and Its Application to Robust Extrinsic Calibration of RGB-D Cameras. Sensors 2019, 19, 1539. [Google Scholar] [CrossRef]
  12. Chen, Z.; Zhang, Z.; Dai, F.; Bu, Y.; Wang, H. Monocular Vision-Based Underwater Object Detection. Sensors 2017, 17, 1784. [Google Scholar] [CrossRef]
  13. Zhang, J.; Qiu, Y.; Duan, X.; Xu, K.; Yang, C. An Improved Robust Method for Pose Estimation of Cylindrical Parts with Interference Features. Sensors 2019, 19, 2234. [Google Scholar] [CrossRef] [PubMed]
  14. Lu, K.; Li, J.; Zhou, L.; Hu, X.; An, X.; He, H. Generalized Haar Filter-Based Object Detection for Car Sharing Services. IEEE Trans. Autom. Sci. Eng. 2018, 15, 1448–1458. [Google Scholar] [CrossRef]
  15. Kim, E.; Haseyama, M.; Kitajima, H. Fast and robust ellipse extraction from complicated images. In Proceedings of the IEEE Information Technology and Applications, Bathurst, Australia, 25–28 November 2002. [Google Scholar]
  16. Libuda, L.; Grothues, I.; Kraiss, K.F. Ellipse detection in digital image data using geometric features. In Advances in Computer Graphics and Computer Vision; Springer: Berlin/Heidelberg, Germany, 2007; pp. 229–239. [Google Scholar]
  17. Mai, F.; Hung, Y.S.; Zhong, H.; Sze, W.F. A hierarchical approach for fast and robust ellipse extraction. Pattern Recognit. 2008, 41, 2512–2524. [Google Scholar] [CrossRef]
  18. Chia, A.Y.S.; Rahardja, S.; Rajan, D.; Leung, M.K. A split and merge based ellipse detector with self-correcting capability. IEEE Trans. Image Process. 2011, 20, 1991–2006. [Google Scholar] [CrossRef]
  19. Prasad, D.K.; Leung, M.K.H.; Cho, S.Y. Edge curvature and convexity based ellipse detection method. Pattern Recognit. 2012, 45, 3204–3221. [Google Scholar] [CrossRef]
  20. Nguyen, T.M.; Ahuja, S.; Wu, Q.M.J. A real-time ellipse detection based on edge grouping. In Proceedings of the 2009 IEEE International Conference on Systems, Man and Cybernetics, San Antonio, TX, USA, 11–14 October 2009; pp. 3280–3286. [Google Scholar]
  21. Liu, Z.Y.; Qiao, H. Multiple ellipses detection in noisy environments: A hierarchical approach. Pattern Recognit. 2009, 42, 2421–2433. [Google Scholar] [CrossRef]
  22. Chen, S.; Xia, R.; Zhao, J.; Chen, Y.; Hu, M. A hybrid method for ellipse detection in industrial images. Pattern Recognit. 2017, 68, 82–98. [Google Scholar] [CrossRef]
  23. Fornaciari, M.; Prati, A.; Cucchiara, R. A fast and effective ellipse detector for embedded vision applications. Pattern Recognit. 2014, 47, 3693–3708. [Google Scholar] [CrossRef]
  24. Dong, H.; Sun, G.; Pang, W.C.; Asadi, E.; Prasad, D.K.; Chen, I.M. Fast ellipse detection via gradient information for robotic manipulation of cylindrical objects. IEEE Robot. Autom. Lett. 2018, 3, 2754–2761. [Google Scholar] [CrossRef]
  25. Fitzgibbon, A.; Pilu, M.; Fisher, R.B. Direct least square fitting of ellipses. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 476–480. [Google Scholar] [CrossRef]
  26. Chaudhuri, D. A simple least squares method for fitting of ellipses and circles depends on border points of a two-tone image and their 3-D extensions. Pattern Recognit. Lett. 2010, 31, 818–829. [Google Scholar] [CrossRef]
  27. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–698. [Google Scholar] [CrossRef] [PubMed]
  28. Whelan, D.A.; Adler, E.A.; Wilson, S.B., III; Roesler, G.M., Jr. DARPA Orbital Express program: Effecting a revolution in space-based systems. In Proceedings of the SPIE—The International Society for Optical Engineering, San Diego, CA, USA, 7 November 2000; p. 4136. [Google Scholar]
  29. Goodman, J. Louis.History of Space Shuttle Rendezvous and Proximity Operations. J. Spacecr. Rocket. 2006, 43, 944–959. [Google Scholar] [CrossRef]
Figure 1. Arcs are classified into four quadrants: (a) the gradient direction of edge points and (b) arc segments.
Figure 1. Arcs are classified into four quadrants: (a) the gradient direction of edge points and (b) arc segments.
Sensors 19 05243 g001
Figure 2. The definition of convex arc and concave arc: (a) the arc αi is the convex arc and (b) the arc αi is the concave arc.
Figure 2. The definition of convex arc and concave arc: (a) the arc αi is the convex arc and (b) the arc αi is the concave arc.
Sensors 19 05243 g002
Figure 3. Arc selection constraints: (a) quadrant constraint and position constraint and (b) tangent constraint.
Figure 3. Arc selection constraints: (a) quadrant constraint and position constraint and (b) tangent constraint.
Sensors 19 05243 g003
Figure 4. Ellipse center calculation.
Figure 4. Ellipse center calculation.
Sensors 19 05243 g004
Figure 5. Distance from the point to the ellipse: (a) the definition of the distance and (b) the computation of the distance.
Figure 5. Distance from the point to the ellipse: (a) the definition of the distance and (b) the computation of the distance.
Sensors 19 05243 g005
Figure 6. Detection results of real images.
Figure 6. Detection results of real images.
Sensors 19 05243 g006aSensors 19 05243 g006b
Figure 7. The average running time and the maximum time of the three methods.
Figure 7. The average running time and the maximum time of the three methods.
Sensors 19 05243 g007
Figure 8. The detection results of the challenge images. (ac) are the images under various illumination conditions; (di) are the images with cluttered background and image blur; (jl) show some failures of the method.
Figure 8. The detection results of the challenge images. (ac) are the images under various illumination conditions; (di) are the images with cluttered background and image blur; (jl) show some failures of the method.
Sensors 19 05243 g008
Figure 9. Results of the steps of our algorithm on the Orbital Express project video sequence: (a) the original image, (b) edge image, (cf) classification of edge points by gradient direction, (g) the arcs after selection, and (h) the docking ring detection result.
Figure 9. Results of the steps of our algorithm on the Orbital Express project video sequence: (a) the original image, (b) edge image, (cf) classification of edge points by gradient direction, (g) the arcs after selection, and (h) the docking ring detection result.
Sensors 19 05243 g009aSensors 19 05243 g009b
Figure 10. The detection results of our algorithm on the Orbital Express project video sequence.
Figure 10. The detection results of our algorithm on the Orbital Express project video sequence.
Sensors 19 05243 g010
Figure 11. The detection results of the steps of our algorithm on Atlantis docking with the International Space Station (ISS) video sequence: (a) the original image, (b) edge image, (cf) classification of edge points by gradient direction, (g) the arcs after selection, (h) the docking ring detection result.
Figure 11. The detection results of the steps of our algorithm on Atlantis docking with the International Space Station (ISS) video sequence: (a) the original image, (b) edge image, (cf) classification of edge points by gradient direction, (g) the arcs after selection, (h) the docking ring detection result.
Sensors 19 05243 g011
Figure 12. The detection results of our algorithm on Atlantis docking with the ISS video sequence.
Figure 12. The detection results of our algorithm on Atlantis docking with the ISS video sequence.
Sensors 19 05243 g012
Table 1. The processing time of each step on the Orbital Express project video sequence.
Table 1. The processing time of each step on the Orbital Express project video sequence.
Steps of the AlgorithmTime (ms)
Edge detection4.52
Arc detection and classification1.98
Arc grouping0.09
Parameter estimation0.12
Validity verification0.24
Total time6.96
Table 2. The processing time of each step on Atlantis Docking with ISS video sequence.
Table 2. The processing time of each step on Atlantis Docking with ISS video sequence.
Steps of the AlgorithmTime (ms)
Edge detection10.02
Arc detection and classification8.55
Arc grouping2.72
Parameter estimation8.57
Validity verification11.93
Total time41.79

Share and Cite

MDPI and ACS Style

Zhang, L.; Pan, W.; Ma, X. Real-Time Docking Ring Detection Based on the Geometrical Shape for an On-Orbit Spacecraft. Sensors 2019, 19, 5243. https://doi.org/10.3390/s19235243

AMA Style

Zhang L, Pan W, Ma X. Real-Time Docking Ring Detection Based on the Geometrical Shape for an On-Orbit Spacecraft. Sensors. 2019; 19(23):5243. https://doi.org/10.3390/s19235243

Chicago/Turabian Style

Zhang, Limin, Wang Pan, and Xianghua Ma. 2019. "Real-Time Docking Ring Detection Based on the Geometrical Shape for an On-Orbit Spacecraft" Sensors 19, no. 23: 5243. https://doi.org/10.3390/s19235243

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop