Next Article in Journal
Simulations of the Concentration Fields of Rosette-Type Multiport Buoyant Discharges Using Combined CFD and Multigene Genetic Programming Techniques
Previous Article in Journal
High Levels of Tetrodotoxin in the Flesh, Usually an Edible Part of the Pufferfish Takifugu flavipterus, Caused by Migration from the Skin and the Regional Characteristics of Toxin Accumulation
Previous Article in Special Issue
Physical Welding Factors for Reclassified Welding Positions in Shipbuilding Assembly Process Based on Muscle Activity Measured by Surface Electromyography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Gird Based Line Segment Detector and Application: Vision System for Autonomous Ship Small Assembly Line

1
State Key Laboratory of Ocean Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
2
School of Naval Architecture, Ocean & Civil Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2021, 9(11), 1313; https://doi.org/10.3390/jmse9111313
Submission received: 13 September 2021 / Revised: 10 November 2021 / Accepted: 16 November 2021 / Published: 22 November 2021
(This article belongs to the Special Issue Smart Technologies for Shipbuilding)

Abstract

:
The shipbuilding industry demands intelligent robot, which is capable of various tasks without laborious pre-teaching or programming. Vision system guided robots could be a solution for autonomous working. This paper introduces the principle and technique details of a vision system that guides welding robots in ship small assembly production. TOF sensors are employed to collect spatial points of workpieces. Huge data amount and complex topology bring great difficulty in the reconstruction of small assemblies. A new unsupervised line segment detector is proposed to reconstruct ship small assemblies from spatial points. Verified using data from actual manufacturing, the method of this paper demonstrated good robustness which is a great advantage for industrial applications. This paper’s work has been implemented in shipyards and shows good commercial potential. Intelligent, flexible industrial robots could be implemented with the findings of this study, which will push forward intelligent manufacturing in the shipbuilding industry.

1. Introduction

An industrial robot is the technical trend of intelligent shipbuilding and is expected to replace labor work in actual manufacturing. However, most welding jobs in shipyards are still performed manually because conventional welding robots lack flexibility. Industrial robots could be classified into 2 categories: teach-replay robots and offline-programming robots, which trajectories are achieved through laborious manual teaching or programming. These approaches are impractical in the shipbuilding industry because shipbuilding consists of multi-type and small-batch manufacturing for which the time and labor cost of manual teaching or programming are unacceptable from a commercial point of view. Thus most welding jobs in shipyards are still performed by manual operators. As rising labor cost and lack of skilled workers are posing challenges to shipyards, the demands of intelligent welding robot, which is capable of autonomous welding, is proposed by the shipbuilding industry.
Studies of autonomous robotic welding in shipbuilding have been attempted [1,2,3,4,5]. The major drawback in previous works is that robots’ trajectories depend on guide rails or rollers, thus repetitive installation and uninstallation of rails or robots consumes too much labor and time. Programming of welding robots also leads unfavorable ration of programming time to production time. Approaches to reduce robot programming costs could be classified in 3 ways: CAD-based method, hybrid method, and vision-based method. The CAD-based method employs macros or templates to reduce programming efforts [6,7]. This leads to significant maintenance costs and does not consider the derivation of workpieces. The hybrid method combines the CAD model with visual images, enjoying better flexibility but still requires CAD data or manual input [8]. The vision-based method could be an optimized solution for intelligent robotic welding because it is completely independent of CAD data and routine user inputs [9,10]. Meanwhile, the vision-based method proposes a great challenge, and splitting it into several small tasks could be a better solution [11]. Chen described robot pose control for the weld trajectory of a spatial corrugated web sheet based on laser sensing [12]. Additionally, welding robots with vision-based weld seam tracking modular are developed, which enjoy better tolerance to position deviation than conventional robots [13,14,15,16,17,18]. Research aiming at weld seam recognition are also conducted. Tsai produced welding path plans for golf club heads [19], and Zhang reconstructed a single weld seam using structured light [20]. Tsai and Zhang’s work successfully recognize a single weld seam in laboratory conditions. However, in previous studies, the problem was not solved under actual manufacturing conditions because workpieces in shipyards usually include complicated weld seam structures. The aforementioned studies did not consider multiple workpieces with various shapes, which is common in actual manufacturing in the shipyard.
Grid-based algorithms are generally more computationally efficient than other algorithms [21]. Most grid-based algorithms achieve a time complexity of O(n), where n is the number of spatial points. STING [22], Wave Cluster [23], and CLIQUE [24] are the most commonly used grid-based algorithms. The aforementioned algorithms emphasize point clustering more than line segment detection. However, the topic of the present study was intelligent robotic welding for shipbuilding, which requires segments reconstructed from complicated spatial points.
For the demand of intelligent shipbuilding, a vision-based method is presented to achieve autonomous robotic welding for small assemblies of various shapes without pre-teaching or programming. This method is tested in actual manufacturing data from the shipyard and demonstrated good tolerance, robustness, and accuracy.

2. Hardware Implementation

This study focuses on autonomous welding of small assemblies. Small assembly refers to the basic components in shipbuilding, consisting of plates and stiffeners, as in Figure 1. Most small assemblies are limited within the dimensional size of 4 m × 4 m, and the weight of 1 ton. All stiffeners need to be double-sided welded onto the plates.
The hardware of this study is displayed in Figure 2. Welding robots and time-of-flight (TOF) laser sensors are installed on the gantry. The robot controller and TOF sensors are connected to the PC by LAN cable.
The TOF laser sensors cast laser stripes onto small assemblies and collect depth values. As the gantry moves, the TOF laser sensors scan over small assemblies on rollers, as shown in Figure 3.
Data from the TOF laser sensor are organized in a depth matrix. The depth matrix is visualized in a depth-colored grayscale image, as shown in Figure 4. It could be transformed into a point cloud of the workpiece as the scan parameters are provided.

3. Reconstruction of Small Assembly

Reconstruction of small assembly is the precondition for autonomous robotic welding. The welding robot needs the exact location of each weld seam, which could be calculated from the stiffener’s centerline. Thus, the section introduces the method adopted in the reconstruction of all the stiffeners’ center lines from the depth matrix.

3.1. Spatial Points of Stiffeners

To reduce computation cost, the depth matrix is processed using various approaches to separate profiles from the background. For instance, Zhang used the Canny edge detector and Tsai calculated the geometric center of profiles [19,20]. In this study, we adopted Tsai’s method because it preserves profile information of stiffener. In this method, the recognition of the profile is performed within each scanned point of each laser stripe. Figure 5a shows a TOF sensor casting a laser stripe onto a workpiece with 2 stiffeners, and Figure 5b displays scanned points collected by the TOF sensor, in which the convex bumps in the dashed circle clearly indicate the location of the stiffener profile.
By extracting the profile centers from each column, the spatial points of stiffeners are obtained. Figure 6 gives examples of small assemblies and corresponding spatial points in this study.

3.2. Grid-Based Line Segment Detector

Problems arise as we are trying to reconstruct all the stiffeners from these spatial points. Inevitable noise interferes with conventional line segment detectors (e.g., Hough transformation (HT), RANSAC, LSD). For example, line segments detected by the HT from spatial points of Figure 6b are marked in red in Figure 7. It’s clear that noise results in missing points and false detection. Deliberately adjusting the parameters of the HT may alleviate the interference of noise. However, this is impractical in real-time manufacturing because the entire process is expected to run unsupervised. Moreover, the uncertainty of stiffeners’ count, multi-density of points, and complex topology of small assemblies add more difficulties to this work.
Thus, a new unsupervised approach is required to detect all of the line segments from spatial points. Here we introduce a grid-based line segment detector. It consists of 5 steps: sampling, rotation, convolution, deconvolution, and reconstruction, which are shown in Figure 8.
Before we take the small assembly in Figure 6b to explain the working process of the detector, it is necessary to introduce the coordinate system on the workpiece. The X-axis was parallel to the gantry rails, the Y-axis was perpendicular to the rail. The system origin is located at the corner of the panel, as shown in Figure 9.
The first step is sampling, which greatly reduces the complexity of the following computation. All spatial points are projected into a grid consisting of square cells, and this is also the grid-based line segment detector named by. Marking empty cells as white, and non-empty cells as black, the grid could be converted to a binary feature map. Figure 10 shows the process of sampling 15,066 points from Figure 6b to a feature map of 78 × 63 cells.
The rotation layer aligns the grid’s axis to the point’s major direction. The rotation angle of the axis is given by random sample consensus (RANSAC). RANSAC determines a single line from these points and calculates its angle from the x-axis, which is denoted by θ. The grid axis is then rotated by θ about the origin. Figure 11a shows the results of the rotation. After rotation, spatial points are re-sampled, as shown in Figure 11b.
Re-sampled feature map is converted to a matrix {f(i, j)}, where the matrix entries f(i, j) = 1 for black pixels and f(i, j) = 0 for white pixels. Matrix {f(i, j)} is convoluted to detect all line segments. Here 2 kernels are designed to detect horizontal and vertical segments.
First, the horizontal segment detector Kh [size: 3 × (2n + 1)] is introduced, where n equals half the length of the desired horizontal segment. The convolution operation using Kh differs slightly from the conventional convolution operation. It consists of the following steps:
  • Move kernel’s center to a black pixel (i,j), i is the row index, j is the column index
  • sum = 0;
  • l = jn;
  • Check pixels at (i1,l), (i,l) and (i + 1,l), increase sum by 1 if any of these pixels is black;
  • Increase l by 1;
  • Repeat 3. until l = j + n;
  • Convolution output at pixel g(i,j) = 1 if sum is not less than 2n, or g(i,j) = 0;
  • Repeat 1. until all nonzero entries are convoluted.
Figure 12a shows the convolution output of Figure 10b, which removes all vertical segments. All horizontal segments could be recovered through the deconvolution process, as shown in Figure 12b. Finally, it helps us extract spatial points to reconstruct stiffeners in this direction, as shown in Figure 12c.
Similarly, the vertical segments are detected by introducing kernel Kv. The size of Kv is (2m + 1) × 3, where m is half the length of the desired vertical segment. The convolution using Kv is similar to the algorithm used for the horizontal segments. Figure 13 shows the vertical segment detection process with kernel Kv.
The last step is reconstructing all line segments. As the spatial points are classified into groups corresponding to each stiffener, the Least-Squares Fitting could perform this job without any challenge. Figure 14 shows the centerlines of all stiffeners by combing the results shown in Figure 12c and Figure 13c. More ever, coordinates of stiffeners’ free ends are listed in Table 1.

4. Results and Verification

To investigate the accuracy of the segmented detector, The coordinates of the stiffeners’ free ends were manually measured and compared with the reconstructed position, as shown in Table 2.
In this study, the welding robot equips a range sensor for precise location and welding tracking. Tolerance of stiffener position is expected to be less than 50 mm. As shown in the last column of Table 2, all of the distances were less than 50 mm, satisfying the accuracy requirement.
Based on the demands of ship manufacturing, this algorithm is expected to give a reliable output, regardless of the kernel size or the shapes of the spatial points.
Figure 15a shows spatial points for which the convolutional grid-based clustering approach was adopted to extract all of the horizontal segments from the noise and other segments. Figure 15b shows the feature map sampled from these spatial points, and Figure 15c–e shows the convolution output for different sizes of Kh (n = 3, 4, 5). Figure 15c–e shows similar clustering numbers and positions of the black pixels, which indicated that the convolution output was insensitive to the value of n. As the goal of this step is to determine the number and positions of potential straight segments, this finding indicated the good robustness and tolerance of the algorithm, which are important for industrial applications. Thus, all the horizontal segments could be easily extracted using linear fitting. The results of the extracted points and segments are shown in Figure 15f.
The line segment detector is expected to deal with small assemblies of various shapes in the shipyard, thus spatial points of small assemblies from actual manufacturing are tested. Figure 16a,c,e show the spatial points, and Figure 16b,d,f show the corresponding segments reconstruction results. The comparison of the spatial points and reconstructed segments shows perfect matches.

5. Application

Figure 17 describes a general working flow of autonomous ship small assembly line, in which the line segment detector plays a crucial role. Data volume from the TOF sensor ranges from 500 MB to 5 GB, and spatial points number are from 50,000 to 100,000. The grid-based algorithm contributes to reducing computation complexity, allowing the assembly line welding to multiply workpieces simultaneously.
The line segment detector was implemented on an autonomous ship small assembly line in Guangzhou Shipyard International Company Ltd., belonging to China State Shipbuilding Corporation Ltd. (CSSC), as shown in Figure 18. The assembly line has been put into production since 2019. The vision system guided robots preforming various welding jobs without manual teaching or programming. It is also shown that the assembly line is capable of mass production, taking the leading position in this field.

6. Conclusions

In this paper, a grid-based line segment detector was introduced, which was used to guide robots for autonomous welding in ship-building. Intelligent robotic welding systems were implemented using the proposed algorithm, and the approach was verified under actual manufacturing conditions. The contributions of this study are summarized as follows:
(1)
The method presented in this paper demonstrated good robustness. It successfully clustered the points of stiffeners with complicated structures, despite the interference of noise. The robustness is a great advantage for industrial applications.
(2)
This algorithm was verified under manufacturing conditions and exhibited accuracy and robustness. Based on this work, intelligent ship small assembly lines were implemented and put into production in shipyards.
Curved segments, which are also important in ship-building, were not considered in this study. In future research, a detector for curved segments will be conducted to fill the gaps of this study. Also, we are planning to extend our research to more cases in shipbuilding, like welding jobs for ship blocks, reversing modeling of hull structures, etc.

Author Contributions

Conceptualization, C.N.; software, C.N.; validation, J.D.; writing—original draft preparation, J.D.; writing—review and editing, C.N.; visualization, J.D.; project administration, C.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rooks, B. Robot Welding in Shipbuilding. Ind. Robot 1997, 24, 413–417. [Google Scholar] [CrossRef]
  2. Lee, D. Robots in the shipbuilding industry. Robot CIM Int. Manuf. 2014, 30, 442–450. [Google Scholar] [CrossRef]
  3. Lee, D.; Ku, N.; Kim, T.; Kim, J.; Lee, K.; Son, Y. Development and application of an intelligent welding robot system for shipbuilding. Robot CIM Int. Manuf. 2011, 27, 377–388. [Google Scholar] [CrossRef]
  4. Kam, B.; Jeon, Y.; Kim, S. Motion control of two wheeled welding mobile robot with seam tracking sensor. In Proceedings of the ISIE 2001—2001 IEEE International Symposium on Industrial Electronics Proceedings (Cat. No. 01TH8570); IEEE: Piscataway, NJ, USA, 2001; Volume 2, pp. 851–856. [Google Scholar] [CrossRef]
  5. Lee, J.; Kim, J.; Kim, J.; Park, J. Development of carriage-type welding Robot for double hull assembly line in shipbuilding. IFAC Proc. Vol. 2008, 41, 4310–4311. [Google Scholar] [CrossRef] [Green Version]
  6. Bedaka, A.; Vidal, J.; Lin, C. Automatic robot path integration using three-dimensional vision and offline programming. Int. J. Adv. Manuf. Technol. 2019, 102, 1935–1950. [Google Scholar] [CrossRef]
  7. Ferreira, L.; Figueira, Y.; Iglesias, I.; Souto, M. Offline CAD-based Robot Programming and Welding Parametrization of a Flexible and Adaptive Robotic Cell Using Enriched CAD/CAM System for Shipbuilding. Procedia Manuf. 2017, 11, 215–223. [Google Scholar] [CrossRef]
  8. Kang, S.; Youn, H.; Kim, D.; Kim, K.; Lee, S.; Kim, S. Development of multi welding robot system for sub assembly in shipbuilding. IFAC Proc. Vol. 2008, 41, 5273–5278. [Google Scholar] [CrossRef] [Green Version]
  9. Alexander, Z. Programming of Welding Robots in Shipbuilding. Procedia CIRP 2021, 99, 478–483. [Google Scholar] [CrossRef]
  10. Muhammad, J.; Altun, H.; Abo-Serie, E. Welding seam profiling techniques based on active vision sensing for intelligent robotic welding. Int. J. Adv. Manuf. Technol. 2016, 88, 127–145. [Google Scholar] [CrossRef]
  11. Chen, X.; Yu, J. Acquisition and optimization of weld trajectory and pose information for robot welding of spatial corrugated web sheet based on laser sensing. Int. J. Adv. Manuf. Technol. 2018, 96, 3033–3041. [Google Scholar] [CrossRef]
  12. Zhou, L.; Lin, T.; Chen, S. Autonomous Acquisition of Seam Coordinates for Arc Welding Robot Based on Visual Servoing. J. Intell. Robot Syst. 2006, 47, 239–255. [Google Scholar] [CrossRef]
  13. Chen, S.; Chen, X.; Qiu, T. Acquisition of Weld Seam Dimensional Position Information for Arc Welding Robot Based on Vision Computing. J. Intell. Robot Syst. 2005, 43, 77–97. [Google Scholar] [CrossRef]
  14. He, Y.; Chen, Y.; Xu, Y. Autonomous Detection of Weld Seam Profiles via a Model of Saliency-Based Visual Attention for Robotic Arc Welding. J. Intell. Robot Syst. 2016, 81, 395–406. [Google Scholar] [CrossRef]
  15. Lü, X.; Zhang, K.; Wu, Y. The seam position detection and tracking for the mobile welding robot. Int. J. Adv. Manuf. Technol. 2017, 88, 2201–2210. [Google Scholar] [CrossRef]
  16. Moon, H.; Kim, Y.; Beattie, R. Multi sensor data fusion for improving performance and reliability of fully automatic welding system. Int. J. Adv. Manuf. Technol. 2006, 28, 286–293. [Google Scholar] [CrossRef]
  17. Gu, W.; Xiong, Z.; Wan, W. Autonomous seam acquisition and tracking system for multi-pass welding based on vision sensor. Int. J. Adv. Manuf. Technol. 2013, 69, 451–460. [Google Scholar] [CrossRef]
  18. Tsai, M.; Lee, H.; Ann, N. Machine Vision Based Path Planning for a Robotic Golf Club Head Welding System. Robot CIM Int. Manuf. 2011, 27, 843–849. [Google Scholar] [CrossRef]
  19. Zhang, K.; Yan, M. 3D Reconstruction of Complex Spatial Weld Seam for Autonomous Welding by Laser Structured Light Scanning. J. Manuf. Proc. 2019, 39, 200–207. [Google Scholar] [CrossRef]
  20. Dong, S.; Liu, J.; Liu, Y.; Zeng, L.; Xu, C.; Zhou, T. Clustering based on grid and local density with priority-based expansion for multi-density data. Inform. Sci. 2018, 468, 103–116. [Google Scholar] [CrossRef]
  21. Yue, S.; Wei, M.; Wang, J.; Wang, H. A general grid-clustering approach. Pattern Recogn. Lett. 2008, 29, 1372–1384. [Google Scholar] [CrossRef]
  22. Wang, W.; Yang, J.; Richard, M. STING: A Statistical Information Grid Approach to Spatial Data Mining. In Proceedings of the 23rd International Conference on Very Large Data Bases (VLDB ’97); Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1997; pp. 186–195. [Google Scholar]
  23. Sheikholeslami, G.; Chatterjee, S.; Zhang, A. WaveCluster: A Multi-Resolution Clustering Approach for Very Large Spatial Databases. In Proceedings of the 24rd International Conference on Very Large Data Bases (VLDB ’98); Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1998; pp. 428–439. [Google Scholar]
  24. Agrawal, R.; Gehrke, J.; Gunopulos, D.; Raghavan, P. Automatic subspace clustering of high dimensional data for data mining applications. In Proceedings of the 1998 ACM SIGMOD International Conference on Management of Data (SIGMOD ’98); Association for Computing Machinery: New York, NY, USA, 1998; pp. 94–105. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Examples of small assemblies in the shipyard. (a,b,d) consist of orthogonal stiffeners, and (c) consists of parallel stiffeners.
Figure 1. Examples of small assemblies in the shipyard. (a,b,d) consist of orthogonal stiffeners, and (c) consists of parallel stiffeners.
Jmse 09 01313 g001
Figure 2. (a) Robot and TOF sensor (shroud removed) installed on a gantry; (b) TOF sensor (with shroud).
Figure 2. (a) Robot and TOF sensor (shroud removed) installed on a gantry; (b) TOF sensor (with shroud).
Jmse 09 01313 g002
Figure 3. TOF sensors scan over small assemblies.
Figure 3. TOF sensors scan over small assemblies.
Jmse 09 01313 g003
Figure 4. Depth-colored grayscale image of small assemblies.
Figure 4. Depth-colored grayscale image of small assemblies.
Jmse 09 01313 g004
Figure 5. (a) Laser stripe on the small assembly; (b) Profiles in the row of scanned points.
Figure 5. (a) Laser stripe on the small assembly; (b) Profiles in the row of scanned points.
Jmse 09 01313 g005aJmse 09 01313 g005b
Figure 6. Small assemblies (a,c,e,g) and corresponding spatial points (b,d,f,h).
Figure 6. Small assemblies (a,c,e,g) and corresponding spatial points (b,d,f,h).
Jmse 09 01313 g006aJmse 09 01313 g006bJmse 09 01313 g006cJmse 09 01313 g006dJmse 09 01313 g006e
Figure 7. Line segments were detected (in red) from spatial points.
Figure 7. Line segments were detected (in red) from spatial points.
Jmse 09 01313 g007
Figure 8. Steps of line segment reconstruction from spatial points.
Figure 8. Steps of line segment reconstruction from spatial points.
Jmse 09 01313 g008
Figure 9. Coordinate system on the workpiece.
Figure 9. Coordinate system on the workpiece.
Jmse 09 01313 g009
Figure 10. (a) Spatial Points (15,066 points); (b) Sampled feature map.
Figure 10. (a) Spatial Points (15,066 points); (b) Sampled feature map.
Jmse 09 01313 g010
Figure 11. (a) Points after axis rotation; (b) Re-sampled feature map.
Figure 11. (a) Points after axis rotation; (b) Re-sampled feature map.
Jmse 09 01313 g011
Figure 12. (a) Convolution output with Kh; (b) Horizontal segments recovered (in red bounds); (c) Points of each stiffener (in red, green, and blue).
Figure 12. (a) Convolution output with Kh; (b) Horizontal segments recovered (in red bounds); (c) Points of each stiffener (in red, green, and blue).
Jmse 09 01313 g012
Figure 13. (a) Convolution output with Kv; (b) Vertical segments recovered (in red bounds); (c) Points of each stiffener (in red, green, and blue).
Figure 13. (a) Convolution output with Kv; (b) Vertical segments recovered (in red bounds); (c) Points of each stiffener (in red, green, and blue).
Jmse 09 01313 g013
Figure 14. Reconstruction of all stiffeners on small assembly.
Figure 14. Reconstruction of all stiffeners on small assembly.
Jmse 09 01313 g014
Figure 15. (a) Spatial points of small assemblies; (b) Feature map; (ce). Convolution output with Kh (n = 3, 4, and 5, respectively); (f) Stiffener center line reconstructed.
Figure 15. (a) Spatial points of small assemblies; (b) Feature map; (ce). Convolution output with Kh (n = 3, 4, and 5, respectively); (f) Stiffener center line reconstructed.
Jmse 09 01313 g015aJmse 09 01313 g015b
Figure 16. (a,c,e) Spatial points of small assemblies; (b,d,f) Reconstructed segments of stiffeners.
Figure 16. (a,c,e) Spatial points of small assemblies; (b,d,f) Reconstructed segments of stiffeners.
Jmse 09 01313 g016aJmse 09 01313 g016b
Figure 17. Working flow of autonomous ship small assembly line.
Figure 17. Working flow of autonomous ship small assembly line.
Jmse 09 01313 g017
Figure 18. (a) Vision guided welding robots in production at shipyard; (b) Overview of the assembly line.
Figure 18. (a) Vision guided welding robots in production at shipyard; (b) Overview of the assembly line.
Jmse 09 01313 g018
Table 1. Coordinates of the reconstructed free ends.
Table 1. Coordinates of the reconstructed free ends.
Free EndX (mm)Y (mm)
A1043452
B343128
C16911919
D24891784
E31811656
F3754783
Table 2. Distance between reconstructed and measured positions.
Table 2. Distance between reconstructed and measured positions.
Free EndMeasured Position
(x, y)
Reconstructed Position
(x, y)
Distance (mm)
A(1032, 426)(1043, 452)28.2
B(3410, 25)(3431, 28)21.2
C(1682, 1910)(1691, 1919)12.7
D(2483, 1773)(2489, 1784)10.8
E(3177, 1650)(3181, 1656)7.2
F(3725, 798)(3754, 783)32.6
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ding, J.; Ni, C. Gird Based Line Segment Detector and Application: Vision System for Autonomous Ship Small Assembly Line. J. Mar. Sci. Eng. 2021, 9, 1313. https://doi.org/10.3390/jmse9111313

AMA Style

Ding J, Ni C. Gird Based Line Segment Detector and Application: Vision System for Autonomous Ship Small Assembly Line. Journal of Marine Science and Engineering. 2021; 9(11):1313. https://doi.org/10.3390/jmse9111313

Chicago/Turabian Style

Ding, Jinhong, and Chongben Ni. 2021. "Gird Based Line Segment Detector and Application: Vision System for Autonomous Ship Small Assembly Line" Journal of Marine Science and Engineering 9, no. 11: 1313. https://doi.org/10.3390/jmse9111313

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop