You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

2 September 2022

The Method of Creel Positioning Based on Monocular Vision

,
,
,
and
1
School of Automation, Zhejiang Institute of Mechanical & Electrical Engineering, Hangzhou 310053, China
2
School of Mechanical Engineering & Automation, Zhejiang Sci-Tech University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Navigation and Positioning

Abstract

Automatic replacement of bobbins on the frame is one of the core problems that must be solved in the textile industry. In the process of changing the bobbin, it is faced with problems such as the position offset of the bobbin, the loosening and deformation of the bobbin, which will lead to the failure of replacing the bobbin. Therefore, it is necessary to initialize the creel coordinates regularly, also considering the labor intensity and poor reliability of manual positioning. This paper proposes an automatic creel positioning method based on monocular vision. Firstly, the industrial camera was installed on the drum changing manipulator controlled by the truss system, and each yarn frame in the yarn area was inspected. Secondly, the end face image of the creel was collected at a reasonable distance, and the collected images were transmitted to the computer in real time through the Ethernet bus. Thirdly, the center coordinates (x, y) and radius r of the creel were marked by the improved Hough circle detection algorithm. Finally, the coordinate deviation of the creel was calculated and transmitted to the system controller to realize deviation correction. Before the test, the creel positioning markers were specially designed and the camera was calibrated. Thus, the influence of image complex background, creel end roughness, reflection and other factors can be reduced, and the image processing speed and positioning accuracy can be improved. The results show that the positioning effect of this method is fine when the distance between the center of the camera lens and the center of the end face of the creel is 170~190 mm. Moreover, when the distance is 190 mm, the positioning effect is the best, with an average error of only 0.51 mm. In addition, the deviation between the center coordinate and the radius of the end face of the marker is also very small, which is better than the requirements of bobbin yarn changing accuracy.

1. Introduction

The floor frame is a form of knitting “weft knitting round machine” frame. The frame is divided into two rows and stands on both sides of the machine in a linear or circular shape, and the bobbins are inserted obliquely on the creel of the frame. Although the floor yarn frame covers a large area and has a long yarn supply distance, it has the advantages of small machine load-bearing and convenient yarn change [1]. At present, it is widely used in the production workshop of circular weft knitting machine, and the yarn bobbin yarn is replaced manually in the traditional way for thread continuation. With the deepening of problems such as high labor cost, high labor intensity and difficult recruitment, it has begun to affect the normal development of the textile industry, and enterprises have put forward an urgent need for “machine replacement” [2,3,4]. Therefore, in the context of Intelligent Manufacturing, it is an inevitable trend to use robots to automatically replace the bobbin on the yarn frame [5].
Although the positioning accuracy required by the manipulator can be achieved through the cooperation of the truss mechanism and the servo system when the machine is changing bobbins, there are many problems such as the low manufacturing accuracy of the creel, the looseness of the creel, the inevitable vibration in the process of automatic drum changing, the displacement of the creel, and the sag of the creel caused by the heavy weight of the creel. All of these will affect the normal bobbin replacement, so the coordinates of the creel must be regularly initialized to ensure the manipulator can replace the bobbin normally. However, the manual initialization and positioning of the creel coordinates has a large workload and poor reliability, so it is considered to realize the initialization and positioning of the creel coordinates based on machine vision.
At present, there is a lot of research on visual positioning, such as the accuracy analysis of threaded hole positioning [6], research on robot arm positioning method [7], and object pose measurement and error [8], but studies on the textile industry are relatively few. The reason is that textiles, as a traditional labor-intensive industry, has not paid enough attention to automation and intelligence in the past, and started relatively late. In addition, the textile factory environment is complex, and there are too many background contents of image acquisition, which leads to difficulties in image processing, and the high cost of vision system investment, which is difficult for enterprises to bear [9]. Therefore, based on the existing results of monocular vision positioning, this paper will study the system and algorithm suitable for the coordinate positioning of creels, mainly including principal analysis, system design, algorithm design, experimental testing and other aspects, so as to provide help for the early realization of automatic drum changing of creels and intelligent manufacturing in the textile industry.
In the production workshop of knitting circular weft knitting machine, the structure diagram of the commonly used floor frame is shown in Figure 1. It is composed of two rows of eight-layer creels on the left and right, and a simple support. The creels on the same layer are placed with the same specification of bobbins, which can realize continuous yarn supply without stopping. In this paper, machine vision technology will be used to locate the creel to meet the requirements of “machine replacing bobbin”.
Figure 1. The structure diagram of the commonly used floor frame.
Since the diameter of the end face of the creel is known, the unknown variables of positioning in image processing are reduced. Then, by designing markers, integrating circle detection and monocular vision technology, the recognition and positioning function of the creel can be realized simply and quickly.
The rest of the paper is organized as follows: Section 2 is the related works. Section 3 presents the prepare works. Section 4 is the main part of this paper; it proposes two steps to detect the creel. Section 5 provides the experiential results and analysis of the method. Section 6 presents the conclusion and discussion of this study.

3. Preparations

Before starting, there are two tasks that need to be performed. The first one is to set up the creel test system. The second one is to design the creel positioning markers. The details are as follows.

3.1. System Design

The creel positioning system is mainly composed of a creel yarn changing manipulator, an industrial camera, a creel rod, an industrial computer and other equipment [27,28]. The industrial camera is installed on the special platform of the bobbin yarn changing manipulator, with good field of vision and no interference. The bobbin yarn changing manipulator is suspended on the truss robot and can follow the truss to move to any position of the yarn frame in the yarn using area. Moreover, it is used to realize the function of grasping and placing the yarn bobbin yarn. The industrial computer is configured at a compromise position near the outside of the yarn using area and the yarn preparation area. Ethernet communication is used with industrial cameras, trusses and manipulator controllers. The main hardware composition of the creel positioning mechanism is shown in Figure 2.
Figure 2. This is the physical drawing of creel visual positioning system.
The composition block diagram of creel positioning system is shown in Figure 3, which includes image acquisition, image processing and positioning and rectification.
Figure 3. This is the composition block diagram of creel positioning system.
The positioning steps of creel are as follows. First, according to the instructions issued by the industrial computer, the bobbin yarn changing manipulator is controlled by the delta AS320T to reach the position of the creel that needs to be positioned in the yarn area. Then, the industrial camera is awakened by the industrial computer to capture the creel image, and the image is processed by the industrial computer positioning algorithm to obtain the new creel position coordinates. Finally, it is stored in the database to replace the original coordinate data, which can be called for subsequent bobbin yarn replacement, and updated when the next relocation, so as to ensure the accuracy of bobbin yarn positioning and the success rate of barrel replacement.

3.2. Markers Design

There are the following problems in image recognition of the end face of the bobbin directly: the diameter of the end face is only about 8 mm, the color is silver-white, and the process is slightly rough. If the image of the creel is taken directly with an industrial camera for processing, recognition and positioning, the configuration requirements of industrial computers and other equipment will be very high, which will greatly increase the use cost [29,30]. Therefore, the method of setting markers on the creel is adopted to highlight the characteristics of the creel. Figure 4a,b show the visual positioning markers designed or selected by the team. The disc marker has its own contrast, which makes recognition easier and greatly reduces the requirements of hardware configuration. The cover marker is installed on the creel, which is conducive to improving the flatness of the end face and improving the recognition. Its diameter is 10 mm.
Figure 4. Shows the visual location markers. (a) Disc marker; (b) Cover marker.
By analyzing the layout of the yarn frame in the yarn area and the structure of the yarn frame itself, we determined that the better installation positions of the markers are the support under the creel and the end face of the creel, as shown in Figure 5. The disc marker is directly pasted on the support under the creel, and the object is shown in Figure 5a. Its test recognition effect is fine, but it has high requirements for the camera installation mechanism, which is easy to interfere with. Moreover, due to the uneven end face of the creel, the direct pasting firmness and durability are poor, so it is only suitable for experimental testing. Moreover, since the image processing effect of the end face of the creel does not meet the actual requirements, it is considered to install the sleeve on the creel, as shown in Figure 5b. Moreover, it has the advantages of convenient installation, good durability, wide adaptability of yarn frame type, and the test recognition effect and shooting distance can also meet the requirements. So, it is a good choice for practical application. At the same time, in order to enhance the effect of image processing, we design the end face of the cover as a reflective surface on the basis of Figure 5b, and the actual effect is shown in Figure 5c.
Figure 5. Shows the installation drawing of markers. (a) Physical installation drawing of disc marker; (b) Physical installation drawing of blue surface cover; (c) Physical installation drawing of black surface cover; (d) Physical installation drawing of reflective surface cover.

4. Proposed Methods

In this part, we use two steps to realize the computer vision positioning creel. The first step is the circle detection based on Hough transform, and the second part is the monocular positioning based on industrial camera. The details are as follows.

4.1. Creel Recognition Based on Circle Detection

Hough circle transformation is the process of transforming a circle in the two-dimensional image space into a point in the three-dimensional parameter space, which determined by the radius r of the circle and the coordinates of the center (x, y). Therefore, the circle determined by any three points on the circumference corresponds to a point in the three-dimensional parameter space after Hough transformation [31,32].
The basic idea of Hough circle transformation is to think that every non-zero pixel on the image may be a point on a potential circle. Similarly the Hough line transformation, the Hough circle transformation also generates the cumulative coordinate plane by voting, setting a cumulative weight to locate the circle.
We know that in the Cartesian coordinate system, the equation of the circle is shown as Formula (1).
( x a ) 2 + ( y b ) 2 = r 2
(x, y) is the point on the circumference, (a, b) is the center coordinate, and r is the radius, which can also be expressed as Formula (2).
{ x = a + r cos θ y = b + r sin θ
θ is the angle between the line connecting the center (a, b) and the point (x, y) on the circle and the horizontal line passing through the center (a, b). By sorting out Formula (2), we can obtain the following Formula (3).
{ a = x r cos θ b = y r sin θ
Therefore, in the three-dimensional coordinate system composed of a, b and r, a point can uniquely determine a circle.
In Cartesian x, y coordinate system, all circles passing through a certain point are mapped to a, b and r coordinate system, which is a three-dimensional curve. All circles passing through all non-zero pixels in the x, y coordinate system constitute many three-dimensional curves in the a, b and r coordinate system.
In the x and y coordinate systems, there is only one circular equation for all points on the same circle, which maps to the same point in the a, b and r coordinate systems. Therefore, in the a, b, r coordinate system, this point should have N0 curves intersecting the total pixels of the circle. Then, by judging the number of intersections of each point in a, b and r, the point which is greater than a certain threshold is regarded as a circle.
The above is the implementation algorithm of standard Hough circle transformation, but the problem is that its cumulative surface is three-dimensional space, which means that it requires more computational consumption than Hough line transformation. Therefore, OpenCV optimizes the operation of Standard Hough circle transformation. It adopts the “Hough gradient method” detection idea, which is to traverse and accumulate the center of all non-zero points. Considering that the center of the circle must be on the module vector of each point on the circle, that is, on the vertical line perpendicular to the point and passing through the tangent of the point, the intersection of the module vectors on these circles is the center of the circle.
The schematic diagram of the relationship between the module vector and the center of the circle is shown in Figure 6. The Hough gradient method is to find these centers, and then determine them according to the number of intersections of the modulus vector on the “center” and the appropriate threshold value. As shown in Figure 6, if we calculate the gradient of a circle, the gradient direction of all points on the circle faces the center of the circle.
Figure 6. Schematic diagram of the relationship between module vector and circle center.
Therefore, the steps of circle detection are as follows. Firstly, the Canny algorithm is used to detect the edge of the image. Secondly, the gradient of all pixels is calculated using Sobel operator. Thirdly, after Canny traverses all non-zero pixel points, it will draw lines along the gradient direction. Each point has an accumulator, when a line passes through the point, the accumulator adds 1. Then, all of the accumulated values are sorted, and all possible center points are found according to the threshold value. Fourth, we calculated the distance between all non-zero pixels in canny image and the center of the circle, sort the distance from small to large, and select the appropriate radius. Fifth, we set the accumulator for the selected radius, and add 1 to the accumulator value that meets the radius.
Considering the factor of efficiency, the Hough circle detection implemented in OpenCV is based on image gradient. The first step is to detect the edge, the second step is to find the possible center of the circle and calculate the optimal radius from the candidate center based on the first step. In OpenCV, some codes and analysis of Hough circle detection algorithm based on image gradient are as follows (Algorithm 1).
Algorithm 1. The process of Hough circle detection algorithm based on image gradient
1: HoughCircles;
2: InputArray image;
3: OutputArray circles;
4: Int method;
5: Double dp;
6: Double mindist;
7: Double param1;
8: Double param2;
9: Int minradius;
10: Int maxradius;
“InputArray” indicates input image. The data type is generally mat type. It needs to be an 8-bit single channel gray-scale image.
“OutputArray” stores the output vector of the detected circle.
“Method indicates” the detection method used. At present, only the Hough gradient method is available in OpenCV. Fill in Hough for this parameter_grade is enough.
“dp” is double type data, which refers to the reciprocal of the ratio of the center accumulator to the resolution of the input image. For example, if dp = 1, the accumulator and the input image have the same resolution. If dp = 2, the accumulator has half the width and height of the input image.
“Mindist” is the minimum distance between the centers of circles detected by Hough transform.
“Param1” is the corresponding parameter of the detection method set by the third parameter method. For the current only method, the Hough gradient method CV_ HOUGH_Gradient, which represents the high threshold passed to Canny edge detection operator, while the low threshold is half of the high threshold.
“Param2” is also the corresponding parameter of the detection method set by the third parameter method. For the current only method, Hough gradient method_ Gradient, which represents the accumulator threshold of the center of the circle in the detection phase. The smaller it is, the more non-existent circles can be detected, and the larger it is, the more perfect circles can be detected.
“Minradius” indicates the minimum value of the circle radius.
“Maxradius” indicates the maximum value of the circle radius.
Figure 5a is processed by the above algorithm, and the recognition effect of disc markers is shown in Figure 7. It can be seen from the Figure that this algorithm not only realizes the rapid recognition of the marker, but also accurately locates its center. Therefore, it is feasible to use this method to locate the creel.
Figure 7. Effect drawing of disc marker recognition.

4.2. Creel Positioning Based on Monocular Vision

The schematic diagram of creel positioning based on monocular vision is shown in Figure 8. OwXwYwZw, OcXcYcZc, and OXY are world coordinate system, camera coordinate system and image coordinate system, u, v is the pixel coordinate system. Ow is the origin position for establishing the world coordinate system. Oc is the optical center of the camera; P (Xw, Yw, Zw) is a point in the world coordinate system. p (x, y) is the imaging point in the image, the coordinates in the image coordinate system (x, y), and the coordinates in the pixel coordinate system (u, v), f is the focal length of the camera, equal to the distance between point O and Oc [33,34].
Figure 8. Creel positioning schematic diagram.
We select an image of the creel for processing, and obtain the coordinates of the creel in the image coordinate system as p (x, y), the coordinates in the camera coordinate system as P (xc, yc, zc), the coordinates in the world coordinate system as P (xw, yw, zw), and the focal length of the lens as f. Therefore, according to the geometric relationship in Figure 8, the calculation formula of the center position of the creel is obtained as follows.
x = f X c Z c
y = f Y c Z c
The converted camera pinhole model is shown as (6).
[ x y 1 ] = [ f 0 0     0     f     0     0     0     1     0     0     0 ] [ x c y c z c 1 ]
Convert the image coordinate system to the world coordinate system. The specific formula is shown as (7).
[ x c y c z c 1 ] = [ R   0   T 1 ] [ x w y w z w 1 ]
[ R   0   T 1 ] is the camera external parameter matrix; R is the rotation matrix and T is the translation matrix, which is used to express the relationship between the origin of the world coordinate system and the optical center of the camera.
Next is the conversion from image coordinate system to pixel coordinate system, and the secondary conversion of the object that has been projected on the imaging plane. The specific formula is shown as follows.
[ u v 1 ] = [ 1 d x   0   0 0 1 d y 0     u 0     v 0 1 ] [ x y 1 ]
d x and d y are pixel sizes. u 0 and v 0 are the actual position of the main point.
By transforming the image coordinate system into pixel coordinate system by Formulas (3)–(5), we can obtain the following formula.
[ u v 1 ] = [ f d x 0 0     0   f d y 0     u 0   v 0 1 ] [ 1     0     0     0 0     1     0     0 0     0     1     0 ] [ R   0   T 1 ] [ x w y w z w 1 ]
f d x , f d y are called normalized focal lengths on the x-axis and y-axis, respectively. [ f d x 0 0     0   f d y 0     u 0   v 0 1 ] is the camera internal parameter matrix. The internal parameter matrix reflects the attributes of the camera itself. Each camera is different and needs calibration to know these parameters. The combination of internal parameter matrix and external parameter matrix is the camera matrix, which establishes the transformation from three-dimensional point homogeneous coordinates to two-dimensional point homogeneous coordinates.
Assume that the distortion free industrial camera is applied to the system. Its resolution is m × n, and target size is a × b. The radius of the creel end face in the image coordinate system is r1, and the actually measured radius is r2. Through measurement, the spatial position of the origin of the camera coordinate system under the manipulator coordinate system is P( x 1 , y 1 , z 1 ), and through coordinate conversion, the spatial position coordinate of the creel end face under the manipulator coordinate system is P’( x r 2 r 1 + x 1 , y r 2 r 1 + y 1 , r 2 m f r 1 a + z 1 ).
According to the monocular vision positioning principle, Figure 5a is further processed to obtain the image shown in Figure 9. It can be seen from the Figure that the distance between the camera, the center of the mark and the coordinates of the center of the circle are successfully obtained, which prove the correctness of this method.
Figure 9. Effect drawing of creel positioning.

5. Experimental Test and Analysis

When testing, the image acquisition equipment adopts the MV-CA050-10GM black-and-white industrial camera produced by Hikvision, with a resolution of 2448 × 2048, and the target size is 2/3 inch. It is equipped with Gigabit Ethernet, the lens is distortion free, the focal length is 8 mm, and the shooting effect is good under low illumination. Vs2019 and OpenCV were used as image processing software in the experiment. The process is completed on the computer with 8 GB memory and 2.50 GHz CPU working frequency.

5.1. Camera Calibration

The creel is in the three-dimensional world, and the creel photos that are taken are two-dimensional. If the camera is regarded as a function, the input is a scene, and the output is a gray-scale image [35]. In the process from three-dimensional to two-dimensional, the function is irreversible. The goal of camera calibration is to find a suitable mathematical model, then calculate the parameters of the model, approximately simulate the process from three-dimensional to two-dimensional, and finally find the inverse function [36,37]. Therefore, camera calibration is one of the basic works of visual measurement and positioning. The accuracy of calibration parameters is directly related to the accuracy of creel. Generally, camera calibration methods are divided into four methods: traditional camera calibration method, active vision camera calibration method, camera self calibration method and zero distortion camera calibration method.
Zhang Zhengyou calibration method is adopted in this paper. Strictly speaking, Zhang Zhengyou calibration method is between the traditional calibration method and the self calibration method. The GML camera calibration toolbox is used as the calibration software, in which the number of calibration images is 20, and the checkerboard specification is 10-11-14 mm printing paper. The camera calibration effect is shown in Figure 10. The observation shows that the corners of the black-and-white checkerboard in the image are accurately identified.
Figure 10. Camera calibration diagram. (a) Calibration object; (b) Reproject.
The specific camera parameters obtained after calibration are shown in Figure 11, key parameters are as follows. Focal Length (Focal length in X, Y axis direction f x , f y ) is [2317.659 2314.165] ± [3.266 3.275]. Principal Point ( u 0 , v 0 ) is [1242.338 975.647] ± [2.490 2.656]. Distortion (Lens distortion coefficient) is [−0.149059 0.227607 −0.000687 −0.001450] ± [ 0.009618 0.130363 0.000270 0.000264]. The Camera Matrix is shown as Formula (10).
M = [ f x 0 0 0   f y       0 u 0 v 0 1 ] = [ 2317.659 0 0 0 2314.165         0 1242.338 975.647 1 ]
Figure 11. Camera calibration result diagram.

5.2. Image Processing and Positioning

After the camera calibration was completed, the calibrated camera parameters were directly inputted into the improved circle detection program of OpenCV platform to process and recognize the creel image, and automatically correct the positioning data according to the calibrated camera parameters.
In order to ensure the positioning accuracy of the creel and prevent interference with the yarn frame, it is necessary to determine the optimal distance for camera positioning. We selected the distance between the camera lens center and the marker center 160 mm, 170 mm, 180 mm, 190 mm, 200 mm and 210 mm, respectively, then conducted a large number of image processing at these six distances. We randomly select one image of each interval, as shown in Figure 12a–f.
Figure 12. Image processing results of each interval. (a) Image processing results of 160 mm; (b) Image processing results of 170 mm; (c) Image processing results of 180 mm; (d) Image processing results of 190 mm; (e) Image processing results of 200 mm; (f) Image processing results of 210 mm.
According to the image processing results, marker diameter, pixel size and other parameters, using the pinhole imaging principle, the pixel can be converted to the positioning distance through similar triangle calculation. The diameter of the marker is 10 mm, and the pixel size can be obtained according to Formula (7) or read according to the camera manual. Finally, the pixel size was 3.45 um × 3.45 um.
Pixel size = target size/total pixel size

5.3. Analysis of Experimental Data

We selected six intervals of 160 mm, 170 mm, 180 mm, 190 mm, 200 mm and 210 mm, respectively, for testing. In order to ensure the accuracy of the positioning data, in the experimental test, 10 images were quickly collected each time as a group, and then the 10 images were processed to obtain the 10 data, and the average value of the 10 data was obtained as one data. A total of 13 sets of images were collected in the experiment, and finally 13 data were obtained. After calculation, 13 groups of data at six intervals were drawn into curves, which are shown in Figure 13.
Figure 13. Image processing data curve of each interval.
It can be seen from the Figure 13 that the maximum errors of the five spacing of 160 mm, 170 mm, 180 mm, 190 mm, 200 mm and 210 mm are 6.49 mm, 3.61 mm, 3.37 mm, 1.47 mm, 1.77 mm and 4.25 mm, respectively. So, we can conclude that only 190 mm and 200 mm meet the requirements if the bobbin yarn replacement accuracy is 2 mm.
After the 13 groups of data measured for each interval in Figure 10 are averaged, the results are shown in Table 1.
Table 1. This table shows the creel positioning results under different test spacing.
It can be seen from the Table 1: (1) When the spacing is 190 mm, the average error measured is the smallest, only 0.51 mm, and the relative error is 0.27%, which is far less than the accuracy requirements of 2 mm; (2) The average error of spacing of 180 mm and 200 mm is only 1.17 mm and 0.68 mm, which also meets the requirements; (3) The average errors of 160 mm and 210 mm spacing reached 5.12 mm and 4.88 mm, respectively, which exceeded the accuracy requirements and did not meet the requirements of creel positioning.
Based on the conclusions in Figure 10 and Table 1, the distance between the camera lens center and the center of the marker is an important factor to ensure the positioning accuracy of the creel. When using this monocular vision system to locate the creel, the best spacing range of image shooting is 190 mm~200 mm. Therefore, the positioning method proposed in this paper has a wide range of spacing, which reduces the difficulty of installation and debugging of the positioning system, and it is helpful to promote this method to practical production and application. In addition, the yarn frames in the production workshop of the circular weft knitting machine are in a free state, the impact and touch in the daily production process may cause the overall displacement of the yarn frames. If the distance between the camera lens center and the marker center is set to be 190 mm in the program, and the yarn frame displacement distance is less than 10 mm, it will not have a great impact on the automatic bobbin yarn changing, otherwise manual intervention will be taken, and a lot of labor resources will be occupied. Therefore, in the automatic bobbin yarn changing system, it is recommended to fix the bottom of the yarn frame with the ground.
In order to analyze the cause of the positioning error and further reduce the error, the distance between the camera lens center and the center of the marker is adjusted to 190 mm for testing. During the test, we take the end face of the marker as the coordinate plane, where x, y and r are the X-direction coordinates, Y-direction coordinates and the radius of the end face of the marker on the plane, respectively. Since it is impossible to precisely adjust the optical center of the camera and the center of the end face of the marker on the same straight line. First, we locate the camera position and assume that the center coordinates of the end face of the marker are (0, 0), and the radius r is 5 mm through measurement. Then, we use the improved Hough circle detection algorithm to identify the end face of the marker and calculate the coordinates and r corresponding to the center of the circle through the small hole imaging principle. The obtained data and their average values are shown in Table 2.
Table 2. This table shows the creel positioning data at 190 mm spacing.
It can be seen from the Table 2 that the deviation between the preliminary results of x, y and r, and the positioning results is very small. The maximum deviation in X direction is −0.310 mm, and the average deviation is −0.3036 mm. The maximum deviation in Y direction is 0.889 mm, and the average deviation is 0.6395 mm. The average value of radius r is 4.794 mm, and the average deviation is −0.206 mm. The reason for the deviation may be the inaccurate detection of the end circle, and the center of the circle is not on the optical axis, etc., but the deviation is very small, which will not affect the practical application. In order to ensure the reliability of long-term operation of the system, the creel position can be calibrated regularly according to the actual situation.
In order to further test the reliability of the positioning system, we fine tune the positions of the camera and the creel according to the results obtained in Table 2. In this way, the optical center of the camera and the center of the end face of the marker shall be in a straight line as far as possible. Then we test the X-direction offset of −20 mm, −10 mm, 0 mm, 10 mm, 20 mm at the distance of 190 mm. We conclude that the left offset is negative and the right offset is positive. The specific data are shown in Table 3. Since the Y direction is similar, it will not be described in detail here.
Table 3. This table shows the positioning results under different offset distances.
It can be seen from Table 3 that when the right offset is 20 mm, the maximum positioning error is 1.11 mm, but it is still less than the accuracy requirement of 2 mm. In addition, after the camera position is adjusted, the coordinate deviation of the center of the circle becomes smaller, but it has not been completely eliminated. However, it does not affect the actual use.
To sum up, it is completely feasible to use the end face of the marker as the positioning marker of the creel at a spacing of 190 mm, and the positioning error is very small. When in use, the deviation data is imported into the controller through the Ethernet bus to correct the deviation, and the bobbin automatic yarn change is completed by the bobbin yarn change manipulator according to the updated data movement. At the same time, the time used for image analysis, positioning processing and data communication of the creel was counted during the experiment. The time used for positioning was within 1 s, which could meet the time requirements of creel positioning calculation when the bobbin yarn changing manipulator was working.

6. Conclusions and Discussion

6.1. Conclusions

Aiming at the problem that the coordinate of the bobbin needs to be initialized regularly when the bobbin is automatically changed on the bobbin by the manipulator in the textile production workshop, this paper proposes a bobbin positioning technology based on monocular vision. Through the design of creel identification markers, the improved Hough circle detection algorithm was used for identification and positioning. This method reduces the difficulty of creel positioning, recognition and the requirements of processor configuration. At the same time, the positioning results and detection time can meet the requirements of use. Meanwhile, it has the advantages of low cost, easy installation and easy industrialization. So, it can provide a certain basis for the industrial application of automatic bobbin change in circular weft frames.

6.2. Limination

Since there are many rows of yarn frames in the production workshop of circular weft machine, the background is complex; however, we reduce the difficulty of identification of creels by using markers and realize the positioning of creels in principle. However, due to the small area of the end face of the creel, the vibration of the manipulator during operation has a great impact on the accuracy and stability of the identification. Therefore, it is necessary to shoot after the camera is really stopped. In addition, there are many yarn racks in the yarn area. The current identification method can only locate a single creel at one time. So, there is still a problem of low positioning efficiency for practical application.

6.3. Discussion

Future work will mainly improve the positioning efficiency and accuracy of the system from two aspects. On the one hand, starting with the performance of hardware equipment, we choose industrial computers with a higher resolution and faster processing capacity to overcome the problems of small end area of creel and high recognition accuracy. On the other hand, starting from the recognition and positioning algorithm, machine learning or deep learning is used to improve the accuracy and effect of positioning, realize the recognition of multiple creels at a single time, and promote the practical application of this technology in textile workshops.

Author Contributions

Conceptualization, J.T. and N.D.; Formal analysis, J.T. and L.S.; Funding acquisition, W.S.; Methodology, J.T. and S.H.; Software, J.T. and S.H.; Supervision, W.S.; Writing-original draft, J.T. and S.H.; Writing-review & editing, N.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key R&D Program of China, Grant/Award Number: 2017YFB1304000.

Institutional Review Board Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zou, K.; She, J.; Bai, Y.X. Design of Replacing Yarn System of Intelligent Creel Multi-Robot. Mach. Tool. Hydrau. 2015, 43, 60–63. [Google Scholar]
  2. Jiang, G.M.; Gao, Z.; Gao, Z.Y. Research Advance of Knitting Intelligent Manufacturing. J. Text. Res. 2017, 38, 177–183. [Google Scholar]
  3. Hu, X.D.; Shen, C.Y.; Peng, L.H.; Ru, X. Intelligent Manufacturing and Standard about Interoperability Verification of Knitting Machine. J. Text. Res. 2017, 38, 172–176. [Google Scholar]
  4. Mei, S.Q.; Hu, G.P.; Wang, J.W.; Chen, Z.; Xu, Q. Analysis of Some Key Technology Basis for Intelligent Textile Manufacturing and Its Equipment. J. Text. Res. 2017, 38, 166–171. [Google Scholar]
  5. Hu, X.D.; Shen, C.Y. Intelligent Manufacturing and Practice in Knitting Industry. Knit. Ind. 2019, 8, 1–6. [Google Scholar]
  6. Li, Q.M. Research on Positioning Accuracy of Threaded Hole Based on Monocular Vision. Modul. Mach. T. Aut. Manuf. Techn. 2021, 7, 20–23. [Google Scholar]
  7. Gao, H.; Zhou, Y.; Jiang, Y.; Yu, J.; Liu, J. Measurement of Simulated Lunar Soil information Using Rutting Images. IEEE Access. 2020, 8, 130281–130292. [Google Scholar] [CrossRef]
  8. Hao, R.J.; Wang, Z.Y.; Li, Y.R. Error Analysis Method for Monocular Vision Pose Measurement System. J. Appl. Opt. 2019, 40, 79–85. [Google Scholar]
  9. Zhang, H.; Wei, Y.; Chen, R. Research and development of automatic bobbin yarn changing robot system for warping machine’s bobbin yarn frame. Shanghai Text. Sci. Technol. 2020, 48, 10–13, 16. [Google Scholar]
  10. Zhan, D.H.; Wang, H.; Lin, J.; Zhong, H.Y.; Pan, Y.L.; Chen, X. Phase Laser Ranging System Based on apFFT for Manipulator Positioning. Las. Opt. Pro. 2022, 59, 1–12. [Google Scholar]
  11. Jin, X.J.; Jung, J.; Ko, S.Y.; Choi, E.; Park, J.O.; Kim, C.S. Geometric Parameter Calibration for a Cable-Driven Parallel Robot Based on a Single One-Dimensional Laser Distance Sensor Measurement and Experimental Modeling. Sensors 2018, 18, 2392. [Google Scholar] [CrossRef] [PubMed]
  12. Chen, S.H.; Xie, H.; Liu, Z.H.; Fang, Y.B.; Li, Z.H.; Chen, Z.H.; Yao, T. three-dimensional Grasping System of Manipulator Based on Triangle Laser. Electro. Measur. Technol. 2020, 43, 96–103. [Google Scholar]
  13. XU, H.; SUN, H.T. Research and Simulation of Laser radar Positioning Method for Indoor Mobile Robots. J. Mach. Des. 2022, 39, 90–94. [Google Scholar]
  14. Chen, W.; Yu, C.; Tu, C.; Tu, C.; Lyu, Z.; Tang, J.; Ou, S.; Fu, Y.; Xue, Z. A Survey on Hand Pose Estimation with Wearable Sensors and Computer-Vision-Based Methods. Sensors 2020, 20, 1074. [Google Scholar] [CrossRef]
  15. Wang, Y.; Shen, D.F.; Wang, R.J.; Li, Y.J.; Li, J.Y. Automatic Plaiting Vehicle Trajectory Tracking Control Algorithm. J. Silk 2022, 59, 65–73. [Google Scholar]
  16. Su, J.P.; Zhuang, P.L. AGV Technology and Application at Guangdong Esquel Group. Knit Ind. 2017, 11, 1–5. [Google Scholar]
  17. Chen, X.; Yuan, H.B.; Yao, W.L. Design of Cotton Bucket Handling System in Textile Workshop Based on AGV. Manufact. Atu. 2020, 42, 127–129. [Google Scholar]
  18. Duan, B.; Yu, L.Q.; Fan, F. Design and Research of Control System of Super Long AGV Cloth Falling Car. Automat. Instrum. 2020, 11, 79–84. [Google Scholar]
  19. Zhou, J.Q.; Shi, L.; Meng, F.Y. A Depth Measurement Algorithm of Monocular Vision System Based on Servomechanism. Mach. Des. Res. 2020, 36, 154–159. [Google Scholar]
  20. Luo, Z.; Cai, Z.L.; Jiang, W.S.; Yang, L. Endoscopic Attitude Measurement Method for Industrial Robot Based on Monocular Vision. Modul. Mach. T. Aut. Manuf. Techn. 2020, 11, 60–64. [Google Scholar]
  21. Xie, Y.S.; Wu, Q.C.; Chen, B.; Jiang, M. Grasping Operation Method of Mobile Manipulator Based on Monocular Vision. J. Mech Elect. Eng. 2019, 36, 71–76. [Google Scholar]
  22. Lai, J.W.; He, Y.Q.; Li, X.P.; Liu, M.Q.; Zhai, X.Y. Design of Manipulator Target Positioning System Based on Monocular Vision. Opt. Tech. 2019, 45, 6–11. [Google Scholar]
  23. Shi, Z.W.; Shi, W.M.; Wang, J.R. The Detection of Thread Roll’s Margin Based on Computer Vision. Sensors 2021, 21, 6331. [Google Scholar] [CrossRef] [PubMed]
  24. Wang, W.S.; Li, T.J.; Ran, Y.C.; Lu, Y.; Huang, M. Method for Position Detection of Cheese Yarn Rod. J. Text. Res. 2020, 41, 160–167. [Google Scholar]
  25. Zhang, J.X.; Li, Q. Online Cheese Package Yarn Density Detection System Based on Machine Vision. J. Text. Res. 2020, 41, 141–146. [Google Scholar]
  26. Jing, J.F.; Guo, G. Yarn packages hairiness detection based on machine vision. J. Text. Res. 2019, 40, 147–152. [Google Scholar]
  27. Wang, K.L.; Wu, Y.; Sha, C.Z.; Liu, S.; Wang, C.X.; Wang, Z.T. Research on Spatial Position Measurement Technology Based on Monocular Vision Theory. J. Astro. Metrol. Meas. 2019, 39, 16–21. [Google Scholar]
  28. Wang, J.; Feng, K.R.; Jiang, T.X.; Cheng, Q.C.; Li, W.T. Simple and Convenient Part Pose Measurement Method Based on Monocular Vision. Modul. Mach. T. Aut. Manuf. Techn. 2020, 11, 127–131. [Google Scholar]
  29. Wu, Q.; Shi, L.; Xie, X.Y.; Yue, L.J. An On-Line Calibration Method Between Laser and Camera in Autonomous Driving Vehicle Based on Visual Marker Board. Aut. Technol. 2020, 04, 40–44. [Google Scholar]
  30. Zhang, G.L.; Wu, Y.X.; Guo, W.P. Augmented Reality System Modeling Based on Simple Visual Marker. J. Chin. Com. Sys. 2014, 35, 2507–2510. [Google Scholar]
  31. Shen, X.P.; Peng, G.; Yuan, Z.Q. Insulator Location Method Based on Hough Transformation and RANSAC Algorithm. Electro. Meas. Technol. 2017, 40, 132–137. [Google Scholar]
  32. Liu, A.Y.; Xiong, G.L.; Yao, J.K.; Le, W.H. Marker Localization Method on Multiple Image Features Based on Hough Circle Detection. Compu. Simu. 2021, 38, 341–345. [Google Scholar]
  33. Zhao, J.; Hu, Y.; Tian, M. Pose Estimation of Excavator Manipulator Based on Monocular Vision Marker System. Sensors 2021, 21, 4478. [Google Scholar] [CrossRef]
  34. Yoneyama, R.; Duran, A.J.; del Pobil, A.P. Integrating Sensor Models in Deep Learning Boosts Performance: Application to Monocular Depth Estimation in Warehouse Automation. Sensors 2021, 21, 1437. [Google Scholar] [CrossRef] [PubMed]
  35. Shi, Y.Q.; Chang, C.X.; Liu, X.H.; Li, Z.Y.; Zhang, Z.H.; Gao, N.; Meng, Z.Z. Calibration Methods and Progress for Internal and External Parameters of Area-Array Camer. Laser Optoelectron. Prog. 2021, 58, 2400001. [Google Scholar]
  36. Xia, R.B.; Liu, W.J.; Zhao, J.B.; Xu, J.T. Fully Automatic Camera Calibration Method Based on Circular Markers. Chin. J. Sci. Ins. 2009, 30, 368–373. [Google Scholar]
  37. Liu, Y.; Li, T.F. Reaserch of the Improvement of Zhang’s Camera Calibration Method. Opt. Tech. 2014, 40, 565–570. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.