Next Article in Journal
The Performance Investigation of Smart Diagnosis for Bearings Using Mixed Chaotic Features with Fractional Order
Next Article in Special Issue
YOLO-Weld: A Modified YOLOv5-Based Weld Feature Detection Network for Extreme Weld Noise
Previous Article in Journal
Differential Evolution Optimization of Microwave Focused Hyperthermia Phased Array Excitation for Targeted Breast Cancer Heating
Previous Article in Special Issue
Just Noticeable Difference Model for Images with Color Sensitivity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Yarn-Tension Detection Using Machine Vision Combined with a Tension Observer

1
School of Control Science and Engineering, Tiangong University, Tianjin 300387, China
2
Tianjin Key Laboratory of Intelligent Control of Electrical Equipment, Tiangong University, Tianjin 300387, China
3
School of Electrical Engineering, Tiangong University, Tianjin 300387, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(8), 3800; https://doi.org/10.3390/s23083800
Submission received: 5 March 2023 / Revised: 24 March 2023 / Accepted: 4 April 2023 / Published: 7 April 2023
(This article belongs to the Special Issue Image/Signal Processing and Machine Vision in Sensing Applications)

Abstract

:
Machine vision can prevent additional stress on yarn caused by contact measurement, as well as the risk of hairiness and breakage. However, the speed of the machine vision system is limited by image processing, and the tension detection method based on the axially moving model does not take into account the disturbance on yarn caused by motor vibrations. Thus, an embedded system combining machine vision with a tension observer is proposed. The differential equation for the transverse dynamics of the string is established using Hamilton’s principle and then solved. A field-programmable gate array (FPGA) is used for image data acquisition, and the image processing algorithm is implemented using a multi-core digital signal processor (DSP). To obtain the yarn vibration frequency in the axially moving model, the brightest centreline grey value of the yarn image is put forward as a reference to determine the feature line. The calculated yarn tension value is then combined with the value obtained using the tension observer based on an adaptive weighted data fusion method in a programmable logic controller (PLC). The results show that the accuracy of the combined tension is improved compared with the original two non-contact methods of tension detection at a faster update rate. The system alleviates the problem of inadequate sampling rate using only machine vision methods and can be applied to future real-time control systems.

1. Introduction

Yarn tension during the spinning process is an important parameter that is highly correlated with yarn performance and fabric quality in production links [1]. Yarn tension during actual production is not constant, but fluctuates continuously during the spinning process. The abnormal fluctuations lead to changes in yarn properties, which have strict requirements for real-time and accurate tension detection. Through real-time tracking and control of yarn tension, the tension under each operating condition can be maintained within a constant range [2,3]. Therefore, the failures caused by aberrant tension, such as sticking and shutting down of textile machines, can be avoided, thus improving the automation level of the spinning industry and creating economic benefits [4].
Contact tension sensors are currently the typical devices used for measuring yarn tension in the textile industry. Designing sensors using surface acoustic wave (SAW) oscillators is a hot topic in the research on contact tension detection, and the main principle is that the oscillation frequency of the SAW oscillator varies with yarn tension [5,6,7]. Research on tension sensors has been conducted for many years, with good results on yarn-tension detection. However, this measurement method causes additional friction on the yarn, impacting yarn quality. In addition, in some situations, non-contact measurement is required for textile composite materials [8,9].
In many parts of the textile process, the winding of yarn is mostly carried out in an axial motion to achieve the corresponding workflow [10]. Dynamic analysis of axially moving string has received attention in many studies on the winding process in various engineering areas [11,12]. Yarn tension detection has been achieved using a laser doppler vibrometer (LDV) [13,14]. This method is novel and enables non-contact detection. Machine vision technology can determine characteristics of the tested products and is widely used in the textile industry [15,16,17]. On this basis, Wang et al. proposed an optical method for automatically measuring the microtension of yarns. The moving yarns in a winding system can be illuminated using a line laser, and a line scan industrial camera can be used to acquire images [18]. However, there are some defects which the mathematical model does not take into account the influence of other variables during motor operation. Peng et al. established a yarn-tension detection method based on transverse frequency [19]. Although non-contact detection can be achieved, the results are discontinuous and discrete. It is difficult to meet the demand for real-time tension detection in practical production. Besides, the methods only use computer software to process image data, which is not suitable for on-site control in the spinning industry. Thus, to achieve real-time detection and subsequent use in control systems, it is important to increase the tension sampling rate and complete the project using high-performance processors. Due to their strong data computing capabilities and flexible functions, digital signal processors (DSP) and field-programmable gate arrays (FPGA) have become the two most widely used processors in embedded image processing systems [20]. Herein, the tension estimated using the machine vision is implemented in a heterogeneous processing system based on the axially moving string model and image data processing methods.
In addition, PLCs are usually the main part of automatic industrial systems. They are very efficient and reliable in the manufacturing, chemical and process industries [21]. Tension observer can be realized in programmable logic controller (PLC) and has a high-tension observation rate. It has also been studied by many scholars to measure the tension of winding materials for sensorless control [22,23,24]. Hou et al. designed a decentralized coordinated control scheme with tension observers for a motor winding system. The tension observers were established to observe the winding tension [25]. Kang et al. designed a reduced dimensional observer for estimating tension over a wide frequency range [26] and the performance of the observer was analysed using only simulations. These researches have promoted the development of the tension observer model, but there are many factors that affect yarn winding tension under actual working conditions. Therefore, it will be combined with the tension estimated using the machine vision in a PLC to supplement the insufficient sampling rate of machine vision and improve the mathematical model.
Herein, a dynamic yarn-tension detection system using machine vision combined with a tension observer is made for high sampling rate and detection accuracy. The rest of this paper is organized as follows: Section 2 presents a mathematical model for the entire system and analysis. The image data processing method for machine vision tension detection is introduced in Section 3. The implementation process of tension-observer tension observation and data fusion is shown in Section 4. Section 5 contains the full experimental test platform description, while Section 6 discusses the test results. Finally, the conclusion is presented in Section 7.

2. Mathematical Models

2.1. Axially Moving String Model

The yarn moving between the conveying roller and the winding roller during the spanning process is equated to an axially moving string. The yarn is stretched by the rolls to produce vibrations, the frequency of which is correlated with the characteristics of the web and the tension [27]. The axially moving string model is shown in Figure 1 and enables us to obtain
E k = 1 2 0 l ρ v y x + y t 2 + v 2 d x
E p = 1 2 0 l F y x 2 d x
where Ek is the kinetic energy; Ep is the potential energy; ρ is the yarn density; v is the velocity of axial motion; l is the length stretched along the x-axis; x represents the coordinates along the yarn direction; F is the tension; y (x, t) represents the transverse movement of the yarn at the spatial coordinate x at time t; yx represents partial derivative with respect to x; and yt represents partial derivative with respect to t.
Defining the Lagrangian L = E k E p , Hamilton’s principle can be written as
δ t 1 t 2 L d t = 0
Letting c 2 = F ρ , the partial differential equation for the axially moving string is obtained [28], and the joint boundary conditions are written as
c 2 v 2 2 y x 2 2 v 2 y x t 2 y t 2 = 0 y 0 , t = 0 ,   y l , t = 0
where l is the distance between the conveying roller and the winding roller. A modal solution is assumed as
y x , t = R W x e i ω t
where W(x) is the eigenfunction, ω is the circular frequency, and R[·] denotes the real part. Substituting this solution in Equation (4), the eigenvalue problem for the motion string as
c 2 v 2 W + 2 i ω v W ω 2 W = 0
The eigenfunction W(x) is complex, substituting the solution form W(x) = Beikx in Equation (6), gives
c 2 v 2 k 2 2 ω v k ω 2 = 0 k 1 = ω c + v   or   k 2 = ω c v
Using Equation (7), the general solution can be written as
W x = D e i ω x c + v + E e i ω x c v
where D and E are arbitrary constants. Using the boundary conditions, the characteristic equation for the motion string can be written as
e i ω l 2 c / c 2 v 2 1 = 0
Thus, the eigenvalues for the axially moving string is given as
ω = n π c l c 2 v 2   n = 1 , 2 , ,
Using Equation (10), the relationship between string tension and the circular frequency is written as
c = F s ρ = l ω + l ω 2 + 4 ( π v ) 2 2 π
Since the process of adjusting the tension of the conveying rollers affects the moving yarn, the theoretical relationship between the yarn tension, speed and vibration frequency is subsequently modified as
F s = ρ 2 l 2 ω 2 + 2 l ω ( l ω ) 2 + 4 ( π v ) 2 4 π 2 + F f = ρ 2 l 2 f 2 + v 2 + 2 l f l 2 f 2 + v 2 + F f
where f = ω 2 π is the yarn vibration frequency; Fs is the yarn tension calculated by machine vision and Ff is the additional friction of the conveyor rollers. Its value is determined by fitting the functional relationship of yarn tension, vibration frequency and spinning speed. The key to calculating the yarn tension is to obtain the yarn vibration frequency.
The approximate general estimation model is obtained from the axially moving string model, and the yarn tension is calculated using the yarn’s transverse vibration frequency and the running linear velocity. Only the vibration frequency of the yarn and the speed are considered, without taking into account the disturbance formed by the speed of the motor during operation.

2.2. Design of the Tension Observer

The most basic unit of tension control is the two-motor tension system, which is a strongly coupled non-linear system, and the change in tension is associated with a variety of variables. The basic structure of the winding system is shown in Figure 2.
According to the principle of torque balance, the dynamic system equation is defined as
d d t J w ω w T w F R w T f w B ω w
where Jw is the rotational inertia of the winding roller; Rw is the radius of the winding roller; ωw is the angular velocity of the winding roller; Tw is the electromagnetic torque of the winding motor; Tfw is the coulomb friction; Rc is the radius of the conveyor roller; ωc is the angular velocity of the conveyor roller; Tc is the electromagnetic torque of the conveyor motor; and B is the coefficient of viscous friction. The winding motor is used to control the linear velocity of yarn movement, and the conveyor roller motor is employed to adjust the yarn tension.
Taking the tension, F, as a state variable, the tension generated during the same control cycle can be approximated as a constant value due to the high sampling frequency of the controller. The rate of change of the rotational inertia is much smaller than the rate of change of the angular velocity [29], and is defined as
d d t J w ω w J w d ω w d t T w F g R w T f w B ω w d F g d t = 0
where Fg is the yarn tension observed by the tension observer. The equation for the state for the winding motor is generated from Equation (14) as
x ˙ = A x + B u y = C x
where
A = B w J w R w J w 0   0       B = 1 J r 0       C = 1   0 x = ω w F g T       y = [ ω w ]       u = [ T w T f w ]
Using the angular velocity of the winding roller as a correction quantity for the observer, the tension observer is constructed as
x ^ ˙ = A x ^ + B u + K y y ^ y ^ = C x ^
where: x ^ is the observed state variable; y ^ is the output of the observer; and K is the feedback gain.
x ^ = ω ^ w F ^ g T       y ^ = [ ω ^ w ]       K = [ k 1 k 2 ] Τ
In order to obtain a fast, noise-resistant and relatively stable tension-observation system, a suitable feedback matrix, which is mainly based on the pole configuration, is required. The poles are first located on the left half of the s-plane to make the system stable, and then the poles are chosen to balance the response speed and noise immunity of the observer. Assuming that the eigenvalues of the desired observer are α and β, respectively, we obtain
k 1 = α + β B w J w k 2 = α β J w R w
The obtained yarn tension is then added to a first-order low-pass filter, L(s), to filter out external high-frequency noise disturbances such as motor vibrations. The filter is shown in Equation (18).
L s = 1 T o s + 1
where To is the time constant of the low-pass filter. Different values of To are chosen to correspond to the different filter cut-off frequencies [30]. A block diagram of the tension observer structure is shown in Figure 3.
The tension observer based on the motor-torque balance principle completes the estimation of the yarn tension using the motor electromagnetic torque and angular velocity, without taking into account the effect of the actual yarn oscillation on the observed tension.

2.3. Adaptive Weighted Fusion Model

Two mathematical models of indirect tension estimation sometimes do not take some variables into account. Therefore, the estimated tension values need to be fused with each other to improve measurement accuracy and fault tolerance. The adaptive weighted fusion algorithm is easy to implement in the PLC as the controller in most textile sites. The weight coefficient of each sensor can be adjusted in real-time based on changes in the variance of the tension values, such that the mean square error of the fusion system is kept to a minimum and optimal fusion effect is achieved [31]. The fusion model is shown in Figure 4.
In Figure 4, n is the rotational speed of the motor. F ^ is the tension value after data fusion. The result of the tension fusion is written as
F ^ = W s F s + W g F g
where W is the weight coefficient for each method, and
W s + W g = 1
Based on the theory of multivariate functions for extreme values, the total mean square error is minimized when the weight coefficient is as presented in Equation (21).
W s = 1 σ s 2 1 σ s 2 + 1 σ g 2
The data detected by each method are assumed to be X1, X2, …, Xn and the population variance is written as
σ s 2 = 1 n k = 1 n X k X ¯ 2         k = 1 , 2 , , n
Similarly, σ g 2 can be obtained and the mean value is written as
X ¯ = 1 n k = 1 n X k
Wg can be obtained as shown above. The weight coefficient changes as the variance in the tension values obtained using two detection methods change and current measurements are combined to achieve optimal fusion effect. Since the tension value observed using the tension observer is a supplement to the tension value estimated using machine vision, the initial value is more dependent on the tension value detected by the machine vision. Therefore, the initial value of Ws is determined as 1 and the initial value of Wg is determined as 0. The overall design concept of a dynamic yarn tension detection system is established by combining the vibration model with a tension observer. The system is divided into three parts: axially moving string model to estimate tension, tension observer to observe tension and adaptive-weighted tension fusion. The yarn tension detected by machine vision is combined with the tension obtained by the tension observer to achieve the most stable, real-time results possible.

3. Machine Vision Tension Detection

3.1. Image Acquisition

The winding roller motor runs and the yarn travels at a set velocity through the conveyor rollers. A line scan camera is used to quickly obtain images of the moving yarn. It is installed at a suitable location to obtain images of the yarn at a point of lateral vibration, as shown in Figure 5.
The objective is to capture information on the vibrating yarn image. A large field depth is not required to highlight the subject, but a suitable field of view needs to be selected in order to obtain sufficient yarn vibration. To make the yarn outline clear and reduce interference, the stripe light source—with adjustable brightness—is enabled. After selecting the appropriate object distance, the depth of the field and the field of view are obtained by adjusting the focal length and aperture. The formula for calculating the field of view is
μ N f o = l o d
where μ is the image element size; N is the number of pixels occupied by the object image in a single direction; fo is the camera focal length; lo is the actual field of view in the direction of vibration; and d is the object distance [32]. Based on Equation (24), the camera is mounted at 430 mm, the same level as the yarn, and the line frequency of the line scan camera is set to 10 kHz. The flexible parallel structure of the FPGA is used as the motion yarn-image acquisition unit and pre-processing unit. DSP and FPGA are connected to each other using the SRIO protocol to achieve high-speed connection of the heterogeneous processors. The FPGA quickly transmits the collected and processed yarn greyscale image data to the DSP, resulting in real-time signal processing.

3.2. Yarn Image Processing

The images captured by the line scan camera can be seen as discrete time series. Each frame represents a different point in time. Since the line scan camera is set to capture images at a line frequency of 10 kHz, the time for each image line is determined to be 0.1 ms. To have a suitable range of tension measurements, a set of yarn vibration peaks and troughs needs to be obtained. The line scan camera is chosen to give a tension solution at every full 512 × 512 resolution capture, which is also a sampling period of 0.0512 s. Acquisition of the yarn images and image processing occur simultaneously, i.e., the DSP performs image processing while the line scan industrial camera is still capturing images. This does not affect the speed of detection but requires image processing to occur at a faster rate than image acquisition.

3.2.1. Overall Image Processing Process

Yarn image processing is based on a multi-core DSP processor with embedded real-time cores and parallel image processing, making it capable of meeting the performance requirements of real-time data processing in many industrial fields. The overall flow of yarn image processing is shown in Figure 6.
TMS320C6678 is a high-performance floating point embedded 8-core digital signal processor. Yarn greyscale image data from the FPGA is received by the multi-core DSP processor as a Serial Rapid I/O (SRIO) target. The main core, Core0, responds to hardware interruption, selects a threshold for image segmentation and divides the image equally. The segmented yarn image data is subsequently sent to the DSP’s slave cores via the Inter-Process Communication Message Queue (IPC MQ). The slave cores simultaneously respond to image information and process the images, which reduces the time required to run the single-core-only program. The feature lines of the vibrating yarn are extracted and smoothed based on the processed image data. Finally, the vibration frequency is calculated. The overall design idea for the program is to minimize the running time of the program on the premise that yarn vibration frequency can be calculated.

3.2.2. Image Segmentation Threshold Selection

The regions of interest are obtained conventionally from banded images by scanning the image line by line and determining the upper and lower boundary points or the brightest point [33]. This method increases the algorithm running time as the image resolution increases and the region of interest decreases. The greyscale values of row pixels of the yarn image are shown in Figure 7.
If the upper and lower boundary points or the brightest point are directly selected as thresholds during image processing, the results will be unfavourable for the subsequent calculation of vibration frequency. The more stable the grey value is, the closer it is to the centre point of the rise or fall, thus the brightest point’s grey value, as the reference, is divided in half and used as the threshold value for image segmentation. This method only needs to identify the threshold value of the first image row, hence not only excludes the uncertainty in finding the boundary point or the brightest point’s grayscale value, but also reduces the time required to find the brightest point in each row in cases where there are a lot of operations.
The program is implemented in the DSP, thus the image processing results are viewed with the help of the YUV Player software. The image data is exported directly from the address memory by setting a breakpoint in the program. The image processing results are shown in Figure 8.
The feature-line fast-scanning algorithm is designed to avoid wasting a lot of time scanning from the beginning of each row. First, the position with a grey value of 255 pixels is determined by searching for the first line of the yarn image. The starting value for the subsequent line scans is then decided by floating a fixed value based on the position of the previous moment. And the next boundary will be searched as long as find the brightest grey value. Ultimately, the data that characterise the information in the moving yarn image are obtained. The schematic of the algorithm is shown in Figure 9.

3.2.3. One-Way Extreme Value Search Algorithm

Since the uneven density of the transversely vibrating yarns and micro-angle changes during light source irradiation, the signal noises are formed and the change-point detection results of the yarn image are affected by the noises [34]. The directly extracted feature line is shown in Figure 10a. The figure shows deviations in some points of the extracted feature line. In order to facilitate the determination of wave peaks and valleys and reduce errors, a smoothing filtering algorithm is used to filter out signal noise interference.
The filtered feature line in Figure 10b shows that the extreme values cannot be taken directly as peaks and troughs. Therefore, the height of the yarn vibration is gradually scanned at each moment, and the moments in the middle of the poles are used as the peaks or troughs. To avoid directly determining the position of the starting part as a peak or valley, determination of the peak and valley positions commences when the scanning height changes. Before searching for the poles, it is necessary to confirm whether the peaks and valleys have been determined. When the current height in the wave peak determination process is higher than that of the previous moment, it is initially determined as the peak, and the count is started. If the height changes again to a lower value, the position of the peak is determined based on the count value. Otherwise, the count is cleared and the peak needs to be determined again. The trough is determined in a similar manner. A one-way extreme-value scanning algorithm is designed as shown in Figure 11.
The yarn vibration frequency is calculated directly in half a cycle after scanning the first set of peaks and troughs, based on the line frequency set by the line scan camera, which communicates with the PLC through the serial port. If the estimated frequency exceeds the measurable range, the value obtained is judged to be incorrect and immediately abandoned. Finally, the yarn tension is estimated according to the approximate fitting formula.

4. Tension-Observer Tension Observation and Data Fusion

The rotational inertia of the conveyor-roller motor is not easy to determine, thus a coordinated control system of the winding tension and speed is constructed with the winding-roller motor as the host and the conveyor-roller motor as the slave. The winding-roller motor is controlled by the high-speed pulse signal from the PLC to the servo drive; at the same time, the conveyor-roller motor works in a speed-control mode, where the maximum speed is constrained by the host speed [35]. Through Modbus polling communication, the electromagnetic torque and angular velocity information of the motor are obtained from the servo drive in real time. The tension observer algorithm is then run inside the controller to facilitate observation of the yarn tension. The internal PLC program implementation is shown in Figure 12.
During algorithm fusion, the insufficient sampling rate of machine vision detection makes it difficult to achieve effective adaptive weighted fusion with the tension observer’s observed tension, hence a moving average prediction algorithm is employed to infer missing data due to the machine vision sampling rate based on historical data [36]. A flag bit is set to determine whether the yarn vibration frequency calculated by the machine vision section is transmitted to the PLC to switch the tension value.
The adaptive weighted fusion algorithm improves the measurement accuracy of the multi-sensor data fusion by combining the tension detected using machine vision with the tension data observed using the tension observer; however, the effect of the delay in measured tension values, from acquisition to calculation, also needs to be considered in order to achieve the best fusion results.

5. Experimental Test Platform

An experimental laboratory platform was created to evaluate the performance of the proposed detection system. The system can be divided into three parts: machine vision detection, tension-observer observation and yarn tension control. The implementation of the tension observer is correlated with the control part, which mainly includes a human–machine interface (HMI), programmable logic controller (PLC), servo drive and servo motor, to realise the tension observer and coordinated control of the tension and speed in the spinning process of the fabric-based coils. The machine vision part includes a line scan industrial camera, FPGA and multi-core DSP to complete tension detection based on the yarn vibration image data. A simple winding-control system is built by simulating the spinning process. The yarn-tension detection platform is shown in Figure 13.
The experimental system includes the control part of the HMI device (Smart 700 IE V3), a Siemens PLC (S7-200smart) with two Siemens servo drives and motors and winding materials. A line scan camera (Hikvision MV-CL042-91CM), a stripe light source and a lens (MVL-LF5040M-F) are selected for image acquisition. The image processing part constitutes an FPGA chip (Xilinx Kintex-7) and a floating-point DSP chip (TMS320C6678). In addition, the tension sensor (ZHZL-HL), which has an accuracy of 0.5%, is installed in the experimental platform for comparison. The specific parameters of the experimental system are: line scan industrial camera resolution, 512 × 2; line frequency, 10 kHz; lens aperture, F4.0; servo motor power, p = 0.4 kW; rated speed, n = 3000 r/min; rotational inertia, J = 3.51 × 10−5 kg·m2; torque, 1.27 Nm, where the winding material is Kevlar; yarn linear density, 230.769 tex; and yarn diameter, 0.6 mm.

6. Presentation and Discussion of Test Results

To verify the accuracy of the system’s results, the yarn tension measured by the contact-tension sensor was compared with the tension detected by the non-contact machine vision, the tension observed by the tension observer and the final achieved tension after data fusion. The experiments were carried out at an ambient temperature of 25 °C and a relative humidity of 40%.
Figure 14 shows comparisons between estimated tension based on the machine vision and measured tension using the tension sensor when the tension sampling period is 0.104 s, the pre-set tension is 20 cN and the winding-roll motor is set to different speeds. The relative error between the tension estimated using machine vision and the tension measured using a contact sensor is shown in Table 1. It can be seen that estimated tension based on machine vision fluctuates around the pre-set tension over time and agrees with the tension obtained using the contact tension detector at approximately 10 tension values per second of output; the relative error is generally less than 14%. At the same time, tension detection can also be achieved by changing the motor speed to 150 r/min, as shown in Figure 14b. Discrepancies in the tension detection results of the two methods may be due to the fact that the machine vision method is based on the vibration frequency of the yarn and does not take into account the disturbance formed on the yarn during the operation of the motor. In the theoretical model of machine vision tension detection, the yarn density is considered to be constant under ideal conditions, but the linear density may change due to repeated stretching of the yarn during the experiment, leading to errors. Measured tension values may also be affected by smooth filtering on raw yarn data in the image data processing.
When the tension-detection system needs to output 20 tension values per second, a sampling period of 0.05 s is set. Figure 15 shows the comparisons between estimated tension based on the machine vision and the tension measured using tension sensor when the tension sampling period is 0.05 s and the pre-set tension is 20 cN. The tension of the yarn can still be detected when the sampling rate is faster and the motor speed is 100 r/min, as shown in Figure 15a. However, the processing speed provided by the vision solution within a single sampling time is not sufficient to meet the solution requirements. The tension values estimated using machine vision are often discontinuous, resulting in some tension values not being sampled on time, hence affecting the effectiveness of tension detection. The same procedure is repeated for the yarn when the motor speed is 150 r/min, and the result comparisons are presented in Figure 15b. The detection results are in agreement with the preceding conclusions.
The tension values observed using the tension observer and the tension values measured using the contact tension detector when the motor speed is 100 r/min, the pre-set tension is 20 cN and the tension sampling period is 0.05 s are shown in Figure 16. Table 2 presents the relative error between the tension observed using the tension observer and the tension measured using a contact sensor. The tension observer also fluctuates around the pre-set tension over time, and the sampling rate is not limited. However, the relative error is generally less than 17%, and may be due to the fact that the actual yarn oscillations are not taken into account during the continuous winding of the yarn. Moreover, it may also be impacted by the accuracy of the measurement of motor parameters and the variation in rotational inertia.
Figure 17 shows the comparison between estimated tension based on the machine vision, the tension values observed using the tension observer, the fusion tension and the tension measured using contact tension sensors when the tension sampling period is 0.05 s and the pre-set tension is 20 cN. Relative errors of data fusion tension detection are shown in Table 3. The combined tension values correlate with the tension values measured using the contact tension sensor, with the error generally within 10%. Under the condition that the sampling rate increases, the missing tension values generated using machine vision methods are predicted using the moving average prediction algorithm. The adaptive weighted fusion algorithm solves the sudden change situation where the two individual detection methods are affected and generate large errors, making the detected tension value relatively stable. The differences in the tension test results are mainly caused by the accuracy of measurement of motor parameters, the effect of rotational inertia change and the method used for smooth filtering raw yarn data during image data processing. In addition, the yarn density changes during the multiple experimental stretches. The additional friction in the tension sensor also has unavoidable effects on the tension calculation results. The same procedure is repeated with all conditions except sampling time kept constant, as shown in Figure 18. The comparisons between the results are basically consistent with those of the analysis above.
Finally, the contact tension sensor is replaced with the tension value detected using data fusion to represent the feedback value of the controller to control the conveyor roller motor and check the yarn tension control effect when the motor speed is 100 r/min, the pre-set tension is 20 cN and the tension sampling period is 0.05 s, as shown in Figure 19. The yarn tension value basically fluctuates around the pre-set tension value over time, indicating that the designed detection system can be used for tension control. However, it can be seen that the yarn tension value deviates from the pre-set tension value within 2–3 s. More research is needed to improve the detection system.

7. Conclusions

In summary, a dynamic yarn-tension detection system was developed by combining machine vision with a tension observer. Machine vision is employed to capture and process the images of vibrating yarns, and the calculated yarn tension value is then combined with the value obtained from the tension observer, using an adaptive weighted data fusion method to improve the sampling rate and detection accuracy to achieve the non-contact measurement of dynamic yarns. The performance of the system was investigated using the simulated spinning experiment platform, which showed that the proposed method is effective. Some issues remain for further research, in particular the smaller range of the tension sensors with higher accuracy, the accuracy of measuring motor parameters, the effect of rotational inertia change, and the method for finding peaks and valleys, which should improve the system. At the same time, the actual application results should be highlighted in the spinning production site, and the error rate of the system should be verified and corrective measures proposed. With continuous improvements in the various motor control algorithms, hardware design and signal processing scheme, the accuracy of the system can be improved further, and is expected to be used in more industrial applications.

Author Contributions

Conceptualization, Y.J. and J.M.; methodology, Y.J. and J.M.; software, Y.J. and J.M.; validation, Y.J. and J.M.; formal analysis, Y.J. and Z.Z.; investigation, Y.J. and J.M.; resources, Y.J., L.S. and Z.Z.; data curation, Y.J. and J.M.; writing—original draft preparation, Y.J. and J.M.; writing—review and editing, Y.J., J.L. and J.M.; supervision, Y.J., Z.Z., J.L. and L.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science & Technology Development Fund of Tianjin Education Commission for Higher Education (Grant No. 2019KJ021).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available from the corresponding author (J.Y.) upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhao, B.; Cong, H.; Wu, G. Construction and system realization of the yarn tension model of fully fashioned flat knitting fabric. Text. Res. J. 2021, 91, 1380–1388. [Google Scholar] [CrossRef]
  2. Tang, Z.-X.; Wang, X.; Fraser, W.B.; Wang, L. Simulation and experimental validation of a ring spinning process. Simul. Model. Pract. Theory 2006, 14, 809–816. [Google Scholar] [CrossRef]
  3. Liu, X.; Miao, X. Analysis of yarn tension based on yarn demand variation on a tricot knitting machine. Text. Res. J. 2017, 87, 487–497. [Google Scholar] [CrossRef]
  4. Ali, M.; Ahmed, R.; Amer, M. Yarn tension control technique for improving polyester soft winding process. Sci. Rep. 2021, 11, 1060. [Google Scholar] [CrossRef]
  5. Lei, B.; Lu, W.; Zhu, C.; Liu, Q.; Zhang, H. A novel optimal sensitivity design scheme for yarn tension sensor using surface acoustic wave device. Ultrasonics 2014, 54, 1649–1655. [Google Scholar] [CrossRef]
  6. Ding, Y.; Lu, W.; Zhang, Y.; Feng, Y.; Zhou, Y. Study on the Practical Application of Surface Acoustic Wave Yarn Tension Sensor. IEEE Trans. Ind. Electron. 2021, 69, 13781–13790. [Google Scholar] [CrossRef]
  7. Peng, C.; Yuan, X. Enhanced Frequency Stability of SAW Yarn Tension Sensor by Using the Dual Differential Channel Surface Acoustic Wave Oscillator. Sensors 2023, 23, 464. [Google Scholar]
  8. Bessette, C.; Decrette, M.; Tourlonias, M.; Osselin, J.-F.; Charleux, F.; Coupé, D.; Bueno, M.-A. In-situ measurement of tension and contact forces for weaving process monitoring: Application to 3D interlock. Compos. Part A Appl. Sci. Manuf. 2019, 126, 105604. [Google Scholar] [CrossRef]
  9. Xiang, H.; Jiang, Y.; Qi, Y.; Li, J. Process-Induced Distortions Characterization of MBWK Fabric Reinforced Composite Helmet Shell. Materials 2020, 13, 2983. [Google Scholar] [CrossRef]
  10. Gauri, S.K. Improving yarn quality through multi-response optimisation of electronic yarn clearer in winding machine. Int. J. Adv. Manuf. Technol. 2011, 56, 387–396. [Google Scholar] [CrossRef]
  11. Ma, L. A highly precise frequency-based method for estimating the tension of an inclined cable with unknown boundary conditions. J. Sound Vib. 2017, 409, 65–80. [Google Scholar] [CrossRef]
  12. Bortnowski, P.; Gładysiewicz, L.; Król, R.; Ozdoba, M. Models of Transverse Vibration in Conveyor Belt—Investigation and Analysis. Energies 2021, 14, 4153. [Google Scholar] [CrossRef]
  13. Zhang, D.; Ma, Q.; Tan, Y.; Liao, H.; Lu, C.; Tang, F.; Liu, X.; Fu, Y.; Wang, X.; Gan, X. Non-contact detection of polyester filament yarn tension in the spinning process by the laser Doppler vibrometer method. Text. Res. J. 2022, 92, 919–928. [Google Scholar] [CrossRef]
  14. Zhang, D.; Tan, Y.; Ma, Q.; Liao, H.; Lu, C.; Liu, X.; Fu, Y.; Wang, X.; Han, F.; Gan, X. Real-time tension estimation in the spinning process based on the natural frequencies extraction of the Polyester Filament Yarn. Measurement 2022, 188, 110514. [Google Scholar] [CrossRef]
  15. Haleem, N.; Bustreo, M.; Del Bue, A. A computer vision based online quality control system for textile yarns. Comput. Ind. 2021, 133, 103550. [Google Scholar] [CrossRef]
  16. Kuo, C.-F.J.; Wang, W.-R.; Barman, J. Automated Optical Inspection for Defect Identification and Classification in Actual Woven Fabric Production Lines. Sensors 2022, 22, 7246. [Google Scholar] [CrossRef]
  17. Fang, B.; Long, X.; Sun, F.; Liu, H.; Zhang, S.; Fang, C. Tactile-Based Fabric Defect Detection Using Convolutional Neural Network with Attention Mechanism. IEEE Trans. Instrum. Meas. 2022, 71, 1–9. [Google Scholar] [CrossRef]
  18. Wang, Q.; Lu, C.; Huang, R.; Pan, W.; Li, X. Computer vision for yarn microtension measurement. Appl. Opt. 2016, 55, 2393–2398. [Google Scholar] [CrossRef]
  19. Peng, L.; Li, Y.; Zheng, Q.; Hu, X. Measurement of axial motion yarn tension based on transverse vibration frequency. Text. Res. J. 2023, 93, 1019–1029. [Google Scholar] [CrossRef]
  20. Qi, B.; Shi, H.; Zhuang, Y.; Chen, H.; Chen, L. On-Board, Real-Time Preprocessing System for Optical Remote-Sensing Imagery. Sensors 2018, 18, 1328. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Alphonsus, E.R.; Abdullah, M.O. A review on the applications of programmable logic controllers (PLCs). Renew. Sustain. Energy Rev. 2016, 60, 1185–1205. [Google Scholar] [CrossRef]
  22. Hong, Q.; Shi, Y.; Chen, Z. Dynamics Modeling and Tension Control of Composites Winding System Based on ASMC. IEEE Access 2020, 8, 102795–102810. [Google Scholar] [CrossRef]
  23. Lu, J.-S.; Cheng, M.-Y.; Su, K.-H.; Tsai, M.-C. Wire tension control of an automatic motor winding machine—An iterative learning sliding mode control approach. Robot. Comput. Manuf. 2018, 50, 50–62. [Google Scholar] [CrossRef]
  24. Dai, K.; Zhu, Z.; Shen, G.; Li, X.; Tang, Y.; Wang, W. Modeling and Adaptive Tension Control of Chain Transmission System with Variable Stiffness and Random Load. IEEE Trans. Ind. Electron. 2022, 69, 8335–8345. [Google Scholar] [CrossRef]
  25. Hou, H.; Nian, X.; Chen, J.; Xiao, D. Decentralized coordinated control of elastic web winding systems without tension sensor. ISA Trans. 2018, 80, 350–359. [Google Scholar] [CrossRef]
  26. Kang, H.; Shin, K.-H. Precise tension control of a dancer with a reduced-order observer for roll-to-roll manufacturing systems. Mech. Mach. Theory 2018, 122, 75–85. [Google Scholar] [CrossRef]
  27. Zhao, Z.; Liu, Y.; Guo, F.; Fu, Y. Vibration control and boundary tension constraint of an axially moving string system. Nonlinear Dyn. 2017, 89, 2431–2440. [Google Scholar] [CrossRef]
  28. Hagedirn, P.; Dasgupta, A. Vibrations and Waves in Continuous Mechanical Systems, 3rd ed.; John Wiley & Sons: New York, NY, USA, 2008; pp. 243–288. [Google Scholar]
  29. Yang, Z.; Ding, Q.; Sun, X.; Zhu, H.; Lu, C. Fractional-order sliding mode control for a bearingless induction motor based on improved load torque observer. J. Frankl. Inst. 2021, 358, 3701–3725. [Google Scholar] [CrossRef]
  30. Raphika, P.M.; Abdulla, P.; Jasmine, P.M. Compact lowpass filter with a sharp roll-off using patch resonators. Microw. Opt. Technol. Lett. 2014, 56, 2534–2536. [Google Scholar] [CrossRef]
  31. Zhang, L.; Liu, C. Robust Low-Rank Analysis with Adaptive Weighted Tensor for Image Denoising. Displays 2022, 73, 102200. [Google Scholar] [CrossRef]
  32. Gonçalves, N.; Carvalho, V.; Belsley, M.; Vasconcelos, R.M.; Soares, F.O.; Machado, J. Yarn features extraction using image processing and computer vision—A study with cotton and polyester yarns. Measurement 2015, 68, 1–15. [Google Scholar] [CrossRef]
  33. Li, W.-B.; Lu, C.-H.; Zhang, J.-C. A local annular contrast based real-time inspection algorithm for steel bar surface defects. Appl. Surf. Sci. 2012, 258, 6080–6086. [Google Scholar] [CrossRef]
  34. Ma, L.; Chen, J.; Tang, W.; Yin, Z. Vibration-based estimation of tension for an axially travelling web in roll-to-roll manufacturing. Meas. Sci. Technol. 2017, 29, 015102. [Google Scholar] [CrossRef] [Green Version]
  35. Wang, Q.; Li, A.; Li, Y.; Liu, J.; Shen, H.; Li, G. Microtension control for a yarn winding system with an IMC PID controller. Mech. Ind. 2019, 20, 609. [Google Scholar] [CrossRef] [Green Version]
  36. Minor, B.D.; Doppa, J.R.; Cook, D.J. Learning Activity Predictors from Sensor Data: Algorithms, Evaluation, and Applications. IEEE Trans. Knowl. Data Eng. 2017, 29, 2744–2757. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Axially moving string model.
Figure 1. Axially moving string model.
Sensors 23 03800 g001
Figure 2. Diagram of the winding system structure.
Figure 2. Diagram of the winding system structure.
Sensors 23 03800 g002
Figure 3. Diagram of the tension observer principle block.
Figure 3. Diagram of the tension observer principle block.
Sensors 23 03800 g003
Figure 4. Adaptive weighted fusion algorithm structure.
Figure 4. Adaptive weighted fusion algorithm structure.
Sensors 23 03800 g004
Figure 5. Mechanical structure of the detection system.
Figure 5. Mechanical structure of the detection system.
Sensors 23 03800 g005
Figure 6. DSP multi-core image processing flow chart.
Figure 6. DSP multi-core image processing flow chart.
Sensors 23 03800 g006
Figure 7. Yarn image row pixel greyscale values.
Figure 7. Yarn image row pixel greyscale values.
Sensors 23 03800 g007
Figure 8. Yarn image processing effect. (a) Original image. (b) Threshold image segmentation.
Figure 8. Yarn image processing effect. (a) Original image. (b) Threshold image segmentation.
Sensors 23 03800 g008
Figure 9. Principle diagram of feature-line fast-scanning algorithm.
Figure 9. Principle diagram of feature-line fast-scanning algorithm.
Sensors 23 03800 g009
Figure 10. Yarn feature line results. (a) Before filtering. (b) After filtering.
Figure 10. Yarn feature line results. (a) Before filtering. (b) After filtering.
Sensors 23 03800 g010
Figure 11. One-way extreme-value search algorithm.
Figure 11. One-way extreme-value search algorithm.
Sensors 23 03800 g011
Figure 12. Flowchart of the PLC internal program.
Figure 12. Flowchart of the PLC internal program.
Sensors 23 03800 g012
Figure 13. Construction of the experimental platform.
Figure 13. Construction of the experimental platform.
Sensors 23 03800 g013
Figure 14. The sampling period is 0.104 s. (a) Motor speed is 100 r/min. (b) Motor speed is 150 r/min.
Figure 14. The sampling period is 0.104 s. (a) Motor speed is 100 r/min. (b) Motor speed is 150 r/min.
Sensors 23 03800 g014
Figure 15. The sampling period is 0.05 s. (a) Motor speed is 100 r/min. (b) Motor speed is 150 r/min.
Figure 15. The sampling period is 0.05 s. (a) Motor speed is 100 r/min. (b) Motor speed is 150 r/min.
Sensors 23 03800 g015
Figure 16. Comparisons between tension-observer tension observations.
Figure 16. Comparisons between tension-observer tension observations.
Sensors 23 03800 g016
Figure 17. Experimental comparisons of adaptive-weighted tension fusion.
Figure 17. Experimental comparisons of adaptive-weighted tension fusion.
Sensors 23 03800 g017
Figure 18. Data fusion experiments. (a) The sampling period is 0.06 s. (b) The sampling period is 0.07 s.
Figure 18. Data fusion experiments. (a) The sampling period is 0.06 s. (b) The sampling period is 0.07 s.
Sensors 23 03800 g018
Figure 19. Tension control experiment using data fusion method.
Figure 19. Tension control experiment using data fusion method.
Sensors 23 03800 g019
Table 1. Relative error in machine vision detection.
Table 1. Relative error in machine vision detection.
Tension Measured Using Tension Sensor (cN)Tension Estimated Using Machine Vision (cN)Relative Error (%)
19.5720.947.00
21.0918.79−10.91
19.7120.122.08
19.8421.558.62
19.6522.3013.49
19.0020.306.84
21.1319.62−7.15
19.2420.134.63
20.4018.10−11.28
Table 2. Relative error of tension-observer observations.
Table 2. Relative error of tension-observer observations.
Tension Measured Using Tension Sensor (cN)Tension Obtained Using the Observer (cN)Relative Error (%)
27.4323.95−12.69
21.1224.1414.30
20.1920.521.63
20.1222.6512.45
17.7320.4015.06
18.7418.10−3.42
18.8722.0716.90
23.8926.139.38
24.0720.30−5.66
19.2921.8713.37
Table 3. Relative error of data fusion yarn tension detection.
Table 3. Relative error of data fusion yarn tension detection.
Tension Measured Using Tension Sensor (cN)Tension Obtained Using Data Fusion Method (cN)Relative Error (%)
20.4221.625.88
18.7818.44−1.81
22.1221.24−3.98
18.4618.43−0.16
17.7418.514.34
21.3923.349.12
17.8516.68−6.55
20.5420.22−1.56
19.5819.55−0.15
20.1421.707.75
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ji, Y.; Ma, J.; Zhou, Z.; Li, J.; Song, L. Dynamic Yarn-Tension Detection Using Machine Vision Combined with a Tension Observer. Sensors 2023, 23, 3800. https://doi.org/10.3390/s23083800

AMA Style

Ji Y, Ma J, Zhou Z, Li J, Song L. Dynamic Yarn-Tension Detection Using Machine Vision Combined with a Tension Observer. Sensors. 2023; 23(8):3800. https://doi.org/10.3390/s23083800

Chicago/Turabian Style

Ji, Yue, Jiedong Ma, Zhanqing Zhou, Jinyi Li, and Limei Song. 2023. "Dynamic Yarn-Tension Detection Using Machine Vision Combined with a Tension Observer" Sensors 23, no. 8: 3800. https://doi.org/10.3390/s23083800

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop