Next Article in Journal
Experimental Investigation on the Transpiration Cooling Characteristics of Sintered Wire Mesh in Plain Weave
Next Article in Special Issue
Adaptive Sliding Mode Disturbance Observer and Deep Reinforcement Learning Based Motion Control for Micropositioners
Previous Article in Journal
Coriolis Vibratory MEMS Gyro Drive Axis Control with Proxy-Based Sliding Mode Controller
Previous Article in Special Issue
Study on the Mechanical Properties of Monocrystalline Germanium Crystal Planes Based on Molecular Dynamics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Vision-Based Method for Measuring and Controlling the Angle of Conductive Slip Ring Brushes

1
Ministry of Education Key Laboratory for Cross-Scale Micro and Nano Manufacturing, Changchun University of Science and Technology, Changchun 130022, China
2
Chongqing Research Institute, Changchun University of Science and Technology, Chongqing 401135, China
*
Author to whom correspondence should be addressed.
Micromachines 2022, 13(3), 447; https://doi.org/10.3390/mi13030447
Submission received: 20 February 2022 / Revised: 13 March 2022 / Accepted: 14 March 2022 / Published: 16 March 2022
(This article belongs to the Special Issue Ocean MEMS and Related Technology)

Abstract

:
The conductive slip ring is used for power or signal transmission between two objects rotating relative to each other. It has become an essential part of modern industrial development. In traditional automated production measurements, the typical method is to use calipers, goniometers, or angle gauges to measure a parameter of the workpiece several times and then average it. These inspection means have low measurement accuracy and slow measurement speed, and measurement data cannot be processed in a timely manner. A machine vision-based method for measuring and controlling the angle of the brushes is proposed for this problem. First, the brush angle forming device was built for the conductive slip ring brush wire, forming the principle and rebound characteristics. Then, machine vision and image processing algorithms were applied to measure the key parts of the conductive slip ring brushes. The data of the forming angle value and rebound angle value were obtained during the forming process of the brush wire angle. Finally, a pre-compensation model for the brush filament rebound was developed and validated based on the curve fitting method. The test results show that the error of the angle measurement is within 0.05°. The average error of the measured rebound angle and the calculated rebound angle of the brush filament pre-compensation model was 0.112°, which verifies the correctness of the pre-compensation model. The forming angle can be controlled more precisely, and the contact performance between the brush wire and the ring body can be improved effectively. This method has the potential to be extended to engineering applications.

1. Introduction

Conductive slip rings are key components of various precision rotary tables, centrifuges, and inertial guidance equipment and have been used in sophisticated military applications [1]. In the context of today’s Industry 4.0, there are higher requirements for the signal transmission rate and rotation speed of conductive slip rings. Technical indicators are gradually approaching the direction of zero wear and zero maintenance [2]. Therefore, scholars have conducted in-depth research on the contact form [3,4,5], material wear [6], and performance reliability of conductive slip rings [7]. The position of the brush wire in contact with the slip ring directly affects the amount of pressure in contact with the slip ring. When the contact pressure is too small, the resistance of the contact becomes large, the temperature rise is large, and the loss also increases, which in turn affects the current as well as the signal transmission. If the pressure is too large, it will cause the brush to increase the amount of wear and tear, resulting in a melt welding phenomenon that affects the service life of the conductive slip ring [8]. As the brush wire is mounted on the brush plate, when measuring the brush wire by mechanical contact, it affects the form and surface quality of the brush wire, and the uneven surface on the brush plate will bring some error to the measurement result. Therefore, it is necessary to find an optimal angle measurement method.
Many scholars have conducted a series of studies and analyses on machine vision-based measurement technology because of its non-contact, high stability, high speed, and low-cost characteristics [9]. Some scholars have focused on practicalizing vision detection on the one hand, such as medical detection, fruit and vegetable quality inspection, navigation, and other directions. Sánchez-Jiménez et al. [10] presented research advances in two computer vision techniques for skin wound measurement, using an open-source computer vision library (OpenCV) and SciPy open-source libraries to calculate skin wound dimensions in an objective and reliable manner, allowing the generation of two-dimensional results in real time and the submission of the results as a report to medical experts for interpretation as a complementary test. Blasco et al. [11] presented a quality assessment monitoring system based on machine vision and optical techniques. This system is much faster than any manual nondestructive inspection of the quality of fruits and vegetables. The detection of color, shape, and external defects is based on machine vision for color images and finally provides a better assessment of ripeness and other defects. Eriksen et al. [12] proposed a visual ranging algorithm for underwater navigation based on monocular cameras. He evaluated the performance of the proposed algorithm on underwater video sequences and discussed how to further improve the robustness and computational performance of the algorithm. On the other hand, some scholars have improved the visual measurement calibration detection algorithm to improve the accuracy and stability of detection. Gadelmawla et al. [13] presented a vision system for the automatic inspection of various threads. The system was calibrated for imperial and metric units, and the measurement results were compared with the standard values by measuring standard metric thread plug gauges. The results show that the vision system has good inspection accuracy. Du et al. [14] proposed a fast and effective method for detecting passive components and providing complete measurements and sufficient accuracy in automatically detecting the physical characteristics of passive components. Kewei et al. [15] proposed a calibration-based method for optical planimetry. He used photogrammetry in computer vision to display six different orientations of the inspection graphics on another Liquid Crystal Display (LCD) and calibrated the pinhole camera. He obtained the intrinsic parameters of the pinhole camera by processing the images of the inspection graphics. Hacini et al. [16] proposed a one-dimensional numerical fractional-order Charef differentiator (FCD). It was generalized to 2D by means of multi-directional operators, and the 2D operator was eventually improved by defining x and y multi-directional operators. He proposed 2D-IFCD-based edge extraction and conducted numerical experiments to evaluate the performance of 2D-IFCD-based edge extraction methods, and the obtained results proved its effectiveness. Othman et al. [17] proposed a method to calculate pixel thresholds for foreground and background images from global and local image analysis. According to this method, an image is divided into blocks according to different resolution levels and then an improved sampling method is applied globally and locally. The results show that the method outperforms the Canny method and other adaptive methods.
Some other scholars have also researched the advantages of vision inspection for processing and manufacturing as well as practical applications, optimizing and improving the accuracy of inspection algorithms and measurement methods. Dong et al. [18] has made great progress in the field of structural health monitoring (SHM) by combining machine vision-based sensing technology with digital image processing algorithms. Sun et al. [19] used this technology to effectively identify and classify weld defects of thin-walled metal canisters. A weld defect detection and classification algorithm based on machine vision is proposed. Li et al. [20] proposed a novel defect extraction and classification scheme for mobile phone screens based on machine vision. A clustering algorithm is proposed to avoid the false detection or missed detection of cluster defects. Xie et al. [21] proposed an improved image sequence velocity detection algorithm combined with the Kalman prediction equation to predict the coordinate position of the workpiece in the nonlinear motion state in order to compensate the time delay error of the visual servo system. Finally, an experiment was carried out on a visual tracking robot system with a conveyor belt. The results show that the maximum tracking speed can reach 90 mm/s, and the tracking accuracy error is less than 1 mm. Wang et al. [22] presented a new deep learning-based machine vision inspection method to identify and classify defective products without a loss of accuracy. Dong et al. [23] developed a machine vision-based system for detecting the position of micro-slots on the surface of conductive slip rings in order to accurately detect the position of micro-slots on the surfaces of conductive slip rings.
In the following part of this paper, we first explain in detail the built brush wire angle forming and measuring device and image pre-processing. A machine vision-based brush angle measurement and control method is proposed. Then, we measure the brush wire forming angle and rebound angle value data under the autonomous control of the forming angle value, build a data pre-compensation model, and experimentally verify this model. It can control the forming angle more precisely and improve the contact performance of the brush filament and the ring body effectively. A detailed discussion based on the results and observations is presented. The accuracy and applicability of the machine vision-based brush wire angle measurement method are demonstrated by comparisons with experiments. Lastly the paper summarizes its conclusions, and potential further studies are suggested.

2. Principles and Methods

2.1. Brush Wire Angle Forming and Measuring Device Construction

A brush wire angle forming device was designed for the phenomenon of rebound and uneven force during angle forming of brush wire. The device consists of a clamping plate (1), a deflection plate (2), an angle table (3), a rotation table (4), a brush plate (5), a displacement table (6), a positioning plate (7), and an optical platform (8). As shown in Figure 1, the angle of the brush wire is shaped by designing the brush wire clamping device using the torque transmitted by the deflecting angle stage and the accuracy of the deflecting angle.
The brush wire angle measuring device is shown in Figure 2. The device consists of a camera (1), two light sources (2), a support frame (3), a light source controller (4), and an angle forming device (5). In this paper, we chose the model MER-500-14U3M CMOS black-and-white surface array camera of the Daheng Mercury (Beijing, China) series with a resolution of 2592 × 1944 and a mechanical size of 29 mm × 29 mm × 29 mm, which is compact in size and convenient to install. The pixel size is 2.2 μm × 2.2 μm. The selected focal length of the lens should have been 8 mm, and positive illumination was selected. In order to avoid the generation of positive lighting shadows, we selected symmetrical lighting to offset the generation of shadows. The LED light source was selected, and the light source consisted of a controller and two strip light sources.

2.2. Camera Calibration

2.2.1. Camera Aberration Model

The photographed object in the imaging process will produce a certain amount of aberration, resulting in deviations in visual measurement results. The aberration of the camera is mainly divided into radial aberration and tangential aberration [24].
(1) Radial distortion
The radial distortion in the optical center of the camera is 0. As it moves outward, the radial distortion increases symmetrically in either direction. Therefore, radial distortion is more easily observed at the edge of the image. The straight lines near the image edge are bent due to radial distortion, a phenomenon of expansion known as the barrel or fisheye effect. Radial distortion can be expressed in the form of a Taylor series, with the first two terms of the series typically used to approximate the radial distortion phenomenon. The pixel coordinates of the image (x, y) will be distorted according to the following factors:
x d = x ( 1 + k 1 r 2 + k 2 r 4 + k 3 r 6 )
y d = y ( 1 + k 1 r 2 + k 2 r 4 + k 3 r 6 )
where r = ( x x p ) 2 + ( y y p ) 2 ,   k 1 , k 2 , k 3 are the radial distortion coefficients of the camera lens. x and y are the coordinates of the pixel. x p and y p are the coordinates of the center of the aberration. Generally, the k 1 accuracy can satisfy the aberration correction application for a narrow field of view. In this paper, the third-order accuracy of k 1 , k 2 , and k 3 was chosen to ensure the accuracy of the acquired graphics.
(2) Tangential aberration
During the manufacturing process of a camera, the lens may not be perfectly parallel to the imaging plane. As a result, this manufacturing defect usually causes tangential lens aberrations, where the coordinates of the pixels are distorted.
x d = x + ( 2 p 1 y + p 2 ( r 2 + 2 x 2 ) )
y d = y + ( p 1 ( r 2 + 2 y 2 ) y + 2 p 2 x ) )
Where p 1 and p 2 are the tangential aberration parameters of the lens.

2.2.2. Calibration Methods

The general camera calibration is divided into the linear calibration method, nonlinear calibration method, two-step method, and tensor calibration method [25]. The Zhang calibration method combines the advantages of linear calibration and nonlinear calibration, and the method is simple to operate and has high accuracy, which is suitable for most occasions. In this paper, Zhang’s calibration method was be used for camera calibration. The aberrations of the acquired feature images are corrected by the OpenCV tessellation calibration function.
As shown in Figure 3 and Figure 4, a 12 × 9 float glass tessellation calibration template with a processing accuracy of 0.01 mm and a cell side length of 10 mm was selected. Twenty tessellated grid images were acquired according to different tilt angles. By detecting the coordinate values of sub-pixel corner points, the internal reference matrix and aberration coefficient of the camera were calculated. The original image matrix was transformed to eliminate the effects of radial and tangential aberrations. The final correction processing of the images was realized.
The internal reference matrix of the camera was derived after calibration.
M t x = [ f x 0 c x 0 f y c y 0 0 1 ] = [ 3273.9272 0.0000 581.4176 0.0000 3265.1508 519.8155 0.0000 0.0000 1.0000 ]
where f x and f y are the products of the physical focal length and cell size, and c x and c y are the pixel differences in the x and y directions between the image center coordinates and the origin coordinates.
The aberration parameter matrix was calculated to be:
D   =   [ K 1 , K 2 , P 1 , P 2 , K 3 ]   =   [ 0.40098 7.98644 0.00578 0.00412 0.02106 ]
where K 1 , K 2 , and K 3 are the radial distortion parameters, P 1 and P 2 are the tangential distortion parameters.
The average reprojection error is generally considered to be less than 0.5 pixels, which indicates a high calibration accuracy. The average reprojection error of this calibration was 0.0128 pixels, which meets the requirements of camera calibration. As shown in Figure 5, the calibration plate images before and after calibration were compared. It can be seen that the corrected images obviously reduce the influence of distortion.

2.3. Pixel Equivalent

After correcting the aberrations of the above camera, the image features could be acquired more accurately. At this point, the unit of distance detection for the object to be measured was the pixel value. In order to obtain the true size of the feature distance in the subsequently acquired image, a conversion between the pixel size and true size was, therefore, required. The imaging model of the camera is shown in Figure 6. This leads to H2/h2 = D/f, where H2 is the physical length, h2 is the pixel length, D is the distance between the camera and the photographed object, and f is the focal length of the camera. When the distance is fixed, the ratio of D to f is a fixed value, and the correspondence between the physical length and the pixel can be derived as a constant, which is the pixel equivalent, noted as K. When the pixel distance of h2 is known, the true distance of H2 can be derived from h2 × K.
In this paper, a standard block measuring 9 mm × 9 mm × 30 mm was used, and the number of edge pixels of the block at the current shooting distance was obtained by image processing. The calibration value of the pixel equivalent was obtained by calculating the ratio of the real size of the standard block to the acquired pixel size. As shown in Figure 7, the figure takes the results of one, four, and eight measurements during the measurement process. The average pixel length of the block in the current state can be calculated as 84.9634 pixels from the 8 measured experimental data points, which further results in the pixel equivalent value K = 9/84.9634 = 0.1059.
To further verify the accuracy of the obtained pixel equivalent values, the upper edge of the block was measured by changing the pose of the block, and the block size was measured in reverse by the above calculated pixel equivalents. The details are shown in Table 1. Eight experiments were conducted separately. The experimental results show that the average error of the block size values obtained by pixel equivalence was 0.027, and the error value was less than 0.05 mm. The K value was relatively stable. For brush angle measurement, we adjusted the lift table to put the surface of the measuring block and the surface of the brush plate positioning block at the same horizontal position. If there was a large movement in the z-direction, it was necessary to recalibrate to obtain the pixel equivalents.

2.4. Image Pre-Processing

2.4.1. Image Filtering

The general filtering methods median filtering, mean filtering, Gaussian filtering, and bilateral filtering were used [26]. As shown in Figure 8 below, the bilateral filtering ensured that the image edge features were preserved while the noise was suppressed. The remaining three filtering methods started to blur the edges of the brushes and were not applicable to the experimental study in this paper. In summary, the bilateral filtering method was selected for the noise reduction of the acquired bristle images.
Bilateral filtering is an improved filtering method based on Gaussian filtering. This method maintains the edges of the image and removes Gaussian noise from it. Bilateral filtering is a kind of nonlinear weighted Gaussian filtering that uses the estimated value of large pixels at the edges of the image as the values of edge points. Bilateral filtering has a good edge processing effect and can retain the edge features well. As shown in Figure 9 and Figure 10 below, bilateral filtering was applied to the acquired bristle images, and it can be seen that the edge details of the bristles are obvious and the noise is effectively reduced by bilateral filtering.

2.4.2. Image Edge Detection

In order to ensure accuracy in image recognition and measurement, the detection of feature edges is required for preprocessed images. In the detection process, a pixel point with a gray value greater than a threshold is an edge, and a pixel point with a value less than that threshold is not an edge [27,28]. The more commonly used edge detection algorithms are the Sobel, Laplace, and Canny algorithms. Sobel edge detection is more effective in cases where there is more noise in the image. Laplace edge detection can be used to determine the bright and dark areas of the image. Canny edge detection can effectively resist the interference of noise and detect the true image. Therefore, the Canny edge detection algorithm was chosen to detect the edges of the image in this paper. From Figure 11, it can be seen that after Canny edge detection, the edges of the brush wire and the positioning block of the brush plate were well reflected.

2.4.3. Corner Point Detection

Shi-Tomasi corner point detection is an improved algorithm based on Harris corner point detection. The principle is to select the smallest eigenvalue in the matrix and then judge this eigenvalue. If the eigenvalue is greater than a threshold, this value is judged to be a strong corner point.
The weighted autocorrelation function of the algorithm:
F ( x , y ) = x i , y i W w ( x i , y i ) [ I ( x i , y i ) I ( x i + Δ x , y i + Δ y ) ] 2
where w ( x i , y i ) are the weights, and the autocorrelation matrix, A, is given by the following equation:
A = [ w w I x 2 w w I x I y w w I x I y w w I y 2 ]
Target validation relationship equation:
m i n ( λ 1 , λ 2 ) > λ
where λ 1 and λ 2 are eigenvalues, and λ is a predefined threshold value, which is calculated to yield a strong corner point if this above equation is satisfied.
Compared with straight line detection in images, corner point detection has less influence under different lighting conditions and has the advantage of being almost invariant when the image is rotated at different angles. The corner points retain important local detail information during image acquisition, and the corner-point-based detection information has a certain degree of accuracy. As shown in Figure 12, the position of the corner point of the positioning block on one side of the brushboard was detected, and the detection results yielded the pixel coordinates of (pixel 731, pixel 377) for the left corner point and (pixel 796, pixel 377) for the right corner point. The edge straight line detection was also performed by constructing a mask for the brush wires to be detected. Based on the Hoff straight line detection algorithm, the left edge of the brush filament was a detected straight line, and the detection results are shown in Figure 12.

3. Results and Discussion

3.1. Principle of Bristle Angle Measurement

As shown in Figure 13, by obtaining the values of the four coordinate points in the image, the slope k 1 of the line at the detection edge of the brush wire and the slope k 2 of the line where the two corner points of the positioning block were located were calculated, respectively. The line where these two corner points were located was used as the reference line. The angle value between the two lines was further calculated by the tangent formula, and the resulting angle value was the actual angle at which the brush wire was tilted. As shown in Figure 14, the angle value of the brush wire was tested, and the angle of the brush wire was 85.13° at this time.
k 1 = y 2 y 1 x 2 x 1
k 2 = y 4 y 3 x 4 x 3
k = t a n α = t a n | α 1 α 2 | = t a n α 1 t a n α 2 1 + t a n α 1 t a n α 2 = | k 1 k 2 1 + k 1 k 2 |
a n g l e = a r c t a n ( k ) × 57.3

3.2. Principle of Bristle Angle Measurement

In order to test the above brush wire angle detection algorithm, three different angles of photosensitive resin triangles were made in this paper to verify the angle value detection. As shown in Figure 15, according to the calculation method of the brush wire angle. When measuring angle A, first, the linear coordinates of the adjacent side, m, of angle A were obtained by the Hoff linear detection, and then the endpoint of the other adjacent side, n, of angle A was detected to obtain the endpoint coordinate value of n. Using the neighboring edge, n, as the reference line, the angle value between the two lines m and n was obtained, and this value was the degree of angle A.
As shown in Figure 15, the outline of the triangle was first obtained using Canny edge detection. Then, the angle values were detected for three different angles of the triangle, respectively. The measurement results are shown in Table 2. The pixel coordinate values of each triangle vertex were obtained separately to determine the angle positions, and the angle detection was performed for 10°, 20°, 30°, 40°, 50°, and 60°, respectively. The data in Table 2 show that the measurement error is within 0.05°, which meets the requirements of this paper for brush wire angle detection.

3.3. Brush Wire Springback Control

The deformation of the brush wire during the bending process can be divided into two parts, namely, the elastic deformation stage and the plastic deformation stage. As shown in Figure 16a, the initial angle of the brush wire before forming is β, and the bending deformation is produced by the action of the deflector plate and the clamping plate. The forming angle is α and the deflection angle of the deflector plate and the clamping plate is γ. This is the angle forming stage of the brush wire. When the angle forming stage was completed, the deflection plate and the clamping plate were unloaded, and the plastic deformation of the brush wire disappeared gradually due to the high elasticity of the brush wire itself. As the elasticity was restored, the material on the pulled side of the brush wire shrank and the material on the other side elongated, resulting in the forming angle of the brush wire being smaller than the theoretical forming angle, which is called the springback phenomenon of the brush wire. As shown in Figure 16b, when the deflection plate and clamping plate were unloaded, the brush wire produced a certain amount of springback. θ is the forming angle after the brush rebound, and the springback angle is δ, which seriously affects the accuracy of brush wire forming.

3.3.1. Brush Wire Rebound Angle Value Detection

According to the algorithm of brush wire angle measurement, the angle of the brush wire after forming was measured separately. As shown in Figure 17 below, after bending the wire at a certain angle with the deflector plate, the angle of the wire was measured by unloading the deflector plate and the clamping plate. In order to compare the angle value of the rebound of the brush wire, the angle values after the rebound of the brush wire were 102°, 113°, and 119°, respectively, which were 94.238°, 103.247°, and 108.145°. It is obvious that the forming angle of the brush wire deviates greatly from the angle after the rebound, so the control of the rebound of the brush wire is very important to improve the accuracy of the forming angle of the brush wire.

3.3.2. Brush Wire Rebound Angle Value Detection

The center of deflection is located at the neutral layer to build a pre-compensation model for brush wire rebound and to measure the brush wire rebound value to obtain the correspondence between the target forming angle and the overbending pre-compensation angle by experimental method. The range of wire forming was 0–30°, and the angle of wire forming was measured in 1° steps.
α is the angle of the brush wire after rebound, and β is the forming angle of the brush wire. If the forming angle of the brush wire is 10°, the overbend of the forming angle of the brush wire is Δγ, as shown in Figure 18, and the forming angle of the brush wire should be controlled at 10° + Δγ to ensure that the forming angle of the brush wire remains around 10°. As shown in Figure 19, it is the rebound compensation value of the brush wire that is used as a reference for the calculation of the angle compensation value of the brush wire.
The function model was obtained by fitting a quadratic polynomial to the data obtained for the angle of the formed brush wire. In this model, the independent variable x is the forming angle value of brush wire, and y1 is the angle value of the forming after the spring back of the brush wire. The actual rebound y value of the brush wire was obtained by subtracting the formed angle value from the rebound angle value, i.e., the model of the function model (15) of the brush wire rebound precompensation, by which the angle precompensation value of the brush wire can be calculated at any angle from 0° to 30°. The functional model is shown as follows.
y 1 = 0.17159 + 0.38742 x 0.02446 x 2 + 0.0026 x 3 0.00005036 x 4
y = x y 1 = 0.17159 + 0.61258 x + 0.02446 x 2 0.0026 x 3 + 0.00005036 x 4

3.4. Brushing Angle Verification

Based on the brush wire pre-compensation model derived above, the model was experimentally verified by forming the brush wire angles of 5°, 10°, 15°, 20°, 25°, and 30°. The angle forming process was performed using the above proposed overbending pre-compensation model. The experiment was repeated five times for each angle. The angle values were measured by visual measurement after the brush wire angle forming was completed. Table 3 shows the experimental data from the five repetitions of the experiment. The average value was calculated for each forming angle. It can be seen that the average value of the rebound increases with an increase in the forming angle.
In statistics, a confidence interval (CI) for a probability sample is an estimate in the form of an interval for an unknown parameter value in the parametric distribution of the aggregate that produced the sample. In contrast to point estimation, which uses a sample statistic to estimate a parameter value, a confidence interval also contains information about the precision of the estimate. The 95% confidence interval is an interval extrapolation indicator of the overall mean by the mean of the specimen data.
P ( μ 1.96 σ n M μ + 1.96 σ n ) 0.95
where μ is the mean value of brush rebound for each group, σ is the standard deviation, and n is the number of tests.
As shown in Table 4, the rebound angle of the brush wire was calculated according to the pre-compensation model given in this paper. Combined with the mean value of the springback calculated in Table 3, the angle forming compensation model proposed in this paper was verified. The calculated value of the compensation model has a 95% confidence interval, which is more accurate in predicting the rebound angle of the bristles and reduces the angle error caused by the rebound phenomenon during the angle forming process of the bristles. However, when the angle is too large, the rebound angle is smaller than the theoretical calculated value. This is due to the plastic deformation of the bristles when the angle is too large during the angle forming process.
The method for calculating the angle of the brush filament was established by filtering and edge detection of the image. After testing, the error of the measurement result was within 0.05°. Based on this method, an experimental device of brush wire angle forming was built. Finally, a pre-compensated model of brush wire rebound was established to measure the angle of brush wire forming and rebound angle based on visual measurement. The corresponding data relationship between the angle forming value and the rebound value was obtained, the model was experimentally analyzed, and it was verified that the influence of the forming angle error caused by the rebound of the brush wire can be eliminated.

4. Conclusions

In this paper, the process of forming the angle of the brush wire was studied and a method of forming the angle of the brush wire was proposed. Based on this method, a brush angle forming and measuring device was built, which realizes the accurate control of the angle output value during the forming process of the brush angle. The details are as follows:
1. The camera was calibrated before the measurement experiment. The average reprojection error by calibration was 0.0128 pixels, which effectively reduced the influence of camera distortion on angle measurement. The molding angle was precisely controlled by deflecting the angular stage. In the visual measurement process, the acquired images were filtered and edge detected, and a method of calculating the brush angle was proposed. After the measurement test, the error of the measured brush angle value based on this method was within 0.05°, which meets the requirements of brush angle detection.
2. A pre-compensation model of brush filament rebound was established based on curve fitting. Based on machine vision, we measured the forming angle and rebound angle of the brush filament and obtained the corresponding relationship between the forming angle and rebound value. The correctness of the model was verified by comparing the experimental data with the real wire rebound angle. The calculated value of the compensation model had a 95% confidence interval, which was more accurate in predicting the rebound angle of the bristles. The model can be implemented reliably to optimize the brush wire forming process and improve slip ring life and contact reliability.

Author Contributions

Conceptualization and methodology, J.L. (Junye Li); data curation, and writing—original draft preparation, J.L. (Jun Li); software and visualization, X.W.; validation and formal analysis, G.T.; supervision and resources, J.F. All authors have read and agreed to the published version of the manuscript.

Funding

The authors would like to thank the national natural science foundation of China No. NSFC 51206011, Jilin province science and technology development program of Jilin province No.20200301040RQ, and Changchun science and technology program of Changchun city No. 18DY017.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, J.; Zhang, X.; Zhu, C.; Song, K.; Zhou, T. Research on Electromagnetic Compatibility of Non-Contact Slip Ring. In Proceedings of the 2020 International Conference on Intelligent Computing, Automation and Systems (ICICAS), Chongqing, China, 11–13 December 2020; pp. 318–321. [Google Scholar]
  2. Grandin, M.; Wiklund, U. A Wear Tolerant Slip-Ring Assembly. In Proceedings of the ICEC 2014; The 27th International Conference on Electrical Contacts, Dresden, Germany, 22–26 June 2014; pp. 1–6. [Google Scholar]
  3. Houenouvo, A.T.H.; Hofmann, W. Finite Element Analysis of the Contact Problem between Graphite-Brushes and—Slip Rings in Double-Fed Asynchronous Generators. In Proceedings of the International Multi-Conference on Systems, Signals Devices, Chemnitz, Germany, 20–23 March 2012; pp. 1–6. [Google Scholar]
  4. Hong, E.; Krishnamurthy, A.; Gregory, D.A.; Anderson, S. Optical Slip Ring for Sensor Interface Applications. In Sensors for Propulsion Measurement Applications; SPIE: Bellingham, WA, USA, 2006; Volume 6222, p. 62220E. [Google Scholar] [CrossRef]
  5. Wcislik, M.; Suchenia, K.; Cyganik, A. Vectorized Mathematical Model of a Slip-Ring Induction Motor. Energies 2020, 13, 4015. [Google Scholar] [CrossRef]
  6. Zhang, J.; Ma, X.; Cheng, B.; Zhao, Y. Storage Life Modeling and Analysis for Contacting Slip Ring Based on Physics of Failure. IEEE Trans. Compon. Packag. Manuf. Technol. 2017, 7, 1969–1980. [Google Scholar] [CrossRef]
  7. Liu, X.; Zhu, Y.; Liu, H.; Zhao, L.; Xie, Z. Transmission reliability assessment of star-loaded long-life conductive slip rings. Opt. Precis. Eng. 2019, 27, 2028–2035. (In Chinese) [Google Scholar]
  8. Jia, Q.; Yao, P. Analysis of vacuum high and low temperature friction torque characteristics of conductive slip rings for space applications. Lubr. Seal. 2020, 45, 123–127, 140. (In Chinese) [Google Scholar]
  9. Jain, R.; Kasturi, R.; Schunck, B.G. Machine Vision; McGraw-Hill series in computer science; McGraw-Hill: New York, NY, USA, 1995; ISBN 978-0-07-032018-5. [Google Scholar]
  10. Sánchez-Jiménez, D.; Buchón-Moragues, F.; Escutia-Muñoz, B.; Botella-Estrada, R. Development of Computer Vision Applications to Automate the Measurement of the Dimensions of Skin Wounds. Proceedings 2019, 19, 18. [Google Scholar] [CrossRef] [Green Version]
  11. Blasco, J.; Munera, S.; Aleixos, N.; Cubero, S.; Molto, E. Machine Vision-Based Measurement Systems for Fruit and Vegetable Quality Control in Postharvest. In Measurement, Modeling and Automation in Advanced Food Processing; Hitzmann, B., Ed.; Advances in Biochemical Engineering/Biotechnology; Springer International Publishing: Cham, Switzerland, 2017; pp. 71–91. ISBN 978-3-319-60111-3. [Google Scholar]
  12. Voisin-Denoual, M. Monocular Visual Odometry for Underwater Navigation: An Examination of the Performance of Two Methods. Master’s Thesis, KTH Royal Institute Of Technology, Stockholm, Sweden, 2018. [Google Scholar]
  13. Gadelmawla, E.S. Computer Vision Algorithms for Measurement and Inspection of External Screw Threads. Measurement 2017, 100, 36–49. [Google Scholar] [CrossRef]
  14. Du, W.Y.; Dickerson, S.L. Passive Component Inspection Using Machine Vision. In Proceedings of the 1998 International Conference on Multichip Modules and High Density Packaging (Cat. No.98EX154), Denver, CO, USA, 17 April 1998; pp. 74–79. [Google Scholar]
  15. Kewei, E.; Li, D.; Yang, L.; Guo, G.; Li, M.; Wang, X.; Zhang, T.; Xiong, Z. A New Optical Flat Surface Measurement Method Based on Machine Vision and Deflectometry. In Optical Measurement Technology and Instrumentation; SPIE: Bellingham, WA, USA, 2016; Volume 10155, p. 101551X. [Google Scholar] [CrossRef]
  16. Hacini, M.; Akram, H.; Hachouf, F.; Akdag, H. A 2D-Fractional Derivative Mask for Image Feature Edge Detection. In Proceedings of the 2017 International Conference on Advanced Technologies for Signal and Image Processing, Fez, Morocco, 22–24 May 2017. [Google Scholar]
  17. Othman, Z.; Abdullah, A. An Adaptive Threshold Based on Multiple Resolution Levels for Canny Edge Detection. In Proceedings of the Recent Trends in Information and Communication Technology; Saeed, F., Gazem, N., Patnaik, S., Saed Balaid, A.S., Mohammed, F., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 316–323. [Google Scholar]
  18. Dong, C.Z.; Ye, X.W.; Jin, T. Identification of Structural Dynamic Characteristics Based on Machine Vision Technology. Measurement 2018, 126, 405–416. [Google Scholar] [CrossRef]
  19. Sun, J.; Li, C.; Wu, X.-J.; Palade, V.; Fang, W. An Effective Method of Weld Defect Detection and Classification Based on Machine Vision. IEEE Trans. Ind. Inform. 2019, 15, 6322–6333. [Google Scholar] [CrossRef]
  20. Li, C.; Zhang, X.; Huang, Y.; Tang, C.; Fatikow, S. A Novel Algorithm for Defect Extraction and Classification of Mobile Phone Screen Based on Machine Vision. Comput. Ind. Eng. 2020, 146, 106530. [Google Scholar] [CrossRef]
  21. Xie, X.; Zhu, C.; Xie, M. Visual Tracking Control of SCARA Robot System Based on Deep Learning and Kalman Prediction Method. Int. J. Hydromechatron. 2021, 4, 384–396. [Google Scholar] [CrossRef]
  22. Wang, J.; Fu, P.; Gao, R.X. Machine Vision Intelligence for Product Defect Inspection Based on Deep Learning and Hough Transform. J. Manuf. Syst. 2019, 51, 52–60. [Google Scholar] [CrossRef]
  23. Dong, F.; Zheng, Z.; Yuan, S.; Li, H.; Shi, Y. Development of machine vision-based micro-slot position detection system for conductive slip ring surfaces. Manuf. Technol. Mach. Tools 2018, 126–130. [Google Scholar]
  24. Sun, J.; Cheng, X.; Fan, Q. Camera Calibration Based on Two-Cylinder Target. Opt. Express 2019, 27, 29319–29331. [Google Scholar] [CrossRef] [PubMed]
  25. Park, K.; Chae, M.; Cho, J.H. Image Pre-Processing Method of Machine Learning for Edge Detection with Image Signal Processor Enhancement. Micromachines 2021, 12, 73. [Google Scholar] [CrossRef] [PubMed]
  26. Zhu, J.; Li, B.; Zhang, Y.; Cao, J.; Wang, Y. Experimental Study on Image Filtering Algorithm. In Proceedings of the 2020 5th International Conference on Advanced Robotics and Mechatronics (ICARM), Shenzhen, China, 18–21 December 2020; pp. 7–11. [Google Scholar]
  27. Mittal, M.; Verma, A.; Kaur, I.; Kaur, B.; Sharma, M.; Goyal, L.M.; Roy, S.; Kim, T.-H. An Efficient Edge Detection Approach to Provide Better Edge Connectivity for Image Analysis. IEEE Access 2019, 7, 33240–33255. [Google Scholar] [CrossRef]
  28. Paravolidakis, V.; Ragia, L.; Moirogiorgou, K.; Zervakis, M.E. Automatic Coastline Extraction Using Edge Detection and Optimization Procedures. Geosciences 2018, 8, 407. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Design of the brush wire forming device. (a) Three-dimensional diagram of the brush wire angle forming device; (b) Brush wire angle forming device.
Figure 1. Design of the brush wire forming device. (a) Three-dimensional diagram of the brush wire angle forming device; (b) Brush wire angle forming device.
Micromachines 13 00447 g001
Figure 2. Brush angle measuring device.
Figure 2. Brush angle measuring device.
Micromachines 13 00447 g002
Figure 3. Different angle checkerboard grid calibration diagram.
Figure 3. Different angle checkerboard grid calibration diagram.
Micromachines 13 00447 g003
Figure 4. Checkerboard grid calibration diagram.
Figure 4. Checkerboard grid calibration diagram.
Micromachines 13 00447 g004
Figure 5. Camera calibration image: (a) before calibration and (b) after calibration.
Figure 5. Camera calibration image: (a) before calibration and (b) after calibration.
Micromachines 13 00447 g005
Figure 6. Camera imaging model.
Figure 6. Camera imaging model.
Micromachines 13 00447 g006
Figure 7. Measurement results of block length: (a) The 1st experiment, (b) the 4th experiment, and (c) the 8th experiment.
Figure 7. Measurement results of block length: (a) The 1st experiment, (b) the 4th experiment, and (c) the 8th experiment.
Micromachines 13 00447 g007
Figure 8. Brush filtration comparison images: (a) Original image and the (b) median, (c) mean, (d) Gaussian, and (e) bilateral filtering methods.
Figure 8. Brush filtration comparison images: (a) Original image and the (b) median, (c) mean, (d) Gaussian, and (e) bilateral filtering methods.
Micromachines 13 00447 g008
Figure 9. Bilateral filtering diagram.
Figure 9. Bilateral filtering diagram.
Micromachines 13 00447 g009
Figure 10. Bilateral filtering local comparison image: (a) Original partial picture of the brush wire, (b) Partial picture of the bilateral filter brush wire.
Figure 10. Bilateral filtering local comparison image: (a) Original partial picture of the brush wire, (b) Partial picture of the bilateral filter brush wire.
Micromachines 13 00447 g010
Figure 11. Edge detection comparison image: (a) Original image and (b) Canny edge detection.
Figure 11. Edge detection comparison image: (a) Original image and (b) Canny edge detection.
Micromachines 13 00447 g011
Figure 12. Brush wire corner point and edge straight line inspection diagram.
Figure 12. Brush wire corner point and edge straight line inspection diagram.
Micromachines 13 00447 g012
Figure 13. Principle diagram of brush wire angle measurement.
Figure 13. Principle diagram of brush wire angle measurement.
Micromachines 13 00447 g013
Figure 14. Brush wire clamping angle value detection diagram.
Figure 14. Brush wire clamping angle value detection diagram.
Micromachines 13 00447 g014
Figure 15. Angular value detection diagram: (a) First edge detection, (b) 10° angle detection, (c) 20° angle detection, (d) Second edge detection, (e) 30° angle detection, (f) 40° angle detection, (g) Third edge detection, (h) 50° angle detection, (i) 60° angle detection.
Figure 15. Angular value detection diagram: (a) First edge detection, (b) 10° angle detection, (c) 20° angle detection, (d) Second edge detection, (e) 30° angle detection, (f) 40° angle detection, (g) Third edge detection, (h) 50° angle detection, (i) 60° angle detection.
Micromachines 13 00447 g015
Figure 16. Schematic diagram of brush wire rebound: (a) Wire forming process, (b) wire rebound process.
Figure 16. Schematic diagram of brush wire rebound: (a) Wire forming process, (b) wire rebound process.
Micromachines 13 00447 g016
Figure 17. Measured angular values of bristles after rebound: (a) 102° springback value, (b) 113° springback value, (c) 119° springback value.
Figure 17. Measured angular values of bristles after rebound: (a) 102° springback value, (b) 113° springback value, (c) 119° springback value.
Micromachines 13 00447 g017
Figure 18. Fitting diagram of brush wire rebound: (a) Measured angle of brush wire rebound, (b) Fitted angle of rebound.
Figure 18. Fitting diagram of brush wire rebound: (a) Measured angle of brush wire rebound, (b) Fitted angle of rebound.
Micromachines 13 00447 g018
Figure 19. Rebound compensation value image.
Figure 19. Rebound compensation value image.
Micromachines 13 00447 g019
Table 1. Block pixel size measurement results (mm).
Table 1. Block pixel size measurement results (mm).
Number12345678Average
Block size9.0169.0328.9799.0879.0048.9839.0279.0949.027
Error0.0160.0320.0210.0870.0040.0170.0270.0940.027
Table 2. Angle value detection results.
Table 2. Angle value detection results.
NumberTheoretical Value/°Vertex X Coordinate/PixelVertex Y Coordinate/PixelDetection Value/°Error/°
110127119100
220120481320.040.04
3301559529.950.05
440123079640.020.02
550118177550.030.03
66061410659.970.03
Table 3. Results of the brushing angle forming rebound experiments.
Table 3. Results of the brushing angle forming rebound experiments.
Angle Forming Value/°12345Mean Value of Rebound/°
53.6643.5253.4923.6673.7013.609
106.8116.7716.9036.8456.7796.821
158.6678.5628.9788.5238.6598.677
209.5119.4039.5789.5939.4999.516
259.7719.8219.7169.6689.7099.737
3010.88210.92910.88711.17210.93610.961
Table 4. Experimental results of the brush angle forming pre-compensation model.
Table 4. Experimental results of the brush angle forming pre-compensation model.
Angle Forming Value/°Average Value of Brush Wire ReboundCalculated Value of the Compensation Model95% CI
53.6093.552[3.527, 3.691]
106.8216.782[6.774, 6.868]
158.6778.638[8.520, 8.8339]
209.5169.465[9.449, 9.583]
259.7379.821[9.597, 9.877]
3010.96111.154[10.936, 10.986]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, J.; Li, J.; Wang, X.; Tian, G.; Fan, J. Machine Vision-Based Method for Measuring and Controlling the Angle of Conductive Slip Ring Brushes. Micromachines 2022, 13, 447. https://doi.org/10.3390/mi13030447

AMA Style

Li J, Li J, Wang X, Tian G, Fan J. Machine Vision-Based Method for Measuring and Controlling the Angle of Conductive Slip Ring Brushes. Micromachines. 2022; 13(3):447. https://doi.org/10.3390/mi13030447

Chicago/Turabian Style

Li, Junye, Jun Li, Xinpeng Wang, Gongqiang Tian, and Jingfeng Fan. 2022. "Machine Vision-Based Method for Measuring and Controlling the Angle of Conductive Slip Ring Brushes" Micromachines 13, no. 3: 447. https://doi.org/10.3390/mi13030447

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop