Next Article in Journal
Determination of Tumor Marker Carcinoembryonic Antigen with Biosensor Based on Optical Quantum Weak Measurements
Previous Article in Journal
Application and Comparison of the MODIS-Derived Enhanced Vegetation Index to VIIRS, Landsat 5 TM and Landsat 8 OLI Platforms: A Case Study in the Arid Colorado River Delta, Mexico
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Development and Performance Evaluation of Image-Based Robotic Waxing System for Detailing Automobiles

Department of Mechanical Engineering, National Taiwan University of Science and Technology, No. 43, Keelung Rd., Sec. 4, Taipei 106, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(5), 1548; https://doi.org/10.3390/s18051548
Submission received: 2 April 2018 / Revised: 9 May 2018 / Accepted: 12 May 2018 / Published: 14 May 2018
(This article belongs to the Section Physical Sensors)

Abstract

:
Waxing is an important aspect of automobile detailing, aimed at protecting the finish of the car and preventing rust. At present, this delicate work is conducted manually due to the need for iterative adjustments to achieve acceptable quality. This paper presents a robotic waxing system in which surface images are used to evaluate the quality of the finish. An RGB-D camera is used to build a point cloud that details the sheet metal components to enable path planning for a robot manipulator. The robot is equipped with a multi-axis force sensor to measure and control the forces involved in the application and buffing of wax. Images of sheet metal components that were waxed by experienced car detailers were analyzed using image processing algorithms. A Gaussian distribution function and its parameterized values were obtained from the images for use as a performance criterion in evaluating the quality of surfaces prepared by the robotic waxing system. Waxing force and dwell time were optimized using a mathematical model based on the image-based criterion used to measure waxing performance. Experimental results demonstrate the feasibility of the proposed robotic waxing system and image-based performance evaluation scheme.

1. Introduction

As a common procedure in auto detailing, waxing removes the fine scratches in automotive sheet metal and reduces surface roughness, which in turn prevents impurities such as dust or moss from sticking to the outer shell of sheet metal, reduces resistance and undesired attachments, and thereby decreases fuel consumption. However, waxing by hand means long durations of applying force, being subjected to machine vibrations, and remaining in poor posture, which cause muscle aches, bone injuries, and damage to the joints in the hand [1]. There are also the issues of uneven force application and inconsistent quality during the waxing process, so waxing quality cannot be guaranteed. This study therefore developed a robotic system for waxing in auto detailing to increase the efficiency and cost-effectiveness of waxing procedures.
At present, most of the existing literature associated with waxing automation lies in patent reports [2,3,4,5], and there are few technical articles that provide data reference. Thus, we based our design on existing robotic systems developed to perform motions similar to those used in waxing, such as grinding [6], deburring [7,8], and polishing [9,10,11,12,13,14]. The objective of those engineering applications was not to alter the geometric shape of the workpiece, but rather to adjust the surface roughness to within set standards. This means that the need for precision in force control far exceeds that required for position control. In previous studies, industrial robotic arms were equipped with force sensors to achieve force control and surface finishing procedures [15,16]. These same devices were also used to measure surface roughness as an indicator of surface quality [10,17]. The researchers in [18] investigated how various parameters pertaining to grinding and polishing processes, including the size of abrasives, the linear velocity of the belt, contact force, and feed rate affected the final surface roughness of the workpiece. Their results revealed that surface roughness is positively correlated with machining force and feed rate. We therefore adopted waxing force and waxing speed (defined as the amount of time spent working on a given spot) as our primary parameters of waxing quality.
One prerequisite for waxing sheet metal is that the surface roughness of the underlying material must surpass a given threshold. Coating quality can be determined using three indices: durability, appearance, and protection against the environment [19]. Protection can be quantified by measuring the contact angle of water drops on the surface of the workpiece [20], wherein a higher angle means that water drops are less likely to adhere to the surface. The measurement of durability involves observing the severity of rust formation following exposure to test environment for extended durations. Appearance (in terms of the numerical classification of paint coatings) can be gauged using tele-spectroradiometers and multi-angle spectrophotometers [21]. Protection and durability depend on the type of coating, but are beyond the scope of this study. We therefore based our assessment of the proposed automated waxing system solely on appearance.
Most auto detailers manually assess and adjust the appearance quality of waxing jobs. Optical microscopes could be used to evaluate the status of workpiece surfaces [10,17,18]; however, this would be time-consuming and expensive. Thus, for the quantification of waxing surface quality, we opted for non-contact image sensing technology, which is easily implemented and inexpensive to install and operate. In [22], image processing algorithms, including thresholding, noise removal, edge detection, and segmentation were used to identify surface defects such as holes, scratches, coil breaks, and rust with a reasonable degree of accuracy. In [23], multilevel thresholds were used to divide grayscale images into various regions representing the roughness of workpiece surfaces as an indicator of surface quality. Although several advanced machine vision based measurement methods have been proposed for assessing surface roughness recently [24,25,26,27], they all suffer from the drawbacks of computational burden or high-cost hardware. In [28], a computer vision system was integrated with a robotic arm to measure surface defects in polished workpieces before and after processing. However, this automated polishing system did not use image feedback to adjust system parameters, such as force or speed to improve processing quality.
In this study, we incorporated computer vision technology within a robotic arm to enable automated waxing while simultaneously monitoring the quality and adjusting system parameters. We developed an image-based assessment method for waxing quality based on image analysis results of wax jobs performed manually by an auto detailer, and used this feedback information to optimize system parameters. Comparing to the existing image-based assessment methods, the proposed method has the advantages of high computational efficiency and easy implementation for the development of autonomous robotic waxing systems. The feasibility of the image assessment criteria and overall performance of the proposed system were evaluated using an engine hood as a waxing target. A depth camera was used to establish point cloud data of the sheet metal in order to plan the waxing path of the system. Waxing procedures were performed using a three-axis robotic arm with an xyz platform. Waxing force control was achieved using impedance control and feedback information from a multi-axis force sensor at the end of an effector attached to the robotic arm. The waxing speed was adjusted by altering the dwell time along identical waxing paths. Experimental results confirm the effectiveness of the proposed image assessment criteria in detecting differences in waxing quality. Parameter optimization using this feedback information enabled the proposed waxing system to achieve waxing quality on par with that of professional auto detailers.

2. System Description

Figure 1 displays the overall framework of the proposed system. We adopted an xyz platform and a self-built three-axis robotic arm equipped with a waxing motor as the end-effector to wax sheet metal. The xyz platform is responsible for large-stroke movement. The effective travel range of the platform along the x- and y-axes is 500 mm × 500 mm, with a screw lead of 10 mm/rev, whereas the effective travel range along the z-axis is 400 mm, with a screw lead of 5 mm/rev. The resolution of the motor encoders for each axis is 10,000/rev. The three-axis robotic arm served to wax the sheet metal and exercise force control. DC motors (models IG-42CGM and IG32RGM) manufactured by Sha Yang Ye Industrial Co. Ltd. (Taoyuan, Taiwan) were used for the joints. For the end-effector, we used the DC motor (model GBP30) made by Soho Precision Machinery Co., Ltd. (New Taipei City, Taiwan) to meet the need for greater rotational speed and torque. To use a robotic arm to wax the target sheet metal, we first had to understand the external environment, which includes elements such as posture, position, and color. To do so, we installed an Intel RealSence F200 RGB-D camera (Santa Clara, CA, USA) (color camera with the resolution of 1920 × 1080 pixels; depth camera with the resolution of 640 × 480 pixels) on the second axis of the robotic arm (eye-in-hand setup) to estimate the 3D point cloud and world coordinates of the sheet metal to be waxed and capture post-waxing images to assess waxing quality. We installed a HEX-58-RE-400 N force sensor from OptoForce Ltd. (Budapest, Hungary) on the end-effector to measure waxing force and achieve force control. The system uses a computer to change the position command settings of the xyz platform and process the images from the depth camera. The measurement information derived from the force sensor is sent to an Arduino control board via a USB port to control the position of the robotic arm and perform the waxing process.

3. Stereo Vision and Path Planning

To obtain the spatial relationship among the target sheet metal, the camera, and the robotic arm and facilitate the waxing process, we first had to establish a coordinate system to calculate the correct location of the sheet metal and plan the waxing path [29]. We employed a depth camera to obtain the color and depth information of the sheet metal. The parts that do not belong to the sheet metal are removed via image preprocessing, and then the infrared reflection principle of depth cameras is utilized to achieve stereo vision, estimate the point cloud plot of the target, and plan the system’s waxing path.

3.1. Image Processing

Estimation of the sheet metal’s location is based on color images from the depth camera, which are integrated with the information measured from the camera plane to establish the point cloud plot of the sheet metal. However, each point cloud plot contains three channels of image information and depth information for all the coordinates in the images, so the information content is substantial. To reduce the computation time, the color images from the depth camera are subjected to color filtering, thresholding, and morphology processing. The resulting black-and-white images serve as masks to remove the areas that do not belong to the sheet metal in the images, which reduces the system’s computation time and decreases the chance of miscalculations. The mask operation procedure is as shown in Figure 2. Before capturing images, we covered the sheet metal with non-reflective khaki paper to prevent incomplete depth information extraction due to reflections, as shown in Figure 2a. Figure 2b displays the depth image information that was captured. The shade of the color indicates the distance of a spot from the camera. The scattered back spots were due to the diffusion of infrared light caused by the sheet metal, so the camera could not receive the light reflected back from these posts, which resulted in incorrect depth values. Figure 2c presents the mask created via image processing. The white area is the part that the system needs to wax and the target of coordinate conversion, whereas the black area does not need to be processed. After placing this mask on the depth plane, what remains is the part needed by the system, as shown in Figure 2d. Converting the coordinates of the depth information and location of each spot then produces the point cloud plot of the sheet metal.

3.2. Stereo Vision

The locations of the robotic arm, camera, and target are not fixed, so the relationships of the three with the world coordinates have to be established [30]. We installed a depth camera on the second axis of the robotic arm to create a stereo vision system. Based on different descriptive backgrounds, the system has to define three different coordinate systems to describe the relationships among the target, the camera, and the robotic arm. These coordinate systems include the world coordinate system (xw, yw, zw), the camera coordinate system (xc, yc, zc), and the pixel coordinate system (up, vp). The conversion relationship is as shown in Figure 3. The world coordinate system refers to a specific origin in the workspace, so any object in the space has coordinates in relation to the defined origin so that the world coordinate system and robotic arm coordinate system can be defined with an identical origin and directions. The camera coordinate system is a coordinate system that describes other objects based on the location of the camera. Generally speaking, there is no deflection in the light that travels through the center of the lenses. For this reason, we set the center of the lenses as the origin of the camera coordinate system. The origin of the pixel coordinate system is the upper left corner of the image plane and describes the location of the target on the image plane. The system must convert the location of the target into world coordinates so that the robotic arm can track the planned path. The coordinate conversion is achieved by first converting the world coordinates into camera coordinates and then converting them into pixel coordinates.
The conversion relationship between the pixel coordinates and the world coordinates is as presented in Equation (1):
S [ u p v p 1 ] = [ f 0 0 u i 0 f 0 v i 0 0 1 0 ] [ R 3 × 3 T 3 × 1 0 1 × 3 1 ] [ x w y w z w 1 ] ,
where S is a magnification factor; (up, vp) are the target coordinates as described by the pixel coordinate system, and (xw, yw, zw) are the coordinates of the target in the world coordinate system; f, ui, and vi are intrinsic parameters of the camera, which are fixed upon manufacturing; f denotes the focus of the camera; (ui, vi) are the coordinates of the camera center. Note that these intrinsic parameter values can be obtained via a standard camera calibration process [31]. R3×3 and T3×1 are the extrinsic parameters of the camera, including a translation matrix and a rotation matrix. However, extrinsic parameters vary with world coordinate definitions and the position in which the camera is installed. We installed the camera on the second axis of the robotic arm in this study, so the conversion relationships between coordinate systems will change as the robotic arm moves. Thus, a camera coordinate system had to be defined to calculate the extrinsic parameters.
The axes of the robotic arm and the camera coordinate system are as shown in Figure 4. Frames 0 and 1 are the coordinate systems of the first and second axes of the robotic arm and change with motor rotations. Frames 2 and 3, respectively, represent the position of the camera and the center coordinates of the camera and do not change with the posture of the robotic arm. According to this coordinate system, we can derive the conversion relationship in Equation (2), where E is the extrinsic parameter of the camera; Pr and Pc are the locations of the target as described by the coordinate systems of the robotic arm and the camera, respectively. With this coordination matrix, we can convert the target location estimated by the camera coordinate system into coordinates in the robotic arm coordinate system, perform 3D reconstruction using the depth and color planes obtained using the depth camera, and construct a point cloud plot of the target sheet metal to plan the waxing path:
P r = E P c .

3.3. Planning of Waxing Path

The planar dimensions of the target plane on the engine hood to be waxed in this study was 400 mm × 300 mm, as shown in Figure 5. During the waxing procedure, each swipe overlapped the previous swipe to a certain degree so as to prevent missing any spots. In order to evenly cover each spot on the target area, the system divided the waxing path into six paths, and waxing was performed from the bottom upwards. The relative positions of each path were at 40 mm, 105 mm, 170 mm, 235 mm, 300 mm, and 360 mm, as shown in Figure 5, and waxing was performed with a waxing sponge 80 mm in diameter, so the paths overlapped about 1/5 of each other. We conducted curve fitting of the different paths three times based on the side view of the yz plane of the point cloud plot converted from the coordinates (as shown in Figure 6). The results are as shown in Table 1. We then calculated the derivative of the curve equation:
z ( y ) = tan θ ,
with the equation above, we could calculate the angle θ between the horizontal plane and the tangents at different spots on the paths on the surface. With inverse kinematics, we established the posture limitations of the robotic arm to ensure that the wax can coat the sheet metal evenly without angle deviations that prevent the robotic arm from applying force vertically to the curved surface of the sheet metal.

4. Force Control of Robotic Arm

For the robotic arm to maintain constant force while waxing and avoid damaging the sheet metal and the structure of the robotic arm, we adopted an impedance control method that is commonly seen in the literature to control the waxing force of the robotic arm [32]. The concept is to achieve force tracking control with the expected force of the system (Fd) and an impedance control algorithm. Furthermore, if the sponge at the end-effector of the robotic arm is not completely flat on the surface of the sheet metal, it will affect the evenness of force application, increase the chance of structural damage in the robotic arm, and create an additional moment Mx in the robotic arm, as shown in Figure 7. In view of this, we also considered orientation impedance control to keep the sponge flat on the surface based on the external moment. Figure 8 displays a block diagram of the overall force control system. Considering the position terms py and pz of the end-effector in the y- and z-directions and rotation angle θx around the x-axis, x = [py pz θx]T (position of end-effector); x ˙ and x ¨ denote the velocity and acceleration of the end-effector, xd= [pyd pzd θxd]T is the desired position of the end-effector. The mass and moment of inertia matrix of the end-effector is M = diag(My, Mz, Jθ), and the matrices of damping and stiffness between the robotic arm and the environment are B = diag(By, Bz, Bθ) and K = diag(Ky, Kz, Kθ), respectively. The force that comes into contact with the external environment is the external force/torque Fext= [Fy Fz Mx]T. As the force sensor is installed on the end-effector, conversion relationships Fy = Fssinθ and Fz = −Fscosθ exist between the measured waxing force Fs and the coordinate system of the point cloud. Note that θ can be derived using Equation (3). Let the waxing force being applied to the target be Fw; its components Fyd = Fwsinθ, Fzd = −Fwcosθ, and Mxd = 0 can be expressed using expected force vector Fd= [Fyd Fzd Mxd]T.
The principle behind Figure 8 is to use the difference between Fext and Fd to change the position responses of the overall system. When the difference is zero, it means that the force applied by the robotic arm to the sheet metal has reached the system's target value. Furthermore, impedance control is used to alleviate the impact of the force created by the end-effector when it comes into contact with the sheet metal.

5. Image Based Performance Evaluation

To effectively test the waxing results and verify the feasibility of the proposed system, we divided the automotive sheet metal used in our experiment into two parts as shown in Figure 9. The left part was waxed by the robotic arm, while the right part was waxed by an experienced auto detailer, which served as the control group. Using the auto detailer’s waxing results, we applied an external light source for analysis to develop assessment criteria for waxing quality.

5.1. Selection of External Light Source

White point light sources have a condensation effect that shows any fine scratches on paint surfaces. For this reason, we adopted eight 5050 LED strips 50 cm in length for the light source. Each strip contained 36 LED point lights, so there were 144 point lights on either side, as shown in Figure 9. The right shows the reflection results of the waxing done by the auto detailer. As the fine scratches on the surface of the sheet metal have been filled in, the result is an almost mirror-like reflection of the white point light sources, and each point light can be distinguished clearly. In contrast, the fine scratches in the unwaxed area on the left are still very visible, so the reflection results are dim, and some reflected light spots are indistinguishable with only a glimmer remaining.

5.2. Waxing Assessment Criteria

To evaluate the quality of waxing in the waxed areas, we utilized the condensation effects of the round LED point lights. Image preprocessing was conducted on the area waxed by the auto detailer to calculate the radius distribution of the reflected light spots. The preprocessing included grayscale conversion, Canny edge detection [33], and the Hough circle transform [34]. Next, we calculated the radius of each reflected light spot for statistical analysis to establish the assessment criteria of the proposed waxing system.

5.2.1. Image Preprocessing

Once the camera captures a color image of the waxed surface, it must be converted to grayscale for edge detection. Grayscale conversion merges the three-color channels into a single channel, and it is also a way of expressing the brightness values of each pixel. The image processing results are as shown in Figure 10. Figure 10a shows the image captured by the camera of the surface waxed by the auto detailer; Figure 10b displays the image following grayscale conversion; Figure 10c presents the results of edge detection. As can be seen, the halos in the image were completely removed after edge detection. Next, the Hough circle transform determines whether the reflected light spots are uniform in size and presents their distribution conditions.
Figure 11 presents the radius distributions of the reflected LED light spots following the Hough circle transform. All 144 reflected light spots were circular and 4 pixels to 5 pixels in diameter. The maximum error was less than 1 pixel, which means that the auto detailer had filled in the fine scratches on the surface of the sheet metal, so the resulting surface facilitated reflection. The influence of the poor surface conditions on the image analysis results was also reduced.

5.2.2. Establishment of Image Assessment Criteria

To effectively analyze the radius distributions resulting from circle detection, we used the mean and standard deviation to construct the probability density function of the normal distribution and express the degree of dispersion in the radiuses of the reflected LED light spots. The distribution conditions of this function served as the assessment criteria. If the image analysis results of the area waxed by the proposed robotic arm meet the preset threshold, it means that the proposed system has successfully completed the waxing process.
Figure 12 presents the normal distribution function plotted based on the radius distribution from the area waxed by the auto detailer. The mean ( x ¯ ) and standard deviation (σ) of the function are 4.40 and 0.20, respectively. As the standard deviation is very small, it is clear that the function distribution is very close to the mean of the series. This means that the area waxed by the auto detailer was very smooth and could completely reflect the LED light source above to the camera plane; no fine scratches caused light diffusion.
The data above was obtained from a camera angle of 30°.To confirm whether these statistical characteristics can serve as the assessment criteria of the system, we analyzed the area waxed by the auto detailer using images taken from camera angles of 15°, 30° and 45° and used the data from the 30° images for quality screening. In the following experiments, we used an intermediate value of camera angle (30°) to ensure desired reflection effects and avoid unexpected overexposure for consistent image analysis results. Table 2 displays the radius analysis results from the different angles, and Figure 13 shows the corresponding image analysis results. The red circles indicate that the circle meets the threshold standard. As can be seen, the results from the different angles are very similar, which demonstrates that the data can serve as the assessment criteria for waxing quality. We set the mean plus and minus two standard deviations (±2σ) as the acceptable range for reflected light spot radiuses. Waxing results that are closer to the average value of the ones in Table 2 (134 samples) indicate greater quality, which is the objective of the proposed system.

6. Experiments and Performance Analysis

The experiment in this study was divided into two parts: force tracking and wax image analysis. Due to the fact that force tracking performance will affect the evenness of the wax covering the sheet metal while the robotic arm is waxing, the impedance control parameters must be adjusted in advance so that the waxing force of the robotic arm will remain constant and thereby prevent excessive error from affecting waxing quality or damaging the robotic arm or target sheet metal. Once suitable impedance parameters are obtained, a waxing experiment is conducted using these parameters with waxing force and dwell time as the impact factors of waxing quality so as to analyze and investigate the waxing results derived from different parameters and the pre-established assessment criteria.

6.1. Analysis of Force Tracking Performance

The impedance control parameters (M, B and K) in Figure 8 all influence the force tracking error and response speed. By adjusting the impedance parameters, we change the response behavior of the system to achieve force tracking performance with stability, a short settling time, and no overshoot. If the damping ratio (ζ) is set as 1 and the settling time as less than 0.5 s, then, with the transient responses of a standard second-order dynamic system [35], we can derive that the natural frequency (ωn) of the system must be greater than 8 rad/s to fulfill this constraint. In this experiment, we set the natural frequency as 10 rad/s; in terms of Fy, if the mass coefficient selected for the impedance controller is M = 3.5 kg, then the damping and stiffness coefficients are B = 70 N s/m and K = 350 N/m, respectively. We also chose the same set of parameters to achieve the impedance control of components Fy and Mx and used a single target point to perform waxing force tracking of the robotic arm and analyze its response results.
As previously shown in Figure 7, when the system begins the waxing process, the end-effector must maintain vertical contact with the surface of the sheet metal to ensure that the waxing sponge is completely flat on the sheet metal. Thus, we must consider whether Fs, the force that is vertical to the sheet metal, reaches the target value Fw set by the system. At the same time, we must suppress the torque Mx created when the end-effector is not vertical to the sheet metal. Figure 14 presents the response results of force tracking when Fw is set as 15 N. As can be seen, the responses resulting from this set of parameters can quickly reach the target value, the settling time is significantly shorter than 0.5 s, no overshoot exists, and the absolute value of the steady state error is less than 0.3 N. The robotic arm showed some posture deviations during the time from 0.1 s to 0.4 s, which prevented the end-effector from being completely flat on the sheet metal surface and created a torque of approximately 3 N mm. However, posture compensation eliminated this after 0.4 s, so it did not affect the results of the waxing experiment. When Fw is set between 10 N and 20 N, the force tracking results of other fixed values present an identical trend.

6.2. Examination of Waxing Parameters

In this experiment, we divided the area to be waxed on the sheet metal into six paths. Based on the previously established waxing paths, we adjusted the dwell time of each spot to change the waxing speed. We also used the impedance parameters established in Section 6.1 for waxing force control. After the waxing task was completed, we waited for 5 to 10 min for the wax to dry and solidify. To simplify the overall experimental process, a normal cotton buffing cloth was used to remove the excess wax (the dewaxing/polishing process). With the camera angle set at 30°, images of the reflections of external point lights resulting from different experimental conditions were processed, and the radius of each reflected light spot was calculated. The results were then compared with the assessment criteria that we established to verify the feasibility of the proposed system. The accepted range of circle radiuses and the target number are indicated in Table 3, the latter being 134 circles and the threshold set at a difference under 10 circles.

6.2.1. Experiment Conditions and Parameter Settings

We used Carnauba wax, which is the most common type of wax used in the auto detailing industry, for the waxing experiment. The waxing sponge, a fine waxing sponge specifically made for auto detailing, was 25 mm thick to prevent the vibrations caused by waxing from damaging the structure of the robotic arm, facilitate heat dissipation needed for high-speed motor rotations, and reduce the possibility of burning the wax and the sheet metal surface. Table 4 presents the waxing parameter settings of the system. As the force required in most waxing procedures ranges from 10 N to 20 N, we experimented with waxing forces 10 N, 13 N, 15 N, 17 N, and 19 N and chose a motor with rotational speed 850 rpm and torque output reaching 30 N-m for the waxing motor. For dwell time, we experimented with 0.1 s/pt, 0.2 s/pt, 0.3 s/pt, and 0.4 s/pt to facilitate subsequent investigations. In the experiment, we divided the y-direction into 500 equally spaced points (as shown in Figure 6) and used the curve fitting equation and Equation (3) in Section 3.3 to derive the target location of the end-effector, xd. Via inverse kinematics, we obtained the motor position command and then realized point-to-point motion control in the motor with the proportional-derivative (PD) control algorithm. Waxing of each fixed point was then performed from the bottom upwards according to each of the planned paths. The movements between paths were achieved using the xyz platform, coordinated with impedance control to prevent the robotic arm from applying excessive force when it comes into contact with the sheet metal.

6.2.2. Discussion of Waxing Results

Table 5 presents the image analysis results from the five different waxing forces above, and Figure 15 displays the corresponding 3D plot. Observation revealed that, as the waxing force increased from 10 N to 15 N, the number of reflected light spots increased significantly because the wax filled the fine scratches more fully as the waxing force increased. Once the waxing force reached 17 N; however, the number of assessment criteria that met standards stagnated and began to decline. This is because the optimal waxing force for the Carnauba wax chosen for this system is around 15 N; excessive force affects the stacking effect of the wax on the sheet metal, preventing the wax from filling in the fine scratches normally and leading to the diffusion of the reflected light spots on the sheet metal surface. As a result, the image analysis results were poorer than those at 15 N.
As can be seen in Figure 15, the number of circles that meet standards increases with dwell time and peaks at 127 when waxing force 15 N is paired with 0.4 s/pt. This number is less than the result shown in Table 3 by seven circles and within the threshold of 10 circles, which means that the quality is already very close to that achieved by the auto detailer. When the waxing force is greater than 15 N, the image analysis results do not improve as the dwell time increases. The reason is similar; excessive force affects the stacking effect of the wax on the sheet metal, and even if the waxing speed decreases, the wax cannot fill in the scratches on the sheet metal, which reduces the repairing effects of the wax. Thus, the results of increased dwell time paired with waxing forces 17 N and 19 N were poorer than those with waxing force 15 N.

6.3. Waxing Parameter Optimization

The surface formed by the results in Figure 15 presents a convex function. To derive the optimal parameters, we conducted 3D surface fitting with the experimental results from different waxing forces using least squares regression, thereby producing a mathematical model for optimization.

6.3.1. Problem Formulation

The mathematical model for optimization comprises two parts: an objective function and constraints. Both parts are functions of the design variables. We adopted a fifth-order polynomial to fit the surface using least squares regression, as shown in Equation (4). Figure 16 presents the surface fitting results of the experimental data. The R-squared value (R2 = 0.996) indicates that the fitted surface is very close to the distribution of the experimental results, so the quantic equation with two variables served as the objective function of optimization in the proposed system. The various coefficients are as shown in Table 6:
z = f ( x , y ) = n = 0 5 i = 0 n p ( n i , i ) x n i y i .
Table 7 presents the parameter constraints of the optimization problem. As mentioned in the previous section, excessive force is not suitable for the car wax chosen for this system. For this reason, we limited the waxing force to between 10 N and 17 N. In consideration of the overall efficiency of the system, we limited the dwell time to between 0.1 s/pt and 0.5 s/pt so as to prevent overly long dwell time from burning or wearing down the wax, which would then prevent the image analysis results from corresponding to the optimized parameters.

6.3.2. Optimization Results and Verification

We chose constrained nonlinear minimization [36] for parameter optimization and used the Matlab Optimization Toolbox (version 8.0, The MathWorks, Inc., Natick, MA, USA) to derive the optimal combination of waxing parameters. Constrained nonlinear minimization searches for the minimum value of the objective function within the constraints. Thus, we multiplied Equation (4) by −1 to flip the surface, as shown in Equation (5), to facilitate optimization:
Objective   function :   z = f ( x , y ) = n = 0 5 i = 0 n p ( n i , i ) x n i y i .
The optimization process above included six iterations and 26 calculations. The results indicate that waxing force 15 N and dwell time 0.5 s/pt is the optimal parameter combination of the system. The number of reflected light spot radiuses that reached the assessment criterion was 132. Next, we performed the waxing experiment using this parameter combination, conducted image analysis of the results, and compared them with the assessment criteria.
Figure 17 presents the image analysis results of waxing with the optimal parameters. As can be seen in Figure 17a, the resulting reflection is very close to that in the area waxed by the auto detailer. The sheet metal displayed a mirror-like reflection of the LED light spots, and the number of circles that reached the assessment criterion was 130, which is extremely close to the 132 derived in optimization. Table 8 compares the execution results of the system with those of the auto detailer. Although the optimal parameters were used for waxing, it is still noticeable that the execution time needed by the system was far from that needed by the auto detailer. This is because we used a smaller robotic arm and smaller waxing sponge (diameter 80 mm) for the waxing procedure so as to prevent overburdening the xyz platform during operation. In contrast, the sponge used by the auto detailer was 150 mm in diameter, which covered a much larger area, and this resulted in the difference in execution time. The experimental results verify the feasibility of the proposed system and demonstrate that the parameter optimization analysis can further improve actual waxing quality.

7. Conclusions

This study developed a robotic waxing system for auto detailing and an image-based assessment method for performance evaluation and improvement. We adopted a depth camera to obtain the color and depth information of the sheet metal and then established waxing paths using image processing and coordinate conversion. In order for the robotic arm to maintain constant waxing force and better compliance, we adopted a position-based impedance control framework to realize waxing force control. With the waxing results of an experienced auto detailer as the standard, we analyzed images of LED point light reflections on the waxed sheet metal to develop a set of assessment criteria for the waxing system. We performed experiments and parameter optimization analysis with two parameters, waxing force and dwell time, which served as the primary impact factors of waxing quality. The results indicate that the proposed system can successfully achieve the waxing quality of an experienced auto detailer and provide reference for further developments on whole automated waxing systems in auto detailing.

Author Contributions

C.-Y.L. conceived and designed the robotic waxing system, analyzed the results, and wrote the paper. B.-C.H. performed the automatic waxing experiments and the data collection for performance evaluation.

Funding

This research was funded by Ministry of Science and Technology, Taiwan, under grant number MOST-105-2628-E-011-004.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kumar, S. Theories of musculoskeletal injury causation. Ergonomics 2001, 44, 17–47. [Google Scholar] [CrossRef] [PubMed]
  2. Wagner, H.R. Apparatus for Washing and Waxing Cars. U.S. Patent 3447505, 3 June 1969. [Google Scholar]
  3. Falls, J.W. Automatic Car Waxer. U.S. Patent 5076202, 31 December 1991. [Google Scholar]
  4. Hung, V. Car Waxing Machine with Driving Handle. U.S. Patent 6725491, 27 April 2004. [Google Scholar]
  5. Wu, X.; Wu, X.; Guo, H. Combination Automobile Detailing Machine. U.S. Patent 11518864, 11 September 2006. [Google Scholar]
  6. Ning, F.; Cong, W.; Wang, H.; Hu, Y.; Hu, Z.; Pei, Z. Surface grinding of CFRP composites with rotary ultrasonic machining: A mechanistic model on cutting force in the feed direction. Int. J. Adv. Manuf. Technol. 2017, 92, 1217–1229. [Google Scholar] [CrossRef]
  7. Sallinen, M.; Heikkila, T.; Sirvio, M. Robotic Deburring System of Foundry Castings Based on Flexible Workobject Localization; SPIE: Boston, MA, USA, 2001; pp. 476–487. [Google Scholar]
  8. Hsu, F.Y.; Fu, L.C. Intelligent robot deburring using adaptive fuzzy hybrid position/force control. IEEE Trans. Robot. Autom. 2000, 16, 325–335. [Google Scholar] [CrossRef]
  9. Mohsin, I.; He, K.; Cai, J.; Chen, H.; Du, R. Robotic polishing with force controlled end effector and multi-step path planning. In Proceedings of the IEEE International Conference on Information and Automation, Macau, China, 18–20 July 2017; pp. 344–348. [Google Scholar]
  10. Chen, F.; Hao, S.; Miao, X.; Yin, S.; Huang, S. Numerical and experimental study on low-pressure abrasive flow polishing of rectangular microgroove. Powder Technol. 2018, 327, 215–222. [Google Scholar] [CrossRef]
  11. Han, G.; Sun, M. Compound control of the robotic polishing process basing on the assistant electromagnetic field. In Proceedings of the IEEE International Conference on Mechatronics and Automation, Changchun, China, 9–12 August 2009; pp. 1551–1555. [Google Scholar]
  12. Sharma, K.; Shirwalkar, V.; Das, A.P.; Pal, P.K. Robotic polishing of pilger-die. In Proceedings of the Conference on Advances in Robotics, Goa, India, 2–4 July 2015. [Google Scholar]
  13. Jamisola, R.; Ang, M.H.; Oetomo, D.; Khatib, O.; Ming, T.; Lim, S.Y. The operational space formulation implementation to aircraft canopy polishing using a mobile manipulator. In Proceedings of the IEEE International Conference on Robotics and Automation, Washington, DC, USA, 11–15 May 2002; pp. 400–405. [Google Scholar]
  14. Zhang, D.; Yun, C.; Song, D.; Lan, X.; Wu, X. Design and application for robotic polishing system. In Proceedings of the IEEE International Conference on Mechanic Automation and Control Engineering, Wuhan, China, 26–28 June 2010; pp. 2470–2474. [Google Scholar]
  15. Guvenc, L.; Srinivasan, K. Force controller design and evaluation for robot-assisted die and mould polishing. Mech. Syst. Signal. Process. 1995, 9, 31–49. [Google Scholar] [CrossRef]
  16. Lu, Y.Y.; Dong, J.H. The study of force control with artificial intelligence in ceramic grinding process. In Proceedings of the IEEE International Symposium on Intelligence Information Processing and Trusted Computing, Huanggang, China, 28–29 October 2010; pp. 208–211. [Google Scholar]
  17. Liu, H.; Wan, Y.; Zeng, Z.; Xu, L.; Zhao, H.; Fang, K. Freeform surface grinding and polishing by CCOS based on industrial robot. In Proceedings of the International Symposium on Advanced Optical Manufacturing and Testing Technologies, Suzhou, China, 26–29 April 2016. [Google Scholar]
  18. Zhao, T.; Shi, Y.; Lin, X.; Duan, J.; Sun, P.; Zhang, J. Surface roughness prediction and parameters optimization in grinding and polishing process for IBR of aero-engine. Int. J. Adv. Manuf. Technol. 2014, 74, 653–663. [Google Scholar] [CrossRef]
  19. Akafuah, N.K.; Poozesh, S.; Salaimeh, A.; Patrick, G.; Lawler, K.; Saito, K. Evolution of the automotive body coating process—A review. Coatings 2016, 6, 24. [Google Scholar] [CrossRef]
  20. Placido, F.; Birney, R.; Kavanagh, J. Investigation of automotive detailing products by ellipsometry and contact angle analysis. Acta Phys. Pol. A 2009, 116, 712–714. [Google Scholar] [CrossRef]
  21. Gomez, O.; Perales, E.; Chorro, E.; Burgos, F.J.; Viqueira, V.; Vilaseca, M.; Martinez-Verdu, F.M.; Pujol, J. Visual and instrumental assessments of color differences in automotive coatings. Color Res. Appl. 2016, 41, 384–391. [Google Scholar] [CrossRef] [Green Version]
  22. Sharifzadeh, M.; Alirezaee, S.; Amirfattahi, R.; Sadri, S. Detection of steel defect using the image processing algorithms. In Proceedings of the IEEE International Multitopic Conference, Karachi, Pakistan, 23–24 December 2008; pp. 125–127. [Google Scholar]
  23. Besari, A.R.A.; Zamri, R.; Rahman, K.A.A.; Palil, M.D.M.; Prabuwono, A.S. Surface defect characterization in polishing process using contour dispersion. In Proceedings of the IEEE International Conference on Soft Computing and Pattern Recognition, Malacca, Malaysia, 4–7 December 2009; pp. 707–710. [Google Scholar]
  24. Samtaş, G. Measurement and evaluation of surface roughness based on optic system using image processing and artificial neural network. Int. J. Adv. Manuf. Technol. 2014, 73, 353–364. [Google Scholar] [CrossRef]
  25. Khoo, S.W.; Karuppanan, S.; Tan, C.S. A review of surface deformation and strain measurement using two-dimensional digital image correlation. Metrol. Meas. Syst. 2016, 23, 461–480. [Google Scholar] [CrossRef]
  26. Simunovic, G.; Svalina, I.; Simunovic, K.; Saric, T.; Havrlisan, S.; Vukelic, D. Surface roughness assessing based on digital image features. Adv. Prod. Eng. Manag. 2016, 11, 93–140. [Google Scholar] [CrossRef]
  27. Liu, E.; Liu, J.; Gao, R.; Yi, H.; Wang, W.; Suo, X. Designing indices to measure surface roughness based on the color distribution statistical matrix (CDSM). Tribol. Int. 2018, 122, 96–107. [Google Scholar] [CrossRef]
  28. Besari, A.R.A.; Prabuwono, A.S.; Zamri, R.; Palil, M.D.M. Computer vision approach for robotic polishing application using artificial neural networks. In Proceedings of the IEEE Student Conference on Research and Development, Kuala Lumpur, Malaysia, 13–14 December 2010; pp. 281–286. [Google Scholar]
  29. Luh, J.Y.S.; Lin, C.S. Optimum path planning for mechanical manipulators. J. Dyn. Syst. Meas. Control 1981, 103, 142–151. [Google Scholar] [CrossRef]
  30. Tsai, L.W. Robot Analysis: The Mechanics of Serial and Parallel Manipulators; Wiley: New York, NY, USA, 1999. [Google Scholar]
  31. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  32. Jung, S.; Hsia, T.C.; Bonitz, R.G. Force tracking impedance control of robot manipulators under unknown environment. IEEE Trans. Control Syst. Technol. 2004, 12, 474–483. [Google Scholar] [CrossRef]
  33. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–698. [Google Scholar] [CrossRef] [PubMed]
  34. Ballard, D.H. Generalizing the Hough transform to detect arbitrary shapes. Pattern Recognit. 1981, 13, 111–122. [Google Scholar] [CrossRef]
  35. Norman, S.N. Control Systems Engineering; Wiley: New York, NY, USA, 2011. [Google Scholar]
  36. Luenberger, D.G.; Te, Y. Linear and Nonlinear Programming; Springer: New York, NY, USA, 2008. [Google Scholar]
Figure 1. Proposed robotic waxing system.
Figure 1. Proposed robotic waxing system.
Sensors 18 01548 g001
Figure 2. Mask operation. (a) color image; (b) depth image; (c) mask computing; (d) mask to depth image.
Figure 2. Mask operation. (a) color image; (b) depth image; (c) mask computing; (d) mask to depth image.
Sensors 18 01548 g002
Figure 3. Coordinate conversion.
Figure 3. Coordinate conversion.
Sensors 18 01548 g003
Figure 4. Camera coordinate system (Frame 0–1: motor rotation axis, Frame 2: position of camera, Frame 3: center of camera).
Figure 4. Camera coordinate system (Frame 0–1: motor rotation axis, Frame 2: position of camera, Frame 3: center of camera).
Sensors 18 01548 g004
Figure 5. Path planning for the sheet metal. (a) area to be waxed; (b) waxing trajectories.
Figure 5. Path planning for the sheet metal. (a) area to be waxed; (b) waxing trajectories.
Sensors 18 01548 g005
Figure 6. Posture constraint for the robot end-effector. (a) point cloud 3D plot; (b) slope estimation.
Figure 6. Posture constraint for the robot end-effector. (a) point cloud 3D plot; (b) slope estimation.
Sensors 18 01548 g006
Figure 7. Use of robotic arm in waxing force control. (a) normal waxing force control; (b) end-effector not flat on sheet metal.
Figure 7. Use of robotic arm in waxing force control. (a) normal waxing force control; (b) end-effector not flat on sheet metal.
Sensors 18 01548 g007
Figure 8. Block diagram of robot force control system in this study.
Figure 8. Block diagram of robot force control system in this study.
Sensors 18 01548 g008
Figure 9. Comparison of LED lighting results.
Figure 9. Comparison of LED lighting results.
Sensors 18 01548 g009
Figure 10. Results of image preprocessing (area waxed by auto detailer). (a) captured RGB image; (b) RGB to grey image; (c) after edge detection.
Figure 10. Results of image preprocessing (area waxed by auto detailer). (a) captured RGB image; (b) RGB to grey image; (c) after edge detection.
Sensors 18 01548 g010
Figure 11. Histogram of reflected LED light spot radiuses (area waxed by auto detailer).
Figure 11. Histogram of reflected LED light spot radiuses (area waxed by auto detailer).
Sensors 18 01548 g011
Figure 12. Normal distribution of reflected light spot radiuses (area waxed by auto detailer).
Figure 12. Normal distribution of reflected light spot radiuses (area waxed by auto detailer).
Sensors 18 01548 g012
Figure 13. Image analysis results from different camera angles. (a) circle detection results from camera angle 15°; (b) circle screening results from camera angle 15° (130 samples); (c) circle detection results from camera angle 30°; (d) circle screening results from camera angle 30° (135 samples); (e) circle detection results from camera angle 45°; (f) circle screening results from camera angle 45° (137 samples).
Figure 13. Image analysis results from different camera angles. (a) circle detection results from camera angle 15°; (b) circle screening results from camera angle 15° (130 samples); (c) circle detection results from camera angle 30°; (d) circle screening results from camera angle 30° (135 samples); (e) circle detection results from camera angle 45°; (f) circle screening results from camera angle 45° (137 samples).
Sensors 18 01548 g013
Figure 14. Force tracking control results of the robotic waxing system (Fw = 15, M3×3 = diag(3.5, 3.5, 3.5), B3×3 = diag(70, 70, 70), K3×3 = diag(350, 350, 350)).
Figure 14. Force tracking control results of the robotic waxing system (Fw = 15, M3×3 = diag(3.5, 3.5, 3.5), B3×3 = diag(70, 70, 70), K3×3 = diag(350, 350, 350)).
Sensors 18 01548 g014
Figure 15. Experimental results of different waxing parameters—3D plot.
Figure 15. Experimental results of different waxing parameters—3D plot.
Sensors 18 01548 g015
Figure 16. Fifth-order surface fitting results.
Figure 16. Fifth-order surface fitting results.
Sensors 18 01548 g016
Figure 17. Waxing image analysis following system parameter optimization (waxing force 15 N, dwell time 0.5 s/pt). (a) waxing results comparison (proposed system vs. car detailer); (b) captured image with the proposed system; (c) Hough circle transform results (130 circles).
Figure 17. Waxing image analysis following system parameter optimization (waxing force 15 N, dwell time 0.5 s/pt). (a) waxing results comparison (proposed system vs. car detailer); (b) captured image with the proposed system; (c) Hough circle transform results (130 circles).
Sensors 18 01548 g017aSensors 18 01548 g017b
Table 1. Curve fitting results from the side view of sheet metal (Paths 1 to 6).
Table 1. Curve fitting results from the side view of sheet metal (Paths 1 to 6).
# of PathPath Function (after Cubic Curve Fitting)
Path 1 z 1 ( y ) = 2.86 × 10 6 y 3 + 0.0055 y 2 2.135 y 126.0215
Path 2 z 2 ( y ) = 1.02 × 10 5 y 3 + 0.0223 y 2 15.037 y 3154.08
Path 3 z 3 ( y ) = 5.49 × 10 6 y 3 + 0.0117 y 2 6.9988 y 1134.38
Path 4 z 4 ( y ) = 7.69 × 10 6 y 3 + 0.0167 y 2 10.716 y 2051.36
Path 5 z 5 ( y ) = 7.29 × 10 6 y 3 + 0.0156 y 2 9.7094 y 1746.57
Path 6 z 6 ( y ) = 1.17 × 10 5 y 3 + 0.0256 y 2 17.241 y 3607.75
Table 2. Radius distributions of reflected light spots from different camera.
Table 2. Radius distributions of reflected light spots from different camera.
Samples in Different Range ( x ¯ = 4.40 , σ = 0.20 )
Camera View (θ)Total Samples x ¯ ± σ x ¯ ± 2 σ
15°14496 samples130 samples
30°14498 samples135 samples
45°14497 samples137 samples
Table 3. Waxing image assessment criteria adopted in this study.
Table 3. Waxing image assessment criteria adopted in this study.
Estimation of RadiusCriterion Definition
Mean value ( x ¯ )4.4 pixel
Standard deviation (σ)0.2 pixel
Range of circle radius x ¯ ± 2 σ
Average result from Table 2134 samples
Threshold10 samples
Table 4. Waxing experiment parameter settings.
Table 4. Waxing experiment parameter settings.
Waxing Motor SpecificationsWaxing Force (From 10 to 19 N)Dwell TimeExecuting Time for all Paths (Approx.)
24 V—850 rpm
(30 N-m)
10 N0.1 s/pt
0.2 s/pt
0.3 s/pt
0.4 s/pt
5~6 min
10~11 min
15~16 min
20~21 min
13 N
15 N
17 N
19 N
Table 5. Experimental results of different waxing parameters.
Table 5. Experimental results of different waxing parameters.
Waxing ForceDwell TimeExecuting TimeEstimated Result (Total: 144 Samples)
10 N0.1 s/pt5 min 28 s70 samples
0.2 s/pt10 min 35 s82 samples
0.3 s/pt15 min 24 s89 samples
0.4 s/pt20 min 19 s93 samples
13 N0.1 s/pt5 min 36 s92 samples
0.2 s/pt10 min 21 s102 samples
0.3 s/pt15 min 18 s110 samples
0.4 s/pt20 min 30 s115 samples
15 N0.1 s/pt5 min 23 s110 samples
0.2 s/pt10 min 29 s116 samples
0.3 s/pt15 min 34 s121 samples
0.4 s/pt20 min 36 s127 samples
17 N0.1 s/pt5 min 31 s112 samples
0.2 s/pt10 min 33 s117 samples
0.3 s/pt15 min 29 s115 samples
0.4 s/pt20 min 24 s110 samples
19 N0.1 s/pt5 min 36 s115 samples
0.2 s/pt10 min 21 s117 samples
0.3 s/pt15 min 18 s115 samples
0.4 s/pt20 min 30 s113 samples
Table 6. Coefficient table of surface equation.
Table 6. Coefficient table of surface equation.
p00−7917p21−76.58p0418140
p103081p12−1923p500.01935
p01−3663p03−8127p41−0.008048
p20−471.6p40−1.324p32−4.448
p11915.7p312.224p2312.24
p029295p22159.7p14−255.9
p3035.63p13−83.42p05−11600
Table 7. Constraints of optimization.
Table 7. Constraints of optimization.
Optimization Parameters of the Waxing SystemLower and Upper Bounds
Force (N)From 10 N to 17 N
Dwell time (s/pt)From 0.1 s/pt to 0.5 s/pt
Table 8. Comparison of waxing results (proposed system: waxing force 15 N, dwell time 0.5 s/pt).
Table 8. Comparison of waxing results (proposed system: waxing force 15 N, dwell time 0.5 s/pt).
Executing TimeEstimated Result (Total: 144 Samples)
Car detailerAbout 15 min134 samples
The waxing system25 min 16 s130 samples

Share and Cite

MDPI and ACS Style

Lin, C.-Y.; Hsu, B.-C. Development and Performance Evaluation of Image-Based Robotic Waxing System for Detailing Automobiles. Sensors 2018, 18, 1548. https://doi.org/10.3390/s18051548

AMA Style

Lin C-Y, Hsu B-C. Development and Performance Evaluation of Image-Based Robotic Waxing System for Detailing Automobiles. Sensors. 2018; 18(5):1548. https://doi.org/10.3390/s18051548

Chicago/Turabian Style

Lin, Chi-Ying, and Bing-Cheng Hsu. 2018. "Development and Performance Evaluation of Image-Based Robotic Waxing System for Detailing Automobiles" Sensors 18, no. 5: 1548. https://doi.org/10.3390/s18051548

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop