A Representation Generation Approach of Transmission Gear Based on Conditional Generative Adversarial Network

: Gear reliability assessment of vehicle transmission has been a challenging issue of determining vehicle safety in the transmission industry due to a signiﬁcant amount of classiﬁcation errors with high-coupling gear parameters and insufﬁcient high-density data. In terms of the preprocessing of gear reliability assessment, this paper presents a representation generation approach based on generative adversarial networks (GAN) to advance the performance of reliability evaluation as a classiﬁcation problem. First, with no need for complex modeling and massive calculations, a conditional generative adversarial net (CGAN) based model is established to generate gear representations through discovering inherent mapping between features with gear parameters and gear reliability. Instead of producing intact samples like other GAN techniques, the CGAN based model is designed to learn features of gear data. In this model, to raise the diversity of produced features, a mini-batch strategy of randomly sampling from the combination of raw and generated representations is used in the discriminator, instead of using all of the data features. Second, in order to overcome the unlabeled ability of CGAN, a Wasserstein labeling (WL) scheme is proposed to tag the created representations from our model for classiﬁcation. Lastly, original and produced representations are fused to train classiﬁers. Experiments on real-world gear data from the industry indicate that the proposed approach outperforms other techniques on operational metrics.


Introduction
Given the significant increase in the number of vehicles in recent years, there has been an increase in attention paid towards the safety evaluation of vehicles in both the industrial sector as well as in academia. The transmission is a core component of vehicles, providing power transfer and direction change. It is also a major source of vehicle failure and noise. In transmission, gears are the main parts and are significant factors in transmission safety. As a result, gear reliability assessment is a significant and direct indicator in terms of vehicle safety.
Regarding vehicle transmission gears, the data acquisition process is complicated and expensive, resulting in insufficient gear data collection, which increases the difficulty in evaluating gear reliability [1]. Traditionally, the solutions of gear reliability assessment can be categorized into two parts: model-driven and data-driven techniques. For the former, gear reliability cannot be calculated directly; however, it can be estimated by two gear safety factors-bending and contact safety factors [2]. To address both of these factors, the mechanical structure and operation process of transmission gears are both modeled to calculate gear safety factors [3][4][5][6][7][8]. Obviously, the accuracy of gear reliability assessment is determined by the precision of modeling. A lot of assumed conditions are required in the modeling process to guarantee solvability and availability. However, these conditions, with regard to theoretical physical equations, do not reflect transmission gears under realistic operating conditions, leading to objective deviations [9]. For instance, the autoregressive model (AR) [10] evaluates data using the autocorrelation function, but is vulnerable to data noise. The moving average (MA) [11] model assesses data according to the weighted summation of present and past inputs, which is necessary to ensure the difference stationarity of data. However, gear data are unable to converge everywhere. The autoregressive moving average (ARMA) [12] model uses the least square method to appraise current data. ARMA requires linear data, yet gear data are usually nonlinear; therefore, model-driven methods are not effective in gear reliability assessment.
For the latter, instead of assumed conditions, intrinsic relations between gear reliability and monitored parameters are learned from collected data. Data-driven techniques are motivated by implicit and explicit characteristics of collected data without constraint conditions and specific models, which are widely used to evaluate gear reliability. For example, a hybrid data-driven method combined by support vector data description and extreme learning machine is proposed to monitor the unhealthy status of wind turbines gears [13]. Meanwhile, a deep structure of a denoising autoencoder is designed to assess wind turbines gears through analyzing the monitored vibration data [14,15], and an adaptive signal resampling model is established for the fault diagnosis of wind turbine gears with current signals [16,17]. A time series-histogram method is presented to predict the remaining useful life of aero-engine gears by extracting features of event data [18]. In these methods, sample generation is the key; however, on the one hand, oversampling methods that are widely used to produce sufficient samples by learning the location relationship of the original data [19], like random oversampling [20] and synthetic minority over-sampling technique [21], generate samples inside the ranges of the original data without consideration of the correlations among dimensions of the original data and instead deemed as independent. On the other hand, because the initial values of gear parameters in a test rig are empirically based on an engineer's experience, collected real-world gear data are highly dense. Furthermore, one gear parameter has a high coupling relationship with other parameters [22,23]. The distance between any two samples in the gear data of vehicle transmission has lost correlation with their corresponding reliability, which indicates that general distance measurement (e.g., Euclidean distance and cosine distance) in oversampling methods cannot work on gear data effectively [24]. Thus, oversampling methods are unable to produce reliable gear data for vehicle transmission.
Rather than producing samples based on location calculations of the original data in oversampling methods, another type of generative technique learns the inherent distribution from the original data and creates new samples under the estimated distributions. GAN [25] is an attractive deep generative architecture that estimates the probability density or mass functions in a minimax game. A generator and a discriminator in the game confront each other. The generator tries to forge samples as real as the original data to confuse the discriminator. At the same time, the discriminator aims to distinguish the produced samples from the original data. With the estimated distribution, new samples from the generator are produced over the whole data space and are not constrained in the ranges of the original data.
Nevertheless, there exist three issues for expanding gear data using GAN to improve the effectiveness of gear reliability assessment. First, the training process of a traditional GAN in estimating the distribution without considering any class information may cause an over-generation of one class and an under-generation of other classes. This could cause imbalance issues and decrease the precision of reliability assessment [26]. Second, according to the high density and coupling of collected gear data, GAN collapses easily and produces samples with close properties [27]. Finally, produced samples from GAN and its variations have no label and cannot be used in reliability evaluation operation.
To address these issues, a novel approach is presented to acquire sufficient data for transmission gears, which improves the classification accuracy of gear reliability. In the model, we establish a CGAN-based model by implementing label information to produce representations with high diversity. According to this model, we can transform the unsupervised training process into a supervised training process by adding the label information of the sample, which will greatly improve the generation ability of the network, and not only learn the mapping between collected gear data and the degree of gear reliability, but also reflect the real situation of the gearbox under actual working conditions through the generated samples. Additionally, we propose a Wasserstein labeling scheme to label generated representations according to the characters of gear data. This labeling method based on Wasserstein distance can avoid the situation that there is no duplication between different categories of generated samples. By measuring the probability density relationship between the generated sample set and the real sample set, the generated sample can be correctly classified, thereby producing better label information. The main contributions of this paper are summarized as follows.

1.
A novel approach is proposed as a pretreatment for the gear reliability assessment of vehicle transmissions. With an estimated global distribution, this approach produces credible transmission gear representations to expand existing space and raises the efficiency of the gear reliability assessment.

2.
In the CGAN-based model, label information gains access to the distribution estimation to generate representations guided by label distribution. Furthermore, we introduce a mini-batch strategy to randomly sample original and forged representations from the generator and send these representations into the discriminator for differentiation, strengthening the diversity of generated representations. 3.
The proposed WL scheme names the generated representations based on the measurement between these representations and Wasserstein barycenter of gear reliability degrees. This scheme offsets the unlabeled ability of GAN and provides available labels for classifiers.
The rest of this paper is organized as follows. In Section 2, a brief view of GAN and gear reliability is introduced. In Section 3, we describe the proposed approach in detail. Experimental results are presented and discussed in Section 4. Finally, in Section 5, we conclude the paper, detailing the advantages of the proposed approach.

Transmission Gear Reliability
In the production process of vehicle transmissions, one of the main reasons for gear tooth failure is excessive bending stress at the tooth root fillet of the loaded gear. These stresses often shorten the overall life of the gear, and under the action of peak load, the gear teeth suddenly break, causing engineering accidents. The reliability of gears is conventionally mainly determined through gear reliability tests to obtain the probability statistical distribution law of gear bending fatigue life and strength. However, in engineering applications, conventional gear fatigue life tests often require a lot of manpower, material and financial resources. They take a long time and often involve commercial secrets; therefore, there are no public datasets related to gear tooth data. It was announced that most of the research on gear teeth is based on simulation systems such as automatic dynamic analysis of mechanical systems (ADAMS).
Meanwhile, after collecting the consent of relevant companies and the data of gear teeth in the production process of vehicle transmissions, we analyzed these data. According to the actual situation, the reliability of the gear teeth is determined by two safety factors: bending safety factor and contact safety factor, and divided into four levels (i.e., higher, high, standard and low reliability). The relations between gear reliability and safety factors are shown in Figure 1. In each degree of gear reliability, the minimum values of gear reliability and safety factors are both defined, and one safety factor having a smaller value than another other safety factor means that the degree is higher. We preprocessed the existing gear data with four degrees of gear reliability as class 1 to 4 in the simulation.

Generative Adversarial Networks
GANs are unsupervised and semi-supervised learning techniques used as a means of producing generative models. The purpose of a generative model is to explore the statistical distribution of training/testing data, and forge samples from this distribution. The distribution of training data (i.e., real data) is denoted as , and the distribution of produced data is denoted as . The advantage of GAN compared to common generative models, e.g., variational autoencoder (VAE) and PixelRNN, is that and in GAN are not required to be explicit descriptions. Instead, GANs can obtain the distribution implicitly by training two networks in competition [28].
These two networks are the discriminator D and the generator G. G aims to learn the representation of real data ∼ by discovering a mapping from a noise variable ∼ ( ) to real data, where ( ) is the distribution of the noise and usually initialized as simple distributions (e.g., Gaussian and uniform distribution). Meanwhile, D tries to distinguish real data and produced data ∼ after receiving them simultaneously. Many learning objectives have been proposed for adversarial training, such as those based on F-divergences [29,30]. For the standard cross-entropy GAN, the critic outputs a probability of a data-point being real and optimizes the following objective: The generator and the critic are both parameterized by deep neural networks and trained via alternating gradient updates. Because adversarial training only requires samples from the generative model, it can be used to train generative models with intractable or ill-defined likelihoods [31]. Hence, adversarial training is likelihood-free and in practice, it gives excellent performance for tasks that require data generation. However, these models are hard to train due to the alternating minimax optimization and suffer from issues such as mode collapse [32].

Materials and Methods
To address reliability assessment of vehicle transmission gears without mechanic modeling and particular conditions, the proposed approach contains two components to prepare for the evaluation of gear reliability with existing gear data (illustrated as in Figure 2). First, we need to process the gear data collected from the factory. When this kind of highly dimensional and small amount of data are directly put into the generator for generation, it will often cause non-convergence of the generation model. Therefore, this step is indispensable. In this step, our input is the original dataset with the same feature dimensions but different amounts of data under different categories. After data processing, the amount of data in different categories will remain balanced to reduce the imbalance between the data. It is convenient for the generation model.

CGAN-based model
Afterwords, we input low-dimensional noise with label information into the generator of our CGAN-based model. After the convolutional layer mapping, we get the generated data with the same dimension as the original dataset, then put them together into the discriminator. With the help of back propagation, the weight of the convolutional layers will be updated. In order to raise the diversity of generated representations in accordance with the characteristics of the gear data for the vehicle transmission, the mini-batch scheme is designed to sample certain amounts of original data and produced representations instead of training the discrimination with all representations. When the training is completed, our CGAN-based model will generate data with label information. We take out the generated data and remove the labels.
Finally, we input the generated data without label information into the k-nearest neighbor (KNN) model for classification based on Wasserstein distance. Through this model, we can make reasonable annotations for generated data, regardless of whether they coincide in different categories. The flow-process diagram of our proposed method is shown in Figure 3. In addition, the definition of symbols in our article is scattered. We have listed those central and confusing symbols in Table 1.  Compared with image data, structured data are close to orthogonal between the various features contained in the data; therefore, for the CGAN-based model, the discriminator cannot pass the gradient back to the generator according to the result of the generator for the iterative update in the processor. Before data generation, we extract and integrate the same features of all data to form a new dataset, as shown in Figure 4, and after our model generates new features, we will merge these features to form new data again; as a result, we obtain an original dataset, with its objects { | = 1, 2, ..., }, its number of instances , with its attribute set { | = 1, 2, ..., }, its number of attributes , with its categories { | = 1, 2, ..., } and its number of categories . For the input of the network, a new dataset is constructed, and use to represent this dataset, with its objects { | = 1, 2, ..., }, its number of instances equal to , with its attribute set { | = 1, 2, ..., }, its number of attributes equal to , with its categories { | = 1, 2, ..., } and its number of categories equal to ; ( , , ) addresses real gear data and ( , , ) represents produced data from our trained generator.

Model Structure
Due to the discrete nature of gear tooth data, when a single GAN is used to generate gear tooth data, the boundaries between different types of generated data are often blurred according to unconditional constraints, especially when similar to gear wheels. The gear data forms a dataset with a small gap between classes, so our generative model uses CGAN with conditional information.
G and D in our CGAN-based model are both a neural network with multiple hidden layers. The minimax game is denoted as: Mathematically, the solution of this game is to learn the joint probability function other than the probability function in GAN.
In terms of G, noise variables and label information are inputs and forged data are outputs. is the iteration time. In the process of generating CGAN, the noise is used as an input for the purpose of making the network random, which can generate very complex distributions. The goal is to make it close to the distribution generated by real data. For the gear data we used, although we have combined these discrete data into a continuous distribution, the randomly added noise will not purposefully make our data distribution more in line with the real data distribution. When the dimension of the input noise is smaller than the amount of data contained in a single gear dataset, this noise will make our data more confusing. Therefore, in the process of CGAN generation, we placed a restriction on the noise according to the characteristics of the data, that is, the dimension of the noise must not be lower than the amount of data contained in a single gear dataset. For D, gear samples from and are inputs, and the computation result of objective function (Equation (3)) is an output.
The details of our CGAN-based model are in Algorithm 1.

Mini-Batch Scheme
Typically, G would collapse due to the parameter settings, leading to the production of representations in one model. Thereby, to guarantee the diversity of generated representations, we adopt the mini-batch strategy on D as shown in Figure 5. Suppose that is the mini-batch size, we stochastically sample representations from as ( , , ) and . Instead of accessing the entire and , we load into D to compare with part of generated representations .
where and are the number of representations in and . The advantage of this strategy is that is confronted with different real representations in every iteration. The generated representations at each iteration is compared with a portion of real representations, which enhances the diversity of generated representations.  (4); while < t do Input: + + ; 8: Train D using Adam optimizer; 9: Compute D in Equation (3); 10: = + 1. end 11: Feedback average to the generator; 12: = + 2 * . end 13: Generate representations using the trained G. 14: Obtain qualified without label.

Network Optimization
Due to the high density and small amount of gear data, the generated representations are easily trapped into a group of a similar sample to the original. Although the minibatch scheme is designed to deal with the overfitting issue, one category of gear data has an extremely small size, which aggravates the non-convergence problem. Therefore, instead of gradient descent method for ∇ Φ( G) and ∇ Φ( D) in GAN, we implement the Adamax optimizer to train both G and D, which computes exponential moving averages of gradients {∇ Φ( G) , ∇ Φ( D) } and Hessian matrices { Φ( G) , Φ( D) } and provides a simpler range for the upper limit of the learning rate. Exponential decay rates are controlled by coefficients { 11 , 12 , 21 , 22 } ∈ [0, 1), which are updated at each iteration. 1 and 2 represent the learning rates of the gradient with first order, and Hessian matrices are defined as: where and are initial learning rates of G and D, respectively. The estimations of the first moment gradient and Hessian matrices at iteration are given as: where D and G are gradients, computed with the following formulas: Φ(D) and Φ(G) are updated as follows.
The 11 , 12 , 21 , 22 and are hyperparameters initialized with empirical evidence in the simulation. In the training process of the network, the reason why we use Adamax instead of the Adam that comes with CGAN is because Adam's ability to adjust the learning rate changes based on a simpler range for the upper limit of the learning rate, as shown in (9). The definition of this range allows our network to process discrete data without modifying the initialization deviation, and has a more flexible adjustment method and a smaller magnitude of change.

Wasserstein Labeling Scheme
Considering that from our CGAN-based model has no labels, it is necessary to tag the generated representations for classification. With regards to one specific property of gear data [22], general distance measurement (e.g., Euclidean and cosine distance) cannot display well on transmission gear data. To label generated gear representations, the Wasserstein labeling scheme is proposed with a three-step process. First, a Wasserstein barycenter for gear data in each category of gear reliability is discovered by the k-nearest neighbor Wasserstein clustering algorithm. Then, the Wasserstein distance between each sample in the generated data and each Wasserstein barycenter is estimated by the Wasserstein critic. The details are shown in Algorithm 2.

Wasserstein Barycenter
We use Π( , ) to represent all possible joint distribution combinations of and distribution combinations for each joint distribution that can be sampled, and the distance between the samples can be calculated. Under the joint distribution , the expected value ( , )∼ [|| − ||] of the sample to the distance is obtained. At this time, we can define the earth-mover (EM) distance between the two samples as the lower bound of the expected value, as shown in the following formula. The solution of finding the Wasserstein barycenter is transformed into optimizing the following formula: where is the order of the Wasserstein distance, and is initialized as = 1. Thus, is termed as in this paper. Let Ω be the transmission matrix in EM distance and be the distance matrix, Integrate Equation (12) into Equation (11), the optimal problem is changed into: Suppose ( , ) as discrete description of a distribution, is the value of samples and is the frequency of samples. Real gear data in each reliability are denoted as { 1 , 2 , 3 , 4 }, respectively. To resolve Equation (13), the Sinkhorn iteration is designed to obtain Ω of : Then, the location of the Wasserstein barycenter in is solved as: Accordingly, a set of Wasserstein barycenters for gear reliability evaluation with existing gear data is obtained as { | = {1, 2, 3, 4}}.

Labeling Generated Data
Wasserstein distance between two datasets, take and for example, is given by . . = = (16) where = (− · ( , )), ∈ and ∈ ; is a 1 * matrix with all values of 1. Suppose as th sample in , the Wasserstein distance between and Wasserstein barycenter of th reliability of transmission gear is denoted as: After estimating the distance from each generated sample and each Wasserstein barycenter, the reliability with the minimum distance is used to tag the generated sample. * = min ( , ) where * is the label of the generated sample .

The Necessity of Data Processing
When analyzing unbalanced data, we use a set of gear data for specific analysis. In this dataset, occurrences of each reliability is 40/44/97/31 from class 1 to 4 and the dimension of each sample is 85. It can be seen that high reliability having the maximum samples is the most common operating condition. Although the number of samples between different categories is quite different, we can find that all samples have the same dimensional characteristics. The difference between gear data and other data is that the same dimensions represent the same characteristics. Therefore, we reorganize data in the same category according to feature dimensions. We also select them through the mini-batch scheme and form four categories with the same number of samples. For example, the samples in the first category are 85 and each has 40 dimensional features. Samples in the second class are also 85 and each has 40 dimensional features. We will discard the extra samples and repeat part of the data. In this method, CGAN improves performance in generating different classes of samples.
In order to analyze the performance of data processing, we observe the values of loss functions by using both our CGAN-based model and traditional GAN for transmission gear data. The results are shown in Figure 6. When data processing is not used in the training process, loss values fluctuate with training epochs, which means that GAN and CGAN are unable to learn the right mapping relationship directly from the transmission gear data. After data processing is used in CGAN, the loss values are steadily declining with training epochs, which shows that the imbalance class issues are lightened.

Algorithm Performance
The method consists of Algorithms 1 and 2. In the CGAN-based model in Algorithm 1, we can generate new samples without label information in different class spaces. Then, generated data are filtered according to the Wasserstein distance in Algorithm. In order to explore the stability and convergence of these two algorithms, we observe the values of loss functions [33] by using both our CGAN-based model and traditional GAN for transmission gear data. From Figure 7a, discriminator loss values in Algorithm 1 are gradually decreasing with a consistent trend, whereas those values of the traditional GAN have uncertain fluctuations. This indicates that the convergence of our CGAN-based model is more stable on gear data than the traditional GAN. Furthermore, to avoid training randomness, we trained our CGAN-based model five times. Loss function curves of discriminator loss (as illustrated in Figure 7b) have parallel trends. The loss values are decreasing with training epochs, which shows that our model has effective stability and required convergence for transmission gear data.

Simulation Settings
Experimental gear data are provided by a transmission company named Qingshan industry, and is a real-world gear dataset including 212 items. In the dataset, the occurrences of each reliability is 40/44/97/31 from class 1 to 4 and the dimension of each sample is 85. It can be seen that high reliability having the maximum samples is the most common operating conditions.
The CGAN-based model in our approach was simulated on a single NVIDIA TITAN Xp GPU. All simulations ran on an Intel i7-6800K CPU. The output layer of the generator and the input layer of the discriminator both contain 212 neurons, manipulated by the size of the gear parameters. Other sizes of the generator are {128, 1024, 256} and that of the discriminator are {256, 1024, 128}. The input of the CGAN generator is a random noise and its dimension is 212. Through the convolution layers of the generator, random noise was mapped into higher dimensions and judged by discriminator. After noise passes through generator convolutional layers, which are {128, 1024, 256}, noise becomes a 256 * 1 feature vector. The convolutional layers that are designed as {256, 1024, 128} are used for feature extraction. At the end, it passes through a softMax layer for classification.
Moreover, weight initialization in these two networks follows the distribution The learning of parameters in the training process is optimized based on the gradient descent algorithm. This algorithm needs to assign an initial value at the beginning of training. In our experiment, one of the problems was how to choose the way of random initialization. Considering the softMax layer in our network, we decide to use uniform distribution initialization. Hence, there is a good distinction between neurons in different layers [34].

Model Parameter Analysis
We investigate the effect of a hyperparameter, i.e., mini-batch size , described in the mini-batch strategy of our CGAN-based model. For a clear mathematical description, = / is sampled from {60%, 70%, 80%, 90%, 100%}. The number of generated samples in each degree is obtained with different . Simulation results are shown in Figure 8, in which the diversity is preferable when = 80%. While the value of is small, the training dataset contains less samples in the smallest group of reliability degree (e.g., Degree 4), leading to the inadequate learning of this degree and the generation of infrequent samples in Degree 4. Interestingly, when the value of increased to 90% and 100%, the imbalance among the number of all degrees becomes greater, which suffers from mode collapse in GAN. As such, in this paper, to ensure the diversity of generated data from our approach, is set to 80%.

Comparisons of Different Labeling Strategy
To validate the effectiveness of proposed WL scheme, we compared the approach with other labeling strategies: a labeling scheme based on cosine and Euclidean distance. Two different classifiers (i.e., decision tree and multilayer perceptron) and four indicators (i.e., precise, recall, F-measure and G-mean) were implemented to evaluate the performance of expanded data. The results are shown in Figure 9, where red lines are averaged results within 10 runs and blue boxes are performance intervals of these 10 runs.  Note that the proposed CGAN-WL plays effectively on gear data in all indicators even while using different classifiers. This discovery provides further verification on the theory that gear parameters have strong coupled relations between each others while general distance measurements (e.g., Euclidean distance) cannot display effectively on gear data.

Comparisons of Different Generation Methods
To observe the credibility of generated samples from our approach, we compared our approach to six other techniques for expanding gear data. Five compared techniques were sourced from the imbalanced-learn API toolbox [35]: random oversampling (ROS), synthetic minority oversampling technique (SMOTE), adaptive synthetic sampling (ADASYN), synthetic minority oversampling technique-edited nearest neighborhood (SMOTEENN) and synthetic minority oversampling technique-Tomek link (SMOTETomek). Furthermore, we make the contrast of giving labels to produced samples from GAN using WL (GAN-WL). In the simulation, (representing the number of nearest neighbors that identify if minority samples are on the spot) is set to 15, and (addressing the number of nearest neighbors that are synthetic samples) is initialized as 10.
Both simple metrics (i.e., recall (R) and precise (P)) and comprehensive metrics (i.e., Gmean (G-M) and F-measure (F-M)) are used to score the results of gear reliability assessment. Higher values of these indicators imply better performance. These metrics are denoted with the given Equation (19), where is the amount of negative instances that are misclassified; is the amount of positive instances that are classified properly; is the amount of positive instances that are misclassified; is the amount of negative instances that are classified properly.
Considering the specifications of various classifiers, we implement three kinds of classifiers in the scikit-learn toolbox [36] to discover the performance of gear reliability evaluation using different classification techniques: decision tree (DT), random forest (RF) and multilayer perceptron (MLP). Four-fold cross-validation is used where real-world gear data are stochastically segmented into four folds. One fold is used for testing and the remaining three folds are used for training. In order to decrease the influence of sampling randomness in the training-testing process, we ran all classifiers 10 times, and all metric values were averaged over these 10 runs. Table 2 examines the performance of classification with various generative techniques. Clearly, the proposed CGAN-WL outperforms other compared techniques both using the three classifiers. It can be seen that other compared techniques display unstably with different classifiers. Take GAN-WL for example, it operates well with DT but works bad with MLP. SMOTE and its variations (i.e., ADASYN, SMOTETomek and SMOTEENN) based on Euclidean distance measurement are not capable to discovery the relation between gear data and its reliability. To further prove the enhancement in statistical ways, we implement both Welch's T-test [37] and Mann-Whitney U test [38] to assess the significance of the improvement. Comparisons between the proposed approach and compared techniques are observed with three classifiers. With a significance level at 0.05, test results of all metrics are illustrated in Table 3. The values of all metrics are all lower than 0.5, which indicates the proposed approach observably outperforms the compared methods.   In order to verify the effectiveness of our method more effectively, we ran our test on a standard dataset from UCI (called refractive errors dataset (RED)) [41]. The aim of this dataset is to study the impact of personal lifestyle and genetics on eye refractive errors. This dataset is gathered from forms filled out by 467 individuals. The first sheet contains the information of 210 people suffering from eye refractive errors and the second sheet contains information of the remaining 257 participants, which had a healthy eye condition. This discrete data are similar to our gear dataset, and the test results are shown in Table 4. We can see that other compared techniques are unstable with different classifiers. It can be found that compared to other methods, CGAN-WL can learn better conditional mapping from discrete dataset, and use WL to make reasonable annotations.

Conclusions
Gear reliability evaluation was found to be remarkable at guaranteeing the safety of vehicle operations. This paper proposed a CGAN-WL approach to deal with gear reliability assessment when collected gear data were insufficient. To find the prior value of hyperparameter in the approach, the diversity of generated representations with different was observed. Furthermore, in order to demonstrate the effectiveness of our approach, different labeling schemes were conducted to work on gear data. Simulation results revealed the effectiveness of our approach and validated the characteristics of gear data that show that the relations among gear parameters were tight. Finally, with three different classifiers in the experiments, CGAN-WL outperformed other popular generative techniques in both simple and complicated metrics. Statistical tests further proved the significant improvement of our approach. In the future, we intend to work on the transmission gear data of electric vehicles. The transmission of an electric vehicle adopts a fixed rate in gear data. Gear clearance tends to larger, and noise generated by the transmission is weakened. So the overlap in different classes of gear data becomes smaller, which brings great challenges to traditional identification methods.

Institutional Review Board Statement:
The study did not involve humans or animals.
Informed Consent Statement: Informed consent was obtained from all subjects involved in the study.