Next Article in Journal
Innovative Multi-Target Estimating with Clutter-Suppression Technique for Pulsed Radar Systems
Next Article in Special Issue
Robust Semi-Supervised Traffic Sign Recognition via Self-Training and Weakly-Supervised Learning
Previous Article in Journal
Design of Multi-Mode Antenna Array for Use in Next-Generation Mobile Handsets
Previous Article in Special Issue
An Angle Error Compensation Method Based on Harmonic Analysis for Integrated Joint Modules
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Clustering Algorithm to Solve Driving Cycle Prediction Problems Based on Unbalanced Data Sets: A Chinese Case Study

School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(9), 2448; https://doi.org/10.3390/s20092448
Submission received: 7 February 2020 / Revised: 18 April 2020 / Accepted: 21 April 2020 / Published: 25 April 2020
(This article belongs to the Special Issue Data Processing of Intelligent Sensors)

Abstract

:
Vehicle evaluation parameters, which are increasingly of concern for governments and consumers, quantify performance indicators, such as vehicle performance, emissions, and driving experience to help guide consumers in purchasing cars. While past approaches for driving cycle prediction have been proven effective and used in many countries, these algorithms are difficult to use in China with its complex traffic environment and increasingly high frequency of traffic jams. Meanwhile, we found that the vehicle dataset used by the driving cycle prediction problem is usually unbalanced in real cases, which means that there are more medium and high speed samples and very few samples at low and ultra-high speeds. If the ordinary clustering algorithm is directly applied to the unbalanced data, it will have a huge impact on the performance to build driving cycle maps, and the parameters of the map will deviate considerable from actual ones. In order to address these issues, this paper propose a novel driving cycle map algorithm framework based on an ensemble learning method named multi-clustering algorithm, to improve the performance of traditional clustering algorithms on unbalanced data sets. It is noteworthy that our model framework can be easily extended to other complicated structure areas due to its flexible modular design and parameter configuration. Finally, we tested our method based on actual traffic data generated in Fujian Province in China. The results prove the multi-clustering algorithm has excellent performance on our dataset.

1. Introduction

Car parameters are an important indicator to guide consumers in purchasing cars, and they quantify users’ driving experience as comparable figures. However, the direct use of other countries’ test standards will lead to some car parameters that cannot accurately reflect the actual driving experience of consumers in China’s automotive test projects, which can cause the information to be false or misleading and thus lead to confusion among consumers. At present, developed countries such as Europe, the United States, and Japan generally accept driving cycle standards and energy consumption/emission certifications that are adapted to their respective driving cycles. Due to the vast territory and complicated structure area, more customized algorithms are required for driving cycle prediction in China.
Car fuel consumption is one of the indicators of most concern for car buyers. However, in the past, with the New European Driving Cycle (NEDC), published in Europe for energy consumption/emission evaluation and widely used in China, it is difficult to accurately illustrate the actual performance of cars in the various regions of China. Thus, the gap between the manufacturer’s published data and the actual test data of the users is extremely obvious, which seriously decreases the overall buying experience of users. The reason for the above result is China’s geographical location and economic development are vastly different from those in Europe, not only in many various car types but also in the complicated terrain of China. There are obvious three-level ladders in geographical trends and the economic level is high in the west and low in the east. The number of cars in developed cities is large but the urban area is generally small, as shown in Figure 1a. Therefore, the congestion delay index [1,2] is higher, as shown in Figure 1b. At the same time, the average elevation in Tibet, Yunnan and other places is very high, and there are problems, such as insufficient gasoline combustion. The above various reasons explain why the NEDC driving cycle evaluation system proposed by Europe cannot reflect the actual performance of cars in China.
In recent years, the rise of big data and artificial intelligence has provided a strong theoretical foundation for solving driving cycle prediction problems with deep personalization, and computer advancements have provided the necessary computing power.
In the driving cycle prediction research field, the predecessors were based on K-means clustering algorithms for unsupervised clustering analysis. K-means is a partitioned clustering technique that is widely known for its low computational complexity. Representative algorithms include hard k-means and fuzzy k-means, but these approaches always suffer from bad performance and are not suitable for skewed data distributions (i.e., unbalanced data). Even if the input data has different cluster sizes, they usually produce clusters of relatively uniform size, which is called "uniform effect" [3]. As we know intuitively, during normal vehicle driving, people usually keep driving at a medium speed (10–90 km/h). There are relatively few cases of low-speed driving (0–10 km/h) and ultra-high-speed driving (>90 km/h). Therefore, in a normal vehicle driving data set, the amount of data is obviously tilted toward the medium speed section. Therefore, the “uniform effect” largely affects the driving cycle prediction results, resulting in a large gap between the driving cycle map parameters and the actual vehicle driving experience.
To address the above limitations, we draw inspiration from ensemble learning to design the multi-clustering algorithm, a novel unsupervised learning algorithm. Usually, boosting, bagging and stacking are three effective algorithms used in the field of ensemble learning. This method is considered to have good performance in the processing of unbalanced data sets [4]. In the supervised learning algorithms of machine learning, the goal setting for the stable model that performs well in all aspects is often so ideal that it’s hard to achieve, and sometimes we can only get multiple models with preferences (weakly supervised models perform better in some respects). Ensemble learning combines multiple weak monitoring models in order to get a better and more comprehensive strong monitoring model. The underlying idea of ensemble learning is that even if a weak classifier gets a prediction wrong, other weak classifier can correct the error. In this paper, our main contributions are summarized as follows:
  • A multi-clustering algorithm framework is proposed based on further improvement of the stacking algorithm framework. The algorithm framework integrates k-means-type, K-means++, DBSCAN and other clustering algorithms, and uses stacking to strengthen each individual model. Multi-clustering can solve potential vulnerabilities such as data imbalances and random initial states. We performed multiple experiments on the matching order and crossover method of the algorithm to obtain the best results. Through experiments, we have verified that our proposed multi-clustering framework performs well in the data set and offers a relatively high improvement over other algorithm.
  • We use T-SNE algorithm instead of the principal component analysis (PCA) algorithm for data dimensionality reduction. In the field of traditional driving cycle map research, the PCA method is used for feature extraction and dimensionality reduction. This approach creates the problem of unclear division between categories. Therefore, we select the T-SNE algorithm instead of the PCA algorithm to reduce the dimensionality of high-dimensional features. After a comparison between the two different dimension reduction algorithms, it can be clearly seen from the experimental results that T-SNE performs better given the characteristics of our working data set.
The remainder of this paper is organized as follows: In Section 2, we investigate the methods used by predecessors in the field of driving cycle prediction and compare them with our research. In Section 3, we first introduce the overall flow of the experiments, and then detail the theoretical derivation and actual implementation of the basic algorithm used in the experiments. In Section 4, we describe the proposed multi-clustering algorithm, show the internal structure of the algorithm, and theoretically explain and deduct our proposed algorithm. In Section 5, we elaborate on the implementation of the actual experiments and a comparison of the effects of each algorithm with the multi-clustering algorithm. Furthermore, we conduct a detailed analysis of the experimental results in Section 6. Finally, we give the conclusions in Section 7.

2. Related Work

2.1. Unbalanced Datasets

In life, questions about the classification of unbalanced data sets are ubiquitous. For example, pathological analysis, traffic accident analysis, crime analysis, and so on. Predecessors have reported many achievements in the field of unbalanced dataset research [4,5,6,7].
Some researchers use sampling methods to balance data sets. Chawla et al. proposed the oversampling data processing method SMOTE [8]. Tahir et al. handled data imbalances by down sampling [9]. Later, Wang et al. proposed a problem that cannot be solved by simply using up sampling or down sampling, and tried to balance the dataset by using multiple sampling [10].
With the advancement of technology, some works have proposed a new loss function that calculates the loss of positive and negative classes separately, making the machine more sensitive to small classes [11]. To better measure losses, Khan et al. proposed a dynamic loss function. The parameters of the loss function are constantly modified with the iteration of the training to obtain better prediction results [12].
At the same time, some scholars are studying the problem of unbalanced data in the fields of unsupervised learning and ensemble learning [4]. The most widely used ensemble learning algorithms are AdaBoost [13] and bagging [14] whose applications in several classification problems have led to significant improvements [15]. At the same time, many scholars have combined data preprocessing and ensemble learning to propose algorithms such as SMOTEBoost [16], RUSBoost [17], SMOTEBagging [18] and EasyEnsemble [19].

2.2. Driving Cycle Prediction

Many scholars have conducted research in the field of driving cycle prediction [20,21,22]. They have proposed different prediction methods from the perspectives of engineering and mathematics. Huang et al. using kinematic segment classification to establish a basic driving cycle database including 4704 different transmission matrices. Based on the inter-station driving characteristic equations and the basic driving cycle database, a driving cycle prediction model was developed, generating drive cycles by the iterative Markov chain for assigned bus lines [23]. Chen et al. proposed an online correction algorithm based on a backup control strategy and a fuzzy logic controller [24]. Huang et al. proposes an intelligent multifeatured statistical approach to automatically discriminate the driving conditions of the hybrid electric vehicles (HEVs) [25]. Chinese researchers have also discussed the differences in vehicle driving characteristics between China and Europe and the United States, and have studied the roads in China. Wang et al. used 11 driving parameters to develop a formal cycle for China’s national conditions, and compared it with the standards proposed by Europe and the United States [26]. Li et al. used principal components analysis and cluster analysis of the features of driving speed, idling, maximum speed, acceleration and travel mileage, as well as their distribution features in three typical cities in China [20].
While the above approaches for driving cycle prediction have been proven effective and used in many countries, these algorithms are difficult to achieve the effect in china with complex traffic environment and increasingly high traffic jams. Meanwhile, we found the unbalanced data will have a huge impact on the performance when building driving cycle maps, and the car parameters generated by above algorithms deviate quite a lot from actual ones. In order to address the above limitations, this paper proposes a new algorithm framework for the driving cycle prediction problem from the perspective of machine learning and unbalanced data analysis. An experimental study was conducted in a Fujian (China) vehicle data set, and a driving cycle was proposed for Fujian. After our comparative analysis, we found that this problem has been studied from a fresh perspective and has a good performance. The multi-clustering algorithm is simple and easy to implement, and it is suitable for migration to different regions for practical applications, especially in China, where the terrain is complex and the urban development gap is large.

3. Preliminary

In this section, we firstly briefly introduce the general situation, classification and basic processing of the driving cycle map generation. Then, we describe in detail how to build the working condition map of our experiments and provide technical support for the subsequent experiments. At the same time, we also carried out theoretical derivation and verification of the algorithm used in the experiments, and this theoretical analysis proved the correctness of the subsequent experiments.
Driving cycles are often used to determine a vehicle’s 100-kilometer fuel consumption, pollutant emissions, and for the development, evaluation, or evaluation of traffic risks of new vehicles. It is a core technology in the automotive-related field.

3.1. Brief Description of the Driving Cycle

Vehicle driving cycles are also known as car operating cycles. The refer to the characteristics of a certain type of vehicle (such as light vehicles, buses, heavy vehicles, etc.) under certain circumstances. They are usually displayed in the form of a speed-time curve (usually the total time is 1800 seconds, but there is no standard limit). Because the vehicle driving cycle must reflect the actual running state of the car, we usually need to collect driving data in an actual environment with sensors and use various mathematical, physical, computer methods to process, screen and analyze it.
Driving cycle reflect the kinematics of road driving, which is an important and common basic technology for the automotive industry. As the main benchmark for the calibration and optimization of various performance indicators of automobiles, it is also the basis of vehicle energy consumption/ emission test methods and limit standards. For example, it can be used to determine vehicle fuel consumption and pollutant emissions per 100 km, to determine whether the car really saves energy as advertised. In addition, as a core technology of automobile-related fields, it is also widely used in the development of new cars, performance evaluation, and assessments of traffic risk [27].

3.2. Driving Cycle Classification

In the field of driving cycle research, the maps are generally divided into two categories: transient conditions and modal conditions, according to the characteristics of the speed-time curve.
FTP75, proposed by the United States based on FTP72 in 1975, is a typical transient condition cycle. Such working condition maps are directly constructed from real road traffic data, with the advantage of easily reflecting the various driving behaviors, such as rapid acceleration and emergency stops, but they cannot fully illustrate the detailed information of driving behaviors during the test, because the driving behavior is complicated and in fact there are many details.
The NEDC driving cycle proposed by Europe has all the basic characteristics of modal conditions. It synthesizes a representative working condition map by performing modal analysis on real traffic data. The modal condition map does not have too many sudden changes in acceleration, and each state has a longer holding time, which can be better used for scenes such as factory tests of the automobile. In general, the velocity-time image of the driving cycle map is a continuous smooth curve while the velocity-time diagram of the modal map is a line graph.

3.3. Method for Constructing Transient Map

When constructing the transient driving cycle map, we mainly analyze and construct it according to the flow shown in Figure 2, and use the short stroke method for data analysis [28]. The short stroke method is a common method for constructing driving conditions. The short stroke, also known as the kinematics segment, refers to the range of speeds between the start of the idling state and the start of the next idle state, including an idle period and a period of exercise. The method first divides the data into multiple kinematic segments. Then the computer method is used to explore the correlation between each motion segment, to select the most representative short-stroke segment. Finally, according to the 36 features and data ratios, such as speed, acceleration, displacement, and load ratio, the short-stroke segments are stitched together to obtain the final transient working condition diagram.

3.3.1. Build Step

  • First, we analyze the data given, study the data content, quality, analyze the causes of noise and find ways to denoise the data. The data set is processed into a high quality, low error state by means of completion, screening and the like.
  • Secondly, find the appropriate way and rule to divide the kinematics after the data screening is completed. Compare the different division methods, and choose the division method with the most practical application meaning and the most analytical competitiveness for later use.
  • Thirdly, we extract and process all the features in the data set, merge the redundant information, delete the invalid information, and finally get 36 features. Then, through the T-SNE algorithm, the features are reduced in dimension to facilitate subsequent calculations.
  • Fourthly, we first try to cluster the group feature data using the K-means algorithm. Several improved algorithms are proposed for the problem of K-means algorithm. In order to concentrate the advantages of several algorithms and reduce the model bias, a multi-clustering algorithm based on unsupervised integrated learning is proposed. Then the clustering result is converted into a condition map by Markov chain method.
  • Finally, we compare and verify the correctness of the proposed driving cycle diagram and compare it with the working condition diagram proposed by the predecessors. At the same time, the effect of our proposed multi-clustering algorithm framework on traditional algorithms is improved, and the final analysis leads to the conclusions.

3.3.2. Drop Feature Dimension

In the field of driving cycle prediction research, most of the existing papers use the principal component analysis (PCA) algorithm [29] to reduce the dimensionality of high-dimensional features.
The main job of PCA is to sequentially search for a set of mutually orthogonal axes from the original high-dimensional space. The first axis created is selected from the one with the largest variance in the original data. The second coordinate axis selects the direction in which the original data variance is the largest in the plane orthogonal to the first coordinate axis. The third coordinate axis selects one direction in which the original data variance is the largest in a plane orthogonal to the first and second coordinate axes, and iterates according to this step until the k (target dimension) coordinate axes are selected. In the coordinate axis found in this way, the top k coordinate axes with the largest variance are selected, and the variances of the remaining coordinate axes are very small and can be ignored. Therefore, PCA achieves data feature dimensionality reduction by ignoring the dimension with very small variance [30].
PCA is a linear dimensionality reduction algorithm. However, it cannot explain the complex polynomial relationship between features. Therefore, we use T-distributed stochastic neighbor embedding (T-SNE) algorithm to reduce the dimension of the 36 features data.
The T-SNE algorithm is an improved version of the stochastic neighbor embedding (SNE) algorithm. It is based on the probability distribution of random walks on the neighborhood graph to find the structure within the data. The dimensionality reduction effect is excellent, but the calculation amount is large, and the system resources are high [31,32]. Our experimental team has cluster computing resources that are enough to support the T-SNE algorithm, so we use the more effective T-SNE algorithm for dimensionality reduction. The T-SNE algorithm maps data points to probability distributions by affine transformation. The main process is divided into two parts. Firstly, construct a probability distribution between high-dimensional objects. This makes similar objects more likely to be selected and dissimilar objects less likely to be selected. Secondly, the probability distribution of these points is constructed in a low dimensional space such that the two probability distributions are as similar as possible. From an algorithm perspective, the overall T-SNE algorithm is as follows:
  • The first step: Using the SNE algorithm, starting by converting the high-dimensional Euclidean distance between data points into a conditional probability of similarity, the conditional probability between the calculated data points x i and x j is as shown in Equation (1):
    p j | i = exp ( | | x i x j | | 2 2 σ i 2 ) k i exp ( | | x i x j | | 2 2 σ i 2 )
    where σ i is the Gaussian variance centered on the data point x i
  • The second step: For the low-dimensional corresponding points y i and y j of the high-dimensional data points x i and x j , a similar conditional probability can be calculated as shown in Equation (2):
    q j | i = exp ( | | y i y j | | 2 ) k i exp ( | | y i y j | | 2 )
  • The third step: minimize the difference between p j | i and q j | i , and use the gradient descent method to minimize the Kullback-Leibler Divergence (KL-Divergence) distance [31]. The cost function uses a heavy-tailed distribution to alleviate congestion problems and SNE optimization problems.

3.3.3. Kinematic Fragment Clustering

A road segment may contain different traffic characteristics, but a short trip contains fairly specific traffic characteristics. In order to meet the characteristics of China’s mixed traffic, different short trips need to be classified according to the road traffic characteristics. Short trips with the same traffic characteristics are grouped into one category, and segments of different categories are combined into typical representative conditions. Therefore, it is necessary to construct a classification model to classify different short strokes according to driving conditions (using characteristic parameter representations).
Clustering is a commonly used classification method in data analysis. It is a method of grouping samples according to data similarity without given a classification category. Because the category of each short trip is unknown, the short trip is classified using an unsupervised clustering method. Commonly used clustering methods mainly include classification (split) method, analytic hierarchy process, density-based method, grid-based method and model-based method. In addition, there are semi-supervised, spectral clustering, graph clustering and other clustering methods.

4. Multi-Clustering Algorithm

In this section, we will introduce in detail the structure and actual implementation steps of the multi-clustering algorithm proposed by us, explain the advantages and feasibility of the model from a theoretical perspective, and provide theoretical support for subsequent experiments.

4.1. Model Construction

Considering that, as mentioned above, the use of a model alone cannot achieve a good clustering effect, and will be affected by the skew of the data set distribution, making the classification inaccurate. Therefore, we consider designing a ensemble model to further improve the clustering effect of a single model based on the principle of using integrated learning. The model construction of unsupervised ensemble learning is mainly considered from the following two perspectives:
  • Multi-Dataset: Drawing on the idea of bagging algorithm, we need to divide the kinematics dataset of the whole vehicle into multiple sub-datasets for clustering. However, if the sub-data sets are randomly selected in all kinematic segments directly, the kinematic segment categories in each sub-data set are unevenly distributed, and the clustering results are abnormal. Therefore, we adopted a method of randomly extracting sub-data sets according to the average speed.
  • Multi-model: Based on the idea of boosting algorithm, the clustering results of the two single models described above are fused in a certain way (such as voting method, weighted average method, etc.) when we get the clustering results of multiple data sets on multiple models.
In this regard, we need to consider how to fuse multiple clustering results. We have adopted the following two options.
  • Directly integrate the drawings obtained by each model.
  • Drawing on the idea of stacking algorithm, the kinematic segments of each model with the most typical features are grouped into a new set of kinematics segments, and then clustered again to obtain the merged work images.
Among them, the first model fusion method is to first match the corresponding points of the driving cycle map obtained by each model, and then perform a weighted average and other fusion operations. Although this method is intuitive and effective, it is found in the actual operation that even if the driving cycle maps obtained by different models are not significantly different from the perspective of the overall feature vector, the corresponding points tend to be different or even have many phase offsets (one working condition). The acceleration segment of the graph corresponds to another condition map and becomes the deceleration segment. Thus, direct weighted averaging causes problems in the trend of each small kinematic segment of the drawing. Therefore, this paper uses the second model fusion method to build multi-clustering algorithm. The proposed multi-clustering overall model framework is shown in Figure 3.

4.2. Algorithm Implementation

We first give the detailed steps of the overall multi-clustering algorithm, as Algorithm 1 shows:
  • Classify according to the average speed of short-stroke, and then randomly extract sub-data sets. Short-stroke is divided into three categories with average speeds below 10 km/h, above 70 km/h, and between the two. The number of fragments in the first two is significantly less than the third. Each time a sub-data set is extracted, it is extracted at a ratio of 1:1:2 so that the distribution of the sub-data set is relatively even. Each sub-data set is then clustered separately using two models to obtain the results.
  • Select a strong feature dataset. For the clustering results of each model, the correlation analysis method is used to select the most representative kinematic fragments in each category.
  • Cluster the strong feature dataset again. The data set has obvious data characteristics (small intra-class differences and large inter-class differences), so the effect of clustering will be clearer. Thereby, the final clustering result is obtained.
Then, we present the correlation analysis algorithm used in the algorithm, as Algorithm 2 shows.
Through Algorithm 1, we can obtain short strokes with strong features that better characterize the driving characteristics of the current city. The driving cycle map obtained by clustering and splicing through strong feature segments will be more representative than the results obtained by analyzing all the data. A comparison of the specific algorithm effects will be given in the next section.

5. Experiments

In this section, we will detail the actual data analysis, processing steps, and results and visualize and display indicators and results, and compare the different benefits and effects of different algorithms. The optimal algorithm is chosen as the basis for the classification and aggregation of short stroke segments.
Algorithm 1: Multi-Clustering Algorithm
Input:
  All short-stroke clips. The short-stroke segment is written as X = { x 1 ,   x 2 ,   x 3 ,   ,   x t } .
  Where t is the total number of short-stroke segments. X represents the characteristic parameter P i = { T i ,   T a i ,   T d i ,   ,   R P A i ,   R m a x i ,   ,   I m i n i ,   I m i } of the short-stroke segment.
Output:
  Strong Feature Set Y .
1: Sub data   sets   X = { X 1 ,   X 2 ,   X 3 ,   ,   X m } I n i t i a l i z e _ r a n d o m ( n )
2: for x X 1 , X 4 , ,   X 3 p + 1           p N do
3:   K m e a n s + + ( x ,   k = 4 ) ,   D B S C A N ( x , e p s )
4:   F i n a l _ s c o r e C o r _ s c o r e ( K m e a n s + + ,   x ) × w 6 p + 1 + C o r _ s c o r e ( D B S C A N ,   x ) × w 6 p + 2
5:   ( Initialize )   T e m p { }     R a n k _ d o w n ( F i n a l _ s c o r e ( x ) )
6:   Y 3 p + 1 = { Temp [ 0 ] ,   Temp [ 1 ] ,   ,   Temp [ 0.8 n ] }
7: end for
8: for x X 2 , X 5 , ,   X 3 p + 2 do
9:   K m e a n s t y p e ( x ,   k = 4 ) ,   D B S C A N ( x , e p s )
10:  F i n a l _ s c o r e C o r _ s c o r e ( K m e a n s t y p e ,   x ) × w 6 p + 3 + C o r _ s c o r e ( D B S C A N ,   x ) × w 6 p + 4
11:  T e m p { }     R a n k _ d o w n ( F i n a l _ s c o r e ( x ) )
12:  Y 3 p + 2 = { Temp [ 0 ] ,   Temp [ 1 ] ,   ,   Temp [ 0.8 n ] }
13: end for
14: for x X 3 , X 6 , ,   X 3 p do
15:  K m e a n s + + ( x ,   k = 4 ) ,   k m e a n s t y p e ( x , k = 4 )
16:  F i n _ s c o r e C o r _ s c o r e ( K m e a n s + + ,   x ) × w 6 p + 5 + C o r _ s c o r e ( k m e a n s t y p e ,   x ) × w 6 p
17:  T e m p { }     R a n k _ d o w n ( F i n a l _ s c o r e ( x ) )
18:  Y 3 p = { Temp [ 0 ] ,   Temp [ 1 ] ,   ,   Temp [ 0.8 n ] }
19: Y   { Y 3 p ,   Y 3 p + 1 ,   Y 3 p + 2 }
20: end for
Algorithm 2: Correlation analyze Algorithm
Input:
  The result of different cluster algorithm Y . Each short-stroke segment has a characteristic parameter P i = { T i ,   T a i ,   T d i ,   ,   R P A i ,   R m a x i ,   ,   I m i n i ,   I m i } .
  Each cluster center point parameter Q j = { T j ,   T a j ,   T d j ,   ,   R P A j ,   R m a x j ,   ,   I m i n j ,   I m j } .
Output:
   C o r _ s c o r e ( P ,   Q ) .
1: for y Y do
2:   C o v ( P i ,   Q j ) = E [ ( P i E [ P i ] ) ( Q j E [ Q j ] ) ] = E [ P i Q j ] E [ P i ] E [ Q j ]
3:   V a r ( T ) = E { [ T E ( t ) ] 2 }
4:   Correlation _ Score ( Qj , Pi ) = c o v ( P i , Q j ) V a r ( P i ) V a r ( Q j ) = c o v ( P i , Q j ) σ P × σ Q
5: end for

5.1. Data Preprocessing

In this section, we analyze and evaluate data from the perspectives of motion data processing and analysis, geographic data processing and analysis. From the perspective of actual data, it provides a reasonable basis and methodology for subsequent algorithms. Among them, the sports data mainly includes related information of motor sports, such as vehicle motion data record completion, idle duration analysis, kinematic segmentation, etc.; geographic data mainly analyzes vehicle geographic trajectories from latitude and longitude information.

5.1.1. Vehicle Motion Data Processing

We used the data of three light vehicles in Fuzhou collected for 7 days and 168 h for continuous research. The sampling interval was 1 s. The basic data statistics are shown in Table 1. After analysis, we found that in the 7-day record, the three cars had data losses due to equipment failure, GPS signal disappearance, etc., and the amount of missing data was not equal. The missing amount of the vehicle 1 is the smallest, and the missing amount of the vehicle 2 is the largest. The specific missing information is shown in Table 1. For this phenomenon, we use the interpolation method for data completion, and the interpolation method is calculated according to Equations (3) and (4):
k = b x b y n + 1
a ( i ) = b x + k × i
where b x represents the last value before the missing segment. B y represents the first value after the missing segment. K represents the slope. a ( i ) is the insertion value at the i t h of the missing position.
After statistics, as shown in Figure 4, more than 70% of the missing points are less than 10s in length, so we have completed the missing fragments below 10 s. The number of records after completion is shown in Table 1.
After completing the record, we segmented according to the continuous idle duration. We try to segment in 5, 10, 20, 30, 40, 50, and 60 s, and the number of segments is shown in Figure 5. We then select three kinematic speed- time image records separately, and divide them according to different continuous idle durations, and get 9 images as shown in Figure 6.
The three maps in each row are the speed-time images of the same segment of the travel record divided by different division rules. Figure 6a,d,g are divided into continuous idle speeds of 20 s. In Figure 6b,e,h the pictures are divided into 10 s at continuous idle speed, and in Figure 6c,f,i the data is divided into 5 s segments according to the continuous idle speed. According to the continuous idle time of 20 s, a kinematic segment with a duration of several hundred seconds can be obtained, and a kinematic segment with a duration of several tens of seconds can be obtained by dividing by 5 s. Of course, the length of the kinematics segment is related to the speed of the vehicle and the road conditions. In general, we can see that the kinematics segmentation according to the continuous idle duration of 5 s can obtain a complete and independent velocity-time curve with representative features, which will greatly facilitate the accurate operation of the subsequent clustering algorithm. Therefore, we choose to divide the kinematics according to the continuous idle time of 5 s. The length and quantity distribution of the kinematics after the division are shown in Figure 6. The number of kinematic segments after segmentation is shown in Table 1.

5.1.2. Geographic Data Processing

After combining the time information of the data set with the latitude and longitude information, we use the Tencent Map API to draw the motion track information as shown in Figure 7. As can be seen from the figure that the three vehicles providing the data set have driving records on urban roads and expressways, and the number of miles driven on each road is basically the same. However, the speed on the city road is slow, so it takes a long time.

5.2. Feature Selection and Processing

In the field of machine learning, features are the only way for machines to obtain information. Good feature selection, construction, and processing can greatly improve the accuracy of machine judgment. In this section, we will show you how to handle features in a dataset.

5.2.1. Feature Building

In the study of vehicle operating conditions, it is very important to extract effective features as learning and clustering parameters. We designed and constructed 36 digital features based on the characteristics of the data set and the features commonly used in driving cycle prediction. Parameters are shown in Table 2. Among the 36 parameters we designed, there are five time-related parameters, 14 speeds, acceleration-related parameters, 10 vehicle stress situations and power-related physical parameters. These parameters can more comprehensively represent the vehicle’s operating conditions, vehicle driving conditions, and help with subsequent feature processing.

5.2.2. Feature Processing

The similarity calculation formula (Euclidean distance, cosine distance, etc.) of the clustering algorithm is affected by the dimension and abnormal data. Therefore, we need to correct the outliers and normalize them. In our study, the abnormal point is corrected using the long lower bound, as shown in Equation (5):
x = { x 3 δ           ( x u < 3 δ ) x + 3 δ           ( x + u < 3 δ )
Taking the acceleration density distribution as an example, as shown in Figure 8, the Figure 8a is the acceleration distribution map. The red point in the Figure 8b is the abnormal point and needs to be corrected.
Figure 9 shows the comparison results before and after the correction of the outliers. As can be seen from the figure, the distribution of the corrected data is more uniform over the interval. If the outliers are not processed, normalization will cause most of the data to be compressed into a small interval, causing the distance formula to fail.

5.3. Feature Dimension Reduction

Since we have 36 features for the kinematics segment, the dimensions are high and some of them have redundancy features. In order to improve the efficiency of our algorithm and more accurately reflect the important features of kinematics, we need to use algorithms to reduce dimensionality of high-dimensional features. The PCA algorithm is a linear dimensionality reduction algorithm. Although PCA works well in many areas, in some complex areas, such as complex polynomial relations, nonlinear transformation methods are needed to reduce dimensionality.
First, we use the Python sklearn.decomposition library to analyze the contribution rate and cumulative contribution rate of the 36 features proposed in the previous section. Obtain reliable indicators of dimensionality reduction from the perspective of data analysis. It is generally believed that when the cumulative contribution rate of the principal component exceeds 85%, this part of the principal component contains most of the information in the original indicator.
Through statistical analysis, we obtain the principal component contribution rate and the cumulative contribution rate as shown in Table 3 (arranged in descending order). As can be seen from Table 3, the cumulative contribution rate of the components M 1 , M 2 , and M 3 has reached 85%, so we choose to keep the three components of M 1 , M 2 , and M 3 , and reduce the dimension to 3 dimensions. Then we used the T-SNE and PCA algorithms to try in the vehicle 1 dataset. The comparison proves that the T-SNE algorithm performs better on our dataset.
Figure 10 shows the data distribution characteristics when the high-dimensional feature of the vehicle 1 data set is reduced to three dimensions using the PCA algorithm. It can be seen from the observations of two different angles that after the dimensionality reduction using the PCA algorithm, the feature points are not grouped clearly, there are many outliers, and there is no clear classification center. Figure 11 shows the data distribution characteristics when the high-dimensional feature of the vehicle 1 data set is reduced to three dimensions using the T-SNE algorithm. The three graphs from left to right show the changes in the dimensionality reduction effect as the number of iterations increases. It can be seen from Figure 11a,b that the algorithm clearly divides the feature points into two left and right aggregates, and the similar points are close to each other, and there are clear isolation bands between the different types of points. Figure 11c is an algorithmic effect of iteration 1000 times, and the results given by the algorithm have basically stabilized. The algorithm explicitly divides the feature points into four categories, each with a relatively clear separation band, and the outliers are very few relative to the PCA algorithm results, and are within the engineering tolerance. Therefore, the T-SNE algorithm performs well in the current data set. Therefore, the T-SNE algorithm is used to reduce the dimension of the full data set.
After individual verification, we use the T-SNE algorithm to reduce the dimensionality of high-dimensional features in the full amount of data. The characteristics after the T-SNE algorithm are shown in Figure 12. As the number of iterations increases, we can see a clear four-category state, and there is an isolation band between the classes, and the feature points within the class are close.

5.4. Classification and Analysis of Kinematics

In this section, we will detail how to classify and analyze kinematic segments by the features obtained after dimensionality reduction. From the perspective of practical application, compare the performance of different algorithms. And the results obtained by the algorithm with the best performance are selected as the basis for the construction of the subsequent driving condition map.

5.4.1. K Value Selection

Since most clustering algorithms need to determine the number of categories, the advantages and disadvantages of the number of categories will have a greater impact on the experimental results. Therefore, based on the clustering effect, we explore the K value selection that K applies to this problem. There are two options for the selection of the number K of cluster centers in K-means:
  • Make a selection based on an actual business background.
  • Select based on the elbow method or contour factor, where K is selected using the elbow method. That is, the graphical tool method can be used to visualize the within-cluster sum of squared errors (SSE) according to the number of clusters. Through the graph, the influence of k on the error variance in the cluster can be visually observed to select an appropriate K value.
It can be seen from Figure 13 that when the number of clusters is 4, the elbow type appears, which means that k is suitable. Through the analysis of the business background, the kinematics of the car can be generally divided into four states: low speed, medium speed, high speed and extremely high speed. Therefore, whether it is analysis from the business background or using the elbow method, K = 4 is a good choice.

5.4.2. Clustering Results Visualization

Due to the high-dimensional features of the original features, visualization is not straightforward. Therefore, the T-SNE algorithm is used to reduce the original data and visualize the clustering results.
Compared with the K-means algorithm, the DBSCAN algorithm has the advantage of recognizing clusters of arbitrary shapes and recognizing noise. However, the number of clusters is uncontrollable, and the results of clustering are often difficult to interpret. As shown in Figure 14a, the DBSCAN results are sparse, the cluster distribution is not uniform, but some abnormal points are filtered out.
Figure 14b shows the K-means++ clustering visualization results. It can be seen that K-means++ basically divides the motion segment into four clusters. And through the visualization of the T-SNE algorithm, it can be seen that the results of the same cluster are relatively similar, but the results of the light green cluster are scattered.
Figure 14c shows the clustering effect of Multi-Clustering. It can be seen that the clustering results are divided into four categories, and the data distribution of the same cluster (same color) is concentrated, and the results of the four clusters are evenly distributed. From the results of the visualization, it is significantly better than the single model K-means++ and DBSCAN algorithms. Table 4 also validates this conclusion from the perspective of data evaluation.

6. Analysis of Experiments

In this section, we will perform detailed analysis based on the output of algorithms such as multi-clustering. The algorithms are quantitatively compared and conclusions are obtained.

6.1. Transient Driving Cycle Results and Analysis

From the engineering point of view, the construction of the driving cycle map should basically conform to the characteristic laws of the overall data, that is, the parameters should be consistent with the actual statistical measurement data. From the perspective of life, the driving cycle map is used for vehicle fuel consumption and emissions testing, and its characteristics should be consistent with the situation of the living vehicle. Vehicle condition maps that meet these characteristics are in line with reality and can be put into use. Therefore, in this section we present the driving cycle map from our two perspectives of data matching and practical application.

6.1.1. Practical Application Analysis

The output obtained by Multi-Clustering can be obtained as shown in Figure 15. After analyzing and studying the speed, we can see from the transient condition chart that the complete speed time curve can be divided into four parts (low speed, middle speed, high speed and ultrahigh speed). From the data point of view, we can analyze that the total duration of the transient condition map is 1225 s, and the duration of each part is as shown in Table 4.
We disassemble the overall condition diagram shown in Figure 15 for analysis. Combined with the actual situation, the actual meaning of each segment of speed segmentation is explained separately. In order to analyze whether this condition map is reasonable.
As shown in the low-speed stage details in Figure 16a, we can see that the low-speed segment represents the speed characterization of a “crash” in life. The obvious acceleration-deceleration-parking alternation is a typical trend of daily traffic jams. As shown in the detail of the mid-speed section shown in Figure 16b, we can see that the mid-speed section represents the driving condition of the city center road during the period of heavy traffic pressure, with large traffic volume and low average speed. There is a red-light waiting time that obeys a negative exponential distribution and an acceleration/deceleration fluctuation with a minimum speed of 10 km/h. As shown in the high-speed section detail shown in Figure 16c, we can see that the section indicates the daily driving conditions of urban traffic or non-congested sections at night. The maximum speed is 60 km/h for most sections, but due to traffic lights, intersections, etc., there are frequent speed increases and decreases, and continuous driving time is not very long. As shown in the detail of the super-high-speed section shown in Figure 16d, we can see that the super-high-speed section represents the driving state of suburban roads or highways and expressways. The maximum speed is 90 km/h, the acceleration process is stable and the acceleration is high, the continuous high-speed time is long, and there is a tendency to drive for a long time. This is consistent with the analysis results of the geographic data processing part, that is, the vehicle condition map proposed by us can well match the driving situation in Fuzhou.

6.1.2. Data Matching Analysis

After obtaining the working condition map, we made a detailed comparison between the proposed working condition map and the actual data set, and calculated the error rate of the measured parameters separately, as shown in Table 5. It can be seen from the table that the error rate of each parameter is within 10%. The highest error rate is the error rate of the idle time ratio, which is 9.4%, which is within the engineering tolerance. Therefore, the transient condition map we proposed is consistent with the actual data set in the data.

6.2. Advantages of Multi-Clustering

We use different algorithms to study the output of the worksheet for this dataset. The error rates of the results of each algorithm are shown in Table 6. We found that the error rate of the parameter map parameters and the actual data set parameters obtained by the K-means algorithm is the largest. This is due to the initial value sensitivity of the K-means algorithm, which we have already demonstrated in the algorithm analysis. Therefore, we use the improved algorithm K-means++ and DBSCAN algorithm, respectively, in the experiments. It can be seen from Table 6 that the K-means++ algorithm has low error rates such as idle time ratio, acceleration time ratio, deceleration time ratio, speed standard deviation, and acceleration standard deviation. The DBSCAN algorithm has a low error rate of average running speed, average acceleration, and average deceleration. Therefore, we try to use the integrated learning method to capture the advantages between the models, and combine the results of each algorithm to get the final vehicle condition map. As can be seen from Table 6, the error rate of each parameter in the multi-clustering algorithm results is lower than the lowest value in other algorithms. Therefore, the stacking algorithm performs well in our application.

6.3. Advantage Analysis of Local Cycle Map

At the beginning of this century, China directly adopted the NEDC driving conditions in Europe, and the certification of energy consumption and emissions of automobile products effectively promoted the development of energy-saving emission reduction and technology. In recent years, with the rapid growth of car ownership, China’s road traffic conditions have changed greatly. Governments, enterprises, and the public have increasingly found that cars that are calibrated and calibrated based on NEDC conditions have become increasingly biased in actual fuel consumption and regulatory certification results, affecting the government’s credibility.
The vehicle condition map we proposed is based on the data collected by Fuzhou Automobile, which is closer to China’s national conditions and is more consistent with our driving situation. We compare the working condition parameters proposed in this paper with the NEDC and WLTC working condition parameters, as shown in Table 7.
It can be seen from the table that the working conditions presented by Europe and the United States are far from our Fuzhou working conditions. The average acceleration of the NEDC condition map is significantly lower, the idle time is longer, and the acceleration time is shorter. This is because European countries generally have less road congestion. The traffic flow in Fuzhou is concentrated on the main roads, the efficiency of the distributed traffic is insufficient, and traffic congestion is prone to occur. Therefore, the situation of acceleration and deceleration is bound to be more. Thus, only the data obtained through field testing can better represent the current situation of the driving conditions of our country.

7. Conclusions

In this paper, we study the problem of driving cycle prediction. Two key findings found by analysis actual data are: first, the vehicle dataset used by the driving cycle prediction problem is usually unbalanced in real cases. Second, China’s geographical location and economic development are vastly different from those in Europe, not only in the various car types but also due to the complicated terrain in China. Thus, it is difficult for traditional clustering methods to achieve the effect in actual data of China. Then, we propose a novel driving cycle map algorithm framework based on ensemble learning, named multi-clustering algorithm, to improve the performance of traditional clustering algorithms on unbalanced data set. It is noteworthy the algorithm framework integrates k-means-type, K-means++, DBSCAN and other clustering algorithms, and uses stacking to strengthen each individual model. Multi-clustering can solve potential vulnerabilities such as data imbalances and random initial states. In the experimental part, we conducted multiple experiments to compare the matching order and crossover method of the algorithm to obtain the best results. It can be verified through experiments that our proposed multi-clustering framework performs well in real data sets and has a relatively high improvement over other algorithms.
Our multi-clustering algorithm is a simple and effective framework for the driving cycle prediction research. It will greatly improve the performance of traditional clustering algorithms on unbalanced data set for driving cycle prediction problems. In the future, we will continue to use the multi-clustering algorithm to examine other car parameter generation problems in complicated structure areas. We have good reason to believe that this kind of multi-clustering algorithm will play a significant guiding role in the field of unsupervised learning.

Author Contributions

Data curation, Y.W., W.Z. and C.C.; Formal analysis, Y.W., L.Z. and C.C.; Funding acquisition, Y.Q. and J.Y.; Investigation, Y.W. and W.Z.; Methodology, Y.W., L.Z. and C.C.; Software, L.Z.; Visualization, Y.W., W.Z. and C.C.; Writing–original draft, Y.W.; Writing–review & editing, Y.Q. and J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by the National Natural Science Foundation of China (61671078, 61701031), Funds of the Beijing Laboratory of Advanced Information Networks of BUPT, Funds of the Beijing Key Laboratory of Network System Architecture and Convergence of BUPT, the National Key Research and Development Program of China (2017YFB0802701), the Fundamental Research Funds for the Central Universities (2018XKJC04), the Science and Technology on Space Intelligent Control Laboratory 6142208180205, the 111 Project of China (B17007) and the Data Intelligence Innovation and Entrepreneurship Platform conducted in the cooperation between the Beijing University of the Post and Telecommunications and China Telecom Corporation Limited Cloud Computing Branch Corporation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sun, D.J.; Liu, X.; Ni, A.; Peng, C. Traffic congestion evaluation method for urban arterials: Case study of changzhou, China. Transp. Res. Rec. 2014, 2461, 9–15. [Google Scholar] [CrossRef]
  2. Zuo, Y.N.; He, Q.F. Intersection Approach Road Congestion Index and Application for Beijing, China. In Proceedings of the IEEE International Conference on Information Management & Engineering, Chengdu, China, 16–18 April 2010. [Google Scholar]
  3. Liang, J.; Liang, B.; Dang, C.; Cao, F. The -means-type algorithms ver-sus imbalanced data distributions. IEEE Trans. Fuzzy Syst. 2012, 20, 728–745. [Google Scholar] [CrossRef]
  4. Galar, M.; Fernandez, A.; Barrenechea, E.; Bustince, H.; Herrera, F. A review on ensembles for the class imbalance problem: Bagging-, boosting-, and hybrid-based approaches. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2012, 42, 463–484. [Google Scholar] [CrossRef]
  5. Haixiang, G.; Yijing, L.; Shang, J.; Mingyun, G.; Yuanyue, H.; Bing, G. Learning from class-imbalanced data: Review of methods and applications. Expert Syst. Appl. 2017, 73, 220–239. [Google Scholar] [CrossRef]
  6. Shu, K.; Wang, S.; Tang, J.; Zafarani, R.; Liu, H. User identity link-age across online social networks: A review. ACM Sigkdd Explor. Newsl. 2017, 18, 5–17. [Google Scholar] [CrossRef]
  7. Hoens, T.R.; Polikar, R.; Chawla, N.V. Learning from streaming data with concept drift and imbalance: An overview. Prog. Artif. Intell. 2012, 1, 89–101. [Google Scholar] [CrossRef] [Green Version]
  8. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. Smote: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  9. Tahir, M.A.; Kittler, J.; Mikolajczyk, K.; Yan, F. A multiple expert approach to the class imbalance problem using inverse random under sampling. In International Workshop on Multiple Classifier Systems; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  10. Wang, D.; Pedreschi, D.; Song, C.; Giannotti, F.; Barabasi, A.L. Human Mobility, Social Ties, and Link Prediction. In Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Diego, CA, USA, 21–24 August 2011. [Google Scholar]
  11. Wang, S.; Liu, W.; Wu, J.; Cao, L.; Kennedy, P.J. Training Deep Neural Networks on Imbalanced Data Sets. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016. [Google Scholar]
  12. Khan, S.H.; Bennamoun, M.; Sohel, F.A.; Togneri, R. Cost-sensitive learning of deep feature representations from imbalanced data. IEEE Trans. Neural Netw. Learn. Syst. 2015, 29, 3573–3587. [Google Scholar]
  13. Freund, Y.; Schapire, R.E. A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting. In Proceedings of the Second European Conference on Computational Learning Theory, Barcelona, Spain, 13–15 March 1995. [Google Scholar]
  14. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  15. Oza, N.C.; Tumer, K. Classifier ensembles: Select real-world applications. Inf. Fusion 2008, 9, 4–20. [Google Scholar] [CrossRef]
  16. Chawla, N.V.; Lazarevic, A.; Hall, L.O.; Bowyer, K.W. Smoteboost: Improving Prediction of the Minority Class in Boosting. In Proceedings of the European Conference on Principles of data Mining and Knowledge Discovery, Cavtat, Croatia, 22–26 September 2003; Springer: Berlin/Heidelberg, Germany, 2003; Volume 2838, pp. 107–119. [Google Scholar]
  17. Seiffert, C.; Khoshgoftaar, T.M.; Hulse, J.V.; Napolitano, A. Rusboost: A hybrid approach to alleviating class imbalance. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 2010, 40, 185–197. [Google Scholar] [CrossRef]
  18. Wang, S.; Yao, X. Diversity Analysis on Imbalanced Data Sets by Using Ensemble Models. In Proceedings of the IEEE Symposium on Computational Intelligence and Data Mining, CIDM 2009, Part of the IEEE Symposium Series on Computational Intelligence 2009, Nashville, TN, USA, 30 March–2 April 2009. [Google Scholar]
  19. Liu, X.Y. Exploratory under-sampling for class-imbalance learning. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2006, 39, 539–550. [Google Scholar]
  20. Li, M.; Zhang, J.; Zhang, F. A study on real driving cycle of passenger cars in typical cities of China. Automot. Eng. 2006, 28, 554–557. [Google Scholar]
  21. Wang, R.; Lukic, S.M. Review of Driving Conditions Prediction and Driving Style Recognition Based Control Algorithms for Hybrid Electric Vehicles. In Proceedings of the 2011 IEEE Vehicle Power and Propulsion Conference, Chicago, IL, USA, 6–9 September 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1–7. [Google Scholar]
  22. Asus, Z.; Aglzim, E.-H.; Chrenko, D.; Daud, Z.-H.C.; le Moyne, L. Dynamic modeling and driving cycle prediction for a racing series hybrid car. IEEE J. Emerg. Sel. Top. Power Electron. 2014, 2, 541–551. [Google Scholar] [CrossRef]
  23. Huang, D.; Xie, H.; Ma, H.; Sun, Q. Driving cycle prediction model based on bus route features. Transp. Res. Part D Transp. Environ. 2017, 54, 99–113. [Google Scholar] [CrossRef]
  24. Chen, Z.; Xiong, R.; Wang, C.; Cao, J. An on-line predictive energy management strategy for plug-in hybrid electric vehicles to counter the uncertain prediction of the driving cycle. Appl. Energy 2017, 185, 1663–1672. [Google Scholar] [CrossRef]
  25. Huang, X.; Tan, Y.; He, X. An intelligent multifeature statistical approach for the discrimination of driving conditions of a hybrid electric vehicle. IEEE Trans. Intell. Transp. Syst. 2010, 12, 453–465. [Google Scholar] [CrossRef]
  26. Wang, Q.; Huo, H.; He, K.; Yao, Z.; Zhang, Q. Characterization of ve-hicle driving patterns and development of driving cycles in chinese cities. Transp. Res. Part D Transp. Environ. 2008, 13, 289–297. [Google Scholar] [CrossRef]
  27. Tutuianu, M.; Bonnel, P.; Ciuffo, B.; Haniu, T.; Ichikawa, N.; Marotta, A.; Pavlovic, J.; Steven, H. Development of the World-wide harmonized Light duty Test Cycle (WLTC) and a possible pathway for its introduction in the European legislation. Transp. Res. 2015, 40, 61–75. [Google Scholar] [CrossRef]
  28. Zhao, S.; Tian, M.; Hao, R.; Zhang, S.; Li, N. Processes and analysis of vehicle driving data based on short stroke method. Sens. Transducers 2014, 177, 31. [Google Scholar]
  29. Wold, S.; Esbensen, K.; Geladi, P. Principal component analysis. Chemom. Intell. Lab. Syst. 1987, 2, 37–52. [Google Scholar] [CrossRef]
  30. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Inter-Discip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  31. Laurens, V.D.M. Accelerating t-SNE using Tree-Based Algorithms. J. Mach. Learn. Res. 2014, 15, 3221–3245. [Google Scholar]
  32. Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Basic situation of vehicle use. (a) China’s urban area distribution (name location does not represent actual city location, and text size represents the number of cars in each city); (b) Congestion delay index (2016).
Figure 1. Basic situation of vehicle use. (a) China’s urban area distribution (name location does not represent actual city location, and text size represents the number of cars in each city); (b) Congestion delay index (2016).
Sensors 20 02448 g001
Figure 2. Transient driving cycle map construction process.
Figure 2. Transient driving cycle map construction process.
Sensors 20 02448 g002
Figure 3. The multi-clustering algorithm framework diagram.
Figure 3. The multi-clustering algorithm framework diagram.
Sensors 20 02448 g003
Figure 4. Missing duration distribution statistics.
Figure 4. Missing duration distribution statistics.
Sensors 20 02448 g004
Figure 5. The relationship between the number of segments and the duration of continuous idle speed.
Figure 5. The relationship between the number of segments and the duration of continuous idle speed.
Sensors 20 02448 g005
Figure 6. Three cases of dividing kinematic fragments according to different idle durations: (a) Car 1 divided into 20s; (b) Car 1 divided into 10s; (c) Car 1 divided into 5s; (d) Car 2 divided into 20s; (e) Car 2 divided into 10s; (f) Car 2 divided into 5s; (g) Car 3 divided into 20s; (h) Car 3 divided into 10s; (i) Car 3 divided into 5s.
Figure 6. Three cases of dividing kinematic fragments according to different idle durations: (a) Car 1 divided into 20s; (b) Car 1 divided into 10s; (c) Car 1 divided into 5s; (d) Car 2 divided into 20s; (e) Car 2 divided into 10s; (f) Car 2 divided into 5s; (g) Car 3 divided into 20s; (h) Car 3 divided into 10s; (i) Car 3 divided into 5s.
Sensors 20 02448 g006
Figure 7. Captured track information display.
Figure 7. Captured track information display.
Sensors 20 02448 g007
Figure 8. Abnormal point correction: (a) Acceleration distribution map; (b) Abnormal point.
Figure 8. Abnormal point correction: (a) Acceleration distribution map; (b) Abnormal point.
Sensors 20 02448 g008
Figure 9. Abnormal point correction: (a) Comparison results before; (b) Comparison results after.
Figure 9. Abnormal point correction: (a) Comparison results before; (b) Comparison results after.
Sensors 20 02448 g009
Figure 10. Use PCA to reduce the dimension to a 3D rendering. (a) x-y plane viewing angle; (b) y-z plane viewing angle.
Figure 10. Use PCA to reduce the dimension to a 3D rendering. (a) x-y plane viewing angle; (b) y-z plane viewing angle.
Sensors 20 02448 g010
Figure 11. Use the T-SNE algorithm to reduce the dimensional effect. (a) Iteration 100 times; (b) Iteration 300 times; (c) Iteration 1000 times.
Figure 11. Use the T-SNE algorithm to reduce the dimensional effect. (a) Iteration 100 times; (b) Iteration 300 times; (c) Iteration 1000 times.
Sensors 20 02448 g011
Figure 12. Full reduction of dimensions using the T-SNE algorithm. (a) Iteration 100 times; (b) Iteration 300 times; (c) Iteration 1000 times.
Figure 12. Full reduction of dimensions using the T-SNE algorithm. (a) Iteration 100 times; (b) Iteration 300 times; (c) Iteration 1000 times.
Sensors 20 02448 g012
Figure 13. Intra-cluster error variance SSE.
Figure 13. Intra-cluster error variance SSE.
Sensors 20 02448 g013
Figure 14. Clustering results visualization. (a) DBSCAN; (b) K-means++; (c) Multi-Clustering.
Figure 14. Clustering results visualization. (a) DBSCAN; (b) K-means++; (c) Multi-Clustering.
Sensors 20 02448 g014
Figure 15. Transient driving cycle map.
Figure 15. Transient driving cycle map.
Sensors 20 02448 g015
Figure 16. Driving cycle map detail display. (a) Low Speed; (b) Middle Speed; (c) High Speed; (d) Ultrahigh Speed.
Figure 16. Driving cycle map detail display. (a) Low Speed; (b) Middle Speed; (c) High Speed; (d) Ultrahigh Speed.
Sensors 20 02448 g016
Table 1. Data analysis and statistic.
Table 1. Data analysis and statistic.
Vehicle 1Vehicle 2Vehicle 3
Original records185,725145,825164,914
Missing fragments72423751097
After interpolation (<10 s)185,836147,568165,255
Short strokes (idle > 5 s)384918172636
Unfilled missing segments69130127
Acceleration anomalies strokes535198386
Short strokes too short11904861012
Final Short strokes205510031111
Table 2. Vehicle driving data feature setting.
Table 2. Vehicle driving data feature setting.
No.ParameterMeaningNo.ParameterMeaning
1T (s)Operation hours19RmAverage engine speed
2Ta(s)Acceleration time20NmaxMaximum torque percentage
3Td(s)Deceleration time21NminMinimum torque percentage
4Te(s)Uniform time22NmAverage engine speed
5Ti(s)Idle time23FmaxMaximum instantaneous fuel consumption
6STravel distance24FminMinimum instantaneous fuel consumption
7VmaxMaximum operating speed25FmAverage instantaneous fuel consumption
8AmaxMaximum acceleration26OmaxMaximum throttle opening
9AminMaximum deceleration27OminMinimum throttle opening
10VmAverage speed28OmAverage throttle opening
11VmeAverage speed of operation29RmaxMaximum air-fuel ratio
12AaAcceleration section average acceleration30RminMinimum air-fuel ratio
13AdDeceleration section average deceleration31LmaxMaximum engine load
14VsdSpeed standard deviation32LminMinimum engine load
15AsdAcceleration standard deviation33LmeAverage engine load
16RPARelative positive acceleration34ImaxMaximum intake air volume
17RmaxMaximum engine speed35IminMinimum intake air volume
18RminMinimum engine speed36ImAverage intake air volume
Table 3. PCA algorithm parameter statistics.
Table 3. PCA algorithm parameter statistics.
Principal ComponentContribution RateCumulative Rate
M165.172565.1725
M212.344277.5167
M37.845185.3618
M42.432587.7943
M51.675889.4701
M36 1.0215 × 10 10 100
Table 4. Length of each speed state and its proportion.
Table 4. Length of each speed state and its proportion.
Low SpeedMiddle SpeedHigh SpeedUltrahigh Speed
Duration (s)153258403411
Proportion (%)12.5%21.1%32.9%33.6%
Average speed (km/h)5.810.231.940.3
Maximum speed (km/h)11.820.356.590.1
Table 5. Comparison of the driving cycle map and the actual data set parameters.
Table 5. Comparison of the driving cycle map and the actual data set parameters.
Actual Data SetConstruction Cycle MapError Rate (%)
Average speed (km/h)29.8427580129.3575731.6258049
Average travel speed (km/h)26.83671425.2315413.9812576
Average acceleration (m/s2)1.4759291.7667138.1835915
Average deceleration (m/s2)−1.690293−1.7946143.7633836
Idle time ratio (%)10.8331%11.8544%9.4275877%
Acceleration time ratio (%)47.0271%46.5026%1.1153144%
Deceleration time ratio (%)36.8889%36.0751%2.2060837%
Speed standard deviation (km/h)19.89576320.8525634.8090641
Acceleration standard deviation (m/s2)2.9004512.7220456.1509745
Table 6. Comparison of error rates of each algorithm.
Table 6. Comparison of error rates of each algorithm.
Error RateMulti-ClusteringK-MeansK-Means++DBSCAN
Average speed (km/h)1.625804931.913256816.138806620.5892766
Average travel speed (km/h)3.981257642.418896723.90065714.874926
Average acceleration (m/s2)8.183591530.494962829.16115885.9235912
Average deceleration (m/s2)3.763383626.338096428.54629356.3717231
Idle time ratio (%)9.4275877121.067838469.8147345131.3594447
Acceleration time ratio (%)1.115314414.39510414.87888921.1239902
Deceleration time ratio (%)2.20608376.2257215.163070714.5173751
Speed standard deviation (km/h)4.80906416.98924692.637767628.2313274
Acceleration standard deviation (m/s2)6.15097455.93538733.291039914.8070421
Table 7. Comparison of various driving cycle map parameters.
Table 7. Comparison of various driving cycle map parameters.
Actual Data SetFuzhou Driving CycleNEDCWLTC
Average speed (km/h)29.8427580129.35757333.2122.11
Average travel speed (km/h)1.4759291.7667130.5400.392
Idle time ratio (%)10.8331%11.8544%23.9%24.2%
Acceleration time ratio (%)47.0271%46.5026%22.9%33.5%

Share and Cite

MDPI and ACS Style

Wu, Y.; Zhang, W.; Zhang, L.; Qiao, Y.; Yang, J.; Cheng, C. A Multi-Clustering Algorithm to Solve Driving Cycle Prediction Problems Based on Unbalanced Data Sets: A Chinese Case Study. Sensors 2020, 20, 2448. https://doi.org/10.3390/s20092448

AMA Style

Wu Y, Zhang W, Zhang L, Qiao Y, Yang J, Cheng C. A Multi-Clustering Algorithm to Solve Driving Cycle Prediction Problems Based on Unbalanced Data Sets: A Chinese Case Study. Sensors. 2020; 20(9):2448. https://doi.org/10.3390/s20092448

Chicago/Turabian Style

Wu, Yuewei, Wutong Zhang, Long Zhang, Yuanyuan Qiao, Jie Yang, and Cheng Cheng. 2020. "A Multi-Clustering Algorithm to Solve Driving Cycle Prediction Problems Based on Unbalanced Data Sets: A Chinese Case Study" Sensors 20, no. 9: 2448. https://doi.org/10.3390/s20092448

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop