Next Article in Journal
A Multirobot System in an Assisted Home Environment to Support the Elderly in Their Daily Lives
Next Article in Special Issue
Deep-Reinforcement-Learning-Based IoT Sensor Data Cleaning Framework for Enhanced Data Analytics
Previous Article in Journal
A Fast Sparse Decomposition Based on the Teager Energy Operator in Extraction of Weak Fault Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Data Selection-Based Machine Learning Algorithm for Prediction of Component Obsolescence

1
Department of Mathematical Finance, Gachon University, 1342 Seongnamdaero, Sujeong-gu, Seongnam-si 13120, Gyeonggi-do, Korea
2
Department of Mathematics, Korea University, 145 Anam-ro, Seongbuk-gu, Seoul 02841, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(20), 7982; https://doi.org/10.3390/s22207982
Submission received: 18 August 2022 / Revised: 12 October 2022 / Accepted: 18 October 2022 / Published: 19 October 2022
(This article belongs to the Special Issue Machine Learning in Wireless Sensor Networks)

Abstract

:
Product obsolescence occurs in the manufacturing industry as new products with better performance or improved cost-effectiveness are developed. A proactive strategy for predicting component obsolescence can reduce manufacturing losses and lead to customer satisfaction. In this study, we propose a machine learning algorithm for a proactive strategy based on an adaptive data selection method to forecast the obsolescence of electronic diodes. Typical machine learning algorithms construct a single model for a dataset. By contrast, the proposed algorithm first determines a mathematical cover of the dataset via unsupervised clustering and subsequently constructs multiple models, each of which is trained with the data in one cover. For each data point in the test dataset, an optimal model is selected for regression. Results of empirical experiments show that the proposed method improves the obsolescence prediction accuracy and accelerates the training procedure. A novelty of this study is that it demonstrates the effectiveness of unsupervised clustering methods for improving supervised regression algorithms.

1. Introduction

The acceleration of technological development is rendering current technological innovation obsolete. Particularly, innovations in electronic devices based on semiconductors have exacerbated the problem of component obsolescence. However, accurate prediction of obsolescence can significantly reduce costs associated with product design and production. Therefore, preventive and preemptive technology development and management based on the prediction of future obsolescence levels is necessary. Many studies have focused on the concept and management of obsolescence, and the obsolescence problem is referred to as the diminishing manufacturing sources and material shortages (DMSMS) problem (see [1,2]).
Jennings et al. [3] reported that as obsolescence increases, a proactive approach offers more advantages than a reactive approach, which requires more time and materials than the former to manage obsolescence. The key to proactive management is the accurate prediction of the risk of component obsolescence. As outlined in Section 2.1, numerous studies have focused on predicting obsolescence based on mathematical models, and recently, methods using machine learning (ML) have been proposed, as summarized in Table 1.
In this study, we aim to predict the obsolescence date of electronic parts based on the data. In order to accurately predict the obsolescence of parts, we propose an adaptive data selection (ADS)-based ML method. The newly presented ADS method first divides the input dataset into subsets of similar data. Subsequently, when a new data point is inserted, it selects one optimal subset and performs learning and prediction with it. Existing clustering methods, such as the k-means method, do not allow common data between groups, whereas the ADS method classifies data of the same nature into one group while allowing common data between the groups. As a case study, we apply several ML methods, introduced in Section 2.2, with ADS to predict the obsolescence date of electronic diodes. As discussed in Section 5, the prediction results of learning using ADS are considerably more accurate than those of the method that does not use clustering or the k-means method.
The main improvements of this paper are as follows. Firstly, the obsolescence date is accurately predicted based on the discontinued electronic component data using ML methods. Secondly, the accuracy of prediction is improved by dividing the dataset into subsets according to the characteristics, finding parameters suitable for each subset and applying a learning algorithm. Thirdly, in order to offset the shortage of data on discontinued electronic components and to increase the accuracy near the boundary of subsets, a new ADS method is proposed to improve accuracy and efficiency.
The remainder of this paper is organized as follows. Section 2.1 and Section 2.2 present the previous studies on obsolescence prediction problems and the ML algorithms used in the numerical experiments, respectively. The proposed ADS-based method is described in Section 3. The statistics of the data and the hyperparameters are presented in Section 4. The accuracy measures and experimental results are presented and discussed in Section 5. The conclusions are discussed in Section 6.

2. Background

2.1. Literature Review

According to Trabelsi et al. [2], the causes of component obsolescence can be classified by technological development, availability, material shortage, economic feasibility, and environmental factors. It is more economical and efficient to detect and manage component obsolescence problems caused by these various causes in advance. Unlike reactive management, proactive management identifies and prioritizes components at risk of obsolescence. For proactive management, in order to suggest resolutions before components are discontinued, it is necessary to accurately predict the risk of component obsolescence. Many studies on predicting the risk of component obsolescence have been conducted, and they can be divided into two categories, one using a mathematical model based on statistical techniques and the other using ML, as summarized in Table 1.
Mathematical models based on various statistical methods have been studied. Solomon et al. [4] presented a life-cycle forecasting model based on the Gaussian distribution methodology for forecasting the years to obsolescence for electronic parts of integrated circuits. Sandborn et al. [5] proposed combining life-cycle curve forecasting and data mining of historical last-order or last-ship dates. Rojo et al. [6] reviewed the literature on the problem of obsolescence published before 2009. Sandborn et al. [7] proposed a method based on data mining of historical records to forecast procurement lifetime and obsolescence date, and Ma and Kim [8] predicted life cycle curves based on a time series model without evolutionary parametric drivers. Trabelsi et al. [2] developed a method based on probability distribution to model the degree of obsolescence as a function of time. Mastrangelo et al. [9] proposed the conditional Weibull distribution method to determine the likelihood of a component becoming obsolete.
Recently, ML and deep learning (DL) algorithms, which can learn from vast amounts of data, have yielded promising results in various fields [10,11]. However, to the best of the authors’ knowledge, studies using ML to predict the obsolescence of parts remain limited. Among them, Jennings et al. [3] tested the effectiveness of random forest (RF), neural network, and support vector machine algorithms for obsolescence risk and life-cycle forecasting. Grichi et al. [12] used the RF method for obsolescence forecasting and improved it by combining it with a genetic algorithm [13]. The limited success of ML or DL methods for obsolescence prediction can be attributed to the scarcity of available data. To address this problem, Moon et al. [14] proposed using the k-means clustering method before applying a ML algorithm and employed this approach to improve the forecasting accuracy of the obsolescence date of electronic diodes.
Table 1. A summary of recent studies on predicting the risk of component obsolescence. (RF: random forest, NN: neural network, SVM: support vector machine, LR: logistic regression, GB: gradient boosting, GA: genetic algorithm, DT: decision tree, DNN: deep neural network, RNN: recurrent neural network).
Table 1. A summary of recent studies on predicting the risk of component obsolescence. (RF: random forest, NN: neural network, SVM: support vector machine, LR: logistic regression, GB: gradient boosting, GA: genetic algorithm, DT: decision tree, DNN: deep neural network, RNN: recurrent neural network).
CategoryAuthors (Year)Application (Target)Method
MathematicalSolomon et al. (2000) [4]Integrated circuits (years to obsolescence)Gaussian distribution
modelsSandborn et al. (2007) [5]Flash memory/DRAM (life cycle curve)Curve fitting
Sandborn et al. (2011) [7]Electronic parts (obsolescence dates)Linear regression
Ma and Kim (2017) [8]Flash memory (life cycle curve)Time series model
Trabelsi et al. (2021) [2]Smartphones (obsolescence probability density)Statistical test
Mastrangelo et al. (2021) [9]Electronic parts (obsolescence probability)conditional probability method
MachineJennings et al. (2016) [3]Cellphones (obsolescence dates)RF/NN/SVM
learningGrichi et al. (2017) [12]Cellphones (obsolescence dates)RF/Multiple linear regression
Grichi et al. (2018) [13]Cellphones (obsolescence risk)LR/GB/RF with GA
Moon et al. (2022) [14]Electronic diodes (obsolescence dates)DT/RF/GB/DNN/RNN with k-means clustering

2.2. Learning Models

This section briefly reviews the ML and DL methods used in the numerical experiments. ML extracts desired information from data based on statistics and artificial intelligence. It was first proposed in the 1950s, and it has been actively studied since 2000 and used in a variety of fields, including computer science and finance. ML can be classified into supervised and unsupervised learning, and the former can be categorized into classification and regression [15,16].
In this study, we use a regression model to predict future obsolescence dates. Representative regression models include the k-nearest neighbors, linear regression, support vector machines, and decision trees (DTs). We use DT in Section 2.2.1 and two ensemble models with a DT as the basic element, namely the RF method in Section 2.2.2 and the gradient boosting (GB) method in Section 2.2.3. Additionally, the results are compared with those obtained using a deep neural network (DNN) and recurrent neural network (RNN) in Section 2.2.4, which are DL models with multilayered neural networks.

2.2.1. Decision Tree

The DT method is a method of learning that involves repeated creation of a binary tree or hierarchical division of a domain. Prediction on new data involves determining the location of the given data among the divided features, finding leaf nodes, and using the average value of the training data of the found leaf nodes to make a prediction. Learning proceeds in the direction that maximizes the information gain in each branch, which can be expressed as follows:
I n f o = E ( p a r e n t ) i = 1 n N i N p E ( c h i l d r e n i ) ,
where parent and children denote before and after branch, respectively; and N i , N p , and n represent the number of samples of the i-th child node, number of samples of the p-th parent node, and number of children, respectively. Here, E is the entropy, which is defined as
E = i p i log 2 ( p i ) ,
where p i is the proportion of data belonging to category i among the data in an area.
In this process, if a leaf node proceeds until it becomes a pure node comprising a single target, the model may become complex and result in overfitting. Therefore, the max_depth and max_leaf_nodes variables are used to limit the maximum depth of the tree and the maximum number of leaves, respectively. We mitigate the overfitting of the training data by pre-pruning using the min_samples_split and min_samples_leaf variables (Table 6).
DTs are easy to understand by model visualization and do not require preprocessing steps such as the regularization of the data size and characteristics. However, even with pre-pruning, the training data can be overfitted, and thus, the ensemble methods introduced in the following sections are preferred.

2.2.2. Random Forest

To compensate for the overfitting problem associated with DT, an RF randomly creates many DTs and averages the results. First, we determine the number of trees, randomly create bootstrap samples, and create a DT for each generated dataset. Here, not all of the features are considered; instead, the best test among randomly selected features is determined. The number of DTs and number of features are modified using the n_estimator and max_features variables, respectively, as listed in Table 6.
In previous studies such as [3,12,13], the RF method is the most extensively used ML method. It addresses the limitation of DT and offers the advantages of high performance and non-requirement of scaling data. Although averaging more trees can reduce overfitting and create a stabler model, this requires more memory and computation time.

2.2.3. Gradient Boosting

GB is an ensemble method that creates an improved model by combining multiple DTs. However, unlike the RF method, it creates a tree to compensate for the error of the previous tree. GB requires less memory and prediction time than the RF method; however, it is more sensitive to parameter values than the latter. In addition to the increase in n_estimator, which denotes the number of trees, that in learning_rate, which controls the degree of error correction of the previous tree, increases the model complexity.

2.2.4. Deep Learning

DL is a method inspired by the structure and function of the human brain, which essentially repeats tasks while making minor adjustments based on learning to improve results. It is a class of ML methods. DNN expresses various concepts in a hierarchical structure and offers excellent capabilities and flexibility. However, there are still limitations arising from the repeatability of DL algorithms, complexity of increasing the number of layers, and the large amount of data required for network training. Additionally, compared to conventional ML, a DL algorithm requires more data to understand a certain problem, and if the amount of data is insufficient, it performs poorly.
The most basic element in DL is a neural network, which represents the neuronal connection structure among the biological characteristics of the human brain. A network structure modeled after this neural network is called an artificial neural network. An artificial neural network receives learning data from the input layer, processes them using several hidden layers, and outputs the final result through the output layer. A structure in which these neural networks are superimposed is called a DNN, as shown in Figure 1a. A structure that uses an internal cyclic structure to reflect past learning in current learning via weights is called an RNN, as shown in Figure 1b.

3. Proposed Method

Given a dataset, an ML algorithm identifies its parameters and hyperparameters and builds one model with them. However, some data show various characteristics or patterns in it, and one set of parameters and hyperparameters may not be sufficient for proper representation of the data, which leads to inaccurate prediction. The current study thus proposes to split the dataset into several subsets and to identify parameters and hyperparameters for each subset. Then one can construct a group of ML models, each of which best fits one of the subsets. We introduce two unsupervised clustering methods in Section 3.1 to split the dataset to its subsets. One is k-means clustering in Section 3.1.1, whose subsets are mutually disjoint, and the other is ADS in Section 3.1.2, whose subsets have non-empty intersections. We subsequently propose an approach to combine the regression-based supervised learning methods presented in Section 2.2 with unsupervised clustering methods in Section 3.2. For each data point in the test dataset, one identifies a subset that best describes its characteristic or pattern and applies the ML model trained with that subset for the prediction. This strategy is shown to improve the prediction accuracy in Section 5.

3.1. Unsupervised Clustering

3.1.1. k-Means Clustering

A partition of a set X = { x 1 , x 2 , , x n } in mathematics implies a grouping of the elements in X into non-empty disjoint subsets { S 1 , S 2 , , S k } such that every element in X is included in one and only one subset S i for some i { 1 , 2 , , k } . The k-means method is an unsupervised clustering method that partitions data into k clusters to minimize the variance of each group and distance difference. The k-means method first selects k data randomly and sets them as the centroids of k clusters. Subsequently, the data are assigned to groups corresponding to the closest centroids. Once all data are grouped, the centroids of the clusters are recalculated. This procedure is repeated until the clusters of data no longer change. Specifically, the k-means method finds k clusters that minimize the following variance:
min S i i = 1 k x j S i | x j O i | 2 ,
where O i is the centroid of the i-th cluster, S i . Note that { S 1 , S 2 , . S k } is a partition of X; therefore, S i S j = for i j and i = 1 k S i = X . For example, when data similar to that shown in Figure 2a are provided, the k-means method partitions them and introduces k clusters, as illustrated in Figure 2b.
Although the k-means method is efficient, it has certain limitations. First, the number of clusters, k, is a key hyperparameter and should be specified before model training. Section 4.2 introduces a preprocessing method to determine the optimal k from empirical data. Second, the error may converge to a local minimum instead of the global minimum, and the method is easily affected by outliers [17,18]. Third, and most importantly, because the k-means method partitions data into disjoint clusters, some data near the boundaries of the clusters may not be assigned to appropriate clusters. For instance, in Figure 2b, one blue rhombus in the left-most cluster and two sky blue squares in the right-most cluster match that in the middle better. In addition, the two pink and orange triangles in the cluster in the middle match the right-most cluster.

3.1.2. Adaptive Data Selection

To address the limitations of the k-means method, we consider an ADS method. Given a dataset X (called the universe) and a family Ω of the subsets of X, ADS introduces a cover and a subfamily { C i , i = 1 , 2 , } Ω of the subsets of X, whose union is the entire universe X. The cover for X in ADS is constructed as follows.
  • For an arbitrary but fixed point x 1 X , x 1 is set as O 1 , and the distance between O 1 and each point in X is measured. A subset C 1 X is defined by the points in X within the r-neighborhood of O 1 for r > 0 . In this study, the Euclidean L 2 distance is used.
  • For an arbitrary but fixed point x 2 C 1 c = X \ C 1 , the complement of C 1 , x 2 , is set as O 2 , and the distance between O 2 and each point in X is measured. A subset C 2 X is defined by the points in X within the r-neighborhood of O 2 . Because all points in X except C 1 c are considered, C 1 and C 2 may not be disjoint.
  • For an arbitrary but fixed point x 3 C 1 C 2 c = X \ C 1 C 2 , the complement of C 1 C 2 , x 3 , is set as O 3 , and the distance between O 3 and each point in X is measured. A subset C 3 X is defined by the points in X within the r-neighborhood of O 3 . Because all points in X are considered, C 1 C 2 and C 3 may not be disjoint.
  • Generally, when i sets { C 1 , C 2 , , C i } are constructed, for an arbitrary but fixed point x i + 1 k = 1 i C k c = X \ k = 1 i C k , x i + 1 is set as O i + 1 , and the distance between O i + 1 and each point in X is measured. A subset C i + 1 X is defined by the points in X within the r-neighborhood of O i + 1 . Note that C i + 1 may not be disjoint with k = 1 i C k .
  • This procedure is repeated until the union of the subsets includes X. If n subsets C 1 , C 2 , , C n are constructed and the union of these n subsets is X, then the collection, C = { C 1 , C 2 , , C n } , is the cover of X.
Figure 2c shows the cover constructed from ADS for the data presented in Figure 2a. A preliminary version of ADS was proposed in [19].

3.2. Construction of a Group of ML Methods

A dataset X is split into training data X t r and test data X t s . In this study, 80% of X is set as X t r . A typical ML or DL method constructs one model that best describes the data in X t r . The novelty of this study is the construction of a group of models for the data in X t r , which is realized as follows. Suppose that the training data X t r is split into C = { C 1 , C 2 , , C n } using the k-means or ADS method. Specifically, C is the partition of X t r if the k-means method is used, and it is a cover of X t r if the ADS method is used.
Subsequently, for each subset C i , where i = 1 , 2 , , n , a ML algorithm finds the parameters and hyperparameters using only the data in C i and constructs one model M i corresponding to C i . Finally, n models M = { M 1 , M 2 , , M n } are constructed from C = { C 1 , C 2 , , C n } . Note that the data in C i and C j for different i j may have different optimal parameters and hyperparameters, and this procedure allows different models M i and M j for different subsets, which is another novelty of this study.
When the training process is completed, the following procedure is performed on the test data.
For each data point x in the test dataset, X t s ,
  • The distance between x and the centroid, O i , of each C i C ,
    d ( x , O i ) = x O i
    is measured for i = 1 , 2 , , n
  • The subset with the closest distance, e.g.,  C i * , is identified, which minimizes the distance as follows:
    d ( x , O i * ) = min i { d ( x , O i ) , i = 1 , 2 , , n }
  • The ML model, M i * , trained with C i * is used to perform regression to predict the output of x.
Figure 3 illustrates the procedure of the proposed method. In this paper, regression using k-means clustering is called learning with k-means clustering, whereas regression using ADS clustering is referred to as learning with ADS clustering. Regression without clustering, i.e., regression with all training data and a single training model, is called learning without clustering.
The algorithms of learning with k-means clustering and ADS clustering are summarized in Algorithms 1 and 2, respectively.
Algorithm 1:k-means clustering-based learning algorithm
Sensors 22 07982 i001
Algorithm 2: ADS clustering-based learning algorithm
Sensors 22 07982 i002

4. Data and Hyperparameters

In this study, a case study is performed based on an actual diodes dataset in Section 4.1 to test the performance of the proposed method in obsolescence forecasting. Section 4.2 shows the hyperparameters used in the experiments. Those for unsupervised learning are considered in Section 4.2.1 and those for supervised learning in Section 4.2.2.

4.1. Data Collection and Problem Description

Because the accurate prediction of component obsolescence reduces maintenance costs, many electronic component manufacturers and the defense industry have developed obsolescence forecasting systems. Several companies such as RAC, Z2Data, IHS, and QTEC provide their own obsolescence-related information; however, the detailed methodologies and algorithms are not disclosed. Additionally, information on the expected discontinuation dates has large error ranges. Thus, we propose the construction of a suitable prediction algorithm.
First, we build an actual dataset for this study as follows. The part numbers of the discontinued parts were obtained from QTEC, and their detailed characteristics and specifications were obtained from Z2Data. Among the available data, 2366 parts from the category of Zener diodes were selected and used as the dataset. Although parts from the other categories, such as varactors and bridge rectifier diodes, are available, the amount of information is insufficient for ML. Because the parts produced by different manufacturers have different specifications and characteristics in terms of the content and format, detailed technical information in the data sheets and test reports was thoroughly reviewed and standardized. The data of Zener diodes include 16 numeric and 21 categorical features. Table 2 lists their features and data types.
The last time buy (LTB) is the last order that is placed with a component supplier when the production of a component is to be terminated, and the obsolescence date in this study is the LTB date. Since the LTB date is an absolute date, the ML regression algorithms in this study use the duration between the part introduction date and the LTB date as the target variable. Once this relative date is predicted, the predicted obsolescence date is obtained by adding the predicted value to the part introduction date. The information about the part introduction dates is obtained from several services such as QTEC, IHS, or Z2Data. There are some parts whose introduction dates from those services are different, and the most recent date is selected as the introduction date in such cases.
Table 3 summarizes an example of Zener diode data. For simplicity, the values of only a few important features are presented.
Table 4 presents the descriptive statistics of the numeric features. count, mean, and std represent the number, mean, and standard deviation of the data, respectively. min and max are the minimum and maximum values, respectively; 25%, 50%, and 75% are quartiles, which divide the data points into four parts.
The features in Table 2 have their own significance when used in various ML methods. In feature importance, techniques assign scores to features based on their contributions to the prediction of the target variable. In this study, it is measured based on permutation feature importance, which represents the decrease in the model score when one feature value is randomly shuffled. This procedure breaks the relationship between the target and a feature, and thus, a decrease in the model score indicates the degree of model dependence on that feature. The features are standardized to zero mean and unit variance before the feature importance process, and the R 2 score is used as the scoring parameter of the importance function. Figure 4a shows the top ten importance features when the DT method is employed. The importance features are different for different methods, as shown in Figure 4b,c, which presents the top ten importance features when the RF and GB methods are used, respectively. Because the amount of data is small, dimensionality reduction based on the feature importance is not used in this study.

4.2. Hyperparameters

4.2.1. Hyperparameters for Unsupervised Learning Algorithms

Both the unsupervised clustering methods—k-means and ADS—have hyperparameters. For the k-means method, the appropriate number of clusters, k, needs to be estimated. It can be determined using the elbow method based on the sum of the squared errors or the silhouette method based on the silhouette score. Because the use of the k-means methods is aimed at helping in prediction, we select the optimal k based on the following preliminary regression stage. We use the DT method with the training data and observe the prediction accuracy as k increases. The DT method is employed in preprocessing because it is rapid, intuitive, and reasonably accurate. We determine the change in the error, | e k e k 1 | , where e k is the mean relative error (MRE) defined in Section 5 with k clusters, and select the k for which the change is sufficiently small. Specifically, the optimal k = k * is chosen as follows:
k * = arg min k { e k , k 1 < h } ,
where
e k , k 1 α ( e k e k 1 )
and h is a predefined threshold. α = ( e 2 e 1 ) 1 is introduced to avoid dataset dependency. The minimum value is considered in (5) because a small k reduces the computational cost. Figure 5 shows e k , k 1 in (6) for several k values. In this study, h = 0.06 is used, and the corresponding optimal k for the dataset is k * = 5 .
In ADS, the radius, r, of a cluster is an important hyperparameter. Let R be the maximum distance between two arbitrary points in the training data, X t r , i.e.,
R max { d ( x i , x j ) , x i , x j X t r } .
If r = R , then a single cluster includes all data in X t r regardless of the centroid. Therefore, we let r = ϵ R for 0 < ϵ < 1 and calculate the MRE for various values of ϵ . The value minimizing the MRE is selected as the optimal value. Similar to the k-means method, DT is used to estimate the prediction accuracy, and the results are presented in Figure 6. The MRE is minimized when r = ϵ * R with ϵ * = 0.03 , which is the value of ϵ used for the ADS method in this study. Table 5 lists the hyperparameters for the unsupervised clustering algorithms.

4.2.2. Hyperparameters for Supervised Learning Algorithms

The hyperparameters of the supervised ML and DL methods used in this study are listed in Table 6. The left-most column contains the names of the parameters defined in the Scikit library [20]. The column in the middle provides the definition of each hyperparameter, and the right-most column lists the values of the hyperparameters considered in this study. Four hyperparameters are considered for the DT and GB methods and three for the RF, DNN, and RNN methods. For instance, m a x _ d e p t h for the DT method represents the maximum depth of the tree and has one of five values: None, 2, 4, 6, and 8. If None, then the nodes are expanded until all leaves are pure or until all leaves contain fewer than m i n _ s a m p l e s _ s p l i t samples. A grid of all the possible hyperparameter combinations is created, and the best hyperparameters are obtained. Model tuning with grid search is performed for all models using the hyperparameter values in Table 6.
Figure 7 compares the optimal hyperparameters of DT without clustering and 5 sets of optimal hyperparameters for 5 subsets from k-means. Each subset from k-means has different values of hyperparameters. For instance, max_depth for the leftmost data is 8 while that for the data in the middle is 6. Different parameters and hyperparameters for each subset confirm that a single DT model may not be appropriate for the given dataset and that a group of DT models from the proposed method will produce better predictions.

5. Results and Discussion

We show the performance of the proposed method for the obsolescence prediction of Zener diodes. Section 5.1 defines the metrics used for the measurements of prediction errors. Section 5.2 validates the reliability of the proposed method by considering the prediction of obsolescence dates of non-discontinued parts. The errors of the proposed method are compared with those of naive statistics in Section 5.3 and with those of other ML methods in Section 5.4. Confidence intervals of the errors are considered in Section 5.5.

5.1. Metrics

In this study, two distance metrics are considered. The first is the MRE defined in (7).
MRE = 1 n i = 1 n y i y ˜ i y i ,
where y i is the actual target value, and y ˜ i is its predicted value; and n is the number of points used to make the predictions. The second metric is the root mean squared relative error (RMSRE) defined in (8).
RMSRE = 1 n i = 1 n y i y ˜ i y i 2
The MRE and RMSRE are similar to the mean absolute error (MAE) and the root mean squared error (RMSE), respectively; however, the errors from the latter two might be misleading because they depend on the scale of the values. By contrast, the former metrics measure relative errors without depending on the value ranges and, thus, are used in this study.
Instead of ML methods, standard statistical models may be used to perform regression analysis of the obsolescence date. The sample mean of the observed obsolescence dates in the training data is defined as
S TATISTIC = 1 N t r i = 1 N t r y i ,
where N t r = | X t r | is the cardinality of the training data. This is one of the possible statistics and used as a benchmark in this study.

5.2. Reliability of the Proposed Method

In order to check whether the proposed method is reliable, the method has been applied to the obsolescence prediction of non-discontinued Zener diodes as well. We do not know when the correct obsolescence date will be for the non-discontinued parts, but if the obsolescence dates are predicted to be in the past, the proposed method will be questionable. We collected 198 non-discontinued Zener diodes, and Figure 8 shows the histogram of their predicted obsolescence dates when the proposed method is applied with ADS and DT. All the obsolescence prediction dates are in the future, and none are in the past. Similar results are obtained when the proposed method is applied with other methods, which confirms that the proposed scheme is reliable and not questionable.

5.3. Comparison with Naive Statistics

First, we compare the prediction accuracy of the ML methods with that of naive statistics. Figure 9 shows the average E [ r i ] of the relative errors of the prediction, which is defined as
r i y i y ˜ i y i
when the DT method and the naive statistic are used. The prediction error of the DT method without clustering is 0.069, and the error of the naive statistic is 0.928, which is 13 times larger than the former. The error of the DT method with k-means clustering is 0.030, and that of the naive statistic is five times larger. Similar improvements are achieved using ADS clustering. As is evident from the figure, clustering improves the regression. The prediction error of the DT method without clustering is 0.069, which decreases to 0.030 with the k-means clustering and further reduces to 0.011 with the ADS clustering. Moreover, the error of the naive statistic is improved from 0.928 without clustering to 0.175 with the k-means clustering and 0.016 with the ADS clustering.
Figure 10 shows the distribution accuracy in terms of the RMSRE, and the results are similar to those of the MRE in Figure 9. The RMSRE error of the DT method without clustering is 0.319 and the error of the statistic is 2.751, which is eight times larger than the former. The error of the DT method with the k-means clustering method is four times smaller than that of the naive statistic. Additionally, the error of the DT method without clustering decreases to 0.183 with k-means clustering and further reduces to 0.073 with ADS clustering. The naive statistic is also improved by the clustering-based regression.
Figure 11 shows the distributions of the MRE and RMSRE of the RF method. Similar to the corresponding distributions of the DT method presented in Figure 9 and Figure 10, the prediction accuracy using the ML RF method is better than that of the naive statistic, and the prediction accuracy with clustering is better than that without clustering. For instance, the MRE of the statistic without clustering is approximately 13, 4, and 16 times larger than those of the RF method, RF method with the k-means clustering, and RF method with ADS clustering, respectively. Additionally, the MRE of the RF method without clustering is 0.072, which is reduced to 0.042 with the k-means clustering and further decreased to 0.010 with the ADS clustering. Similar trends are observed for the RMSRE. The results from other ML methods, which are not presented herein, showed similar trends and are omitted. These results indicate that ML methods yield considerably better results than the naive statistic, and clustering improves the regression.

5.4. Comparison of ML Methods

This subsection discusses the effects of clustering. The prediction accuracies of various ML methods with and without clustering are compared to determine the method yielding the optimal regression result. Because ML methods are superior to the naive statistic, as discussed in the previous section, the results from the statistic are omitted, and only those obtained using the ML methods are presented.
Figure 12 shows the prediction errors in terms of the MREs of various ML methods, namely DT, RF, GB, DNN, and RNN. Figure 12a–c shows the distributions when the methods are used without clustering, with k-means clustering, and with ADS clustering, respectively. The three shallow learning methods—DT, RF, and GB—yield good prediction accuracy, and the results of the two DL methods—DNN and RNN—are slightly inferior. DL typically results in better results than conventional ML methods; however, the insufficient amount of data in this study results in inferior performance.
Figure 13 shows the prediction errors in terms of the RMSRE, and the trends are similar to those of the MRE, as presented in Figure 12. Notably, the RMSREs are relatively larger than the MREs owing to insufficient data, and the RMSRE is more dependent on such values than the MRE. Both Figure 12 and Figure 13 show that clustering improves the prediction accuracy of the ML methods. Moreover, the ADS-based clustering, the results of which are presented in Figure 12c and Figure 13c, outperforms the k-means clustering, the results of which are shown in Figure 12b and Figure 13b.

5.5. Confidence Intervals

Figure 14 presents 95% confidence intervals of the errors of the various methods. The intervals for the ML methods without clustering are larger than those for those with clustering, and the ADS-based clustering reduces the widths more than the k-means clustering.
Table 7 lists the widths of the 95% confidence intervals shown in Figure 14. Evidently, methods with clustering have a smaller confidence interval width than that without clustering, and the confidence interval width with ADS is much smaller than that with k-means. Therefore, clustering yields estimates that are more stable and accurate, and the estimates with the ADS clustering are superior to those with the k-means clustering.

6. Conclusions

Given a labeled dataset, a supervised ML algorithm finds parameters and hyperparameters from the data and builds a single model for classification or regression. When the dataset is highly nonlinear, however, a single model may not be sufficient for accurate modeling or prediction. In this study, we propose a method to split the dataset into its subsets and identify the parameters and hyperparameters for each subset. We apply two unsupervised clustering methods to split the dataset to its subsets. One is k-means clustering, whose subsets are mutually disjoint and partition the whole dataset. The other is ADS, whose subsets have non-empty intersections. Subsequently, supervised learning builds one model for each subset and eventually forms a group of ML models for the whole dataset. For each data point in the test dataset, one identifies a subset that best describes the characteristics or pattern of the data point and applies the ML model corresponding to the subset for the prediction.
Supervised learning is an ML approach for the classification or regression of the labeled data, while unsupervised learning is for clustering, dimensionality reduction, or association of the unlabelled data. It is the novelty of the study to introduce a method to combine unsupervised clustering with supervised regression to develop a prediction model with improved accuracy and efficiency.
Although both k-means and ADS improved the prediction accuracy, the ADS-based regression outperformed the k-means-based method. This is because the covers from the ADS allow non-empty intersections, enabling better handling of the data near the boundaries of the subsets than with the k-means method. The results of this study are novel because they demonstrate that unsupervised clustering methods can improve supervised learning algorithms.
For future research, we can consider several directions. First, the proposed method showed poor results in this study when combined with DNN and RNN methods because the electrical component data considered in this study were insufficient for DL methods. If sufficient data are obtained, the proposed method combined with DL methods is expected to yield improved predictions in terms of accuracy and efficiency.
Combining unsupervised clustering with supervised learning to improve the prediction accuracy can be extended and validated in various fields. Great effects are expected when the dataset is highly nonlinear, as in financial datasets. Thus, the application of the proposed method to financial optimization or derivatives pricing will be an interesting problem. We hope that our study provides a good reference for the predictions not only in the area of DMSMS but also in many other similar areas.
ADS is an unsupervised learning method that was developed very recently. More detailed analysis of its properties and usefulness in clustering or dimensionality reduction should be studied.

Author Contributions

Conceptualization, H.K.; methodology, H.K. and K.-S.M.; software, H.W.L.; validation, H.K., K.-S.M. and H.W.L.; formal analysis, H.K. and K.-S.M.; investigation, H.K.; resources, H.K.; data curation, H.K.; writing—original draft preparation, H.K.; writing—review and editing, H.K. and K.-S.M.; visualization, H.W.L.; supervision, H.K.; project administration, H.K.; funding acquisition, H.K. and K.-S.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grants funded by the Korean government (MSIT) (NRF-2018R1D1A1B07050046 and No. 2021R1F1A1054766).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Acknowledgments

The data used in this study were collected and provided by Leo Innovision Ltd.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DTdecision tree
RFrandom forest
GBgradient boosting
DNNdeep neural network
RNNrecurrent neural network
MREmean relative error
RMSREroot mean squared relative error

References

  1. Bartels, B.; Ermel, U.; Sandborn, P.; Pecht, M.G. Strategies to the Prediction, Mitigation and Management of Product Obsolescence; Wiley Series in Systems Engineering and Management Book 87; John Wiley & Sons: Hoboken, NJ, USA, 2012. [Google Scholar]
  2. Trabelsi, I.; Zolghadri, M.; Zeddini, B.; Barkallah, M.; Haddar, M. Prediction of obsolescence degree as a function of time: A mathematical formulation. Comput. Ind. 2021, 129, 103470. [Google Scholar] [CrossRef]
  3. Jennings, C.; Wu, D.; Terpenny, J. Forecasting obsolescence risk and product life cycle with machine learning. IEEE Trans. Compon. Packag. Manuf. Technol. 2016, 6, 1428–1439. [Google Scholar] [CrossRef]
  4. Solomon, R.; Sandborn, P.A.; Pecht, M.G. Electronic part life cycle concepts and obsolescence forecasting. IEEE Trans. Compon. Packag. Technol. 2000, 23, 707–717. [Google Scholar]
  5. Sandborn, P.A.; Mauro, F.; Knox, R. A data mining based approach to electronic part obsolescence forecasting. IEEE Trans. Compon. Packag. Technol. 2007, 30, 397–401. [Google Scholar] [CrossRef]
  6. Rojo, F.J.R.; Roy, R.; Shehab, E. Obsolescence management for long-life contracts: State of the art and future trends. Int. J. Adv. Manuf. Technol. 2010, 49, 1235–1250. [Google Scholar] [CrossRef] [Green Version]
  7. Sandborn, P.; Prabhakar, V.; Ahmad, O. Forecasting electronic part procurement lifetimes to enable the management of DMSMS obsolescence. Microelectron. Reliab. 2011, 51, 392–399. [Google Scholar] [CrossRef]
  8. Ma, J.; Kim, N. Electronic part obsolescence forecasting based on time series modeling. Int. J. Precis. Eng. Manuf. 2017, 18, 771–777. [Google Scholar] [CrossRef]
  9. Mastrangelo, C.; Olson, K.; Summers, D. A risk-based approach to forecasting component obsolescence. Microelectron. Reliab. 2021, 127, 114330. [Google Scholar] [CrossRef]
  10. Jordan, M.I.; Mitchell, T.M. Machine learning: Trends, perspectives, and prospects. Science 2015, 349, 255–260. [Google Scholar] [CrossRef] [PubMed]
  11. Sezer, O.B.; Gudelek, M.U.; Ozbayoglu, A.M. Financial time series forecasting with deep learning: A systematic literature review: 2005–2019. Appl. Soft Comput. J. 2020, 90, 106181. [Google Scholar] [CrossRef] [Green Version]
  12. Grichi, Y.; Beauregard, Y.; Dao, T.-M. A random forest method for obsolescence forecasting. In Proceedings of the 2017 IEEE International Conference on Industrial Engineering and Engineering Management (IEEM), Singapore, 10–13 December 2017; pp. 1602–1606. [Google Scholar]
  13. Grichi, Y.; Beauregard, Y.; Dao, T.-M. Optimization of obsolescence forecasting using new hybrid approach based on the RF method and the meta-heuristic genetic algorithm. Am. J. Manag. 2018, 2, 27–38. [Google Scholar]
  14. Moon, K.-S.; Lee, H.W.; Kim, H.J.; Kim, H.; Kang, J.; Park, W.C. Forecasting obsolescence of components by using a clustering-based hybrid machine-learning algorithm. Sensors 2022, 22, 3244. [Google Scholar] [CrossRef] [PubMed]
  15. Müller, A.; Guido, S. Introduction to Machine Learning with Python: A Guide for Data Scientists; O’Reilly Media: Sebastopol, CA, USA, 2016. [Google Scholar]
  16. Géron, A. Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems, 2nd ed.; O’Reilly Media: Sebastopol, CA, USA, 2019. [Google Scholar]
  17. Raschka, S.; Mirjalili, V. Python Machine Learning: Machine Learning and Deep Learning with Python, Scikit-Learn, and TensorFlow 2, 3rd ed.; Packt Publishing: Birmingham, UK, 2019. [Google Scholar]
  18. Kanungo, T.; Mount, D.M.; Netanyahu, N.S.; Piatko, C.D.; Silverman, R.; Wu, A.Y. An efficient k-means clustering algorithm: Analysis and implementation. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 881–892. [Google Scholar] [CrossRef]
  19. Kim, H.; Jun, S.; Moon, K.-S. Stock market prediction based on adaptive training algorithm in machine learning. Quant. Financ. 2022, 22, 1133–1152. [Google Scholar] [CrossRef]
  20. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
Figure 1. Schematic of (a) DNN and (b) RNN.
Figure 1. Schematic of (a) DNN and (b) RNN.
Sensors 22 07982 g001
Figure 2. (a) All data X without clustering, (b) partitioning of X by k-means clustering, and (c) covering of X with ADS clustering. Each color represents the data with the same characteristics.
Figure 2. (a) All data X without clustering, (b) partitioning of X by k-means clustering, and (c) covering of X with ADS clustering. Each color represents the data with the same characteristics.
Sensors 22 07982 g002
Figure 3. Schematic of the proposed ML-based prediction method using unsupervised clustering methods.
Figure 3. Schematic of the proposed ML-based prediction method using unsupervised clustering methods.
Sensors 22 07982 g003
Figure 4. Feature importance from (a) DT, (b) RF, and (c) GB.
Figure 4. Feature importance from (a) DT, (b) RF, and (c) GB.
Sensors 22 07982 g004aSensors 22 07982 g004b
Figure 5. The blue solid line represents the accuracy, e k , k 1 in (6), of the DT method for different k values of k-means for Zener diodes. The yellow dotted line represents the e k , k 1 = 0.06 , and the intersecting red dot represents the optimal value k * = 5 .
Figure 5. The blue solid line represents the accuracy, e k , k 1 in (6), of the DT method for different k values of k-means for Zener diodes. The yellow dotted line represents the e k , k 1 = 0.06 , and the intersecting red dot represents the optimal value k * = 5 .
Sensors 22 07982 g005
Figure 6. The blue solid line represents the accuracy (MRE) of DT for different radii of ADS for Zener diodes. The yellow dotted line represents the minimum value of MRE, and the intersecting red dot represents the optimal value ϵ * = 0.03 .
Figure 6. The blue solid line represents the accuracy (MRE) of DT for different radii of ADS for Zener diodes. The yellow dotted line represents the minimum value of MRE, and the intersecting red dot represents the optimal value ϵ * = 0.03 .
Sensors 22 07982 g006
Figure 7. The values of the hyperparameters of DT without clustering and with k-means.
Figure 7. The values of the hyperparameters of DT without clustering and with k-means.
Sensors 22 07982 g007
Figure 8. The histogram of predicted obsolescence dates for non-discontinued data using the proposed method with ADS and DT.
Figure 8. The histogram of predicted obsolescence dates for non-discontinued data using the proposed method with ADS and DT.
Sensors 22 07982 g008
Figure 9. Distributions of MRE values of predictions using the DT method and naive statistic without clustering, with k-means clustering, and with ADS clustering.
Figure 9. Distributions of MRE values of predictions using the DT method and naive statistic without clustering, with k-means clustering, and with ADS clustering.
Sensors 22 07982 g009
Figure 10. Distributions of RMSRE values of predictions using DT method and naive statistic without clustering, with k-means clustering, and ADS clustering.
Figure 10. Distributions of RMSRE values of predictions using DT method and naive statistic without clustering, with k-means clustering, and ADS clustering.
Sensors 22 07982 g010
Figure 11. Distributions of (a) MRE and (b) RMSRE values of predictions using the RF method and naive statistic without clustering, with k-means clustering, and with ADS clustering.
Figure 11. Distributions of (a) MRE and (b) RMSRE values of predictions using the RF method and naive statistic without clustering, with k-means clustering, and with ADS clustering.
Sensors 22 07982 g011
Figure 12. Distributions of MRE values of predictions using DT, RF, GB, DNN, and RNN methods (a) without clustering, (b) with k-means clustering, and (c) with ADS clustering.
Figure 12. Distributions of MRE values of predictions using DT, RF, GB, DNN, and RNN methods (a) without clustering, (b) with k-means clustering, and (c) with ADS clustering.
Sensors 22 07982 g012
Figure 13. Distributions of RMSRE values of predictions using DT, RF, GB, DNN, and RNN methods (a) without clustering, (b) with k-means clustering, and (c) with ADS clustering.
Figure 13. Distributions of RMSRE values of predictions using DT, RF, GB, DNN, and RNN methods (a) without clustering, (b) with k-means clustering, and (c) with ADS clustering.
Sensors 22 07982 g013aSensors 22 07982 g013b
Figure 14. Comparison of 95% confidence intervals of predicted values using various methods. (a) Without clustering, (b) with k-means clustering, and (c) with ADS clustering.
Figure 14. Comparison of 95% confidence intervals of predicted values using various methods. (a) Without clustering, (b) with k-means clustering, and (c) with ADS clustering.
Sensors 22 07982 g014
Table 2. Features of data of Zener diodes.
Table 2. Features of data of Zener diodes.
TypeFeatures
NumericPower Dissipation, Reverse Zener Voltage (Min), Reverse Zener Voltage (Max), Test Current, Zener Impedance (Max), Zener Impedance at IZK, Maximum Zener Current, Reverse Leakage Current at VR, Reverse Voltage, Forward Voltage, Voltage Tolerance, Forward Current, Diode Capacitance, Operating Temperature (Min), Operating Temperature (Max), Number of Terminals
CategoricalPart Number, Mfr Name, Description, Polarity, ESD Protection, Temperature Coefficient, EU RoHS, Halogen Free, Package Code, Soldering Surface Treatment, Mounting Type, JESD-30 Code, Package Body Material, Package Shape, Package Style, Terminal Form, Terminal Position, Temperature Grade, Part Status, Part Introduction, and Obsolete Date (LTB Date)
Table 3. Example of Zener diode data.
Table 3. Example of Zener diode data.
FeatureValue
Zener Impedance at IZK ( Ω )2000.0
Forward Voltage (V)1.2
Part Number1N4761A
DescriptionZENER DIODE, 1 W, 75 V@3 MA, 5%
ESD ProtectionUnknown
Operating Temperature (Max) ( C )200
Mounting TypeThrough Hole
Mfr NameCENTRAL SEMICONDUCTOR CORP.
Power Dissipation (Max) (W)1.0
Package CodeDO-41
Table 4. Statistics of certain numeric features of discontinued Zener diodes.
Table 4. Statistics of certain numeric features of discontinued Zener diodes.
CountMeanStdMin25%50%75%Max
Power Dissipation (Max) (W)23662.522.490.120.501.05.0010.0
Reverse Zener Voltage (Min) (V)105126.7036.661.805.6012.428.50190.0
Reverse Zener Voltage (Max) (V)236642.2352.441.808.5320.053.16270.0
Test Current (mA)236632.6161.260.255.0010.030.00640.0
Zener Impedance (Max) ( Ω )2366110.81240.051.009.0030.0100.002500.0
Zener Impedance at IZK (Max) ( Ω )19361129.161361.7060.00400.00700.01300.008000.0
Maximum Zener Current (mA)1244200.44288.661.5431.6085.0264.002380.0
Reverse Leakage Current @ VR (uA)23668.7322.780.051.002.05.00300.0
Reverse Voltage (V)236631.1639.760.506.0015.038.80206.0
Forward Voltage (V)18111.290.210.901.201.21.501.5
Voltage Tolerance (Max) (%)23664.622.561.002.385.05.0020.0
Forward Current (Max) (mA)1811437.88395.932.00200.00200.01000.001000.0
Diode Capacitance @ f = 1 MHz, VR = 0 V (pF)157178.23148.2419.0070.00130.0225.00450.0
Operating Temperature (Min) ( C )2366−62.524.32−65.00−65.00−65.0−65.00−55.0
Operating Temperature (Max) ( C )2366171.5616.36125.00150.00175.0175.00200.0
Number of Terminals23661.980.490.002.002.02.004.0
Table 5. Hyperparameters for unsupervised learning algorithms.
Table 5. Hyperparameters for unsupervised learning algorithms.
MethodHyperparameterValue
k-meansNumber of clusters (k)5
ADSRadius of cover (r) 0.03 R
Table 6. Hyperparameters for supervised learning algorithms.
Table 6. Hyperparameters for supervised learning algorithms.
DTDefinitionValues
max_depthMaximum depth of treeNone, 2, 4, 6, 8
min_samples_splitMinimum number of samples required to split internal node2, 4, 6, 8
min_samples_leafMinimum number of samples required to be at leaf node2, 3, 4, 5
max_leaf_nodesMaximum number of terminal nodes20, 40, 60, unlimited
RFDefinitionValues
min_samples_splitMinimum number of samples required to split internal node2, 3, 4, 5
n_estimatorsNumber of trees in forest100, 150, 200
max_featuresNumber of features to consider when determining best splitauto, sqrt, log2
GBDefinitionValues
learning_rateLearning rate reduces contribution of each tree0.01, 0.1, 0.2
subsampleFraction of samples to be used for fitting individual base learners0.5, 0.6, 0.7, 0.8, 0.9, 1
n_estimatorsNumber of boosting stages to perform100, 200, 300, 400, 500
max_depthMaximum depth of individual regression estimators2, 4, 6, 8, 10
DNNDefinitionValues
unitProcessors between input and output units in connectionist system32, 64
optimizerAdjusts model weights to maximize loss functionAdam, Nadam, RMSprop
dropoutTechnique by which randomly selected neurons are ignored during training0, 0.1, 0.01
RNNDefinitionValues
unitProcessors between input and output units in connectionist system32, 64
optimizerAdjusts model weights to maximize loss functionAdam, Nadam, RMSprop
dropoutTechnique by which randomly selected neurons are ignored during training0, 0.1, 0.01
Table 7. Comparison of widths of 95% confidence intervals of predicted values using various methods.
Table 7. Comparison of widths of 95% confidence intervals of predicted values using various methods.
Without_Clusteringk-MeansADS
DT1.2491440.7166510.285660
RF1.3849821.1574870.216856
GB1.2532581.0467530.190099
DNN1.6309551.7533720.699891
RNN1.9463441.6147360.266980
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Moon, K.-S.; Lee, H.W.; Kim, H. Adaptive Data Selection-Based Machine Learning Algorithm for Prediction of Component Obsolescence. Sensors 2022, 22, 7982. https://doi.org/10.3390/s22207982

AMA Style

Moon K-S, Lee HW, Kim H. Adaptive Data Selection-Based Machine Learning Algorithm for Prediction of Component Obsolescence. Sensors. 2022; 22(20):7982. https://doi.org/10.3390/s22207982

Chicago/Turabian Style

Moon, Kyoung-Sook, Hee Won Lee, and Hongjoong Kim. 2022. "Adaptive Data Selection-Based Machine Learning Algorithm for Prediction of Component Obsolescence" Sensors 22, no. 20: 7982. https://doi.org/10.3390/s22207982

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop