Next Article in Journal
Sparse Optimistic Based on Lasso-LSQR and Minimum Entropy De-Convolution with FARIMA for the Remaining Useful Life Prediction of Machinery
Previous Article in Journal
Entanglement of Three-Qubit Random Pure States
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Objective Evolutionary Instance Selection for Regression Tasks

1
Department of Computer Science and Automatics, University of Bielsko-Biała, ul. Willowa 2, 43-309 Bielsko-Biała, Poland
2
Institute of Computational Intelligence, Częstochowa University of Technology, 42-201 Częstochowa, Poland
*
Author to whom correspondence should be addressed.
Entropy 2018, 20(10), 746; https://doi.org/10.3390/e20100746
Submission received: 23 August 2018 / Revised: 20 September 2018 / Accepted: 25 September 2018 / Published: 29 September 2018
(This article belongs to the Section Complexity)

Abstract

:
The purpose of instance selection is to reduce the data size while preserving as much useful information stored in the data as possible and detecting and removing the erroneous and redundant information. In this work, we analyze instance selection in regression tasks and apply the NSGA-II multi-objective evolutionary algorithm to direct the search for the optimal subset of the training dataset and the k-NN algorithm for evaluating the solutions during the selection process. A key advantage of the method is obtaining a pool of solutions situated on the Pareto front, where each of them is the best for certain RMSE-compression balance. We discuss different parameters of the process and their influence on the results and put special efforts to reducing the computational complexity of our approach. The experimental evaluation proves that the proposed method achieves good performance in terms of minimization of prediction error and minimization of dataset size.

1. Introduction

Data preprocessing is a crucial step in data mining systems. It is frequently more important than the selection of the best prediction model, as even the best model cannot obtain good results if it learns using poor quality data [1]. A part of data preprocessing is data selection, which comprises feature selection and instance selection.
The purpose of instance selection is to preserve useful information stored in the data and reject the erroneous information, while reducing the data size by selecting an optimal set of instances. This allows for accelerating the predictive model training and to obtain a lower prediction error [2].
Reducing the data size also makes it easier to analyze the properties of the data by humans, as well as allowing the assessment of the expected performance of the prediction models [3]. In other words, by instance selection, we want to “compress the information”. In this work, we consider instance selection in regression problems, which is a more complex task than instance selection in classification problems [4] an much less literature exists on this topic.
Instance selection finds practical application in a range of problems, where the data size can be reduced. For example, it can be applied to the datasets considered in this study, describing real-world problems from various domains. In addition, one of the authors took part in two practical implementations in the industry. The first one was an artificial intelligence-based system for controlling steel production in electric arc process, where there was a lot of data from the previous processes (such as the amount of energy, of different chemical compounds, etc.). The other one was in the electronics industry in a system for predicting the performance of the electronic appliances (power inverters and others) where the amount of data describing the parameters and behavior of the appliances was also very large. In both cases, there were regression problems with many redundant and erroneous data and instance selection was very useful to enable efficient further processing of the data.
The first difference between instance selection for classification and regression tasks that in classification it is enough to determine the class boundaries (the black thick line in Figure 1a), and to select only the instances needed to determine the boundaries [5]. The remaining instances (in the grayed area) can be removed. However, before removing them, the noisy instances, which do not match their neighbor class, must be removed first in order not to introduce the false classification boundaries.
In the case of instance selection for regression tasks, we also need to remove the noisy instances, which do not match their neighbors (instances A and B in Figure 1b) and then we can remove the instances that are very close to some other instances in terms of input and the output space (instances C and D in Figure 1b). However, the reduction cannot be so strong as in classification problems, where we need only the class boundaries because, in regression, each point in the data space is important.
In instance selection for classification tasks, we can obtain strong data reduction and we can also obtain more balanced class distribution and thus higher entropy H calculated as [6]:
H = i = 1 c p ( X i ) log p ( X i ) ,
where c is the number of classes in the initial dataset and p ( X i ) is the proportion of instances of i-th class to all instance in the dataset.
In regression problems, we can estimate the dependent variable y as a deterministic continuous function of the independent variables x and use differential entropy H d [7]:
H d = S f x log f ( x ) d x ,
where S is area covered by x. We discuss the connections between measures of information and loss functions with instance selection performance in Section 3.4.
However, in practice, instance selection is not so simple as in Figure 1, where there are only two attributes and a few instances. We cannot say which instance needs to be rejected without taking into account which other instances are also rejected. This is because the outcome depends on the set on which the predictive models are trained and thus we must consider the set of selected instances as a whole, which makes instance selection an NP-hard problem (see Section 3.1). For that reason, we decided to base the instance selection method on evolutionary algorithms, where the advantage of this approach is that the evolutionary algorithm evaluates the prediction quality on the entire subsets of selected instances and we do not have to explicitly define the relation of an instance to their neighbors (which in regression tasks is not as simple as in classification) to decide upon its selection or rejection.
Another advantage of evolutionary-based methods is the possibility of obtaining solutions with low prediction error (which in case of regression problems is typically expressed by RMSE—root mean square error) and strong data reduction (compression) at the same time.
As the discussed solution is based on multi-objective optimization, a key advantage of it is that we obtain a pool of solutions situated on the Pareto front, where each of them is the best for certain RMSE-compression balance and we can choose one of them, as will be discussed in Section 2.2.
In binary instance selection, each vector (instance) can be either selected or rejected. In instance weighting, the instances can be assigned real value weights between 0 and 1, which reflect the instance importance for building the predictive model. Then, the model includes the contribution of particular instances in the learning process proportionally to the weights assigned to them [8]. Obviously, we always want to select the most representative instances, so that the reduced set contains as much useful information and as low noise as possible.
First, we review the problems and existing solutions in instance selection with non-evolutionary (Section 1.1) and evolutionary approaches (Section 1.2). Then, we introduce our approach, which uses the multi-objective evolutionary NSGA-II algorithm [9] with k-NN as the inner evaluation algorithm and propose several modifications and improvements (Section 2). Finally, the discussion is supported with experimental evaluations (Section 3).

1.1. Non-Evolutionary Instance Selection Algorithms

Non-evolutionary instance selection algorithms usually are based on some local properties of the dataset, as the nearest neighbors or Voronoi cells, in order to assess which instances can be removed as noisy or redundant.
The CNN (Condensed Nearest Neighbor) algorithm was the first instance selection algorithm developed by Hart [10]. The purpose of CNN is to reject these instances that do not bring any additional information into the classification process. A popular noise filter is ENN (Edited Nearest Neighbor) proposed by Wilson [11] to increase classification accuracy by noise reduction. These two algorithms are still among the most useful due to their simplicity. In later years, more instance selection algorithms have been proposed for classification tasks [5,12,13,14,15,16,17,18].
In the literature, DROP-3 and DROP-5 are frequently considered the most effective of them [11].
There were also some works to directly use information theory for instance selection in classification tasks. Son [19] proposed a method, where the data set was segmented into several partitions. Then, each partition was divided continuously based on entropy, until all partitions are pure or no further partitioning can be done. Then one can search for the representative instance in each partition. Kajdanowicz [20] introduced a method for comparison and selection of the training set using entropy-based distance.
Recently, adaptations of instance selection methods for multi-output classification problems were proposed [21]. A taxonomy and comparative study of instance selection methods for classification can be found in [2,22].
It is obvious that fewer papers addressed the problem of instance selection in regression tasks and one of the first approaches was presented by Zhang [23]. Since there are no classes in regression, several approaches to replace the “class” with another concept were used.
In [24,25], the class concept was replaced by some threshold distance. If the distance in the input space between two instances is greater than the threshold, they can be treated by the instance selection algorithm in the same way as different class instances in classification tasks. Another option is to perform discretization and convert the regression task to a multiple class classification task and then use instance selection algorithms for classification problems [26]. In [27], a Class Conditional Instance Selection for Regression (CCISR) was proposed, which was derived from the CCIS for classification [28]. CCIS created two graphs: one for the nearest neighbors of the same class as a given instance and another one for other class instances. A scoring function based on the distances in graphs was applied to evaluate the instances.
Guillen et al. [29] proposed the use of mutual information for instance selection in time series prediction. In the first step, the nearest neighbors of a given point were determined and then the mutual information between that point and each of its neighbors was calculated. If the loss of mutual information with respect to its neighbors was similar to the instances near the examined instance, this instance was included in the selected dataset. The authors of [30,31] extended this idea to instance selection in time series prediction by calculating the mutual information between every instance from the training set and the currently evaluated instance, then to order the training set in descending order by the distances and selected the predefined number of points.
In [32], an instance selection method for regression was based on recursive data partitioning. The algorithm started with partitioning the input space using the k-means clustering. If the ratio of the standard deviation to the mean of the group was less than a threshold, the element closest to the mean of each cluster was marked as a representative. Otherwise, the algorithm continued to split the leaf recursively.
In [25], an adaptation of DROP2 and DROP3 to regression tasks was presented and two solutions were proposed: to compare the accumulated error that occurs when an instance is selected and when it is rejected, and to use the already mentioned concept of the distance threshold. Since both ideas were used to adapt DROP2 and DROP3 to regression, four resultant algorithms were tested. DROP3-RT (Regression-Threshold) definitely worked best of the four methods and thus we use it for comparison in the experiments.
In [4], ensembles of instance selection methods for regression tasks were used. The ensembles consisted of several members of the same instance selection algorithm and by implementing bagging operated on different subsets of the original training set. A final decision was made by voting with a threshold. If an instance obtained more votes than the threshold, it was finally selected. Using a different threshold, a Pareto front of solutions could be obtained, i.e., no solution exists, which can improve both of the objectives more than any solution on the front. Results of this work are also used in the experimental comparison.
The problem with non-evolutionary instance selection algorithms is that in most cases they are based on certain assumptions and observations made by their authors about how the data is typically distributed. For example, the ENN algorithm removes the instances miss-classified by k-NN, considering them noisy. This is not always true, as they may also be boundary instances and indeed ENN has the tendency to smooth the class boundaries. There are also other assumptions in other algorithms that are more frequently true than not, but in some cases they are wrong.

1.2. Evolutionary Instance Selection Algorithms

Evolutionary instance selection algorithms do not make any assumptions about the dataset properties but verify iteratively large number of different subsets in an intelligent way to minimize the search space. This can result in better solutions or better (lower) Pareto front in the case of multiple solutions. On the other hand, this is usually achieved at the expense of much higher computational cost. For that reason, in this work, we pay special attention to limit the computational cost as far as possible, which is discussed in Section 2.3.
Tolvi [33] used genetic algorithms for outlier detection and variable selection in linear regression models, performing both operations simultaneously. However, he evaluated the model on two very small datasets (35 instances with two features and 21 instances with three features); nevertheless, in his experiments, evolutionary instance selection algorithms outperformed the non-evolutionary ones.
Shuning [34] concluded that genetic algorithm based instance selection for classification works best for low entropy datasets and with higher entropy, there will be less benefit from instance selection. In [35], an algorithm called Cooperative Coevolutionary Instance Selection (CCIS) was presented. The method used two populations evaluated cooperatively. The training set was divided into n approximately equal parts and each part was assigned to a sub-population. Each individual of a sub-population encoded a subset of training instances. Every sub-population was evolved using a standard genetic algorithm. The second population consisted of combinations of instance sets.
Tsaia [36] considered jointly instance and feature selection in an evolutionary approach. In addition, in [37], an evolutionary algorithm was presented for instance and feature selection and particular problems were assigned to several populations to handle each one separately and each population was optimizing a part of the problem. Then, the authors tried to join the obtained solutions in an attempt to obtain better results. Czarnowski [38,39] introduced an instance selection method that incorporates several ideas. First clustering was performed on the data and then, within the clusters, the selection was executed by the team of agents. The agents cooperated by sharing a population of solutions and refined the solutions using local search.
We found only two works describing the application of multi-objective evolutionary algorithms to instance selection, both to classification problems and both dated for 2017.
In [40], the MOEA/D algorithm was used in a coevolutionary approach integrating instance selection and generating the hyper-parameters for training an SVM. The two criteria used in that optimization were the reduction of the training set size and the performance with a given set of an SVM’s hyper-parameters. The average results over some classification datasets were provided.
In [41], the authors also considered the over-fitting problem. At each iteration of the genetic algorithm, the training and validation partitions were updated in order to prevent the prototypes from over-fitting a single validation data set. Each time the partitions were updated, all of the solutions in the Pareto set were re-evaluated. However, only the results for 1-NN averaged over several classification problems was reported, similarly, as in the previous work, so it is not possible to compare the results with our method.

2. Multi-Objective Evolutionary Instance Selection for Regression (MEISR)

This section introduces our solution to instance selection in regression tasks called MEISR. MEISR uses a multi-objective evolutionary algorithm, which in the current implementation is NSGA-II [9], to direct the search for the optimal reduced training sets and the k-NN algorithm to determine the error on the training sets. First, particular aspects are discussed and finally the pseudo-code of the whole algorithm is presented.

2.1. Basic Concepts

As it was stated in the introduction, the two objectives of instance selection process are minimization of the number of instances in the reduced training set S and minimization of the error obtained on the test set by predictive models trained on the reduced training set S. The first objective is known as minimization of retention or maximization of reduction or compression. The second one in our case is expressed with root mean square error ( R M S E ) because R M S E is the standard and commonly used measure of regressor performance. Because the words “ R M S E ”, “reduction” and “retention” all begin with “r”, we will denote the various R M S E values on the test set with symbols starting with “r” ( r 0 , r 1 , r 2 , r 3 ) and we will denote the first objective with symbols starting with “c” ( c 0 , c 1 , c 2 , c 3 ), which will stand for retention or 1-compression:
c = r e t e n t i o n = 1 c o m p r e s s i o n = 1 = N s e l N ,
where N is the number of all instances in the training set T , N s e l is the number of selected instances from T , which create the selected set S . We use the standard definitions of root mean square error on the training set ( R M S E t r n ) and on the test set ( R M S E t s t ):
R M S E t r n = 1 N Σ i = 1 N ( y p r e d i c t e d _ i y i ) 2 ,
R M S E t s t = 1 N t s t Σ i = 1 N t s t ( y p r e d i c t e d _ i y i ) 2 ,
where y p r e d i c t e d _ i is the predicted and y i is the actual value of the output variable for the i-th instance (of the training and test set respectively) and where N is the number of all instances in the training and N t s t in the test set. To prevent over-fitting, either validation set or the other stop criterion can be used, as discussed in Section 2.6.
We use a multi-objective evolutionary instance selection method based on the NSGA-II algorithm [9] to maximize compression and minimize R M S E and to obtain a set of solutions on the Pareto front (these are such solutions in which no other solution exists, which can improve both of the objectives more than any solution on the front—see Figure 2).
We chose NSGA-II because it is widely used and, despite the existence of newer algorithms, NSGA-II still gives the best or one of the best results for two-objective problems and its modification NSGA-III for problems with more than two objectives [42]. As the description of NSGA-II can be easily found in the literature [9], we are not going do describe it in detail, but rather discuss the issues specific to the proposed instance selection method, especially that the main focus of our work is on instance selection and not on genetic or evolutionary algorithms.
The NSGA-II algorithm was adjusted to direct the search for solutions for the instance selection task by setting the proper objectives, by using the proper encoding, and by implementing the proposed initialization schemes and mutation and crossover operators, as will be discussed in subsequent sections.
First, an initial population of P individuals is created, where each individual represents one reduced training set S. The length of the chromosome equals the number of instances in the original training set T. At each position of the chromosome, a value w represents the weight of a single instance. If w = 0, then the instance is rejected. If w > 0, then the instance is included in the prediction model (regressor) learning with the weight w. In the simplest binary case, we allow only for two different values of w: 0 and 1. Initially, all weights w have random values generated by the initialization method presented two subsections later.
Then, the evolutionary-based instance selection process starts and by adjusting the weights tries to find a group of the best solutions (reduced training sets S), located on the Pareto front. The quality of the solutions during the process is assessed by the two criteria already mentioned: retention (the lower the better) and the prediction error of the learning model (also the lower the better). A detailed discussion of the objective criteria is provided in the next subsection.

2.2. The Objectives

The first objective is a minimization of the number of instances of the training set (i.e., minimization of retention or maximization of compression).
The second objective is a minimization of R M S E . It must be distinguished between the final objective, which is a minimization of R M S E on the test set ( R M S E t s t ) and the objective used by the instance selection process, which is a minimization of R M S E on the training set ( R M S E t r n ). The final objective ( R M S E t s t ) cannot be minimized directly because the test set is not available while selecting the instances.
During the instance selection process R M S E t r n is internally determined with the leave-one-out procedure, always using the k-NN algorithm as the inner prediction model (the regressor inside the instance selection process). In the final evaluation, R M S E t s t can be determined using any prediction model trained on the reduced training set S. In the experimental section, the final prediction models we used were k-NN with different k parameters and MLP neural networks and the R M S E t s t obtained on the test sets is reported in the experimental results.
Using a single objective genetic algorithm, a fitness function that incorporates all criteria must be defined. The typical definition of the fitness function for instance selection is:
f i t n e s s = α · r e t e n t i o n + ( 1 α ) · R M S E t r n 1 ,
where α is a coefficient indicating the expected balance between the objectives. However, multi-objective solutions do not require the determination of the coefficient α ; instead, they minimize two objectives: retention and R M S E t r n . Moreover, the obtained results consist of a group of non-nominated individuals situated on the Pareto front (i.e., the reduced training sets S) with different trade-offs between objectives.
In our study, the following encoding of individuals is used: each individual (each selected data set S ) is encoded as a vector w = w 1 , , w N , where N stands for the size of the vector (which equals the number of instances in the original training set T ). The vector w can only take specific values assigned from set 0 , σ , 2 σ , 3 σ , , 1 , where σ depends on the algorithm parameter n u m L e v e l s and it is calculated as follows: σ = 1 / n u m L e v e l s 1 . In case of the n u m L e v e l s = 2 vector, w can have only binary values, in case of n u m L e v e l s = 5 vector, w can take five values ( 0.00 , 0.25 , 0.50 , 0.75 , 1.00 ), etc. If n u m L e v e l s = 0 a vector, w can have any real number values assigned. Such a process was aimed at increasing the readability of weights and the ability to test different variants of simulations.
In the experimental section, we present the results for binary instance selection ( n u m L e v e l s = 2 ) and real-value instance weighting ( n u m L e v e l s = 0 ) as these are two most characteristic encodings. Moreover, implementing instance weighting in regression problems frequently allows for some improvement in prediction quality for noisy data [8,43]. Thus, one of the aims of this work is to examine in what conditions implementing real value weights w i is beneficial, in spite of this making the algorithm more complex and interpretation of the results more difficult.
During the instance selection process, the following objectives are used by NSGA-II for binary instance selection:
R M S E t r n w = k n n ( S ( w ) , T ) r e t w = 1 N i = 1 N n r e d w i ,
while for real-value instance weighting the following objectives are directly used by NSGA-II:
R M S E t r n w = k n n ( S ( w ) , T ) r e t w = β · 1 N i = 1 N w i + 1 β · 1 N i = 1 N n r e d w i ,
where R M S E t r n ( w ) is obtained with the k-NN algorithm while predicting output of all instances from the original training set T , using the reduced (selected) training set S given by the weight vector w (each time without the instance currently being predicted). r e t ( w ) in binary instance selection is the sum of instance weights (which equals in this case the number of selected instances), while, in instance weighting, it is a weighted sum of the instance weights w i and the number of instances with non-zero weights n r e d ( w i ) . n r e d ( w i ) returns 1 if the instance is selected and 0 if it is rejected. β is a parameter that balances the sum of the instance weights and the sum of the not rejected instances. The first term, summing the instance weights, is needed to allow crossover and mutation operations to gradually reduce some of the instance weights w i . In case of instance weighting n r e d ( w ) is calculated as follows:
n r e d w = 1 , for x > γ , 0 , for x γ ,
where γ is a parameter that defines a weight, below which an instance is rejected (which is experimentally set as 0.01; however, the exact value is not so crucial, as the algorithm adjusts its behavior to that value by modifying the weights proportionally to γ ). Such an approach allows the optimization algorithm to minimize the instance weights and, consequently, for a reduction of instances. Instances with weights lower than γ get rejected and are not taken into account by the k-NN algorithm, while instances with weights greater than γ are taken into account proportionally to their weights (as well by the inner evaluation as by the final prediction model), as will be discussed in the next section. However, when we report the retention in the experimental results, we take into account the only the number of all instances with non-zero weights.
An example of the obtained Pareto front and the four points of interest are shown in Figure 2. In orange: results on the training set, in green: the corresponding results on the test set. All the points obtained on the Pareto fronts for each dataset are presented numerically and graphically in the supplementary resources. Due to limited space in the paper, we present only four most characteristic points of interest ( r 0 , c 0 ), ( r 1 , c 1 ), ( r 2 , c 2 ) and also ( r 3 , c 3 ) the in cases where r 3 < r 1 . Figure 2 is used to explain the results obtained in one fold of the 10-fold cross-validation. The corresponding values reported in the table are the average values over the whole 10-fold cross-validation:
  • r0—baseline R M S E t s t obtained without instance selection.
  • c0 c 0 is always 1, which means there is no compression on the whole dataset, for that reason the value does not occur in any table.
  • r1 R M S E t s t obtained with instance selection for the point ( c 1 , r 1 ) in Figure 2; this is the R M S E t s t obtained on the test set, while training the model on this reduced training set S, which is the point on the Pareto front with the lowest R M S E t r n on the training set and the weakest compression.
  • c1—retention rate (1-compression) of the point ( c 1 , r 1 ).
  • r2 R M S E t s t obtained with instance selection for the point ( c 2 , r 2 ) in Figure 2; this is the R M S E obtained on the test set, while training the model on this reduced training set, which was represented by the closest obtained point to the point ( r 0 t r n , r e t e n t i o n = 0 ); the brown diagonal line shows this distance d. This point was selected as a representative point because further increasing compression usually leads to sudden increase of R M S E t r n and R M S E t s t , making the area to the left of this point practically unusable.
  • c2—retention rate of the point (c2,r2).
  • r3, c3 R M S E t s t and retention rate obtained with and additional run of the instance selection process with the alternative initialization (90% probability of each instance being included in the initial population). It was aimed to obtain R M S E t s t lower than r1 and r0. However, this was useful only in a few cases; in other cases, r 3 was equal r 0 or to r 1 , which meant that no further decrease of R M S E below r 0 or r 1 (whichever was lower) was possible to obtain. One of the important reasons that it was not always possible to obtain the lowest R M S E with the main instance selection process at the point ( c 1 , r 1 ) was that the Pareto front was extending gradually during the optimization and in some cases. Before it would reach the point corresponding to ( c 3 , r 3 ), the test error for the same compression can already start to increase (we minimize R M S E t r n and report the R M S E t s t ), so we must stop the process earlier (see Section 2.6).
Although the target users of our method are scientists and engineers, who should understand their process and be able to select the appropriate solution from the Pareto front, we can suggest the solution with the lowest R M S E t r n for predictive model learning and the solution marked by the point ( c 2 , r 2 ) in Figure 2 for analyzing the properties of the data. To enable a better choice, it may also be a good idea to display the front graphically (as can be done in the software available in the supplementary resources) so the user can quickly assess all the solutions.

2.3. k-NN as the Inner Evaluation Algorithm

The rationale behind choosing k-NN as the inner evaluation algorithm is the speed of this approach. This is because the full k-NN algorithm has to be performed only once before the optimization starts. In the case of other prediction algorithms, this would be either impossible or much more complex and thus less efficient. Let us assume that there are 96 individuals in the population and that the optimization requires 30 epochs. In this case, the value of the fitness function must be calculated 2880 times. Training any prediction model 2880 times would be computationally very costly.
Although, our previous experiments [24] showed that the best results in terms of R M S E -compression balance can usually be obtained if the inner evaluation model is the same algorithm as the final predictor, in this work, we sacrifice that small improvement in order to shorten the optimization process usually from two to three orders of magnitude. However, as we will show, when the final predictor is an MLP neural network, for the inner evaluation, we use k-NN with parameters, which makes its prediction as close to the prediction of the neural network as possible. In this way, we obtain better results, while still keeping the process time short.
In the case of the k-NN algorithm, we calculate the distance matrix between each pair of instances in the training set. Then, we create one two-dimensional array D i for each instance X i . The first dimension is N—the number of instances in the training set and the second dimension is three. The three values stored in the array are: d i s t ( X i , X j ) , j and y j . d i s t ( X i , X j ) is the distances between the current instance X i and each other instance X j , which in the simplest case is a simple Euclidean distance and in a general case can be expressed by Equation (11). The second value j is the ordinal number of each instance in the dataset T . The third value is the output value y j of the j-th instance.
Then, we sort the arrays D i increasingly by d i s t ( X i , X j ) . At the prediction step, we only go through the beginning of the array, read the instance number j and check if this instance is selected, if not, then we go to the next instance, as long as we find k selected instances. Then, for the k nearest selected instances, we read their weights w j from the chromosome, their output value y j and predict this instance output value y p r e d i c t e d _ i as the weighted average of the k outputs (which is a simple average in case of binary instance selection):
y p r e d i c t e d _ i = j = 1 k y j w j k · j = 1 k w j ,
where y j is the output value of the j-th neighbor of an instance X i and w j is the weight expressing the j-th neighbor importance (it should not be confused with the weight w d j related to the distance between the instances X i and X j as in the standard weighted k-NN algorithm). Only the neighbors with w j > γ are considered. In binary instance selection, w j = 1 always. To prevent an instance from being considered by the algorithm a neighbor of itself, we set the distance to the instance itself to a very large number as the maximum value of the Double type (1.79 × 10 308 ). The prediction step is extremely fast and only these steps are performed each time the R M S E t r n is calculated.
The distance between two instances X i and X j is calculated as:
d i s t ( X i , X j ) = m = 1 A w a m ( x i m x j m ) v m = 1 A w a m 1 / v ,
where A is the number of features (attributes), w a m is the weight of the m-th attribute and x i m and x j m are the values of the m-th attribute of i-th and j-th instance respectively. In the simplest case, all the attribute weights equal 1. However, there are cases where it is useful to use attribute weighting and assign different weights to different attributes, e.g., as the absolute values of the correlation coefficient between each attribute and the output. For example, attribute weighting is useful in improving instance selection results if the final predictor is also k-NN using the attribute weights or when the final predictor is some other algorithm, which internally performs attribute weighting, e.g., an MLP neural network [44]. v is the exponent in Minkovsky distance measure; for v = 2 , the measure becomes the Euclidean distance. There is usually no need to raise the expression in the bracket to the power 1 / v (calculate a square root for v = 2 ) because we need only to find the nearest neighbors and not the distance to them (unless we use weighted k-NN, where the closest neighbors have a greater influence on the prediction).
If the final prediction model is i.e., a neural network, then, during the learning, an error the network makes as a response to presenting a given instance is multiplied by the instance weight:
E r r o r = i = 1 n f ( w i · f ( y p r e d i c t e d _ i y i ) ) ,
where E r r o r is the total error, used in the network learning algorithms to adjust the network weights and f(.) is an error measure, which, for the frequently used mean square error measure, gives:
E r r o r = i = 1 n w i · ( y p r e d i c t e d _ i y i ) 2 .

2.4. Population Initialization

The main goal of population initialization methods is to provide optimal coverage of the search space. This can shorten the evolutionary optimization and enable search for the solutions either more uniformly spread in the solution space or focused in some areas, depending on the needs.
Typical initialization assigns randomly generated values (weights) w i to each position i at the chromosome (to each instance) of the individuals (training datasets). For this purpose, different methods can be used (e.g., Pseudo-Random Generators [45], Quasi-Random Generators [46], population dependent methods [47]). In addition, methods based on value transformations [48], a priori knowledge about the problem or clustering can be used [49]. According to some authors, initialization should be adjusted to a specified problem [50] and it is also the case in instance selection tasks.
One of the best initialization methods that relies on population is Adaptive Randomness [47]. In this method, for each individual in population several candidates are created. Only the candidate for which that smallest distance to the rest of the population is the larger than for all other candidates is added to the population. The advantage of this method is that the created candidates do not have to be evaluated, only the distance to the already assigned individuals is calculated. Population-based methods do not depend on random number generators and number transformation methods. Due to this, they can be combined with other types of initialization methods [51]. In this work, such combinations have been tested with an idea to produce different initial values (e.g., initialize smaller values and examine the effect of such initialization on the obtained results).
Additionally, several methods that aim at the differentiation of initialized values (to achieve different degrees of instance selection reduction) are introduced:
  • Power transformation. In this method, the randomly generated number real number r n d ( 0 < r n d < 1 ) is raised to a certain power v as follows:
    w i = r n d v ,
    where higher v results in lower values (because 0 > r n d < 1 ) obtained after transformation.
  • Fill transformation. This binary transformation method uses a predefined probability f 1 that determines balance between 0 and 1 as a result of initialization:
    w i = 0 , for r n d < f 1 , 1 , for r n d f 1 .
  • Spread initialization. The idea behind this method is to differentiate the individuals in population in terms of the different probabilities of occurrence of small and large values:
    w i = 0 , for r n d < f 2 , r n d , for r n d f 2 ,
    where f 2 = 0.1 + h n / N is a value dependent on particular individual in the population, n stands for index of the individual in the population, h is a parameter.
In case of binary values, the numbers generated by initialization methods are rounded to the closest acceptable value. Based on the preliminary experiments, which are available in the supplementary resources, we decided to use the adaptive randomness combined with power and fill transformations.

2.5. Genetic Operators: Crossover and Mutation

We used the multi-point multi-parent crossover. The number of instances in the original training set T (which equals the chromosome length) was N. The M split points of the chromosome were randomly selected for each child. Then, a parent providing the genetic material for each segment was randomly chosen with the probability proportional to its fitness (thus the same parent could be chosen for more than one segment). The number M influences the speed of the algorithm conversion and the experiments showed that the optimal number M for the fastest conversion can be roughly expressed as:
M = r o u n d ( n / 10 ) for n 1000 , 100 + r o u n d ( ( n 1000 ) / 100 ) for n > 1000 .
However, smaller numbers M than given by Equation (17) also provide similar results in the instance selection, but the optimization takes longer. P children were generated in this way. Then, the P children and P parents were merged together into one group and P individuals with the highest fitness value from the merged group were promoted into the next iteration. (It can be said that the probability of crossover was 100% and because of this way of promoting individuals to the next iteration, it was the only optimal probability).
A broad range of mutation probabilities could be used without significantly influencing the results. We used the probability of mutation of 0.1% per each chromosome position and the following mutation operation:
w i = w i + ( r n d 0.5 ) · m r a n g e for n u m L e v e l s = 0 , R N D ( 0 , n u m L e v e l s 1 ) for n u m L e v e l s > 0 ,
where r n d stands for randomly generated real number from range 0 , 1 , R N D ( 0 , n u m L e v e l s 1 ) stands for random generated integer number from set { 0 , , n u m L e v e l s 1 } , m r a n g e stands for mutation range parameter (set experimentally as 0.2).

2.6. Training Time Minimization

In genetic algorithms, larger populations require fewer epochs of the algorithm to converge. However, some optimal population size exists from the viewpoint of the computational cost of the whole process. The cost can be approximated by the number of fitness function evaluations. According to our tests, this optimal population size was between 60 and 120 individuals. We decided to use 96 individuals because 96 was a multiple of the number of CPU cores in our server, so it scaled well in parallel computation. Larger populations increase computational costs but do not have any other negative impact on the process. However, if the population would be too small, it may limit the diversity of the individuals and prevent the algorithm from finding the best solutions.
Longer chromosomes, which represent lager datasets, require more epochs of the evolutionary algorithm to find the desired solutions. If too few epochs are used, under-fitting can occur and the solutions can be of poor quality, where the model fits neither the training data nor the test data enough well. On the other hand, using too many epochs can cause over-fitting—the model fits the training data too well and thus fails to fit the test data enough well (does not have generalization capabilities), which prevents good performance on the test data [52,53]. When over-fitting occurs, the error on the training set continuously decreases with further model learning, and the error on the test set starts increasing.
The biggest role in creating over-fitting can be attributed to the model getting fitted to the outstanding, noisy instances, when the model learning runs too long and thus the model gets too complex [52]. In our previous works [44,54], we studied various methods of preventing over-fitting in neural network learning. Noise removal by the ENN instance selection algorithm [11] and its extensions proved to be among the best methods. In the experimental section we compare the MEISR method to the extensions of ENN.
To prevent over-fitting in the MEISR algorithm, we used a standard early stopping approach [53]. By watching the error on a validation test (which is a part of the training set not used for model training but for online evaluation), the process can be stopped before the error starts increasing. In the thousands of experiments, we determined that, for the MEISR algorithm with 96 individuals in the population, the safe number of epochs, which always prevented over-fitting and which was high enough to train the model properly can be expressed as:
E = e 0 · log ( N ) ,
where N is the number of instances in the training set T , e 0 = 8 for binary instance selection and e 0 = 24 for real value instance weighting. For example, this gives 20 epochs for 300 instances and 34 epochs for 30,000 instances for binary instance selection. Thus, this can be also used as an early stopping criterion, especially that the process proved highly repeatable. This criterion has an additional advantage that it allows using all available instances for the original training set and thus achieving the best possible selected subsets. On the other hand, controlling the error on a validation set can allow for more training epochs and thus better model adjustment. Based on the experiments, we can conclude that the stopping criterion given by Equation (19) more frequently allowed for better results, especially for smaller datasets (where the difference between the validation and test set was high) and also the process was frequently faster.

2.7. Pseudo-Code of the MEISR Algorithm

The pseudo-code in Algorithm 1 shows the instance selection process. NSGA-II uses the basic operations of the standard genetic algorithm: crossover and mutation, but, additionally, it utilizes a mechanism to generate a wide Pareto front (the solutions situated on the Pareto front are called non-dominated solutions) by calculating the so called crowding distance [9] between the solutions on the fronts (line 19) and uses it to favor the solutions that are far apart from other ones (line 20).
Algorithm 1 Multi-objective evolutionary instance selection
Input: The original data set T
Output: F F reduced non-dominated training sets S i
1:
calculate and sort the distance matrices for T
2:
P : = initialization ( N ) {population P contains P randomly reduced data sets S i }
3:
evaluation ( P ) {calculate R M S E t r n and c o m p r e s s i o n for each S i in P }
4:
F = fast _ nondominated _ sort ( P , N )
5:
crowding _ distance ( F )
6:
while not S t o p C o n d i t i o n ( e p o c h = E or validation R M S E grows) do
7:
P = {new population P will contain children}
8:
for p = 1 to P do
9:
  for m = 1 to M do
10:
    parent ( m ) = select _ parent ( P )
11:
    crossover _ point ( m ) = rnd ( N )
12:
  end for
13:
   child ( p ) = new _ individual ( parent 1 , , parentM )
14:
   P = P child ( p )
15:
end for
16:
evaluation ( P ) {calculate compression and R M S E t r n using the distance metrices}
17:
P = P P
18:
F = fast _ nondominated _ sort ( P , 2 N )
19:
crowding _ distance ( F )
20:
P = selection ( P , F )
21:
end while
22:
return F

3. Experimental Evaluation

This section presents the experimental evaluation of the MEISR algorithm. The experimental process is presented in Figure 3. To provide comprehensive results, the algorithm is evaluated with prediction models (regressors) that belong to three different groups of models, regarding the response to instance selection (1-NN, k-NN, MLP neural network). Finally, a comparison with several other instance selection methods is presented. Most of the analysis is based on the relative R M S E t s t (a ratio of R M S E after instance selection to R M S E before instance selection) because our focus here is not to find the best predictive model, but to evaluate how much the performance of the given model will differ after applying instance selection. However, for a reference, we also present the absolute values of R M S E . The dataset properties are presented in Table A1. The experimental results are presented in Table A2, Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10, Table A11, Table A12, Table A13 and Table A14, explained in Figure 2 and its description and summarized in Figure 4.

3.1. Data Sets

In case of instance selection, we never know the best solution because it is an NP-hard problem, where trying the number of all possible combinations from sets of thousands of instances is out of accessible computational resources (the number of 5000 combinations from 10,000 instances is about 10 3010 and even the number of 500 combinations from 1000 instances is about 10 300 ). Thus, we cannot use a dataset with the best results known and see how much we approached this. To ensure reliable and unbiased results, we used the standard benchmark datasets from the KEEL Repository [55], as shown in Table A1. The output values were standardized in the experiments to enable us to evaluate and compare the results better.

3.2. Experimental Setup

The experimental setup is shown in Figure 3. We conducted the experiments using our own software written in C# language. Very detailed experimental results and the software source code can be found in the supplementary resources at kordos.com/entropy2018. The interested reader can find much more information there than can fit into this paper and replicate the experiments using our software. The experimental process is presented in Figure 3.
Different predictive models display different sensitivity to instance selection. To capture a broad range of model responses to instance selection and at the same time to keep the paper length within a reasonable limit, we carefully chose three models, where each of them is a representative of one model group regarding the sensitivity to instance selection.
The 1-NN algorithm is very sensitive to a change of single neighbor, as the prediction is based on that only neighbor and for this reason it is strongly influenced by instance selection. k-NN with higher k is less sensitive to a single neighbor change and thus the instance selection influences the R M S E t s t to a lower degree than in 1-NN. The distance measures used for the k-NN algorithm are described in details in Section 2.3 and the optimal k values in Table A1.
There is also a group of predictive models, which base the prediction results (function approximation) on a broad neighborhood and thus their R M S E t s t is much less dependent on instance selection; however, instance selection strongly accelerates the learning process. We chose an MLP neural network as a representative of this group, as it is one of the most popular models. We used a network with a typical structure for regression problems: one hidden layer with hyperbolic tangent transfer functions and one neuron in the output layer with the linear transfer function. The hidden layer consisted of six neurons if the number of attributes was below 12 and 12 neurons otherwise and was trained with the VSS algorithm [56] for 15 iterations (as this algorithm does not require more iterations). We also trained it with the well-known Rprop algorithm [57] for 60 epochs and the obtained results were almost the same as for VSS (the differences were statistically insignificant with t-test and Wilcoxon test), thus we do not report them here.
As the best results can be obtained if the evaluation model within instance selection is the same as the final predictive model [24], we used the same models at both positions, with the exception of when the final model was an MLP neural network. In this case, the evaluation model within instance selection was k-NN with optimal k because of the speed of the solution and the fact that an MLP network response to instance selection is closer to the response of k-NN with optimal k than to that of 1-NN.
We were able to find in the literature two papers suitable for comparison ([4,25], which have been presented in Section 1.2), which considered instance selection for regression problems and presented the results on a set of several datasets, reporting the obtained compression and R M S E or r 2 correlation.
In [4], the Pareto front was used. Since the experiments were conducted on most of the same datasets from the Keel Repository using the optimal k in k-NN, we used the detailed results from the online resources to that paper and performed the comparison. There were eight methods in this paper and we included in comparisons the four best of them, namely: threshold-based CNN ensemble (TE-C), threshold-based ENN ensemble (TE-E), discretization-based CNN ensemble (DE-C), discretization-based ENN ensemble (DE-E) and compared with our MEISR method.
In [25], the results were presented only for a single point with 8-NN for each dataset. We obtained these from the author’s detailed experimental results, which were also performed on most of the same datasets from the Keel Repository. Thus, we conducted the experiments with MEISR using 8-NN for a comparison with their results and measured the output additionally in r 2 correlation because that measure was used in the DROP3-RT method evaluation. Four methods were presented in this paper and DROP3-RT was the best one, so we included only DROP3-RT in the comparison with MEISR.
All the tests were performed using two testing procedures. The first one was 50% holdout, where randomly chosen 50% of instances were used for training and the remaining instances for testing (in case of odd instance number, the last instance was randomly assigned to one of the sets). The tests were performed 10 times with different random instances chosen for the test and training set each time. The average results over the 10 tests are reported. The reason of repeating this procedure 10 times is based on the recommendations that an experimental design should provide a sufficiently large number of measurements of the algorithm performance and, based on many analyses, 10 experimental measurements is the recommended standard [58,59].
The second procedure was a standard 10-fold cross-validation [59]. In this case, the dataset was first randomly divided into 10 parts with an equal number of instances (or almost equal it the instances cannot be divided equally into 10 parts). Then, the 10 measurements were performed; each of the 10 times a different part of the data was selected as a test set and the remaining nine parts as a training set.
In both cases, we reported the final goals of instance selection: the average R M S E over the 10 test sets and average retention over the 10 training sets.
The two most widely used statistical tests for determining if the difference between the results of two models over various datasets is non-random are the paired t-test and Wilcoxon signed-ranks test. The t-test assumes that the differences between the two compared random variables are distributed normally and is also affected by outliers which may decrease its power by increasing the estimated standard error. The Wilcoxon signed-ranks test does not assume normal distributions and is less influenced by outliers. However, when the assumptions of the paired t-test are valid, it is more powerful than the Wilcoxon test [60]. We used both of them and, in each case, both of them equally indicated the significance of the difference for the standard p-value of 0.05.
The standard deviations of the R M S E and retention in the experiments can be found in the online supplementary resources. We do not report them here to save the space because they are not used by the statistical tests, either by any other analysis.

3.3. Experimental Results and Discussion

The experimental results are summarized in Figure 4 (the figure presents the results in 10-fold cross-validation, but the results in 50% holdout were very similar). The numerical results for both testing procedures 50% holdout and 10-fold cross-validation are placed in the tables, where Table A2 summarizes all the results. The reduction (compression) is obviously independent of the final prediction model because the model is not used during the instance selection. However, RMSEtst depends on the final prediction model.
The average influence of instance selection on the results in both testing procedures: 50% holdout and 10-fold cross-validation was very similar, in spite that the absolute R M S E values were higher in 50% holdout due to a smaller training set size. The differences were usually below 1% and statistically insignificant according to Wilcoxon test. There were bigger differences between particular datasets, especially between the smallest ones, but over all datasets the average change of R M S E was comparable ( r 1 / r 0 , r 2 / r 0 , r m i n / r 0 ). Thus, it can be said that the MEISR algorithm performed equally well in both testing procedures. There were only bigger differences in the retention at the point c 3 —on average, the retention was higher by 11% in 50% holdout. This can be explained by the fact that in 50% holdout there are fewer instances in the original training set (50% vs. 90%), and as at the point c 2 the instances are sparse and higher percentage of them must remain to allow sufficient instance density to train the model. The differences in c ( r m i n ) can be bigger because, at that point, the Pareto Front is very flat and very little change in R M S E causes a much bigger change in retention.
The biggest improvement in terms of R M S E t s t was observed for 1-NN as the final model (Table A3 and Table A4), where the average R M S E t s t decrease was about 3.5% for average retention 62% and 8.6% for average retention 76%.
In case of k-kNN with optimal k (Table A5 and Table A6), on average, lower R M S E t s t was obtained on the original uncompressed data than at point c 1 . However, at the point c 3 with an average retention rate 85% we were able to reduce R M S E t s t on average by 1.5%.
When the final prediction model was an MLP neural network, on average, the RMSEtst decreased by 0.8% at c1 and by 2.5% at rmin (Table A7 and Table A8). At the point c2 (with the strongest compression), the increase of the RMSEtst was about three times lower than for 1-NN and k-NN. Thus, the conclusion is the MLP neural network is much less sensitive to instance selection. However, the unquestionable benefit of instance selection, in this case, was a reduction of data sizes and the shortening of the network learning process and thus giving the chance to try many different network configurations in a limited time. Nevertheless, as we have already mentioned, further RMSEtst decrease could be improved by using the MLP network also as the evaluation algorithm on the training set during the evolutionary optimization at the cost of much higher computational complexity.
The real-value instance weighting (Table A9 and Table A10) gave better results (lower Pareto front) only in two areas: for very high compression and for noisy datasets, while, in all the other cases, binary instance selection was better.
The MEISR method outperformed the four ensemble based methods: threshold-based CNN ensemble (TE-C), threshold-based ENN ensemble (TE-E), discretization-based CNN ensemble (DE-C), discretization-based ENN ensemble (DE-E) for the two retention values of 0.5 and 0.25, for which we run the tests (see Table A11 and Table A12).
As we had to chose a single point from the Pareto front for comparison with the DROP3-RT method, in a case that we could not find a point with lower 1 − r 2 and stronger compression, we decided to always use a point with stronger compression, even if the 1 − r 2 (and RMSE) would be higher (see the results in Table A13 and Table A14).
Comparing with other instance selection methods, the MEISR method allowed for obtaining significantly better results than all the other methods. Only for the largest datasets, the R M S E obtained with DROP3-RT was in several cases lower, but the compression of MEISR was stronger. The statistical tests can be found in Table A15.

3.4. Information Distribution, Loss Functions and Error Reduction

As mentioned in the introduction, the entropy used as a measure of information can increase after instance selection in classification tasks and some analogy exists for instance selection for regression problems. Unlike in classification problems, differential entropy used for continuous data (Equation (2)) can also be negative, which means that the information contained in the data is small (which can be thought of as if 2 H is small, then H is negative) [7]. However, unlike only decision boundaries in classification, in regression, each point of the data space matters. For that reason, to assess the instance selection performance in regression, a measure that can show the detail differences between closely located points of the data space will be preferred. Namely, we need to know how well one point can be represented by another closely located point.
Several such measures can be used. The first one is cross entropy, which identifies the difference between two probability distributions P and Q. It measures the average number of bits needed to identify a point from the set P, if a coding scheme from another set Q is used. Cross-entropy loss increases when the predicted class frequently differs from the actual label. For discrete data, cross entropy can be written as:
C E = i = 1 n p y i log q y i .
For continuous data, cross entropy can be defined per analogy as
C E c = Y p y log q y d y .
However, unlike in classification, in regression problems, we do not usually know the probability distributions. As cross entropy is used as a loss function in machine learning, other loss functions can also be useful to identify the difference between the two distributions. If the difference is high, we can assume that the data contains much noise. In addition, indeed the experimental evaluations confirmed this assumption, as in this case instance selection allowed for significantly decreasing the R M S E t s t of the final predictor.
In regression tasks, the most common loss function is root mean square error ( R M S E ). Since this measure was used in the optimization, we will also use it to show the dependencies. The loss functions r 0 for 1-NN and the relative R M S E t s t ( r 1 / r 0 ) obtained for r e t e n t i o n = c 1 are shown in Figure 5, where r 0 is the R M S E t s t without instance selection and r 1 is the R M S E t s t obtained for compression c 1 . The Pearson correlation coefficients between these variables were −0.782 for 1-NN and −0.728 for k-NN with optimal k.
We also observed that the optimal k on the selected subset was frequently different than before instance selection, and it tended to converge to the range of 5 to 7. That is, if the optimal k was 2 before instance selection it may increase after and if it was 11 before it is more likely to decrease. It can be explained, as after the selection fewer instances remained in the dataset, so the distances between them were bigger and frequently the previous closest neighbor of the examined instance no longer existed and thus it had to be replaced by the average of some further still existing instances. On the other hand, a high value of the optimal k is characteristic for noisy datasets, as an average value of several neighbors is needed to mask the noise. The correlation is shown in Figure 5. Instance selection removes the outlier and noisy instances, thus no longer so many neighbors are required to mask the detrimental effect of noise. Therefore, in the cases, where the optimal k was above 11, we used k = 11 because otherwise the optimal k would decrease anyway during the instance selection process and starting the optimization with k = 11 allowed for obtaining lower R M S E t r n and R M S E t s t on the selected subset. The dependency between r 0 for 1-NN and the k value that was optimal on the original datasets ( o r g K in Table A1) characterized by correlation coefficient 0.844 is also shown in Figure 5.

3.5. Computational Complexity

In this case, contrary to popular belief, the evolutionary algorithm based solution does not have to be more computationally expensive than the non-evolutionary ones.
The MEISR instance selection process can be decomposed into two steps:
  • The first step—calculating the distance matrices (see Section 2.3) has the complexity O ( n 2 ) .
  • The second step—running the evolutionary optimization has the complexity O ( n log n ) —because of the increasing number of epochs with dataset size.
One operation in the second step takes longer than in the first step. For small datasets, the second step is dominant, but, for the big ones, the first step. The measurements showed that for 900 instances the first step took about 10% of the total time, but, for 36,690 instances, it took about 65%.
Most of the non-evolutionary instance selection algorithms must also calculate the distance matrix or other equivalent matrix. Their complexity is between O ( n 2 ) (ENN, RHMC, ELH) and O ( n 3 ) (DROP1-5, GE, RNGE) [22]. We really observed that, for big datasets, the instance selection time with DROP3 grows much faster than with MEISR.
In the first step, the distances between each instances in the training set are first calculated in O ( n 2 ) and then they are sorted in O ( n log n ) , so the complexity is O ( n 2 ) —the higher of the two. The time spent on calculating the distance matrix also grows with the number of attributes.
The second step consists of several operations. Calculating the fitness function has the complexity O ( n ) because the output value of n instances must be obtained, where n = N is the number of instances in the original training set. Obtaining the output value requires reading on average k non-zero positions from sorted output value arrays, where k is the number of nearest neighbors in the k-NN algorithm, which assuming a reduction rate of 50% requires reading 2 k entries and calculating the average of them. The time spent in this step grows with the number of k, but much slower than linearly because also other operations are performed in this step, which do not depend on k, or depend but weaker than linearly, as crossover, mutation, and selection. The proportions of time spent at each step depend on particular software implementation. The experimental measurements confirmed that the complexity of the step can be considered O ( n log n ) .
In a practical software implementation, there is also a third-factor consuming time: the constant operations independent on the data, as calling functions, creating objects, etc. This factor is most significant with very small datasets and this is the reason that the times per one instance (two last columns in Table A16) are higher for the smallest dataset than for some of the following datasets.
Figure 6 shows the dependency of the MEISR running time on the number of instances (left) for 1-NN and k-NN with optimal k and the percentage of the running time used to calculate the distance matrix used by k-NN. The two points marked as “ 85 a ” denote the t i c k dataset, which has 85 attributes (more than other datasets) and, for that reason, there is so high cost of calculating the distance matrix for this case. The measurements were performed using our software (available from the online supplementary resources) on a computer with two Xeon E5-2696v2 processors. Detailed values can be found in Table A16).

3.6. Performance Metrics for Multi-Objective Evolutionary Algorithms

To measure the behavior of the NSGA-II algorithm used within the instance selection process, we provide the performance metrics for multi-objective evolutionary algorithms for binary instance selection with k-NN with optimal k (for other sets of experiments, the metrics were very similar).
However, first, it must be emphasized that the metrics express only the performance of the evolutionary algorithm in terms of data reduction and R M S E t r n on the training set and not of the whole process of instance selection, where the final objectives are data reduction and R M S E t s t on the test set. Second, as it was discussed that we had to stop the optimization early to prevent over-fitting and, for that reason, we could not achieve performance metrics that were as good as could be obtained if the target objectives would be optimized directly.
We calculated the following popular metrics: Ratio of Non-dominated Individuals—RNI [61], Inverted Generational Distance—IGD [62] (which expresses closeness of the solutions to the true Pareto front), Uniform Distribution—UD [61] (which expresses distribution of the solutions, with the σ for UD metric set to 0.025), Maximum Spread—MS [63] (spread of the solutions) and HyperVolume indicator—HV [64,65] (which applies to several of listed categories, with a reference point for HV metric set to +5%).
The obtained values of metrics can be summed up as follows:
  • The RNI values are high, average 0.560, which means that more than 50% of the population formed a Pareto front.
  • The IGD values are low, average: 0.014 (lower = better), which means that the results were always close to the optimal Pareto front.
  • The UD values are low, average: 0.234 (lower = better), which means that most of the solutions were properly spread.
  • The MS values are high, average: 0.833 (higher = better), which means that obtained Pareto fronts are wide in range in comparison to optimal Pareto front.
  • The average HV values are 0.039, which means that the values of R M S E t r n , even for high compression, did not increase significantly, which is good because lower R M S E is preferred and the very low compression is not present in the results, which is also good because the desired R M S E has already been achieved with stronger compression and thus HV covers a satisfactory part of the objectives’ area.

4. Conclusions

We presented and experimentally evaluated an instance selection method applied for regression tasks, which uses the k-NN algorithm for error evaluation a multi-objective evolutionary algorithm (NSGA-II in the current implementation) for directing the search process. Different aspects of the solution were discussed, many improvements were proposed and different tests were performed. The following main conclusions can be drawn from this work:
  • A key advantage of the MEISR method is that we obtain several solutions from the Pareto front, and can choose one of them depending on our preferences of the R M S E –compression balance. As the solutions create a Pareto front, each of them is best for a given balance between R M S E and compression, as explained in Section 2.2. If someone is not sure which solution to choose, then we suggest the solution with the lowest R M S E on the training set ( R M S E t r n ) for the purpose for machine learning and the solution ( c 2 , r 2 ) for analyzing the properties of the data, as explained in Figure 2 and in the text following it. This is valid for every dataset, as this is the characteristic feature of the multi-objective optimization itself and it is not dependent on the dataset properties.
  • k-NN is very well suited as the inner evaluation algorithm because of its speed—the distance matrix has to be calculated and sorted only once and then the prediction is extremely fast. This makes the computational cost of the method comparable to the cost on non-evolutionary instance selection algorithms (and some of them, as the DROP family have even higher cost), while the results are usually better (see Section 3.5, Figure 6 and Table A16 for details).
  • We were frequently able to preserve all the useful information in the dataset for the purpose of predictive model performance while reducing it size by about one third (see columns 4 and 5 in Table A3, Table A4, Table A5, Table A6, Table A7 and Table A8 and Figure 4).
  • Proper initialization of the population accelerates the instance selection process and helps to find the desired solutions (see Section 2.4).
  • The best results in terms of R M S E -compression balance can be obtained if the inner evaluation algorithm is the same as the final predictor [24]. For that reason, we used 1-NN as the internal regressor when 1-NN was used as the final prediction model and k-NN with optimal k as the internal regressor when k-NN with optimal k was used as the final prediction model.
  • Although our previous experiments showed [24] that when an MLP neural network as the final predictor better results were achieved if also the internal regressor was an MLP network, it would be very time-consuming, especially for bigger datasets, as the MLP network has to be trained each time (at least on the part of the data close to the currently evaluated point). Thus, we decided to use k-NN with optimal k as the internal regressor for the MLP neural network.
  • We noticed that there are two areas where real-value instance weighting can provide better results than binary instance selection: for very high compression, where it was usually able to achieve lower R M S E and for noisy datasets, which required high k value in k-NN (see Table A10).
  • The obtained R M S E t s t with the k-NN algorithm for a given point on the Pareto front (for a given compression) can be approximately assessed by the measures of how well one instances can be substituted by other, e.g., the loss functions of cross entropy or R M S E . The lower values of the loss function correspond with lower possible decrease R M S E t s t .
  • The MEISR method achieved better results than the other 12 instance selection methods for regression, for which we were able to obtain the experimental results to perform the comparison (see Table A12 and Table A14)
  • We can observe that the “front” is less steep for the test set than for the training set. This is the compression strength grows and the R M S E increases on average slower on the test set than on the corresponding training set solutions (when moving from right to left in Figure 2, the green points are getting closer to the orange ones). This allows for choosing a point with higher compression, as the R M S E t s t on that point is likely to grow less than R M S E t r n
  • There were not significant differences between the algorithm performance tested with 50% holdout and 10-fold cross-validation. Only retention at the point c 2 was on average higher 11% in the first case, as there were already fewer instances in the original training set.
We have also noticed two areas of possible improvement and we are going to investigate them in our future work:
  • A single front covered all solutions of interest in many cases, but not in all. One of the reasons is the tendency of multi-objective evolutionary algorithms to not cover the areas on the ends of the front, as was discussed and some solutions were proposed i.e., in [66,67]. The second reason is that the front extends gradually during the optimization and, to prevent over-fitting, we must stop the instance selection process before the front is fully extended.
  • Experimental comparison with other instance selection methods showed that, for the small and medium size datasets, the MEISR method has the greatest advantage over other methods (see Table A12 and Table A14). However, for the largest datasets, while still the compression was always stronger, the obtained 1 r 2 began to became similar to that of DROP3-RT method. Although MEISR optimized R M S E and 1 r 2 was used only for comparison and the relation between R M S E and 1 r 2 is not linear, using 1 r 2 as the objective on the training set would most likely improve the results. This would be true also for the smaller datasets and the tendency would remain. A well known issue here is that, for genetic algorithms with longer chromosomes, the convergence is more difficult. Thus, we are going to investigate the possibilities of alternative encoding of the instances to limit the chromosome length.
To summarize: the presented method of instance selection in regression tasks has proved to work effectively and has several advantages. Thus, we believe it can be helpful for researchers and practitioners in industry. Moreover, there is probably still room for further improvements.

Author Contributions

Conceptualization, M.K.; Formal analysis, K.Ł.; Funding acquisition, M.K.; Investigation, M.K. and K.Ł.; Methodology, M.K. and K.Ł.; Software, M.K. and K.Ł.; Validation, M.K.; Visualization, K.Ł.; Writing—Original Draft, M.K. and K.Ł.

Funding

This research was supported by the NCN (Polish National Science Center) Grant No. 2017/01/X/ST6/00202 “Evolutionary Methods in Data Selection”.

Acknowledgments

M.K. wants to thank Álvar Arnaiz-González and César García-Osorio from the University of Burgos for the scientific visit at their university in February 2018, which was very useful while working on this study.

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

Appendix A. Detailed Experimental Results

This section contains tables with dataset properties and detailed experimental results. The following symbols are used:
  • r0, r1, r2, r m i n —absolute R M S E t s t (see Figure 2 for details),
  • r1/r0, r2/r0, r m i n /r0—relative R M S E (see Figure 2 for details),
  • c1, c2—retention (see Figure 2 for details),
  • c(r)—retention for specified r,
  • r m i n = m i n ( r 0 , r 1 , r 2 , r 3 ) —the lowest obtained R M S E t s t ,
  • IEA—Inner Evaluation Algorithm,
  • FPA—Final Prediction Algorithm,
  • BIS—Binary Instance Selection,
  • RIS—Real-value Instance Selection,
  • Dr2—relative 1 − r 2 with DROP3-RT,
  • Mr2—relative 1 − r 2 with MEISR,
  • Dc1—retention rate with DROP3-RT,
  • Mc1—retention rate with MEISR,
  • TE-C, TE-E, DE-C, DE-E, MEISR (in Table A11 and Table A12)— R M S E t s t for these algorithms,
  • 10-CV—10-fold cross-validation,
  • 50%H—50% holdout; (50% training set, and 50% in test set)—average of 10 runs.
Results for each experimental case, averaged for all datasets, are presented the Table A2 and the statistical significance test results in Table A15. Results for particular configurations fo the MEISR algorithm are presented in Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9 and Table A10. Result comparison with other algorithms is presented in Table A11, Table A12, Table A13 and Table A14.
The calculation times is and its particular components are presented in Table A16.
Table A1. Datasets used in the experiments and their properties: number of instances (Ints.), number of attributes (Attr.), the optimal k—the k in k-NN that gives the lowest R M S E t s t (orgK) and the k used in the experiments as optimal k (optK) for the reason explained further in the text.
Table A1. Datasets used in the experiments and their properties: number of instances (Ints.), number of attributes (Attr.), the optimal k—the k in k-NN that gives the lowest R M S E t s t (orgK) and the k used in the experiments as optimal k (optK) for the reason explained further in the text.
DatasetInst.Attr.orgKoptK
mach. CPU209611
baseball3371677
dee365677
autoMPG8392766
autoMPG6392544
ele-149521111
forestFires517125011
stock950933
steel9601244
laser993433
concrete1030844
treasury10491533
mortgage10491522
friedman1200577
wizmir1461977
wankara1609999
plastic16502311
quake217835011
anacalt4052722
abalone417781311
delta-ail712851711
puma32h8191322111
compactiv81922122
delta-elv951663511
tic9822855011
ailerons13,750401010
pole14,9982644
elevators16,5981888
california20,640899
house22,784161111
mv40,7671099
Table A2. Averaged results for experiments presented in Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9 and Table A10.
Table A2. Averaged results for experiments presented in Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9 and Table A10.
IEAFPAValidationr0r1r1/r0c1r2r2/r0c2Details
1-NN (BIS)1-NN (BIS)50%H0.6040.5370.9550.6190.6461.1760.246see Table A3
1-NN (BIS)1-NN (BIS)10-CV0.5730.5030.9600.6270.5941.1750.214see Table A4
k-NN (BIS)k-NN (BIS)50%H0.4760.4861.0500.5050.5161.1630.249see Table A5
k-NN (BIS)k-NN (BIS)10-CV0.4460.4531.0560.4970.4871.2030.227see Table A6
k-NN (BIS)MLP50%H0.4270.4240.9940.5050.4581.0760.249see Table A7
k-NN (BIS)MLP10-CV0.3930.3910.9920.4970.4181.0680.227see Table A8
k-NN (RIS)k-NN (RIS)50%H0.4760.4961.1160.5570.5551.3070.185see Table A9
k-NN (RIS)k-NN (RIS)10-CV0.4480.4651.1130.5320.5081.2820.165see Table A10
Table A3. Experimental results for MEISR with 50%H, IEA: 1-NN (BIS), FPA: 1-NN (BIS).
Table A3. Experimental results for MEISR with 50%H, IEA: 1-NN (BIS), FPA: 1-NN (BIS).
Datasetr0r1r1/r0c1r2r2/r0c2r min r min /r0c(r min )
mach. CPU0.4500.5171.1490.4590.5411.2020.2490.4280.9510.910
baseball0.7580.7460.9840.6090.7891.0410.6680.7460.9840.609
dee0.5610.5020.8940.4330.5610.9990.5320.5020.8940.433
autoMPG80.4780.4500.9420.5190.5091.0650.1990.4500.9420.519
autoMPG60.4640.4641.0010.6530.6911.4910.2030.4510.9730.925
ele-10.6900.6700.9700.6280.7081.0250.5960.6700.9700.628
forestFires1.6690.8990.5390.4130.9020.5410.2010.8990.5390.413
stock0.1360.1581.1620.4500.2171.5960.1680.1361.0001.000
steel0.3590.4041.1240.4800.4161.1580.3910.3591.0001.000
laser0.2930.3121.0640.5160.3431.1700.2080.2840.9690.898
concrete0.6390.6771.0600.6640.8251.2920.2010.6391.0001.000
treasury0.0920.0951.0330.5300.1451.5760.1800.0900.9780.902
mortgage0.0710.0821.1510.4530.1121.5720.3090.0711.0001.000
friedman0.5100.4670.9150.7350.6471.2680.1560.4670.9150.735
wizmir0.2440.2451.0030.7260.3161.2930.1920.2441.0001.000
wankara0.2390.2290.9590.7270.2951.2360.2090.2290.9590.727
plastic0.6340.5180.8170.7190.6080.9590.2860.5180.8170.719
quake1.3491.1240.8330.7231.3480.9990.1431.1240.8330.723
anacalt0.2630.2721.0360.5210.3041.1570.2170.2480.9440.886
abalone0.9000.7710.8570.7180.9261.0290.1850.7710.8570.718
delta-ail0.7330.6910.9430.7060.7841.0690.2080.6910.9430.706
puma32h1.2271.0360.8440.7161.2791.0420.2131.0360.8440.716
compactiv0.3030.3441.1350.3960.3841.2670.3120.3031.0001.000
delta-elv0.8260.7090.8590.7190.9011.0910.1650.7090.8590.719
tic1.3481.1110.8240.7191.3491.0010.1891.1110.8240.719
ailerons0.6850.6010.8780.7200.7481.0920.1440.6010.8780.720
pole0.2710.2881.0610.7160.3711.3670.2050.2711.0000.925
elevators0.7190.6590.9160.7190.9401.3070.1710.6590.9160.719
california0.6810.6120.8990.7160.7661.1250.1630.6120.8990.716
house0.8930.7870.8810.7170.9831.1010.1990.7870.8810.717
mv0.2350.2080.8840.6770.3121.3250.1620.2080.8840.677
average0.6040.5370.9550.6190.6461.1760.2460.5260.9180.777
Table A4. Experimental results for MEISR with 10-CV, IEA: 1-NN (BIS), FPA: 1-NN (BIS).
Table A4. Experimental results for MEISR with 10-CV, IEA: 1-NN (BIS), FPA: 1-NN (BIS).
Datasetr0r1r1/r0c1r2r2/r0c2r min r min /r0c(r min )
mach. CPU0.3510.4071.1590.3850.4261.2120.2160.3511.0000.914
baseball0.7270.6390.8780.6620.6540.8990.5590.6390.8780.662
dee0.5550.4500.8110.6520.5100.9180.4010.4500.8110.652
autoMPG80.4290.3870.9040.6850.4831.1270.1550.3870.9040.685
autoMPG60.4070.4241.0420.6980.5061.2420.1500.4010.9850.920
ele-10.6890.6400.9290.6780.6660.9660.6190.6400.9290.678
forestFires1.5480.7680.4960.3440.7960.5140.1910.7680.4960.344
stock0.1120.1321.1750.6290.1901.6920.1550.1121.0000.918
steel0.3480.3170.9110.4690.4021.1540.3690.3170.9110.469
laser0.2310.2761.1920.4670.3061.3240.2210.2311.0000.911
concrete0.5330.5951.1170.6960.7681.4410.1550.5331.0001.000
treasury0.0690.0871.2590.6620.1181.6930.1590.0691.0000.916
mortgage0.0540.0691.2840.5450.0961.7840.2870.0541.0000.926
friedman0.4620.4430.9590.6920.5951.2880.1480.4430.9590.692
wizmir0.2300.2110.9190.6930.3011.3110.1600.2110.9190.693
wankara0.2250.2180.9660.6920.2891.2850.1600.2180.9660.692
plastic0.6170.5430.8790.6820.6070.9830.2120.5430.8790.682
quake1.3441.1510.8560.6851.3491.0030.1571.1510.8560.685
anacalt0.2270.2271.0000.4490.2861.2600.1860.2220.9770.887
abalone0.9150.7730.8450.6850.8800.9620.1550.7730.8450.685
delta-ail0.7160.6280.8770.6850.7381.0310.1590.6280.8770.685
puma32h1.2121.0240.8450.6891.2121.0000.1611.0240.8450.689
compactiv0.2540.2991.1770.4290.2971.1720.2370.2350.9260.847
delta-elv0.8280.7050.8510.6850.8371.0100.1570.7050.8510.685
tic1.3661.1300.8270.6851.3300.9730.1561.1300.8270.685
ailerons0.6570.5840.8890.6880.7091.0790.1600.5840.8890.688
pole0.2440.2581.0550.6860.3531.4430.1570.2441.0000.922
elevators0.6860.6380.9300.6870.7641.1140.1620.6380.9300.687
california0.6540.5960.9110.6860.7181.0990.1580.5960.9110.686
house0.8720.7750.8880.6840.9261.0610.1590.7750.8880.684
mv0.2100.1970.9410.6860.2911.3870.1580.1970.9410.686
average0.5730.5030.9600.6270.5941.1750.2140.4930.9100.740
Table A5. Experimental results for MEISR with 50%H, IEA: k-NN (BIS), FPA: k-NN (BIS).
Table A5. Experimental results for MEISR with 50%H, IEA: k-NN (BIS), FPA: k-NN (BIS).
Datasetr0r1r1/r0c1r2r2/r0c2r min r min /r0c(r min )
mach. CPU0.4500.5171.1490.4370.5411.2020.2490.4410.9800.877
baseball0.6480.6711.0350.4090.7181.1080.2870.6481.0000.409
dee0.4720.4540.9620.3780.4620.9790.3140.4420.9370.912
autoMPG80.4220.4120.9760.4280.4811.1400.2990.4070.9640.936
autoMPG60.4110.3980.9680.4890.4521.1000.2610.3770.9170.939
ele-10.5840.5620.9620.4090.5891.0090.3210.5620.9620.409
forestFires0.9010.8470.9400.3240.8700.9660.2300.8470.9400.324
stock0.1290.1481.1470.5420.1681.3020.2930.1240.9610.962
steel0.3640.3911.0760.4810.4901.3480.2470.3641.0000.912
laser0.2740.2841.0380.4720.3241.1840.2390.2580.9430.941
concrete0.5760.6311.0960.4970.6871.1930.2050.5761.0000.922
treasury0.0790.0861.0930.5020.1261.5940.2550.0791.0001.000
mortgage0.0650.0821.2580.5170.1372.1010.1720.0651.0000.941
friedman0.3750.4101.0920.5060.4901.3050.2000.3751.0000.899
wizmir0.1990.2151.0810.3840.2251.1310.2050.1991.0001.000
wankara0.1850.1991.0760.5030.2241.2110.2100.1851.0001.000
plastic0.4630.4600.9940.4910.4580.9890.2110.4600.9940.491
quake1.0241.0080.9840.6211.0301.0050.2151.0080.9840.621
anacalt0.2420.2541.0520.4840.2861.1840.2600.2421.0001.000
abalone0.7130.7040.9870.6550.7211.0110.1970.7040.9870.887
delta-ail0.5820.5881.0100.6760.6881.1810.1870.5821.0000.905
puma32h0.9100.9181.0080.5830.9321.0240.2200.9101.0001.000
compactiv0.2810.3161.1250.4260.3371.2000.2430.2811.0001.000
delta-elv0.6260.6240.9960.6870.6281.0030.2660.6240.9961.000
tic1.0110.9970.9860.6831.0070.9960.2240.9970.9860.683
ailerons0.5220.5331.0210.5230.5681.0880.2720.5221.0000.932
pole0.2440.2701.1070.5160.3011.2340.3020.2441.0000.933
elevators0.5860.6041.0300.5340.6431.0970.2870.5861.0001.000
california0.5480.5661.0330.5350.4460.8140.2990.5481.0000.948
house0.7120.7241.0170.5660.7641.0730.2870.7121.0000.901
mv0.1600.1981.2360.3840.2051.2800.2670.1601.0001.000
average0.4760.4861.0500.5050.5161.1630.2490.4680.9860.861
Table A6. Experimental results for MEISR with 10-CV, IEA: k-NN (BIS), FPA: k-NN (BIS).
Table A6. Experimental results for MEISR with 10-CV, IEA: k-NN (BIS), FPA: k-NN (BIS).
Datasetr0r1r1/r0c1r2r2/r0c2r min r min /r0c(r min )
mach. CPU0.3510.3701.0530.4360.4101.1680.2130.3390.9640.852
baseball0.5840.5820.9980.4500.6491.1120.2170.5820.9980.450
dee0.4240.4271.0080.4030.4391.0350.2290.4180.9860.928
autoMPG80.3720.4011.0780.4580.4221.1340.2180.3640.9780.927
autoMPG60.3660.3821.0430.4240.3931.0740.2280.3510.9210.944
ele-10.5840.5570.9540.5330.5851.0020.3670.5570.9540.533
forestFires0.8640.7250.8390.3400.7260.8400.1620.7250.8390.340
stock0.1050.1241.1810.4860.1451.3860.3400.1040.9900.954
steel0.3230.3301.0220.4440.4211.3020.2240.3231.0000.934
laser0.2040.2101.0310.4350.2611.2800.2080.1950.9550.926
concrete0.5210.5491.0540.4880.6481.2440.1930.5211.0000.929
treasury0.0580.0771.3210.4860.1031.7680.2340.0581.0000.927
mortgage0.0450.0591.3040.4890.1072.3870.1560.0451.0000.923
friedman0.3400.3651.0730.4720.4411.2970.1620.3401.0000.947
wizmir0.1780.1881.0600.4700.2311.2970.2140.1781.0000.779
wankara0.1670.1831.0980.4970.2201.3150.1600.1671.0001.000
plastic0.4680.4530.9690.4460.4660.9960.2120.4530.9690.446
quake1.0251.0080.9830.6851.0251.0010.2191.0080.9830.685
anacalt0.2120.2151.0130.4400.2821.3300.1740.2090.9880.684
abalone0.7020.7091.0110.6820.7501.0680.1580.6880.9810.801
delta-ail0.5600.5791.0330.6860.6041.0780.1580.5601.0000.835
puma32h0.8960.9081.0140.4610.9201.0270.2220.8961.0001.000
compactiv0.2310.2801.2090.4610.2861.2380.2750.2311.0001.000
delta-elv0.6100.6201.0160.6840.6281.0290.2250.6101.0001.000
tic1.0150.9950.9800.6841.0110.9960.1860.9950.9800.684
ailerons0.5040.5191.0310.4730.5501.0920.2630.5041.0000.934
pole0.2140.2411.1250.4910.2611.2190.3070.2141.0000.947
elevators0.5590.5811.0400.4480.6221.1120.2260.5591.0001.000
california0.5270.5461.0370.4880.5711.0830.3320.5271.0000.944
house0.6870.7071.0290.4830.7371.0740.2760.6871.0000.894
mv0.1400.1591.1360.4920.1831.3110.2730.1401.0001.000
average0.4460.4531.0560.4970.4871.2030.2270.4370.9840.843
Table A7. Experimental results for MEISR with 50%H, IEA: k-NN (BIS), FPA: MLP.
Table A7. Experimental results for MEISR with 50%H, IEA: k-NN (BIS), FPA: MLP.
Datasetr0r1r1/r0c1r2r2/r0c2r min r min /r0c(r min )
mach. CPU0.4570.4911.0760.4370.5701.2470.2490.4571.0000.877
baseball0.6180.6591.0670.4090.7631.2350.2870.6181.0000.409
dee0.4390.4531.0330.3780.5001.1410.3140.4280.9760.912
autoMPG80.3660.3430.9370.4280.4321.1820.2990.3661.0001.000
autoMPG60.3670.3440.9380.4890.4161.1320.2610.3440.9380.939
ele-10.5540.5120.9230.4090.5851.0560.3210.5120.9230.409
forestFire1.0841.0260.9460.3241.1241.0370.2301.0260.9460.324
stock0.1950.1900.9740.5420.1981.0110.2930.1900.9740.962
steel0.2510.2190.8720.4810.2240.8930.2470.2190.8720.481
laser0.2040.2030.9940.4720.2030.9950.2390.2030.9941.000
concrete0.4070.4231.0390.4970.4491.1030.2050.4071.0001.000
treasury0.0770.0760.9850.5020.0801.0500.2550.0760.9851.000
mortgage0.0700.0690.9810.5170.0791.1270.1720.0660.9410.941
friedman0.3010.3000.9970.5060.3471.1520.2000.2980.9890.899
wizmir0.0950.0961.0070.3840.0991.0420.2050.0951.0001.000
wankara0.1000.0980.9800.5030.1021.0230.2100.0960.9610.843
plastic0.4350.4371.0040.4910.4421.0160.2110.4351.0000.491
quake1.0000.9990.9990.6211.0541.0540.2150.9990.9990.621
anacalt0.2230.1990.8910.4840.2511.1240.2600.1990.8910.484
abalone0.6520.6420.9840.6550.6460.9910.1970.6420.9840.887
delta-ail0.5540.5520.9970.6760.5510.9960.1870.5500.9930.905
puma32h0.3680.4061.1040.5830.4731.2860.2200.3681.0001.000
compactiv0.1550.1581.0190.4260.1641.0550.2430.1551.0001.000
delta-elv0.6000.5900.9820.6870.6141.0220.2660.5880.9800.910
tic1.0191.0140.9950.6831.0521.0330.2241.0140.9950.683
ailerons0.4140.4251.0260.5230.4471.0800.2720.4141.0001.000
pole0.2410.2511.0410.5160.2821.1700.3020.2411.0001.000
elevators0.6820.6800.9970.5340.7061.0360.2870.6770.9930.922
california0.5320.5310.9980.5350.5501.0340.2990.5321.0001.000
house0.7120.7191.0110.5660.7281.0230.2870.7121.0001.000
mv0.0550.0551.0020.3840.0551.0080.2670.0551.0001.000
average0.4270.4240.9940.5050.4581.0760.2490.4190.9790.835
Table A8. Experimental results for MEISR with 10-CV, IEA: k-NN (BIS), FPA: MLP.
Table A8. Experimental results for MEISR with 10-CV, IEA: k-NN (BIS), FPA: MLP.
Datasetr0r1r1/r0c1r2r2/r0c2r min r min /r0c(r min )
mach. CPU0.3480.3791.0890.4360.4381.2610.2130.3781.0001.000
baseball0.6260.6561.0480.4500.7501.1980.2170.6431.0001.000
dee0.4070.4121.0120.4030.4701.1530.2290.4040.9920.928
autoMPG80.3540.3360.9480.4580.4211.1870.2180.3340.9420.927
autoMPG60.3570.3300.9240.4240.4071.1410.2280.3260.9140.944
ele-10.5310.4880.9200.5330.5521.0400.3670.4860.9170.533
forestFires0.7350.6990.9520.3400.7451.0140.1620.6910.9400.340
stock0.1780.1710.9580.4860.1811.0160.3400.1680.9450.954
steel0.2250.1980.8820.4440.1990.8850.2240.1960.8720.934
laser0.1640.1620.9880.4350.1651.0110.2080.1600.9750.926
concrete0.3790.3861.0170.4880.4161.0970.1930.3791.0000.929
treasury0.0740.0720.9700.4860.0771.0420.2340.0720.9680.927
mortgage0.0550.0540.9750.4890.0601.0880.1560.0530.9580.923
friedman0.3180.3150.9920.4720.3681.1590.1620.3130.9850.947
wizmir0.0850.0861.0130.4700.0871.0270.2140.0851.0001.000
wankara0.0970.0950.9780.4970.0971.0040.1600.0940.9681.000
plastic0.4370.4330.9900.4460.4340.9950.2120.4330.9900.446
quake1.0000.9920.9920.6851.0171.0170.2190.9920.9920.685
anacalt0.2010.1810.8980.4400.2201.0960.1740.1800.8940.684
abalone0.6510.6500.9980.6820.6551.0060.1580.6490.9970.801
delta-ail0.5520.5510.9970.6860.5551.0050.1580.5480.9930.835
puma32h0.3380.3731.1030.4610.4331.2830.2220.3711.0001.000
compactiv0.1520.1551.0180.4610.1581.0420.2750.1531.0001.000
delta-elv0.5990.5980.9980.6840.6041.0090.2250.5960.9951.000
tic1.0171.0171.0000.6841.0681.0500.1861.0110.9930.684
ailerons0.4020.4091.0180.4730.4181.0410.2630.4071.0001.000
pole0.2550.2651.0380.4910.3031.1880.3070.2621.0001.000
elevators0.3440.3471.0110.4480.3481.0130.2260.3471.0001.000
california0.5320.5310.9970.4880.5320.9990.3320.5240.9850.944
house0.7100.7141.0050.4830.7221.0170.2760.7101.0001.000
mv0.0550.0551.0050.4920.0561.0090.2730.0550.9911.000
average0.3930.3910.9920.4970.4181.0680.2270.3880.9740.880
Table A9. Experimental results for MEISR with 50%H, IEA: k-NN (RIS), FPA: k-NN (RIS).
Table A9. Experimental results for MEISR with 50%H, IEA: k-NN (RIS), FPA: k-NN (RIS).
Datasetr0r1r1/r0c1r2r2/r0c2
mach. CPU0.4500.4621.0270.4020.4851.0790.230
baseball0.6480.6611.0200.6500.8201.2650.114
dee0.4720.4710.9980.4290.5771.2220.148
autoMPG80.4220.4601.0900.5250.5461.2950.166
autoMPG60.4110.4421.0760.4020.5111.2430.176
ele-10.5840.5690.9740.5330.5961.0200.184
forestFires0.9010.7480.8300.4900.7910.8780.151
stock0.1290.1761.3610.7530.2251.7460.364
steel0.3640.4101.1270.6570.5021.3810.193
laser0.2740.3301.2050.4840.3651.3340.152
concrete0.5760.6591.1450.6220.7841.3610.137
treasury0.0790.1201.5230.5410.1652.0930.180
mortgage0.0650.0991.5240.7490.1522.3320.306
friedman0.3750.4291.1440.7000.5351.4250.192
wizmir0.1990.2151.0810.4810.2741.3770.140
wankara0.1850.2061.1150.6730.2971.6040.139
plastic0.4630.4520.9750.5510.5121.1070.223
quake1.0241.0090.9850.4011.0361.0110.123
anacalt0.2420.3991.6520.5040.4781.9780.233
abalone0.7130.7381.0350.4440.7801.0940.166
delta-ail0.5820.5851.0040.4750.6301.0810.180
puma32h0.9100.9131.0030.6340.9601.0550.151
compactiv0.2810.3031.0790.4470.3871.3790.105
delta-elv0.6260.6391.0200.5560.6701.0700.150
tic1.0110.9610.9510.5241.0020.9910.125
ailerons0.5220.5281.0110.4370.5811.1120.153
pole0.2440.3071.2580.6150.3251.3320.135
elevators0.5860.5780.9850.7230.6601.1260.099
california0.5480.5531.0090.5770.5681.0370.124
house0.7120.7461.0480.6990.7811.0970.304
mv0.1600.2051.2790.5780.2211.3800.494
average0.4760.4961.1160.5570.5551.3070.185
Table A10. Experimental results for MEISR with 10-CV, IEA: k-NN (RIS), FPA: k-NN (RIS).
Table A10. Experimental results for MEISR with 10-CV, IEA: k-NN (RIS), FPA: k-NN (RIS).
Datasetr0r1r1/r0c1r2r2/r0c2
mach. CPU0.3920.4071.0380.3850.4261.0860.216
baseball0.6060.6231.0290.5910.7531.2440.110
dee0.4240.4301.0150.4030.5011.1820.137
autoMPG80.3640.4031.1080.4990.4801.3200.160
autoMPG60.3660.3961.0820.4010.4571.2480.156
ele-10.5840.5620.9630.4890.5750.9850.178
forestFires0.8640.7290.8440.4560.7320.8470.126
stock0.1050.1401.3380.6890.1811.7290.371
steel0.3230.3591.1100.6650.4311.3330.165
laser0.2040.2431.1930.4890.2711.3300.128
concrete0.5210.5931.1380.5930.6781.3030.129
treasury0.0580.0881.5030.5310.1182.0150.159
mortgage0.0490.0761.5540.6820.1092.2460.263
friedman0.3400.3851.1320.6680.4891.4370.176
wizmir0.1780.1921.0810.4910.2381.3390.117
wankara0.1670.1831.0960.6430.2731.6350.127
plastic0.4680.4510.9650.5040.5201.1120.184
quake1.0251.0000.9760.3741.0030.9790.117
anacalt0.2120.3521.6610.5100.4041.9040.226
abalone0.6900.7091.0280.4350.7301.0590.136
delta-ail0.5600.5661.0110.4770.5801.0350.170
puma32h0.8960.9021.0080.6190.9271.0350.135
compactiv0.2400.2591.0770.4410.3331.3880.097
delta-elv0.6100.6101.0000.5460.6271.0270.123
tic1.0150.9850.9700.4960.9920.9770.121
ailerons0.5040.5111.0150.4210.5381.0690.128
pole0.2140.2671.2490.5830.2821.3190.108
elevators0.5590.5530.9900.6620.6201.1100.081
california0.5270.5311.0080.5320.5551.0530.112
house0.6870.7181.0460.6710.7431.0830.236
mv0.1400.1791.2800.5400.1861.3280.420
average0.4480.4651.1130.5320.5081.2820.165
Table A11. Relative R M S E t s t for retention c 1 = 0.5 and c 2 = 0.25 (compression = 50% and 75%) in 50% holdout (average over 10 runs) for Threshold-Ensemble-CNN (TE-C), Threshold-Ensemble-ENN (TE-E), Discretization-Ensemble-CNN (DE-C), Discretization-Ensemble-ENN (DE-E) and MEISR. Inner evaluation algorithm k-NN with optimal k and binary instance selection. Final prediction algorithm: k-NN with optimal k.
Table A11. Relative R M S E t s t for retention c 1 = 0.5 and c 2 = 0.25 (compression = 50% and 75%) in 50% holdout (average over 10 runs) for Threshold-Ensemble-CNN (TE-C), Threshold-Ensemble-ENN (TE-E), Discretization-Ensemble-CNN (DE-C), Discretization-Ensemble-ENN (DE-E) and MEISR. Inner evaluation algorithm k-NN with optimal k and binary instance selection. Final prediction algorithm: k-NN with optimal k.
c1 = 0.50c2 = 0.25
DatasetTE-CTE-EDE-CDE-EMEISRTE-CTE-EDE-CDE-EMEISR
mach. CPU1.2701.0651.3621.0221.1491.5371.2961.7551.3931.202
baseball1.1181.0871.1551.2661.0351.3461.3101.7341.5981.183
dee1.1511.0961.0771.0660.9621.3481.2651.2171.3741.203
autoMPG81.2551.0861.1851.1210.9761.5381.3081.2801.2481.198
autoMPG61.1701.0651.1021.0920.9681.3771.3091.2311.2101.112
ele-11.0471.0581.0281.0660.9621.0801.1791.0751.1601.014
stock1.5841.3231.4451.6451.1471.8752.4971.7892.2491.433
laser1.4661.1421.5431.1321.0381.5201.6061.7971.2821.184
concrete1.2791.1561.2251.1361.0961.3821.2781.3381.3651.193
treasury1.5441.5891.6941.3511.0932.0723.0622.3603.2601.990
mortgage1.4261.5681.7991.3051.2812.0494.0322.4763.2412.101
friedman1.2131.1991.1471.1731.0921.3761.4201.3161.4151.305
wizmir1.2891.2241.2271.1461.0811.3651.3891.3561.3711.131
wankara1.2141.2201.1451.1511.0761.4221.5881.4011.4891.211
plastic1.0251.0691.0331.5230.9941.0331.1551.2851.9570.989
quake1.2381.0231.1261.0640.9921.4561.0551.1871.1331.005
abalone1.0451.0121.1371.0741.0231.0931.0391.2101.1481.011
compactiv1.1291.2741.2771.3661.1251.3022.8621.3281.4281.200
tic1.2361.0131.0691.0030.9921.4291.0461.2860.9960.996
ailerons1.1261.0331.1981.0481.0281.2381.1261.2361.1721.088
pole1.3511.0661.4201.0491.1171.5011.7471.5981.7961.234
elevators1.1261.0591.2361.0521.0411.2021.1901.2911.1571.097
california1.1031.1151.0771.0801.0371.2591.2161.1971.1461.234
house1.0731.0931.2531.0621.0361.2221.1461.3041.2581.073
average1.2281.1521.2481.1661.0561.4181.5881.4601.5351.224
times best200121200022
t-test p0.00000.0050.00000.0022 0.00850.01580.00700.0128
Wilcoxon p0.00000.00000.00000.0000 0.00010.00000.00000.0000
Table A12. Relative R M S E t s t for retention c 1 = 0.5 and c 2 = 0.25 in 10-fold cross-validation with k-NN with optimal k. All symbols are explained in the caption of Table A11.
Table A12. Relative R M S E t s t for retention c 1 = 0.5 and c 2 = 0.25 in 10-fold cross-validation with k-NN with optimal k. All symbols are explained in the caption of Table A11.
c1 = 0.50c2 = 0.25
DatasetTE-CTE-EDE-CDE-EMEISRTE-CTE-EDE-CDE-EMEISR
mach. CPU1.2671.0481.3601.0131.0521.3731.2271.5471.2531.148
baseball1.0811.0621.1711.2350.9971.1831.1491.6291.4061.094
dee1.1291.1051.0581.0581.0071.1991.1991.1101.2931.031
autoMPG81.2261.0981.1721.1201.0311.3621.2261.1991.1991.138
autoMPG61.1271.0821.0711.0881.0431.2801.1971.1411.1971.074
ele-11.0191.0291.0411.0390.9541.0561.1001.0541.0971.048
stock1.5491.3451.4421.6461.1801.8492.4011.7712.2361.551
laser1.4091.1551.5701.1181.0311.5011.5011.7231.2931.232
concrete1.2401.1811.1921.1111.0541.3461.3081.3461.3081.224
treasury1.4911.5631.7631.3271.3021.9823.0912.3643.2731.372
mortgage1.3741.5551.8581.2731.3042.0204.0812.4043.1112.854
friedman1.1991.2141.1911.1561.0731.3131.4101.2991.3431.234
wizmir1.2431.2001.2171.1221.0601.3621.3621.3621.3621.264
wankara1.2121.2121.1641.1461.0981.3731.4921.3731.4031.254
plastic0.9911.0711.0061.5190.9691.0221.1381.2531.9090.989
quake1.2351.0121.1021.0560.9901.4551.0151.1441.1020.998
abalone1.0261.0261.1211.0791.0311.0361.0551.1661.1031.058
compactiv1.1071.2931.2811.3411.2091.2452.8101.3311.3481.241
tic1.2211.0011.1131.0000.9861.3881.0021.2621.0010.993
ailerons1.0961.0421.1581.0541.0311.1701.1021.2511.1381.096
pole1.3391.0821.4781.0221.1251.4831.6841.6041.6981.261
elevators1.0921.0501.2231.0401.0391.1291.1091.3101.1291.103
california1.0821.0941.1171.0771.0371.2041.1471.1771.1541.116
house1.0751.0811.2081.0341.0291.1591.0851.2501.2501.078
average1.2011.1481.2531.1531.0681.3541.5371.4201.4841.227
times best110319100122
t-test p0.00000.00000.00000.0000 0.02350.00590.00030.0167
Wilcoxon p0.00010.00010.00000.0008 0.00040.00000.00000.0004
Table A13. Relative 1 r 2 in 50% houdout for DROP3-RT and MEISR for retention c 1 = 0.5 and c 2 = 0.25 with 8-NN. Dr2: relative 1 r 2 with DROP3-RT. Mr2: relative 1 r 2 with MEISR, Dc1: retention rate with DROP3-RT, Mc1: retention rate with MEISR, r 2 is the correlation between the predicted and actual output.
Table A13. Relative 1 r 2 in 50% houdout for DROP3-RT and MEISR for retention c 1 = 0.5 and c 2 = 0.25 with 8-NN. Dr2: relative 1 r 2 with DROP3-RT. Mr2: relative 1 r 2 with MEISR, Dc1: retention rate with DROP3-RT, Mc1: retention rate with MEISR, r 2 is the correlation between the predicted and actual output.
DatasetDr2Mr2Mr2/Dr2Dc1Mc1Mc1/Dc1
mach. CPU1.4611.2420.8500.5050.3960.784
baseball1.2051.1820.9800.4720.3410.723
dee1.1201.0740.9590.5230.3150.602
autoMPG81.2111.1160.9210.5390.3870.718
autoMPG61.1091.1711.0560.5360.3450.644
ele-11.0820.8470.7830.5300.3140.593
stock1.4201.5031.0580.5830.4320.742
laser1.3811.0170.7360.6590.4430.673
concrete1.2921.2240.9480.5070.4100.808
treasury1.4191.4701.0360.6360.4210.662
friedman1.1951.1530.9640.5740.4340.756
wizmir1.2601.1010.8730.5690.3240.569
wankara1.1351.1881.0470.5710.3820.669
plastic1.0500.7970.7590.3910.3780.967
quake1.0780.9620.8920.4380.4030.920
abalone1.0311.0811.0490.4550.4320.949
d e l t a a i l 1.0401.0370.9970.4450.4250.955
puma32h1.6421.5370.9350.4180.4020.963
compactiv1.0231.0941.0700.4850.4290.885
d e l t a e l v 1.0511.0430.9920.4690.4380.933
ailerons1.4471.0490.7250.4570.4170.912
pole1.1511.1390.9900.3350.4091.222
elevators1.0661.1501.0790.4700.4160.885
california1.0891.0720.9840.5020.4740.944
house1.1360.9740.8580.4570.4280.938
mv1.2031.3001.0800.6090.5300.870
average1.2041.1350.9350.5050.4050.819
times best818 125
t-test p 0.0773 0.0000
Wilcoxon p 0.0566 0.0000
Table A14. Relative 1 r 2 in 10-fold cross-validation for DROP3-RT and MEISR for retention c 1 = 0.5 and c 2 = 0.25 with 8-NN. All symbols are explained in the caption of Table A13.
Table A14. Relative 1 r 2 in 10-fold cross-validation for DROP3-RT and MEISR for retention c 1 = 0.5 and c 2 = 0.25 with 8-NN. All symbols are explained in the caption of Table A13.
DatasetDr2Mr2Mr2/Dr2Dc1Mc1Mc1/Dc1
mach. CPU1.5411.2630.8190.4950.3740.756
baseball1.1601.1340.9780.4600.3950.859
dee1.1660.9910.8500.5110.3520.689
autoMPG81.2101.1180.9240.4910.4210.857
autoMPG61.1681.1370.9740.5110.3730.729
ele-11.0230.8920.8730.4860.3530.725
stock1.6851.4210.8430.5720.4830.843
laser1.4361.0790.7510.6050.4550.752
concrete1.3511.2350.9140.5020.4650.926
treasury1.3461.5001.1140.6200.4390.707
friedman1.0761.1861.1020.5380.4370.812
wizmir1.3291.1400.8580.5100.4100.803
wankara1.3551.1030.8140.5210.4440.853
plastic0.9640.8330.8640.4190.3900.930
quake1.0581.0440.9860.4200.3870.922
abalone1.0531.0340.9820.4160.3890.935
delta-ail1.0391.0010.9630.4330.3640.841
puma32h1.0321.0621.0280.3810.3750.986
compactiv1.5571.0660.6850.4490.3910.870
delta-elv1.0161.0110.9950.4320.3760.870
ailerons1.0391.1341.0920.4250.2640.620
pole1.4251.6961.1900.2450.2440.996
elevators1.1371.1350.9980.4380.3580.816
california1.0481.1051.0550.4750.4020.845
house1.0841.0410.9690.4300.3180.741
mv1.0611.2971.2220.5310.3920.739
average1.2281.1610.9530.4810.3910.822
times best819 026
t-test p 0.1112 0.0001
Wilcoxon p 0.0588 0.0001
Table A15. Statistical significance test for experiments presented in Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9 and Table A10.
Table A15. Statistical significance test for experiments presented in Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9 and Table A10.
AlgorithmRelationt-test
p
Wilco-xon prelationt-test
p
Wilco-xon p
MEISR, IEA: 1-NN (BIS), FPA: 1-NN (BIS), 50%Hr1/r00.01480.0026r2/r00.47080.0018
MEISR, IEA: 1-NN (BIS), FPA: 1-NN (BIS), 10-CVr1/r00.01480.0026r2/r00.47080.0018
MEISR, IEA: k-NN (BIS), FPA: k-NN (BIS), 50%Hr1/r00.02590.0182r2/r00.00010.0001
MEISR, IEA: k-NN (BIS), FPA: k-NN (BIS), 10-CVr1/r00.23090.0058r2/r00.00010.0001
MEISR, IEA: k-NN (BIS), FPA: MLP (BIS), 50%Hr1/r00.56640.4839r2/r00.00010.0001
MEISR, IEA: k-NN (BIS), FPA: MLP (BIS), 10-CVr1/r00.47820.4593r2/r00.00030.0001
MEISR, IEA: k-NN (RIS), FPA: k-NN (RIS), 50%Hr1/r00.03050.0022r2/r00.00010.0001
MEISR, IEA: k-NN (RIS), FPA: k-NN (RIS), 10-CVr1/r00.03930.0032r2/r00.00010.0001
Table A16. Calculation time of the MEISR algorithm for BIS using our software (available from the online supplementary resources) on a two Xeon X5-2696v2 machine. i n s t —number of instances, a t —number of attributes, d i s t , 1 N N t and k N N t —total time of the process, where the inner regressor was respectively 1-NN and k-NN with optimal k, 1 d % and k d % —percentage of total time used for calculating the distance matrix with respectively 1-NN and k-NN with optimal k as inner regressors, 1 N N t i and k N N t i —total process time per one instance.
Table A16. Calculation time of the MEISR algorithm for BIS using our software (available from the online supplementary resources) on a two Xeon X5-2696v2 machine. i n s t —number of instances, a t —number of attributes, d i s t , 1 N N t and k N N t —total time of the process, where the inner regressor was respectively 1-NN and k-NN with optimal k, 1 d % and k d % —percentage of total time used for calculating the distance matrix with respectively 1-NN and k-NN with optimal k as inner regressors, 1 N N t i and k N N t i —total process time per one instance.
Datasetinstatk1-NNt [s]kNNt [s]dist [s]1d% [%]kd% [%]1-NNti [ms]kNNti [ms]
mach. CPU188610.1190.1190.0119.39.30.6330.633
baseball3031670.1790.2150.0189.88.10.5900.709
dee329670.1810.2190.0147.86.50.5510.667
autoMPG8353760.1850.2060.0168.67.70.5240.584
autoMPG6353540.1860.2020.0158.07.30.5270.573
ele-14462110.2430.3190.0176.85.20.5450.716
forestFire46512110.2930.3430.0248.16.90.6300.737
stock855930.4740.4910.0439.18.80.5540.574
steel8641240.4620.4810.0469.89.50.5350.557
laser894430.4810.4830.0367.67.50.5380.540
concrete927840.4940.5360.0479.58.80.5330.578
treasury9441530.4910.5060.06613.513.10.5200.536
mortgage9441520.5080.5130.05811.411.30.5380.543
friedman1080570.5480.6120.0488.77.80.5070.567
wizmir1315970.6470.7310.07711.810.50.4920.556
wankara1448990.6710.8020.08011.910.00.4630.554
plastic14852110.6590.9050.0659.97.20.4440.609
quake19603110.8511.0700.12314.511.50.4340.546
anacalt3647721.581.610.30919.619.20.4330.441
abalone37598111.682.250.34120.315.20.4470.599
delta-ail64155112.933.310.74925.622.60.4570.516
puma32h737232114.095.061.6740.833.00.5550.686
compactiv73732123.913.931.4537.136.90.5300.533
delta-elv85646114.115.192.0650.139.70.4800.606
tic884085116.417.813.9361.350.30.7250.884
ailerons1237540108.6110.14.8856.748.30.6960.816
pole134982648.5210.34.9758.348.30.6310.763
elevators1493818810.412.75.1349.344.70.6960.850
california185768916.018.29.2757.950.90.8610.980
house20506161118.621.111.360.853.60.9071.029
part of mv2717810928.130.317.763.058.41.0341.115
mv3669010947.850.032.167.164.21.3031.363

References

  1. García, S.; Luengo, J.; Herrera, F. Data Preprocessing in Data Mining; Springer: New York, NY, USA, 2015. [Google Scholar]
  2. Garcia, S.; Derrac, J.; Cano, J.; Herrera, F. Prototype selection for nearest neighbor classification: Taxonomy and empirical study. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 417–435. [Google Scholar] [CrossRef] [PubMed]
  3. Blachnik, M. Instance Selection for Classifier Performance Estimation in Meta Learning. Entropy 2017, 19, 583. [Google Scholar] [CrossRef]
  4. Arnaiz-González, Á.; Blachnik, M.; Kordos, M.; García-Osorio, C. Fusion of instance selection methods in regression tasks. Inf. Fusion 2016, 30, 69–79. [Google Scholar] [CrossRef]
  5. Barandela, R.; Ferri, F.J.; Sánchez, J.S. Decision boundary preserving prototype selection for nearest neighbor classification. Int. J. Pattern Recog. Artif. Intell. 2005, 19, 787–806. [Google Scholar] [CrossRef]
  6. Kirshners, A.; Parshutin, S.; Gorskis, H. Entropy-Based Classifier Enhancement to Handle Imbalanced Class Problem. Procedia Comput. Sci. 2017, 104, 586–591. [Google Scholar] [CrossRef]
  7. Michalowicz, J.V.; Nichols, J.M.; Bucholtz, F. Handbook of Differential Entropy; Chapman and Hall/CRC: London, UK, 2013. [Google Scholar]
  8. Rusiecki, A.; Kordos, M.; Kamiński, T.; Greń, K. Training neural networks on noisy data. In International Conference on Artificial Intelligence and Soft Computing; Springer: New York, NY, USA, 2014; pp. 131–142. [Google Scholar]
  9. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evolut. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
  10. Hart, P. The condensed nearest neighbor rule (Corresp.). IEEE Trans. Inf. Theory 1968, 14, 515–516. [Google Scholar] [CrossRef]
  11. Wilson, D.R.; Martinez, T.R. Reduction techniques for instance-based learning algorithms. Mach. Learn. 2000, 38, 257–286. [Google Scholar] [CrossRef]
  12. Riquelme, J.C.; Aguilar-Ruiz, J.S.; Toro, M. Finding representative patterns with ordered projections. Pattern Recogn. 2003, 36, 1009–1018. [Google Scholar] [CrossRef] [Green Version]
  13. Brighton, H.; Mellish, C. Advances in instance selection for instance-based learning algorithms. Data Min. Knowl. Discov. 2002, 6, 153–172. [Google Scholar] [CrossRef]
  14. Sáez, J.A.; Galar, M.; Luengo, J.; Herrera, F. INFFC: An iterative class noise filter based on the fusion of classifiers with noise sensitivity control. Inf. Fusion 2016, 27, 19–32. [Google Scholar] [CrossRef]
  15. Leyva, E.; González, A.; Pérez, R. Three new instance selection methods based on local sets: A comparative study with several approaches from a bi-objective perspective. Pattern Recogn. 2015, 48, 1523–1537. [Google Scholar] [CrossRef]
  16. Olvera-López, J.A.; Carrasco-Ochoa, J.A.; Martínez-Trinidad, J. Accurate and fast prototype selection based on the notion of relevant and border prototypes. J. Intell. Fuzzy Syst. 2018, 34, 2923–2934. [Google Scholar] [CrossRef]
  17. Suganthi, M.; Karunakaran, V. Instance selection and feature extraction using cuttlefish optimization algorithm and principal component analysis using decision tree. Cluster Comput. 2018, 1–13. [Google Scholar] [CrossRef]
  18. Krawczyk, B.; Triguero, I.; García, S.; Woźniak, M.; Herrera, F. Instance reduction for one-class classification. Knowl. Inf. Syst. 2018, 1–28. [Google Scholar] [CrossRef]
  19. Son, S.H.; Kim, J.Y. Data reduction for instance-based learning using entropy-based partitioning. In International Conference on Computational Science and Its Applications; Springer: New York, NY, USA, 2006; pp. 590–599. [Google Scholar]
  20. Kajdanowicz, T.; Plamowski, S.; Kazienko, P. Training set selection using entropy based distance. In Proceedings of the 2011 IEEE Jordan Conference on Applied Electrical Engineering and Computing Technologies (AEECT), Amman, Jordan, 6–8 December 2011; pp. 1–5. [Google Scholar]
  21. Arnaiz-González, Á.; Díez-Pastor, J.F.; Rodríguez, J.J.; García-Osorio, C. Study of data transformation techniques for adapting single-label prototype selection algorithms to multi-label learning. Expert Syst. Appl. 2018, 109, 114–130. [Google Scholar] [CrossRef]
  22. Grochowski, M.; Jankowski, N. Comparison of instance selection algorithms II. Results and comments. In International Conference on Artificial Intelligence and Soft Computing; Springer: New York, NY, USA, 2004; pp. 580–585. [Google Scholar]
  23. Zhang, J.; Yim, Y.S.; Yang, J. Intelligent selection of instances for prediction functions in lazy learning algorithms. In Lazy Learning; Springer: New York, NY, USA, 1997; pp. 175–191. [Google Scholar]
  24. Kordos, M.; Blachnik, M. Instance selection with neural networks for regression problems. In International Conference on Artificial Neural Networks; Springer: New York, NY, USA, 2012; pp. 263–270. [Google Scholar]
  25. Arnaiz-González, Á.; Díez-Pastor, J.F.; Rodríguez, J.J.; García-Osorio, C. Instance selection for regression: Adapting DROP. Neurocomputing 2016, 201, 66–81. [Google Scholar] [CrossRef]
  26. Arnaiz-González, Á.; Díez-Pastor, J.F.; Rodríguez, J.J.; García-Osorio, C.I. Instance selection for regression by discretization. Expert Syste. Appl. 2016, 54, 340–350. [Google Scholar] [CrossRef]
  27. Rodriguez-Fdez, I.; Mucientes, M.; Bugarin, A. An instance selection algorithm for regression and its application in variance reduction. In Proceedings of the 2013 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Hyderabad, India, 7–10 July 2013; pp. 1–8. [Google Scholar]
  28. Marchiori, E. Class conditional nearest neighbor for large margin instance selection. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 364–370. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Guillen, A.; Herrera, L.J.; Rubio, G.; Pomares, H.; Lendasse, A.; Rojas, I. New method for instance or prototype selection using mutual information in time series prediction. Neurocomputing 2010, 73, 2030–2038. [Google Scholar] [CrossRef]
  30. Božić, M.; Stojanović, M.; Stajić, Z.; Floranović, N. Mutual information-based inputs selection for electric load time series forecasting. Entropy 2013, 15, 926–942. [Google Scholar] [CrossRef]
  31. Stojanović, M.B.; Božić, M.M.; Stanković, M.M.; Stajić, Z.P. A methodology for training set instance selection using mutual information in time series prediction. Neurocomputing 2014, 141, 236–245. [Google Scholar] [CrossRef]
  32. Abdulali, A.; Hassan, W.; Jeon, S. Stimuli-magnitude-adaptive sample selection for data-driven haptic modeling. Entropy 2016, 18, 222. [Google Scholar] [CrossRef]
  33. Tolvi, J. Genetic algorithms for outlier detection and variable selection in linear regression models. Soft Comput. 2004, 8, 527–533. [Google Scholar] [CrossRef]
  34. Wu, S.; Olafsson, S. Optimal instance selection for improved decision tree induction. In Proceedings of the IIE Annual Conference on Institute of Industrial and Systems Engineers (IISE), Pittsburgh, PA, USA, 20–23 May 2006. [Google Scholar]
  35. García-Pedrajas, N.; Del Castillo, J.A.R.; Ortiz-Boyer, D. A cooperative coevolutionary algorithm for instance selection for instance-based learning. Mach. Learn. 2010, 78, 381–420. [Google Scholar] [CrossRef]
  36. Tsai, C.F.; Eberle, W.; Chu, C.Y. Genetic algorithms in feature and instance selection. Knowl.-Based Syst. 2013, 39, 240–247. [Google Scholar] [CrossRef]
  37. Derrac, J.; Triguero, I.; García, S.; Herrera, F. Integrating instance selection, instance weighting, and feature weighting for nearest neighbor classifiers by coevolutionary algorithms. IEEE Trans. Syst. Man Cybern. Part B 2012, 42, 1383–1397. [Google Scholar] [CrossRef] [PubMed]
  38. Czarnowski, I. Cluster-based instance selection for machine classification. Knowl. Inf. Syst. 2012, 30, 113–133. [Google Scholar] [CrossRef]
  39. Czarnowski, I.; Jędrzejowicz, P. Cluster-Based Instance Selection for the Imbalanced Data Classification. In International Conference on Computational Collective Intelligence; Springer: New York, NY, USA, 2018; pp. 191–200. [Google Scholar]
  40. Rosales-Pérez, A.; García, S.; Gonzalez, J.A.; Coello, C.A.C.; Herrera, F. An evolutionary multiobjective model and instance selection for support vector machines with pareto-based ensembles. IEEE Trans. Evolut. Comput. 2017, 21, 863–877. [Google Scholar] [CrossRef]
  41. Escalante, H.J.; Marin-Castro, M.; Morales-Reyes, A.; Graff, M.; Rosales-Pérez, A.; Montes-y Gómez, M.; Reyes, C.A.; Gonzalez, J.A. MOPG: A multi-objective evolutionary algorithm for prototype generation. Pattern Anal. Appl. 2017, 20, 33–47. [Google Scholar] [CrossRef]
  42. Deb, K.; Jain, H. An evolutionary many-objective optimization algorithm using reference-point-based nondominated sorting approach, part I: Solving problems with box constraints. IEEE Trans. Evolut. Comput. 2014, 18, 577–601. [Google Scholar] [CrossRef]
  43. Liu, H.; Motoda, H. Instance Selection and Construction for Data Mining; The Springer International Series in Engineering and Computer Science; Springer US: New York, NY, USA, 2013. [Google Scholar]
  44. Kordos, M.; Rusiecki, A.; Blachnik, M. Noise reduction in regression tasks with distance, instance, attribute and density weighting. In Proceedings of the 2015 IEEE 2nd International Conference on Cybernetics (CYBCONF), Gdynia, Poland, 24–26 June 2015; pp. 73–78. [Google Scholar]
  45. Matsumoto, M.; Nishimura, T. Mersenne twister: A 623-dimensionally equidistributed uniform pseudo-random number generator. ACM Trans. Model. Comput. Simul. (TOMACS) 1998, 8, 3–30. [Google Scholar] [CrossRef]
  46. Santha, M.; Vazirani, U.V. Generating quasi-random sequences from semi-random sources. J. Comput. Syst. Sci. 1986, 33, 75–87. [Google Scholar] [CrossRef]
  47. Pan, W.; Li, K.; Wang, M.; Wang, J.; Jiang, B. Adaptive randomness: A new population initialization method. Math. Probl. Eng. 2014, 2014. [Google Scholar] [CrossRef]
  48. Rahnamayan, S.; Wang, G.G. Center-based sampling for population-based algorithms. In Proceedings of the 2009 IEEE Congress on Evolutionary Computation, Trondheim, Norway, 18–21 May 2009; pp. 933–938. [Google Scholar]
  49. Javed, K.; Gouriveau, R.; Zerhouni, N. SW-ELM: A summation wavelet extreme learning machine algorithm with a priori parameter initialization. Neurocomputing 2014, 123, 299–307. [Google Scholar] [CrossRef] [Green Version]
  50. Kallel, L.; Schoenauer, M. Alternative Random Initialization in Genetic Algorithms; ICGA: San Fransisco, CA, USA, 1997; pp. 268–275. [Google Scholar]
  51. Łapa, K.; Cpałka, K.; Hayashi, Y. Hybrid initialization in the process of evolutionary learning. In International Conference on Artificial Intelligence and Soft Computing; Springer: New York, NY, USA, 2017; pp. 380–393. [Google Scholar]
  52. Reunanen, J. Overfitting in Feature Selection: Pitfalls and Solutions; Doctoral Dissertaion, Aalto University: Espoo/Helsinki, Finland, 2012. [Google Scholar]
  53. Jabbar, H.; Khan, R. Methods to Avoid Over-fitting and Under-fitting in Supervised Machine Learning (Comparative Study). Comput. Sci. Commun. Instrum. Devices 2014. [Google Scholar] [CrossRef]
  54. Kordos, M.; Rusiecki, A. Reducing Noise Impact on MLP Training. Soft Comput. 2016, 20, 49–65. [Google Scholar] [CrossRef]
  55. Alcalá-Fdez, J.; Fernández, A.; Luengo, J.; Derrac, J.; García, S.; Sánchez, L.; Herrera, F. Keel data-mining software tool: data set repository, integration of algorithms and experimental analysis framework. J. Mult.-Valued Logic Soft Comput. 2011, 17, 255–287. [Google Scholar]
  56. Kordos, M.; Duch, W. Variable step search algorithm for MLP training. In Proceedings of the 8th IASTED International Conference on Artificial Intelligence and Soft Computing, Marbella, Spain, 1–3 September 2004; pp. 215–220. [Google Scholar]
  57. Riedmiller, M.; Braun, H. A direct adaptive method for faster backpropagation learning: The RPROP algorithm. In Proceedings of the IEEE International Conference on Neural Networks, San Francisco, CA, USA, 28 March–1 April 1993; pp. 586–591. [Google Scholar]
  58. Refaeilzadeh, P. Cross-Validation. Encycl. Database Syst. 2009, 532–538. [Google Scholar] [CrossRef]
  59. Arlot, S. A survey of cross-validation procedures for model selection. Stat. Surv. 2010, 4, 40–79. [Google Scholar] [CrossRef] [Green Version]
  60. Demšar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  61. Tan, K.C.; Lee, T.H.; Khor, E.F. Evolutionary algorithms for multi-objective optimization: Performance assessments and comparisons. Artif. Intell. Rev. 2002, 17, 251–290. [Google Scholar] [CrossRef]
  62. Czyżak, P.; Jaszkiewicz, A. Pareto simulated annealing—A metaheuristic technique for multiple-objective combinatorial optimization. J. Multi-Criteria Dec. Anal. 1998, 7, 34–47. [Google Scholar] [CrossRef]
  63. Zitzler, E.; Deb, K.; Thiele, L. Comparison of multiobjective evolutionary algorithms: Empirical results. Evolut. Comput. 2000, 8, 173–195. [Google Scholar] [CrossRef] [PubMed]
  64. Riquelme, N.; Von Lücken, C.; Baran, B. Performance metrics in multi-objective optimization. In Proceedings of the 2015 Latin American Computing Conference (CLEI), Arequipa, Peru, 19–23 October 2015; pp. 1–11. [Google Scholar]
  65. Yu, X.; Lu, Y.; Yu, X. Evaluating Multiobjective Evolutionary Algorithms Using MCDM Methods. Math. Probl. Eng. 2018, 2018. [Google Scholar] [CrossRef]
  66. Rudzinski, F.; Alonso, J.; Bustince, H. Finding Sets of Non-Dominated Solutions with High Spread and Well-Balanced Distribution using Generalized Strength Pareto Evolutionary Algorithm. In Proceedings of the Conference International Fuzzy Systems Association and the European Society for Fuzzy Logic and Technology (IFSA-EUSFLAT-15), Asturias, Spain, 30 June–3 July 2015. [Google Scholar]
  67. Rudziński, F. An Application of Generalized Strength Pareto Evolutionary Algorithm for Finding a Set of Non-Dominated Solutions with High-Spread and Well-Balanced Distribution in the Logistics Facility Location Problem. In International Conference on Artificial Intelligence and Soft Computing; Springer: New York, NY, USA, 2017; pp. 439–450. [Google Scholar]
Figure 1. Instance selection in classification (a) and in regression (b). The axes represent the attributes x1 and x2. In classification, the red circle and blue cross represent points of two different classes. In regression, the height of the vertical line represents the output value of an instance and the circle shows its location in the input space.
Figure 1. Instance selection in classification (a) and in regression (b). The axes represent the attributes x1 and x2. In classification, the red circle and blue cross represent points of two different classes. In regression, the height of the vertical line represents the output value of an instance and the circle shows its location in the input space.
Entropy 20 00746 g001
Figure 2. Sample results of the MEISR algorithm run. Each pair of points (orange and green) represent one solution (one training set) with the percentage of selected instances on the horizontal axis and the corresponding R M S E t r n on training set (orange) and R M S E t s t on test set (green) on vertical axis. Only the points that formed the Pareto front are shown. Horizontal orange and green lines show the R M S E t r n and R M S E t s t respectively without instance selection. The additional are points shown in blue (c3, r3 and others close to them) are described in the text.
Figure 2. Sample results of the MEISR algorithm run. Each pair of points (orange and green) represent one solution (one training set) with the percentage of selected instances on the horizontal axis and the corresponding R M S E t r n on training set (orange) and R M S E t s t on test set (green) on vertical axis. Only the points that formed the Pareto front are shown. Horizontal orange and green lines show the R M S E t r n and R M S E t s t respectively without instance selection. The additional are points shown in blue (c3, r3 and others close to them) are described in the text.
Entropy 20 00746 g002
Figure 3. The experimental process.
Figure 3. The experimental process.
Entropy 20 00746 g003
Figure 4. Comparison of the instance selection approaches.
Figure 4. Comparison of the instance selection approaches.
Entropy 20 00746 g004
Figure 5. Left and middle: dependency between loss function R M S E (1-NN) and relative R M S E t s t ( r 1 / r 0 ) for retention c 1 . MEISR with 1-NN inner evaluator and 1-NN final regressor for retention c 1 (left), MEISR with k-NN inner evaluator and k-NN final regressor center) and k = optimal k. Larger and darker circles stand for datasets with higher number of instances. Right: dependency between optimal k and R M S E t s t . Darker and bigger points represent larger datasets.
Figure 5. Left and middle: dependency between loss function R M S E (1-NN) and relative R M S E t s t ( r 1 / r 0 ) for retention c 1 . MEISR with 1-NN inner evaluator and 1-NN final regressor for retention c 1 (left), MEISR with k-NN inner evaluator and k-NN final regressor center) and k = optimal k. Larger and darker circles stand for datasets with higher number of instances. Right: dependency between optimal k and R M S E t s t . Darker and bigger points represent larger datasets.
Entropy 20 00746 g005
Figure 6. Left: MEISR running time as a function of number of instances in the original training dataset. Right: Percentage of MEISR running time used to calculate the distance matrix used by k-NN. Light circles denote 1-NN as the inner evaluator and dark circles k-NN with optimal k.
Figure 6. Left: MEISR running time as a function of number of instances in the original training dataset. Right: Percentage of MEISR running time used to calculate the distance matrix used by k-NN. Light circles denote 1-NN as the inner evaluator and dark circles k-NN with optimal k.
Entropy 20 00746 g006

Share and Cite

MDPI and ACS Style

Kordos, M.; Łapa, K. Multi-Objective Evolutionary Instance Selection for Regression Tasks. Entropy 2018, 20, 746. https://doi.org/10.3390/e20100746

AMA Style

Kordos M, Łapa K. Multi-Objective Evolutionary Instance Selection for Regression Tasks. Entropy. 2018; 20(10):746. https://doi.org/10.3390/e20100746

Chicago/Turabian Style

Kordos, Mirosław, and Krystian Łapa. 2018. "Multi-Objective Evolutionary Instance Selection for Regression Tasks" Entropy 20, no. 10: 746. https://doi.org/10.3390/e20100746

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop