Next Article in Journal
A New Approach for Euler-Lagrange Orbits on Compact Manifolds with Boundary
Previous Article in Journal
An Efficient Parallel Extragradient Method for Systems of Variational Inequalities Involving Fixed Points of Demicontractive Mappings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization by Context Refinement for Development of Incremental Granular Models

Department of Control and Instrumentation Engineering, Chosun University, Gwangju 61452, Korea
*
Author to whom correspondence should be addressed.
Symmetry 2020, 12(11), 1916; https://doi.org/10.3390/sym12111916
Submission received: 21 October 2020 / Revised: 9 November 2020 / Accepted: 18 November 2020 / Published: 20 November 2020
(This article belongs to the Section Computer)

Abstract

:
Optimization by refinement of linguistic contexts produced from an output variable in the construction of an incremental granular model (IGM) is presented herein. In contrast to the conventional learning method using the backpropagation algorithm, we use a novel method to learn both the cluster centers of Gaussian fuzzy sets representing the symmetry in the premise part and the contexts of the consequent part in the if–then fuzzy rules. Hence, we use the fundamental concept of context-based fuzzy clustering and design with an integration of linear regression (LR) and granular fuzzy models (GFMs). This GFM is constructed based on the association between the triangular membership function produced both in the input–output variables. The context can be established by the system user or using an optimization method. Hence, we can obtain superior performances based on the combination of simple linear regression and local GFMs optimized by context refinement. Experimental results pertaining to coagulant dosing in a water purification plant and automobile miles per gallon prediction revealed that the presented method performed better than linear regression, multilinear perceptron, radial basis function networks, linguistic model, and the IGM.

1. Introduction

Owing to the rapid growth of various application problems, several studies have been conducted on fuzzy models. It is now acknowledged that real-world problems require intelligent methods that integrate the methodology of different models. To solve various application challenges, it is important to design complementary hybrid intelligent systems by integrating multiple computing technologies in a synergistic rather than exclusive manner. A representative method is neuro-fuzzy inference modeling [1,2]. Neural networks adapt to changing environments, whereas fuzzy models deal with fuzzy inference and knowledge-based decision making.
The basic concept of fuzzy clustering has been used in several studies to focus on fuzzy modeling, resulting in the concept of granular fuzzy models (GFMs) [3,4,5,6,7,8]. These models is based on information granules. These information granules (IG) comprise linguistic information produced from both input–output variables. The IG in the design of GFMs are obtained using specialized context-based fuzzy c-means clustering. Unlike context-free clustering, the fuzzy c-means clustering without context produces clusters. When the contexts are generated, then clustering occurs within a specified context. A system modeling method using a context-based fuzzy clustering algorithm has been successfully applied. Yeom [6] proposed a performance evaluation method based on the concept of coverage and specificity in developing a GFM. The performance is typically measured by the root mean square error (RMSE). Byeon [7] designed a Takagi–Sugeno–Kang (TSK) linguistic model. In that study, an intelligent predictor was designed to predict the direction of dinghy booms. Park [8] applied a linguistic model to water quality prediction. The main characteristics of the presented model are that it is user-centric and dwells upon the collection of user-oriented entities. Several studies have been conducted using the GFM [9,10,11,12,13,14,15,16,17,18].
In addition to several studies related to information granules [19,20,21,22,23,24], an incremental granular model (IGM) based on an integration of linear regression and a linguistic model has been proposed by Kwak. First, after performing a linear regression (LR), the error caused by the LR was predicted by the local granular model. Yeom [21] proposed an improved IGM based on granular networks with the compensation of error. Li [22] applied a patch-type sensor model based on a radial basis function network (RBFN) with incremental concept to estimate energy expenditure. Lee [23] presented a design method for a RBFN with incremental concept by linear regression and local RBFN for predicting the heating and cooling load in residual buildings problem. Kwak [24] proposed an IGM designed by context-based fuzzy c-means clustering with interval type-2 concept. This model possesses unique characteristics that preserve the homogeneity and includes uncertainties associated with the fuzzification factor. Several studies have been conducted using the IGM [25,26,27].
Although the IGM has been applied successfully in several studies, a method to adjust the premise and consequent parts using the learning method in the IGM has not been considered. Thus, the purpose of this paper is to perform the systematic design of an IGM based on a hybrid learning method. The design process comprises several important steps involving the adjustment of linguistic contexts in the design of local GFMs, estimation of cluster centers of Gaussian fuzzy sets representing the symmetry in context-based fuzzy clustering, and optimization of premise and consequent parameters. Here, the contexts are adjusted by a derivative-based optimization method. In contrast to the conventional derivative-based optimization method, the advantages of the proposed method is that the only linguistic contexts with weight information are adjusted by gradient descent method. After adjusting the contexts, the premise parameters are automatically updated by context-based fuzzy clustering. This method has unique characteristics in comparison to the conventional method.
The remainder of this paper is organized as follows. Section 2 presents the IGM constituents, including linear regression, context-based fuzzy c-means clustering, and a local GFM. Furthermore, this section includes the architecture and design procedure of the IGM. Section 3 provides the refinement of linguistic contexts in the development of the IGM. In Section 4, experiments performed on a coagulant dosing process in a water purification plant and automobile miles per gallon (mpg) prediction are described [28,29]. Finally, conclusions are provided in Section 5.

2. IGM

The components of the IGM include linear regression as the global part, a GFM as the local part, and context-based fuzzy c-means clustering for generating if–then rules. This section highlights the tendency of system modeling in proceeding to the simplest possible model, evaluating its performance, and then completing a series of necessary improvements.

2.1. Linear Regression

The experimental data under discussion are assumed to comprise 2D input–output data pairs. These data pairs are provided in the following format: { x k , y k } , k = 1 , 2 , , N . Here, x k is the k-th input vector and y k is the k-th output. Linear regression is expressed in the following standard form:
z k = a T x k + a 0 ,
where the regression coefficients are expressed by a and a 0 . The coefficients are obtained using the standard least-squares error method. Here, a T  consists of two coefficients [a1, a2]T. The improvement of the model occurs in which the local part is based on the input-error data { x k , e k } , where the error is e k = y k z k . Subsequently, the if–then fuzzy rules and contexts of triangular membership functions are obtained by the context-based fuzzy clustering approach.

2.2. Context-Based Fuzzy C-Means Clustering

The context-based fuzzy clustering algorithm is an approach for estimating the clusters to maintain the association characteristics [30,31,32]. The following reiterates the nature of context-based fuzzy clustering. This clustering is realized for each context W 1 , W 2 , , W p through the statistical distribution of the output space. The contexts in the traditional GFM are created by a set of triangular fuzzy sets spaced equally along the output. It can be difficult to estimate the clusters because some contexts are associated with data scarcity.
It can be difficult to produce if–then fuzzy rules in context-based fuzzy clustering approach. To create flexible contexts, we used a characteristics of probabilistic distribution for the output in shown in Figure 1. Figure 1 shows the generation of linguistic contexts obtained by probability density function (b) and conditional density function (c) based on the statistical distribution of error (a) obtained by linear regression. Finally, contexts with a triangular fuzzy set were generated.
As shown in Figure 1d, the generated contexts are represented as a linguistic label of error.
The partitioning matrix obtained by the t-th context is introduced as follows:
U ( W t ) = { u i k [ 0 , 1 ] | i = 1 c u i k = w t k     k } ,  
where w t k is the membership grade of the k-th data point in the t-th linguistic context; u i k is the grades corresponding to the k-th data point in the i-th cluster center. The loss function can be defined as
J = i = 1 c k = 1 N u i k m | | x k v i | | 2 ,
where v i is the cluster’s center of the i-th cluster. x k v i is computed by distance between the k-th input and the i-th center. Here, the fuzzification factor m is 2. The objective function is accomplished by adjusting the partition matrix and the cluster center’s membership grades. The update of the membership matrix is computed as [30,31,32]:
u t i k = w t k j = 1 c ( x k v i x k v j ) 2 m 1   i = 1 , 2 , , c ,   k = 1 , 2 , , N ,  
where u t i k denotes the i-th center and the k-th data point corresponding the t-th linguistic context. The cluster center v i is computed as follows:
v i = k = 1 N u t i k m x k k = 1 N u t i k m
Figure 2 shows the estimation of cluster centers corresponding to six contexts. After generating the contexts as shown in Figure 1, the cluster centers are obtained by context-based fuzzy c-means clustering. As shown in Figure 2, the data points (dot mark) included in each context and cluster centers (square mark) are distributed.

2.3. Local GFM

Figure 3 shows the architecture of the local GFM and the association between the cluster center and context. In Figure 3, the circle nodes represents the computation of membership degrees obtained in Equation (4). The GFM’s output is obtained as a triangular fuzzy membership function representing a fuzzy set. In other words, the GFM output E ˆ is fully represented by three parameters, E ˆ = e ˆ , e ˆ , e ˆ + . These parameters denotes lower bound, modal value, and upper bound, respectively.
Assuming a triangular fuzzy membership function for the contexts, the GFM’s output is expressed as [4,8]
E ˆ = W 1 ξ 1 W 2 ξ 2 W p ξ p ,
Equation (6) marked by and are used to compute that the basic calculations are performed by fuzzy numbers. ξ k is the summation value of the activation values generated in the k-th linguistic context. Here, activation level denotes the summation of membership degrees corresponding to each context. That is, the three points of triangular fuzzy sets representing contexts and the activation values are multiplied, respectively. Thus, the output includes the lower value, modal value, and upper value with interval prediction.
The activation values were computed using Equation (4). The numeric bias term was used to eliminate possible systematic errors as follows:
W 0 = 1 N k = 1 N ( e k e ˆ k ) ,
Here, the modeling errors e k obtained from LR are obtained by a global model. This value is calculated simply to eliminate errors. Consequently, the model output E ˆ ( x k ) allows the characterization for three parameters:
t = 1 p ξ t W t + W 0 ,   t = 1 p ξ t W t + W 0 ,   t = 1 p ξ t W t + + W 0  

2.4. IGM

The flow chart of the IGM design is visualized in Figure 4. Adopting a linear regression structure, we compensated the modeling errors through local rules of the GFM that represent the local nonlinearity of the model to be considered [9,11,28]. Figure 5 shows the architecture of the IGM.
The construction procedures of the IGM shown in Figure 5 are as follows:
[Step 1]
Use linear regression on the numerical data points. Subsequently, the errors e k = y k z k are obtained by the difference between the desired and linear regression outputs.
[Step 2]
Obtain the input and error pairs { x k , e k } . These error values are employed as the output data in the use of the local GFM. Subsequently, the contexts in the error space are generated. The linguistic contexts are produced as shown in Figure 1.
[Step 3]
Estimate the clusters using context-based fuzzy clustering approach.
[Step 4]
Calculate the aggregation values by the linear summation of the activation levels and the context weight. Consequently, the model output results in fuzzy number with a triangular type.
[Step 5]
Integrate the linear regression output and granular results of the GFM. Hence, the prediction result is expressed as Y ˆ = z E ˆ

3. Refinement of Contexts in IGM Design

We present the optimization of the context refinement in this Section. This optimization method comprises two steps, as shown in Figure 6. First, we used the gradient descent method to adjust the contexts from the error between the desired and GFM outputs. Next, we obtained new cluster centers using context-based fuzzy clustering approach. The number of iteration to operate in the loop is 20.
We can consider the optimization methods where the minimization is performed by the centers of the triangular membership function as the context. The maximization of the average match between the granular output of the GFM and the available numerical desired output is straightforward. We assumed that Y is an output of the GFM for x k , Y ( x k ) .
As described by the triangular fuzzy sets, we calculated the degrees as follows:
max 1 N k = 1 N Y ( x k ) ( target k )
Furthermore, we can consider minimizing the average values of the output for the corresponding inputs as follows:
min 1 N k = 1 N ( b k a k ) ,
where a k and b k are the values with lower and upper bounds of the triangular membership function produced for x k , respectively. We performed the optimization by assuming that the successive contexts can overlap at the half-point level. Therefore, the optimization focuses on the median values of the triangular fuzzy membership function of the context. When these modal values have been updated, this clustering is performed. Subsequently, the iteration loop of the optimization is repeated. In gradient descent methods, the next point is adjusted as follows:
w ( n e w ) = w ( o l d ) + Δ w
The chain rule to obtain the gradient vector is applied as follows:
Δ w i = η · ( E k w i ) = η · E k y ˆ k · y ˆ k w i
E k y ˆ k = 2 ( y ˆ k y k ) , y ˆ k w i = u i
Hence, the update formula is defined as follows:
Δ w i = 2 · η · ( y ˆ k y k ) · u i
Figure 7 shows a flowchart pertaining to context refinement by gradient descent and cluster adjustment via clustering.

4. Experimental Results

We performed experiments using the proposed IGM with two optimization steps for coagulant dosing in a water purification plant and automobile mpg prediction in these experiments. The experimental results were compared with those of previous studies.

4.1. Coagulant Dosing in Water Purification Plant

The test data were collected at a water purification plant, and the water purification capacity was 1,320,000 ton/day [28]. We used 346 consecutive samples. The input comprised the turbidity, temperature, pH, and alkalinity. The output was polyaluminum chloride (PAC). To evaluate the resulting model, the dataset was partitioned into training and test data sets, respectively. In this experiment, 173 training data pairs were used to construct the IGM using the optimization method, and the 173 test datasets were used to validate the presented IGM. Table 1 lists the performance comparison for the training (Trn_RMSE) and test (Tst_RMSE) data for the PAC prediction of the coagulant dosing process. The number of iteration loops was 20, and the learning rate was 0.001. Figure 8 shows the prediction performance for the training and test datasets, respectively. As shown in Figure 8, the experimental results showed good prediction and generalization capability. Figure 9 shows the error distribution of PAC obtained by linear regression. As shown in Figure 9, it can be seen that the error distribution is concentrated near zero. Figure 10 shows the cluster centers corresponding to each context when the number of contexts was 8. As listed in Table 1, the prediction performance revealed that the IGM with the new optimization method performed better than linear regression, multilayer perceptron, linguistic model, and the IGM. As listed in Table 1, the IGM without optimization method showed good performance in comparison to the previous works. Furthermore, the experimental results revealed that the IGM with optimization method enhanced the performance of IGM itself. When p = c = 8, we obtained the best prediction performance in trial and error method. We followed the selection method of the number of node and rules in the previous works [28].

4.2. Automobile MPG Prediction

The automobile MPG prediction data are available from the UCI repository [29]. The dataset consists of 392 data pairs. The input comprised the weight, acceleration, model year, cylinder number, displacement, and horsepower. The dataset was partitioned into training and test datasets. In this experiment, 196 training data pairs and the remaining test datasets were used for model evaluation. The number of iteration loops and the learning rates used were the same as those used previously. Figure 11 shows the performance for the training and test datasets. Figure 12 shows the distribution of error obtained by linear regression. Figure 13 shows the centers corresponding to each context, when the number of contexts was 8. Table 2 lists the performance comparison. As listed in Table 2, the experimental results revealed that the proposed IGM performed better than the linguistic model and IGM. When p = c = 8, we obtained the best prediction performance in trial and error method. We followed the selection method of the number of node and rules in the previous works [28].

4.3. Boston Housing Data

Next, we shall use Boston housing data set that deals with the problem of real estate price prediction. This data are available from the UCI repository [29]. In this example, we used twelve input variables except for one binary attribute. The total data include 506 data pairs. We divided the data set into training and testing data sets of equal size. In this study, we used 253 training data pairs, while the remaining testing data sets were used for model evaluation. Table 3 listed the comparison results for RMSE for Boston housing data set. When p = c = 8, we obtained the best prediction performance in trial and error method. The proposed method outperformed the linguistic model and the basic IGM.

4.4. Discussion

The experimental results revealed that the proposed optimization method showed good performance in comparison to linguistic model and IGM itself for three data sets. The proposed method can be summarized by the following strength and weakness:
Strengths:
-
The incremental granular model has high prediction performance by combining linear regression and local granular fuzzy model.
-
The local granular fuzzy model generates the automatic if-then rules using context-based fuzzy clustering method from numerical data set.
-
The incremental granular model can enhance the prediction performance by combining the derivative-based optimization and context-based fuzzy clustering.
-
In contrast to the conventional back-propagation method, after adjusting the contexts by steepest descent method, the cluster centers in the premise part are estimated by using context-based fuzzy clustering method.
Weaknesses:
-
The number of contexts is obtained by trial and error method.
-
The number of cluster center per context are obtained by trial and error method.
-
As the number of data points increase, the number of rules also increase
-
The specific context can include the small data points, when the distribution of context is uniform.

5. Conclusions

We developed an optimization method for an IGM. This method comprised two stages: context refinement and cluster adjustment. In contrast to the conventional gradient descent method, we performed a hybrid learning method using the gradient descent method and clustering. The experimental results clearly indicated that the proposed IGM with the optimization method presented better generalization capability compared with the conventional methods and IGM. Therefore, we conclude that the new optimization method for IGM design is effective. Future studies will involve the optimal determination of cluster per context and new performance measures based on the concept of coverage and specificity.

Author Contributions

M.-W.L. suggested the idea for this study and performed the experiments. K.-C.K. designed the experimental method. Both authors wrote and critically revised the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by research fund from Chosun University, 2020.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kwak, K.C.; Kim, S.S. Development of quantum-based adaptive neuro-fuzzy networks. IEEE Trans. Syst. Man Cybern. Part B 2010, 40, 91–100. [Google Scholar]
  2. Yeom, C.U.; Kwak, K.C. Performance comparison of AFNIS models by input space partitioning methods. Symmetry 2018, 10, 700. [Google Scholar] [CrossRef] [Green Version]
  3. Pedrycz, W.; Gomide, F. Fuzzy Systems Engineering: Toward Human-Centric Computing; IEEE Press: Piscataway, NJ, USA, 2007. [Google Scholar]
  4. Pedrycz, W.; Kwak, K.C. Linguistic models as a framework of user-centric system modeling. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 2006, 36, 727–745. [Google Scholar] [CrossRef]
  5. Reyes-Galaviz, O.F.; Pedrycz, W. Granular fuzzy models: Analysis, design, and evaluation. Int. J. Approx. Reason. 2015, 64, 1–19. [Google Scholar] [CrossRef]
  6. Byeon, Y.H.; Lee, M.W.; Lee, J.N.; Kwak, K.C. Prediction of dinghy boom direction using intelligent predictor. Int. J. Control Autom. Syst. 2018, 16, 368–376. [Google Scholar] [CrossRef]
  7. Yeom, C.U.; Kwak, K.C. Performance evaluation of automobile fuel consumption using a fuzzy-based granular model with coverage and specificity. Symmetry 2019, 11, 1480. [Google Scholar] [CrossRef] [Green Version]
  8. Park, J.I.; Jung, N.C.; Kwak, K.C.; Chun, M.G. Water quality prediction in a reservoir: Linguistic model approach for interval prediction. Int. J. Control Autom. Syst. 2010, 8, 866–874. [Google Scholar] [CrossRef]
  9. Lu, W.; Shan, D.; Pedrycz, W.; Zhang, L.; Yang, J.; Liu, X. Granular fuzzy modeling for multidimentional numeric data: A layered approach based on hyperbox. IEEE Trans. Fuzzy Syst. 2019, 27, 775–789. [Google Scholar] [CrossRef]
  10. Pedrycz, W. Granular computing for data analytics: A manifesto of human-centric computing. IEEE/CAA J. Autom. Sin. 2018, 5, 1025–1034. [Google Scholar] [CrossRef]
  11. Zhu, X.; Pedrycz, W.; Li, Z. Granular models and granular outliers. IEEE Trans. Fuzzy Syst. 2018, 26, 3835–3846. [Google Scholar] [CrossRef]
  12. Hu, X.; Pedrcyz, W.; Wang, X. Granular fuzzy rule-based models: A study in a comprehensive evaluation and construction of fuzzy models. IEEE Trans. Fuzzy Syst. 2017, 25, 1342–1355. [Google Scholar] [CrossRef]
  13. Zhu, X.; Pedrycz, W.; Li, Z. A development of hierarchically structured granular models realized through allocation of information granularity. IEEE Trans. Fuzzy Syst. 2020. [Google Scholar] [CrossRef]
  14. Pedrycz, W.; Homenda, W.; Jastrzebska, A.; Yu, F. Information granules and granular models: Selected design investigations. IEEE Int. Conf. Fuzzy Syst. 2020, 1–8. [Google Scholar] [CrossRef]
  15. Zhao, J.; Han, Z.; Pedrycz, W.; Wang, W. Granular model of long-term prediction for energy system in steel industry. IEEE Trans. Cybern. 2016, 46, 388–400. [Google Scholar] [CrossRef] [PubMed]
  16. Lu, W.; Pedrycz, W.; Yang, J.; Liu, X. Granular fuzzy modeling guided through the synergy of granulating output and clustering input subspaces. IEEE Trans. Cybern. 2019, 1–14. [Google Scholar] [CrossRef] [PubMed]
  17. Zhu, X.; Pedrycz, W.; Li, Z. A granular approach to interval output estimation for rule-based fuzzy models. IEEE Trans. Cybern. 2020, 1–10. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, D.; Pedrycz, W.; Li, Z. A two-phase development of fuzzy rule-based model and their analysis. IEEE Access 2019, 7, 80328–80341. [Google Scholar] [CrossRef]
  19. Pedrycz, W.; Kwak, K.C. The development of incremental models. IEEE Trans. Fuzzy Syst. 2007, 15, 507–518. [Google Scholar] [CrossRef]
  20. Li, J.; Pedrycz, W.; Wang, X. A rule-based development of incremental models. Int. J. Approx. Reason. 2015, 64, 20–38. [Google Scholar] [CrossRef]
  21. Yeom, C.U.; Kwak, K.C. The development of improved incremental models using local granular networks with error compensation. Symmetry 2017, 9, 266. [Google Scholar] [CrossRef] [Green Version]
  22. Li, M.; Kwak, K.C.; Kim, Y.T. Estimation of energy expenditure using a patch-type sensor module with an incremental radial basis function neural network. Sensors 2016, 16, 1566. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Lee, M.W.; Kwak, K.C. An incremental radial basis function network based on information granules and its application. Comput. Intell. Neurosci. 2016, 2016, 3207627. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Kwak, K.C. A design of incremental granular model using context-based interval type-2 fuzzy c-means clustering algorithm. IEICE Trans. Inf. Syst. 2016, E99D, 309–312. [Google Scholar] [CrossRef] [Green Version]
  25. Lee, M.W.; Kwak, K.C.; Pedrycz, W. An expansion of local granular models in the design of incremental model. IEEE Int. Conf. Fuzzy Syst. 2016, 1664–1670. [Google Scholar] [CrossRef]
  26. Soares, E.A.; Camargo, H.A.; Camargo, S.J.; Leite, D.F. Incremental Gaussian granular fuzzy modeling applied to hurricane track forecasting. IEEE Int. Conf. Fuzzy Syst. 2018, 1–8. [Google Scholar] [CrossRef]
  27. Leite, D.; Santana, M.; Borges, A.; Gomide, F. Fuzzy granular neural network for incremental modeling of nonlinear chaotic systems. IEEE Int. Conf. Fuzzy Syst. 2016, 64–71. [Google Scholar] [CrossRef]
  28. Byeon, Y.H.; Kwak, K.C. A design of genetically oriented rules-based incremental granular models and its application. Symmetry 2017, 9, 324. [Google Scholar] [CrossRef] [Green Version]
  29. UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/ml/datasets (accessed on 18 October 2020).
  30. Pedrycz, W. Conditional fuzzy c-mans. Pattern Recognit. Lett. 1996, 17, 625–631. [Google Scholar] [CrossRef]
  31. Pedrycz, W. Conditional fuzzy clustering in the design of radial basis function neural networks. IEEE Trans. Neural Netw. 1998, 9, 601–612. [Google Scholar] [CrossRef]
  32. Roh, S.B.; Oh, S.K.; Pedrycz, W.; Seo, K.; Fu, Z. Design methodology for radial basis function neural networks classifier based on locally linear reconstruction and conditional fuzzy c-means clustering. Int. J. Approx. Reason. 2019, 106, 228–243. [Google Scholar] [CrossRef]
Figure 1. Generation of contexts by statistical distribution from error obtained by linear regression. (a) Error histogram (b) probability density function (PDF) (c) conditional density function (CDF) (d) six linguistic contexts.
Figure 1. Generation of contexts by statistical distribution from error obtained by linear regression. (a) Error histogram (b) probability density function (PDF) (c) conditional density function (CDF) (d) six linguistic contexts.
Symmetry 12 01916 g001
Figure 2. Cluster estimation corresponding to six contexts.
Figure 2. Cluster estimation corresponding to six contexts.
Symmetry 12 01916 g002
Figure 3. Architecture of local GFM.
Figure 3. Architecture of local GFM.
Symmetry 12 01916 g003
Figure 4. Fundamental flow of IGM.
Figure 4. Fundamental flow of IGM.
Symmetry 12 01916 g004
Figure 5. Architecture of IGM.
Figure 5. Architecture of IGM.
Symmetry 12 01916 g005
Figure 6. Two-step optimization method.
Figure 6. Two-step optimization method.
Symmetry 12 01916 g006
Figure 7. Flowchart for context refinement and cluster adjustment.
Figure 7. Flowchart for context refinement and cluster adjustment.
Symmetry 12 01916 g007
Figure 8. Performance of PAC prediction for training and testing datasets. (a) prediction for training data (b) prediction for testing data.
Figure 8. Performance of PAC prediction for training and testing datasets. (a) prediction for training data (b) prediction for testing data.
Symmetry 12 01916 g008
Figure 9. Error distribution of PAC obtained by linear regression.
Figure 9. Error distribution of PAC obtained by linear regression.
Symmetry 12 01916 g009
Figure 10. Cluster centers corresponding to each context (square marks denote cluster center). (a) context no.1 (b) context no.2 (c) context no.3 (d) context no.4 (e) context no.5 (f) context no.6 (g) context no.7 (h) context no.8.
Figure 10. Cluster centers corresponding to each context (square marks denote cluster center). (a) context no.1 (b) context no.2 (c) context no.3 (d) context no.4 (e) context no.5 (f) context no.6 (g) context no.7 (h) context no.8.
Symmetry 12 01916 g010
Figure 11. Performance of MPG prediction for training and testing datasets. (a) prediction for training data (b) prediction for testing data.
Figure 11. Performance of MPG prediction for training and testing datasets. (a) prediction for training data (b) prediction for testing data.
Symmetry 12 01916 g011
Figure 12. Error distribution of MPG obtained by linear regression.
Figure 12. Error distribution of MPG obtained by linear regression.
Symmetry 12 01916 g012
Figure 13. Cluster centers corresponding to each context (square marks = cluster center). (a) context no.1 (b) context no.2 (c) context no.3 (d) context no.4 (e) context no.5 (f) context no.6 (g) context no.7 (h) context no.8.
Figure 13. Cluster centers corresponding to each context (square marks = cluster center). (a) context no.1 (b) context no.2 (c) context no.3 (d) context no.4 (e) context no.5 (f) context no.6 (g) context no.7 (h) context no.8.
Symmetry 12 01916 g013
Table 1. Comparison results for root mean square error (RMSE) for PAC prediction.
Table 1. Comparison results for root mean square error (RMSE) for PAC prediction.
No. of Rule (*: No. of Node)Trn_RMSETst_RMSE
Linear regression-3.5083.578
Multilayer perceptron45 *3.1913.251
RBFN based on context-based fuzzy c-means clustering [20]45 *3.0483.219
Linguistic model [4] (p = 8)c = 8642.4272.800
IGM [13]p = c = 8641.7902.009
The proposed modelp = c = 6362.1242.271
p = c = 7491.8622.093
p = c = 8641.6401.976
p = c = 9811.6312.105
Table 2. Performance comparison.
Table 2. Performance comparison.
No. of RuleTrn_RMSETst_RMSE
Linguistic model [4]362.7883.337
IGM [13]362.4183.175
The proposed modelp = c = 6362.2743.041
p = c = 7492.2062.984
p = c = 8642.1382.862
Table 3. Comparison results for RMSE for Boston housing data set.
Table 3. Comparison results for RMSE for Boston housing data set.
No. of RuleRMSE (Training)RMSE (Testing)
Linguistic model [4]364.7904.878
IGM [13]363.3713.649
The proposed method(p = c = 7)493.1063.273
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, M.-W.; Kwak, K.-C. Optimization by Context Refinement for Development of Incremental Granular Models. Symmetry 2020, 12, 1916. https://doi.org/10.3390/sym12111916

AMA Style

Lee M-W, Kwak K-C. Optimization by Context Refinement for Development of Incremental Granular Models. Symmetry. 2020; 12(11):1916. https://doi.org/10.3390/sym12111916

Chicago/Turabian Style

Lee, Myung-Won, and Keun-Chang Kwak. 2020. "Optimization by Context Refinement for Development of Incremental Granular Models" Symmetry 12, no. 11: 1916. https://doi.org/10.3390/sym12111916

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop