Next Article in Journal
Analysis Based on a Two-Dimensional Mathematical Model of the Thermo-Stressed State of a Copper Plate During Its Induction Heat Treatment
Previous Article in Journal
Research of Control Systems and Predictive Diagnostics of Electric Motors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Enhanced Fuzzy Time Series Forecasting Model Integrating Fuzzy C-Means Clustering, the Principle of Justifiable Granularity, and Particle Swarm Optimization

1
School of Business, Sichuan Normal University, Chengdu 610101, China
2
School of Economics and Management, University of Science and Technology Beijing, Beijing 100083, China
3
School of Finance, Hebei University of Economics and Business, Shijiazhuang 050061, China
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(5), 753; https://doi.org/10.3390/sym17050753
Submission received: 7 April 2025 / Revised: 2 May 2025 / Accepted: 8 May 2025 / Published: 14 May 2025
(This article belongs to the Section Mathematics)

Abstract

:
In this paper, we propose a novel fuzzy time series forecasting model that integrates fuzzy C-means (FCM) clustering, the principle of justifiable granularity (PJG), and particle swarm optimization (PSO), with a focus on leveraging symmetry in subinterval partitioning to enhance model interpretability and forecasting accuracy. First, the FCM method is employed to partition the universe of discourse, generating an initial division of subintervals. To ensure symmetric information representation, triangular fuzzy information granules are constructed for these subintervals in accordance with the principle of justifiable granularity. Then, an objective function is formulated for the entire universe of discourse, and the PSO algorithm is utilized to optimize the subinterval division, resulting in the final optimal partition. This process ensures that the subintervals achieve a balance between coverage and specificity, thereby introducing a form of symmetry in the partitioning of the universe of discourse. Leveraging the optimized symmetric partition, the framework of the fuzzy time series model is implemented for forecasting. Finally, the proposed approach is carried out on the Taiwan Weighted Stock Index (TAIEX) datasets and the Shanghai Composite Index (SHCI) datasets. The forecasting results demonstrate that the proposed approach achieves higher prediction accuracy and semantic accuracy compared with other methods.

1. Introduction

A time series is a sequence of observations recorded in chronological order, which essentially reflects the trend of one or more random variables changing over time. The primary goal in time series forecasting lies in extracting the inherent patterns from historical data and then utilizing these patterns for future value prediction. In real life, time series can be extensively applied across numerous domains, including stock index forecasting [1], hydrometeorology forecasting [2], enrollment forecasting [3], temperature forecasting [4], etc.
Conventional time series forecasting techniques mainly encompass several statistical models, including the Auto Regressive (AR) model, the Auto Regressive Moving Average (ARMA) model, and the Auto Regressive Integrated Moving Average (ARIMA) model [5]. These approaches center on estimating the parameters of the fitted models, which are then applied for forecasting purposes. Nevertheless, their restrictive assumptions and parametric characteristics restrict their effectiveness. As machine learning continues to advance, methods like Support Vector Machines (SVM) [6,7], Convolutional Neural Network (CNN) [8,9], and Long Short-Term Memory (LSTM) recurrent neural network [10,11] have also demonstrated remarkable effectiveness in time series forecasting. Despite their ability to address most real-world problems, these methods still face unresolved challenges involving fuzzy and uncertain data. Over the past few years, a growing trend has emerged in developing time series forecasting methods grounded in fuzzy set theory to tackle these issues.
In 1965, Zadeh [12] established fuzzy theory to address problems involving uncertain and fuzzy linguistic variables. In 1985, Sugeno and Tanaka [13] began to use fuzzy models to model and predict complex systems. Based on Zadeh’s fuzzy set theory, Song and Chissom [14,15] developed a prediction model for fuzzy time series in 1993, which expounded the framework involving fuzzy relation equations and approximate reasoning processes, thereby initiating the theoretical and practical research in this area. A traditional time series forecasting model comprises these four steps: (1) partitioning the universe of discourse, (2) defining fuzzy sets and fuzzifying the historical time series, (3) establishing fuzzy logical relationships of the fuzzy time series, and (4) forecasting and defuzzifying the fuzzy time series. Among these steps, step (1) serves as the cornerstone for fuzzy time series modeling [16] and represents the current focal point in fuzzy time series forecasting.
In 1993, Song and Chissom [17] were the first to propose the equal-interval partition technique for the universe of discourse, applying it to forecasting in fuzzy time series. Huarng [18,19] conducted extensive studies on how interval length affects forecasting outcomes. They proposed interval partitioning methods based on distribution, average value, and growth ratio, all of which proved to be superior in highlighting data structure and improving forecasting accuracy. Chen [20] carried out interval division according to the density of the sample data distribution following the principle of fine division in dense areas and coarse division in sparse areas. Chen et al. [21] first divided the universe of discourse using an equal-frequency method and then employed an entropy-based discretization iterative technique to granulate the fuzzy time series‘ universe of discourse. Wang et al. [22,23] combined the concept of information granulation in granular computing with fuzzy C-means clustering and Gath–Geva clustering to partition the universe of discourse in fuzzy time series, which improved the model’s accuracy. Yin et al. [24] proposed the interval type-2 FCM algorithm to replace the traditional FCM for dividing the sample domain, thereby enhancing the effectiveness of the FTS model. Recently, the principle of justifiable granularity (PJG) has been utilized in designing interval type-1 and type-2 fuzzy sets [25,26,27,28]. Furthermore, some scholars [29,30,31] have incorporated optimization algorithms to determine the optimal partitioning of subintervals within the universe of discourse. The optimal partitioning values for the intervals within the universe of discourse are frequently identified using the particle swarm optimization (PSO) algorithm [32,33,34,35]. Xian et al. [36] applied a hybrid artificial fish swarm optimization algorithm to determine the lengths of the intervals.
In sum, the above methods are summarized into four types. (1) Partitioning the universe of discourse with equal interval: This approach is easy to implement but lacks interpretability, failing to effectively reflect the distribution characteristics of datasets with uneven data distribution. (2) Partitioning the universe of discourse with equal frequency: This approach can intuitively show the data distribution and make it easy to understand the central tendency of data. However, it is insensitive to extreme values and limited to handling complex distribution data. (3) Partitioning the universe of discourse based on the clustering method: This method can effectively identify the internal structures and patterns within the data. However, it is vulnerable to the influence of outliers, and the number of data points within the intervals is unbalanced. (4) Partitioning the universe of discourse based on an optimization algorithm: This method usually depends on the setting of the objective function. Generally, these methods cannot simultaneously meet the requirements of both the interpretability of the intervals for dividing the universe of discourse and the prediction accuracy. Therefore, it is necessary to design a method for information granulation of the time series’ universe of discourse that can not only have strong interpretability but also achieve high prediction accuracy.
In this study, a new fuzzy time series forecasting model is proposed that integrates fuzzy C-means clustering, the principle of justifiable granularity, and particle swarm optimization. Firstly, we use the FCM method to divide the universe of discourse to obtain the initial subintervals division. Next, fuzzy information granules are constructed for the subintervals based on the principle of justifiable granularity. Then, an objective function is set for the entire universe of discourse, and the PSO algorithm is applied to obtain the final result of the optimal subintervals division. Based on the optimal partition results, we perform the framework of the fuzzy time series model for forecasting. Finally, we apply this approach to the TAIEX and SHCI datasets, comparing the forecasting results with other methods to demonstrate the effectiveness of the proposed forecasting model.
The remainder of this article is organized as follows. In Section 2, we review some relevant prerequisites. Section 3 presents a novel approach to segmenting the universe of discourse for time series data, utilizing fuzzy C-means clustering and information granulation. Section 4 details the specific steps of the fuzzy time series forecasting model, leveraging the optimal partition of the universe of discourse. Section 5 carries out some experiments to demonstrate the performance of the proposed forecasting model. Section 6 concludes this study.

2. Preliminaries

In this section, we offer a succinct review of the fundamental theories, including the fuzzy time series model, the fuzzy C-means clustering, triangular fuzzy information granules, the principle of justifiable granularity, and the particle swarm optimization algorithm.

2.1. Fuzzy Time Series Model

A fuzzy time series is a descriptive form of time series based on fuzzy theory, which simulates human cognition of the natural world. Each data object in the fuzzy time series is a semantic value. By constructing fuzzy logical relationships among these semantic values, the dynamic evolution process of the time series can be described, and its fuzzy change rules can be obtained. Traditional time series are usually recorded with precise values. However, in real life, the changes of many things often cannot be represented by precise numerical values. Moreover, the method of representing with precise numerical values does not conform to the cognitive patterns of human beings. Humans tend to understand and express things in a language form that they can understand. For example, when people perceive the temperature in a certain area, they usually describe the temperature level with words like “extremely cold”, “very cold”, “cold”, “hot”, “very hot”, “extremely hot”, etc., rather than specific temperature values. On the one hand, specific temperature data need to be measured with devices such as thermometers, and in the absence of measuring devices, the current information cannot be recorded in a timely manner. On the other hand, everyone has a different perception of temperature, and usually has their own judgment criteria. Recording the semantic values of temperature in chronological order results in a time series composed of fuzzy semantic values.
Song and Chissom [14] initially introduced the concept of fuzzy time series. The fundamental definition is provided as follows.
Definition 1.
Fuzzy Set (FS).
Let U denote the universe of discourse and U = { u 1 , u 2 , , u n } be an order segmentation set. Define A as a fuzzy set on U, which is expressed as:
A = f A ( u 1 ) u 1 + f A ( u 2 ) u 2 + + f A ( u n ) u n
where fA denotes the fuzzy membership function of the fuzzy set A; ui is an element of the fuzzy set Ai; and fA(ui) indicates the degree to which ui belongs to Ai, where 0 f A ( u i ) 1 and 0 ≤ in.
Definition 2.
Fuzzy Time Series (FTS).
Let Y ( t ) ( t = , 0 , 1 , 2 , ) be a subset of U and f i ( t ) be a collection of fuzzy sets defined on Y(t). If F ( t ) = { f 1 ( t ) , f 2 ( t ) , } is an ordered set consisting of fuzzy sets f i ( t ) , then F(t) is referred to as a fuzzy time series on Y(t).
Definition 3.
Fuzzy Relationship (FR).
Assume that R(t,t − 1) is the fuzzy relation from F(t − 1) to F(t), satisfying F(t) = F(t − 1)∘R(t,t − 1). In this case, F(t) is obtained from F(t − 1) via the fuzzy relation R(t,t − 1), which can be denoted as F(t − 1) → F(t). Here, “∘” represents the composition operation, F(t − 1) and F(t) are fuzzy sets, and R is the first-order fuzzy relation defined on F(t).
Definition 4.
Logical Relationship (LR).
Suppose F(t − 1) = Ai and F(t) = Aj; then, a fuzzy logic relation AiAj can be employed to depict the consecutive observations F(t − 1) and F(t). Here, Ai is referred to as the left-hand part (antecedent), while Aj is termed the right-hand part (consequent) of the fuzzy relation.

2.2. Fuzzy C-Means Clustering

Fuzzy C-means (FCM) clustering, introduced by Bezdek et al. [37], is a clustering algorithm that incorporates fuzzy theory. Unlike traditional hard clustering algorithms like K-Means, the clustering results of FCM do not completely belong to or completely not belong to a certain cluster but are represented by membership degrees to indicate the extent of belonging to a certain class. The clustering results are more flexible and have found extensive application across diverse fields. The essence of the FCM clustering algorithm is to transform the clustering problem of a dataset into a constrained nonlinear programming problem. By optimizing and solving, the corresponding data partitioning and class prototypes are obtained. The algorithm is straightforward to implement and offers good semantic interpretation.
The optimization goal and constraints for FCM clustering are presented as follows:
J m ( U , V ) = i = 1 N j = 1 C u ij m x i v j 2
s . t .   { 0 u i j 1 , i , j j = 1 C u i j = 1 , j 0 i = 1 N u i j N , i
where xi denotes the data value, N is the total number of data points, C represents the number of clusters, m ∈ (1, +∞) is the fuzziness coefficient, uij indicates the membership degree of the i-th data point in the j-th cluster, V is the set of cluster centers, vj describes the center of the j-th cluster, and ||∙|| represents the Euclidean norm.
Solving the objective function under the constraint conditions mainly consists of two steps: solving the membership degree uij and calculating the cluster center vj. The FCM clustering algorithm employs an iterative method to minimize the objective function, with the corresponding formulas given as follows:
u i j = [ k = 1 C ( x i v j x i v k ) 2 m 1 ] 1
v j = i = 1 N u i j m x i i = 1 N u i j m

2.3. Triangular Fuzzy Information Granules

Fuzzy sets are usually used to represent information granules. Common methods for representing fuzzy information granules (FIGs) [38] include interval fuzzy information granules (IFIGs), triangular fuzzy information granules (TFIGs), gaussian fuzzy information granules (GFIGs), trapezoidal fuzzy information granules (TIGs), etc.
The membership function for a TFIG can be specified as follows:
A ( x ; a , m , b ) = { 0 , x < a x a m a , a x m b x b m , m < x b 0 , x > b
where x is a data point in the dataset X; a and b are the lower and upper supports of the triangular fuzzy set, respectively; and m is the core of the triangular fuzzy set. Figure 1 displays an example of a TFIG with a = 0.2, m = 0.5, b = 0.7.

2.4. Principle of Justifiable Granularity

The principle of justifiable granularity (PJG), proposed by Pedrycz [39], focuses on creating an information granule based on empirical evidence found in one-dimensional numeric datasets. This PJG encompasses two primary metrics: coverage and specificity.
(1) Coverage: This metric reflects the quantity of numeric evidence accumulated within the bounds of the information granule. A high coverage value suggests that the information granule is well supported and accurately represents the original data.
(2) Specificity: This metric measures the precision of the constructed information granule. A shorter length of the information granule implies greater specificity, indicating that the resulting information granule possesses well-defined semantics (meaning).
Obviously, the two indicators are in conflict. When the amount of data covered by the information granule increases, the coverage of this information granule becomes higher, while its specificity becomes lower. The essence of PJG is to seek a balance between the coverage and specificity of the information granule. That is, the information granule should have a certain degree of specificity while covering as much data as possible. Therefore, we use Q to represent the optimal balance, and its expression is as follows:
Q = c o v × s p
Given a one-dimensional numeric dataset X = { x 1 , x 2 , , x n } , we aim to construct a triangular fuzzy information granule Ω (a, m, b) according to PJG. The mean or median of the dataset X is usually used as the value of m. Then, we focus on solving the lower and upper bounds a and b, respectively.
The coverage of Ω is expressed as:
c o v ( Ω ) = F 1 ( c a r d { x k X | a x k b } )
where F1 is an increasing function, and c a r d { } represents the number of data points contained in the information granule Ω.
The specificity of Ω is expressed as:
s p ( Ω ) = F 2 ( | b a | )
where F2 is a decreasing function, and |ba| represents the length of Ω.
The value of m can divide Ω into two parts. The left part is used to determine the optimal value of the lower bound a, while the right part is used to determine the optimal value of the upper bound b. Next, we first discuss how to find the optimal lower bound a. The coverage and specificity of the left part of Ω are expressed as:
c o v a = F 1 ( c a r d { x k X | a x k m } )
s p a = F 2 ( | m a | )
Maximizing Q yields the optimal upper bound value of Ω, which is expressed as follows:
a o p t = arg max a < m Q ( a )
The coverage and specificity of the right part of Ω are expressed as:
c o v b = F 1 ( c a r d { x k X | m x k b } )
s p b = F 2 ( | b m | )
The optimal lower bound value of Ω is derived in an analogous way:
b o p t = arg max b > m Q ( b )
The commonly used increasing function F1 and decreasing function F2 are set as the following functions, respectively:
F 1 ( x ) = x
F 2 ( x ) = exp ( α x )
where α ≥ 0 represents the level of the information granularity. As α changes, the information granule Ω will also change accordingly. That is, the parameter α affects the specificity of the information granule. If α = 0, F2(x) = 1, the information granule constructed encompasses all the data in the dataset. At this time, the resulting information granule is exactly the same as the one obtained by traditional interval granulation, losing its specificity. The larger the value of α, the more specific the constructed information granule will be. For each value of α, an information granule Ω α = [ a o p t α , b o p t α ] can be obtained by optimizing Q ( b α ) and Q ( a α ) . In sum, the level of the information granulation is closely related to the width of the generated information granule.

2.5. Particle Swarm Optimization Algorithm

The particle swarm optimization (PSO) algorithm is inspired by the foraging behavior of birds, where only one food source exists within a flock’s territory [40]. While the birds do not know the precise location of the food, they can sense the distance between their current position and the food source. The method mimics the individual’s search activity, and each particle’s position indicates a possible solution to the optimization problem. The steps are as follows:
(1) First, establish the maximum velocity of the particles, their position information, the number of independent variables for the objective function, and the maximum number of algorithm iterations.
(2) To update the velocity and position, define the fitness function, find the global ideal solution, compare it to the previous optimal solution, and identify the individual extreme value as the optimal solution for each particle.
(3) Update the particles’ position and velocity continuously.
(4) When the maximum number of iterations is reached or the error between generations satisfies the predetermined criterion, the PSO optimization process comes to an end.

3. Partition of the Universe of Discourse Based on FCM, PJG, and PSO

In fuzzy time series forecasting, the way the universe of discourse is partitioned impacts the model’s accuracy. In this section, we propose an approach to partitioning the universe of discourse for time series based on FCM, PJG, and PSO. First, we apply the FCM clustering to obtain the initial partition of the universe of discourse of the time series. Then, based on PJG, triangular fuzzy information granules are constructed on the initially partitioned subintervals. Finally, we set up an objective function for the entire universe of discourse space and use the PSO algorithm for the optimization and solution to obtain the final partition result of the universe of discourse.

3.1. Initial Partition of the Universe of Discourse Based on FCM

Given a time series dataset X = { x 1 , x 2 , , x n } , we define X m i n = min { x i | x i X } and X m a x = max { x i | x i X } . Let U = [ U l , U u ] = [ X min l 1 , X max + l 2 ] denote the universe of discourse, l1 and l2 are called trimming factors, which are two appropriate positive numbers. Assume the universe is partitioned into p intervals (typically p > 2). We utilize the FCM method on the time series X to obtain the cluster prototypes V of the universe of discourse, which are denoted as the clustering centers. Then, we arrange these clustering centers in ascending order and calculate the medians of two adjacent clustering centers as the boundary values of the subintervals of the partitioned universe of discourse as follows:
h i = c i + c i + 1 2 ,   i = 1 , 2 , , n 1
where hi represents the boundary value of the partition, and ci denotes the clustering center obtained by the FCM.
Then, the subintervals after initial partitioning are expressed as follows:
u 1 = { x X | U l x < h 1 } u 2 = { x X | h 1 x < h 2 } u p = { x X | h p 1 x U u }

3.2. Optimized Partition of the Universe of Discourse Based on PJG and PSO

The subintervals after the initial partitioning of the universe of discourse are regarded as information granules, and we use triangular fuzzy sets to describe the information granules. The parameters of the information granules are related to the level of information granularity.
Given a subinterval set X i = { 6.5 , 8 , 1.2 , 3.4 , 0.6 , 2.1 , 2.3 , 3.7 , 4.5 , 5 , 1.6 } , we need to represent it with triangular fuzzy information granule. The specific solution steps are as follows:
(1) Sort Xi from smallest to largest to obtain the value of m.
After sorting Xi, we get X i = { 8 , 6.5 , 3.4 , 2.3 , 1.6 , 0.6 , 1.2 , 2.1 , 3.7 , 4.5 , 5 } . If m = 0.6 is the median of Xi, then X i is divided into two parts: X i l = { 8 , 6.5 , 3.4 , 2.3 , 1.6 } and X i r = { 1.2 , 2.1 , 3.7 , 4.5 , 5 } . X i l is used to solve the lower bound a, and X i r is used to solve the upper bound b.
(2) Assume that the information granularity level α = 0.5, and solve for the optimal lower bound a o p t 0.5 and the optimal upper bound b o p t 0.5 under this information granularity level.
First, solve the optimal lower bound a o p t 0.5 according to X i l and Formula (17).
a i = 8 ,   V ( a i 0.5 ) = 5 × exp ( 0.5 × | 8 0.6 | ) = 0.0678
a i = 6.5 ,   V ( a i 0.5 ) = 4 × exp ( 0.5 × | 6.5 0.6 | ) = 0.1149
a i = 3.4 ,   V ( a i 0.5 ) = 3 × exp ( 0.5 × | 3.4 0.6 | ) = 0.4060
a i = 2.3 ,   V ( a i 0.5 ) = 2 × exp ( 0.5 × | 2.3 0.6 | ) = 0.4691
a i = 1.6 ,   V ( a i 0.5 ) = 1 × exp ( 0.5 × | 1.6 0.6 | ) = 0.3329
As a result, we obtain a o p t 0.5 = 2.3 . Likewise, we determine the optimal upper bound b o p t 0.5 according to X i r and Formula (17).
b i = 1.2 ,   V ( b i 0.5 ) = 1 × exp ( 0.5 × | 1.2 0.6 | ) = 0.7408
b i = 2.1 ,   V ( b i 0.5 ) = 2 × exp ( 0.5 × | 2.1 0.6 | ) = 0.9447
b i = 3.7 ,   V ( b i 0.5 ) = 3 × exp ( 0.5 × | 3.7 0.6 | ) = 0.6367
b i = 4.5 ,   V ( b i 0.5 ) = 4 × exp ( 0.5 × | 4.5 0.6 | ) = 0.5691
b i = 5 ,   V ( b i 0.5 ) = 5 × exp ( 0.5 × | 5 0.6 | ) = 0.5540
As a result, we obtain b o p t 0.5 = 2.1 . In summary, at the information granularity level α = 0.5, we obtain the optimal triangular fuzzy information granule Ω 0.5 = [ 2.3 , 2.1 ] .
Since α [ 0 , 1 ] , the optimal information granule Ωα will change as α varies, as shown in Table 1.
For each subinterval Xi, its information granule Ω i α can be determined through the above fuzzy information granulation operation. In order to make the divided intervals have clearer semantic characteristics, we introduce the following indicator to describe the information granules constructed in the current interval:
V o l ( Ω i ) = | h i h i 1 | 0 1 | Ω i α | d a = d i 0 1 | Ω i α | d a
where d i = | h i h i 1 | is the width of the interval Xi, and | Ω i α | = | b i α a i α | represents the size of the constructed information granule. From Formula (20), it can be seen that the evaluation of the partitioned intervals takes into account the information granule Ω i α ( α = 0 , 0.1 , , 1 ) at all levels of information granularity, rather than considering the information granules at a single level of information granularity.
Still taking the above dataset Xi as an example, based on the information granules formed under different levels of information granularity, we calculate the value of | Ω α | = | b α a α | , as shown in Figure 2. The solution process of 0 1 | Ω i α | d a is as follows:
As shown in Figure 2, as the information granularity level α increases, the corresponding information granule size | Ω α | is constantly decreasing, and its changing trend is in a polyline shape. Therefore, solving for 0 1 | Ω i α | d α can be transformed into calculating the sum of the integrals of several piecewise linear functions. The evaluation indicator for the information granulation of the entire dataset X is as follows:
V = V o l ( Ω 1 ) + V o l ( Ω 2 ) + + V o l ( Ω p )
Since the value of V varies with each subinterval Xi, the optimization problem for the partitioning of the universe of discourse is thus transformed into the following optimization problem [41]:
min d 1 , d 2 , , d p i = 1 p V o l ( Ω i )
For this optimization problem, we select the PSO algorithm to derive the optimal partitioning intervals of the universe of discourse. Among them, the initial solution is the initial partitioning intervals of the universe of discourse using the FCM algorithm.
In summary, we propose a partition algorithm of the universe of discourse based on FCM, PJG, and PSO.
In Algorithm 1, we first apply the FCM algorithm to partition a time series consisting of n points into p clusters, obtaining the initial partition intervals. If the number of iterations for the FCM algorithm is N1, the time complexity is O(p2nN1). Secondly, we calculate the information granule representation at each level of information granularity according to the PJG, with a time complexity of approximately O(n). Finally, we construct the objective function based on Formulas (21) and (22) and solve it using the PSO algorithm. Assuming the number of iterations for the PSO algorithm is N2, the time complexity is O(pnN2). Therefore, the overall time complexity of Algorithm 1 is O(p2nN1 + n + pnN2).
Algorithm 1: Partition of the Universe of Discourse Based on FCM, PJG, and PSO
Input: Time series X = { x 1 , x 2 , , x n } and the number of partition intervals p
Output: The optimal partition intervals of the universe of discourse U = { u 1 , u 2 , , u p }
1. Utilize the FCM clustering method to divide the universe of discourse into p clusters, obtain the clustering centers, then sort the clustering centers in ascending order, and calculate the initial partition subintervals of the universe of discourse according to Formula (19).
2. For the initial partition subintervals, use triangular fuzzy sets to construct information granules Ω = [ a , m , b ] , and solve the parameters of TFIGs according to PJG.
3. For the TFIGs, use the PSO algorithm to solve it according to Formulas (21) and (22) to obtain the final optimal partition intervals U = { u 1 , u 2 , , u p } of the universe of discourse.

4. Fuzzy Time Series Forecasting Based on the Novel Partition of the Universe of Discourse

The fuzzy time series forecasting model usually contains four basic steps: (1) defining and partitioning the universe of discourse; (2) defining fuzzy sets and fuzzifying historical time series; (3) establishing fuzzy logical relationships of the fuzzy time series; (4) forecasting and defuzzifying the fuzzy time series. Utilizing the optimal partitioning results of the universe of discourse from Section 3, the framework for fuzzy time series forecasting is presented in Figure 3.
The exact steps of the fuzzy time series forecasting based on the optimal partition of the discourse universe are presented in this section using the Taiwan Stock Exchange Capitalization Weighted Stock Index (TAIEX) as an example. Among them, the data from 4 January 1991 to 30 October 1991, spanning the first ten months, is designated as the training set, while the data from 1 November 1991 to 28 December 1991, covering the last two months, is used as the test set.

4.1. Defining and Partitioning the Universe of Discourse

Step 1: Converting the original time series into a rate of change series.
R ( t ) = x ( t ) x ( t 1 ) x ( t 1 ) × 100
where R(t) denotes the rate of change of the stock index on the t-th trading day, x(t) is the stock index at the current moment, and x(t − 1) is the stock index at the previous moment.
The original stock index time series can be transformed into a stock index change rate sequence using Formula (23). Table 2 and Figure 4 show the original time series and the transformed change rate sequence of the TAIEX training set.
Obviously, the transformed sequence has removed the influence of trends and is more convenient for subsequent operation and processing [42].
Step 2: Determining the universe of discourse.
Define the universe of discourse U = [ U l , U u ] = [ r min l 1 , r max + l 2 ] , where rmin represents the minimum value of the sequence R, rmax denotes the maximum value of the the sequence R, and l1 and l2 are called trimming factors, which are two appropriate positive numbers.
For the above change rate sequence, its minimum and maximum values are rmin = −6.66 and rmax = 6.76, respectively. Therefore, we set the trimming factors to l1 = 0.34 and l2 = 0.24, resulting in the universe of discourse U = [ 7 , 7 ] to be partitioned.
Step 3: Dividing the universe of discourse.
Step 3.1: Initial partitioning of the universe of discourse based on FCM.
Many scholars have already conducted research on TAIEX datasets, dividing the universe of discourse into seven intervals [23], which can yield relatively high-quality and strongly interpretable forecasting results. Therefore, in this study, we also assume that the number of fuzzy intervals is seven, and the number of clustering centers is seven. We subsequently perform the FCM method on the change rate sequence R(t) to obtain the cluster centers. After that, we sort the cluster centers in ascending order and calculate the median of each pair of adjacent cluster centers to serve as the boundary values for dividing the universe of discourse into subintervals, resulting in the following subintervals: u1 = [−7.00, −4.13), u2 = [−4.13, −1.79), u3 = [−1.79, −0.38), u4 = [−0.38, 0.95), u5 = [0.95, 2.54), u6 = [2.54, 4.96), u7 = [4.96, 7.00].
As shown in Figure 5, the subintervals using FCM are consistent with the distribution characteristics of the data points. In areas with dense data distribution, the intervals of the divided subintervals are smaller, whereas in areas with sparse data distribution, the intervals of the divided subintervals are larger. However, the number of data points within each subinterval differs significantly. Specifically, subintervals with dense data distribution have many data points, whereas those with sparse data distribution have few data points. Therefore, it is necessary to optimize the current subintervals.
Step 3.2: Optimized partitioning of the universe of discourse based on PJG and PSO.
Taking the subintervals obtained in the previous step as the initial solution, we optimize them using the PSO algorithm, and the objective function is min d 1 , d 2 , , d 7 i = 1 7 V o l ( Ω i ) , resulting in the following optimized subintervals: u 1 = [ 7.00 , 3.80 ) , u 2 = [ 3.80 , 2.24 ] , u 3 = [ 2.24 , 0.57 ) , u 4 = [ 0.57 , 1.17 ] , u 5 = [ 1.17 , 2.60 ) , u 6 = [ 2.60 , 4.40 ) , u 7 = [ 4.40 , 7.00 ] .
As illustrated in Figure 6, the subintervals derived from the optimized partitioning algorithm based on PJG and PSO exhibit clearer data point distribution features compared to those from the initial partitioning method shown in Figure 5. Moreover, the optimized subintervals show a more balanced distribution of data point quantities.

4.2. Defining Fuzzy Sets and Fuzzifying Historical Time Series

Step 1: Defining fuzzy sets based on the optimized partition.
Based on the partition subintervals obtained in Section 4.1, we define fuzzy sets for each subinterval as A 1 , A 2 , , A 7 , where each fuzzy set is determined as a semantic value, as shown in Table 3.
Then, we define fuzzy sets on U, and each Ai is expressed in terms of the subintervals A 1 , A 2 , , A 7 .
A 1 = 1 / u 1 + 0.5 / u 2 + 0 / u 3 + + 0 / u 7 A 2 = 0.5 / u 1 + 1 / u 2 + 0.5 / u 3 + + 0 / u 7 A 6 = 0 / u 1 + 0 / u 2 + + 1 / u 6 + 0.5 / u 7 A 7 = 0 / u 1 + 0 / u 2 + + 0.5 / u 6 + 1 / u 7
Step 2: Fuzzifying historical time series.
According to the fuzzy set definitions in Step 1, we perform fuzzification on historical time series. When fuzzifying, the “maximum membership principle” is adopted, which means choosing the fuzzy set with the maximum membership degree corresponding to the sequence value as the fuzzy description of that data point. If the value of the data point belongs to the subinterval ui, then the data point is fuzzified as Ai. Table 4 presents the fuzzification results for some data points of the change rate sequence.

4.3. Establishing Fuzzy Logical Relationships

Step 1: Establishing first-order fuzzy logical relationships.
Leveraging the fuzzified data in Section 4.2, we establish first-order fuzzy logical relationships (FLRs). Assume that the fuzzy values at time t − 1 and t are Ai and Aj, respectively, then the resulting first-order FLR is AiAj. The first-order FLRs of some data of the rate change sequence in Table 4 are shown in Table 5.
Step 2: Establishing first-order fuzzy logical relationships.
Merging FLRs with the same antecedent into the same group, we can establish a fuzzy logical relationship group (FLRG). Assuming the existence of FLRs AiAj and AiAz, we then construct the FLRG: AiAj, Az. Utilizing the partial first-order FLRs listed in Table 5, we obtain the first-order FLRGs shown in Table 6.
Utilizing the derived FLRGs, we can formulate fuzzy rules. The expression for the fuzzy rules of the first-order FLR is presented as follows:
Rule Ri: If Ft−1 is Ai then Ft is Aj.
Step 3: Establishing the fuzzy logical relationship matrix.
Leveraging the FLRGs established in Step 2, we calculate the occurrence frequency of each fuzzy rule and establish the fuzzy logical relationship matrix, which is depicted in Table 7.
Convert the occurrence frequencies in the above fuzzy logical relationship matrix into frequencies with the formula given below:
w i ( t ) = [ w 1 , w 2 , , w m ] = [ w 1 j = 1 m w j , w 2 j = 1 m w j , , w m j = 1 m w j ]
Calculate the frequencies according to Formula (24) to obtain the following weight matrix of fuzzy logical relationships, which is presented in Table 8.

4.4. Forecasting and Defuzzification

Step 1: Forecasting the fuzzy time series.
After establishing the fuzzy logical relationship matrix for historical time series, we can predict the future time series with the formula provided below:
R ( t ) = M d f × w i ( t 1 )
where R(t) indicates the predicted change rate value at time t; M d f = [ m 1 , m 2 , , m h ] denotes the defuzzification matrix; m 1 , m 2 , , m h represent the mid-values of the corresponding subintervals, respectively; and w i ( t 1 ) is the weight vector of the fuzzy logical relationship matrix.
Step 2: Defuzzification of the predicted value.
Translate the predicted stock index rate of change into the final stock index prediction value according to Formula (26):
x ( t ) = x ( t 1 ) × ( 1 + R ( t ) )
where x(t) signifies the forecasted stock index value at time t, while x(t − 1) corresponds to the actual stock index value at the previous moment.
By performing calculations according to Formulas (25) and (26), we obtain the predicted values for the test set of TAIEX. As depicted in Figure 7, the actual values and the predicted values exhibit a fundamentally consistent trend.
In sum, we propose the fuzzy time series forecasting algorithm based on the optimized partition of the universe of discourse, outlined below.
In Algorithm 2, we integrate Algorithm 1 into the prediction framework of fuzzy time series to forecast the time series. By simply inputting the time series X and the number of partition intervals p of the universe of discourse, we can obtain the predicted value x(t).
Algorithm 2: Fuzzy Time Series Forecasting Based on the Novel Partition of the Universe of Discourse
Input: Time series X = { x 1 , x 2 , , x n } and the subintervals p
Output: Predicted value x(t)
1. Define and partition the universe of discourse. Convert the original time series X into a rate of change series R, and define the universe of discourse of R is U = [ U l , U u ] = [ r min l 1 , r max + l 2 ] , then use Algorithm 1 to divide U into p subintervals.
2. Define fuzzy sets and fuzzify historical time series. Based on the subintervals, define a fuzzy set for each subinterval, and fuzzify the historical data to obtain the fuzzy time series.
3. Establish fuzzy logical relationships. Based on the fuzzified data, establish fuzzy logical relationships. Combine the FLRs with the same antecedents into the same group to construct fuzzy logical relationship groups, thereby obtaining fuzzy logic rules and building the fuzzy logical relationship matrix.
4. Perform forecasting and defuzzification. For the fuzzy logical relationship matrix, calculate the fuzzy predicted values according to Formula (25), and perform defuzzification according to Formula (26) to derive the final predicted value x(t).

5. Experiments

To illustrate the effectiveness of the suggested forecasting method, a few experiments are conducted in this section using MATLAB R2022a. The experimental data are selected from the Taiwan Weighted Stock Index (TAIEX) dataset and the Shanghai Composite Index (SHCI) dataset. In Section 5.1, the Root Mean Square Error (RMSE) and Linguistic Accuracy (LA) are used as performance measures to evaluate the forecasting accuracy of the suggested approach. To compare the accuracy of the prediction results of the suggested technique with those of analogous algorithms in the body of existing research, tests are conducted using the TAIEX datasets in Section 5.2. The SHCI datasets are utilized for real-world application research in Section 5.3.

5.1. Evaluation Metrics

5.1.1. Evaluation Metric of Prediction Accuracy

To measure the prediction accuracy of the proposed model, the Root Mean Square Error (RMSE) is chosen as the evaluation metric, which calculates the deviation between the actual values and the predicted values. The calculation formula is as follows:
R M S E = t = 1 n ( x f v ( t ) x a v ( t ) ) 2 n
where n represents the size of the test set, x f v ( t ) is the predicted value at time t, and x a v ( t ) is the actual value at time t. A smaller RMSE value signifies that the predicted values are closer to the actual values, reflecting the higher accuracy of the model’s predictions.

5.1.2. Evaluation Metric of Predicted Linguistic Accuracy

To evaluate the model’s performance in semantic prediction, we use the Linguistic Accuracy (LA) [43] as the assessment metric to measure the difference between the predicted and genuine linguistic values. The following is the formula for computation.
L A ( % ) = t = 1 N g ( t ) N × 100 g ( t ) = { 1 ,   if   L ^ t = L t   0 ,   if   L ^ t L t
where N denotes the total number of the dataset, L t represents the true linguistic value at time t, and L ^ t indicates the predicted linguistic value at time t.

5.2. Experiment A: TAIEX Forecasting

The TAIEX datasets are frequently applied as experimental data for fuzzy time series forecasting models [43,44,45,46,47,48,49,50,51,52,53]. For experimentation and comparison with existing related prediction methods, 14 TAIEX datasets from 1991 to 2004 are selected in this section.
Each year’s TAIEX is treated as a dataset. The first ten months of each dataset are designated as the training set, while the final two months serve as the test set. Table 9 provides the information on training and test sets for TAIEX from 1991 to 2004.
In all experiments, the universe of discourse is divided into seven subintervals, with the parameters for the PSO algorithm listed in Table 10.
The forecasting model developed in this research is used to conduct experiments on the datasets mentioned above. The test set is predicted by first processing the training set to extract the fuzzy logical relationships. Figure 8 illustrates the forecasting results.
Figure 8 shows how the TAIEX’s annual change trends vary from one another. In essence, the trend of the original data and the expected values produced by the suggested model match.
Since most of the existing research methods only use part of the TAIEX datasets from 1991 to 2004 for experiments, for the convenience of comparison with the existing research methods, we divide the TAIEX datasets from 1991 to 2004 into two parts for experiments. Experiment 1 uses the TAIEX datasets from 1991 to 1999, and Experiment 2 uses the TAIEX datasets from 2000 to 2004.
Table 11 presents the RMSE comparison of the proposed model in this paper with other representative prediction models on the TAIEX datasets from 1991 to 1999. Among them, AR(1) and AR(2) [44] are classic time-series prediction models. Chen [45] introduced a traditional fuzzy time-series prediction model. Huarng [46], Yu [47], and Huarng and Yu [48] focused on partitioning the universe of discourse and determining sub-interval lengths to enhance prediction accuracy. Chen and Wang [49] emphasized extracting more effective fuzzy logical relationships from time series. Chen and Chen [50] concentrated on multi-factor fuzzy time-series prediction models using various strategies. Wang [51] developed several fuzzy theory-based time-series prediction models, including models based on automatic clustering and axiomatic fuzzy sets, trend prediction and autoregressive models, and fuzzy data mining.
To facilitate a more intuitive comparison of prediction accuracy among various methods, we computed the average RMSE for different prediction models applied to TAIEX data from 1991 to 1999. As depicted in Figure 9, the proposed model achieves an average RMSE of 80.4. This value is the lowest among the compared methods, signifying superior prediction accuracy.
Table 12 and Figure 10 offer a comparison of the RMSE values between the proposed model and other representative prediction models using the TAIEX datasets from 2000 to 2004.
As shown in Figure 10, the proposed model achieves an average RMSE value of 82.4, which is the lowest among all the prediction methods, indicating the highest prediction accuracy. Compared with Chen’s model [45]; Huarng, Yu, and Hsu’s model [52]; Yu and Huarng’s model [53]; Chen and Chang’s model [54]; Chen and Chen’s model [50]; Wang’s model [51]; and LSTM [9], the proposed model in this chapter has an absolute advantage in prediction accuracy.
The experiments clearly demonstrate that the proposed method in this article achieves higher prediction accuracy than existing models on the TAIEX datasets.

5.3. Experiment B: SHCI Forecasting

The Shanghai Stock Composite Index (SHCI) is a significant indicator of the overall fluctuations in China’s stock market. In this section, a total of 10 datasets of SHCI from 2011 to 2019 are selected as experimental data for practical application. Each year’s SHCI is considered a dataset. The initial ten months of the dataset are used for the training set, whereas the final two months are allocated for the test set. Table 13 lists the information on the training and test sets for SHCI from 2011 to 2019. In all experiments, the universe of discourse is partitioned into seven subintervals, and the parameters for PSO are detailed in Table 14.
Subsequently, we apply the proposed model to Experiment B, with the prediction results illustrated in Figure 11.
Figure 11 clearly shows that the trend of the predicted values generated by the proposed model aligns closely with the direction of the original datasets. We calculate the RMSE of the prediction results of SHCI from 2011 to 2019, as shown in Table 15, and the average RMSE value is 29.69.
According to the prediction results, we transform the numerical predicted values into the corresponding semantic values. Then, we calculate the LA value of the prediction results, as shown in Table 16.
The proposed model achieves an average LA value of 73.07% for SHCI predictions from 2011 to 2019. Since the semantic value corresponds to the degree of increase or decrease in the change rate sequence of the stock index, the obtained average LA indicates that the proposed model can accurately predict the increased or decreased changes in about three-quarters of the dataset.
To discuss the impact of the number of partition intervals p on the forecasting accuracy of our method, we conduct experiments to calculate the RMSE values for different p values, as shown in Table 17. It can be observed that the RMSE value is minimized when p = 7, indicating that the model’s forecasting accuracy is optimal. The RMSE value is maximized when p = 3, and the model’s accuracy begins to decline when p > 7. Therefore, the setting of p should not be too small, nor is it better when it is too large.
In summary, the proposed model in this paper can accurately forecast the rising and falling trends of the future stock market. Investors can adjust their investment strategies according to the forecasting results, thereby increasing returns and reducing risks.

6. Conclusions

In this article, we provide a new fuzzy time series forecasting model that integrates fuzzy C-means clustering, principle of justifiable granularity, and particle swarm optimization. Firstly, the FCM approach is employed to divide the universe of discourse, thereby obtaining the initial subinterval division. Next, triangular fuzzy information granules are created for the subintervals in accordance with the principle of justifiable granularity. Then, an objective function is defined for the entire universe of discourse, and the PSO algorithm is applied to achieve the final result of the optimal subinterval division. Based on the optimal partition results, we implement the framework of the fuzzy time series model for forecasting. Finally, this approach is applied to both the TAIEX and SHCI datasets, and its forecasting performance is evaluated by comparing the results with those of other methods. The experimental results demonstrate that the proposed model in this paper achieves higher forecasting accuracy than other models. It not only exhibits good numerical forecasting performance but also achieves good semantic forecasting performance. This paper focuses on improving the partitioning technique of the universe of discourse to enhance the forecasting accuracy of time series. The forecasting approach is grounded in the established framework of fuzzy time series, so the prediction accuracy is also limited by the processing of fuzzy relationships. In the future, we will consider improving the prediction framework of fuzzy time series to enhance forecasting accuracy. Furthermore, we plan to apply this method to multi-step forecasting in fuzzy time series and utilize it for time series data in other domains to achieve better forecasting outcomes.

Author Contributions

Conceptualization, H.C. and X.G.; methodology, H.C. and X.G.; software, H.C.; validation, H.C. and Q.W.; formal analysis, H.C. and Q.W.; writing—original draft preparation, H.C.; writing—review and editing Q.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant No. 71272161).

Data Availability Statement

The data used to support the findings of this paper are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yao, Y.; Zhang, Z.Y.; Zhao, Y. Stock index forecasting based on multivariate empirical mode decomposition and temporal convolutional networks. Appl. Soft Comput. 2023, 142, 110356. [Google Scholar] [CrossRef]
  2. Palash, W.; Akanda, A.S.; Islam, S. A data-driven global flood forecasting system for medium to large rivers. Sci. Rep. 2024, 14, 8979. [Google Scholar] [CrossRef] [PubMed]
  3. Xie, W.; Liu, C.; Wu, W.-Z. A novel fractional grey system model with non-singular exponential kernel for forecasting enrollments. Expert Syst. Appl. 2023, 219, 119652. [Google Scholar] [CrossRef]
  4. Meng, X.; Zhao, H.; Shu, T.; Zhao, J.H.; Wan, Q.L. Machine learning-based spatial downscaling and bias-correction framework for high-resolution temperature forecasting. Appl. Intell. 2024, 54, 8399–8414. [Google Scholar] [CrossRef]
  5. Box, G.E.P.; Jenkins, G.M. Time Series Analysis: Forecasting and Control; Holden-Day: San Francisco, CA, USA, 1990. [Google Scholar]
  6. Aasim, S.N.; Singh, A. Mohapatra, Data driven day-ahead electrical load forecasting through repeated wavelet transform assisted SVM model. Appl. Soft Comput. 2021, 111, 107730. [Google Scholar] [CrossRef]
  7. Wei, X.K.; Li, Y.H.; Zhang, P.; Lu, J.M. Analysis and applications of time series forecasting model via support vector machines. Syst. Eng. Electron. 2005, 27, 529–532. (In Chinese) [Google Scholar]
  8. Wibawa, A.P.; Utama, A.B.P.; Elmunsyah, H.; Pujianto, U.; Dwiyanto, F.A.; Hernandez, L. Time-series analysis with smoothed Convolutional Neural Network. J. Big Data 2022, 9, 44. [Google Scholar] [CrossRef]
  9. Kirisci, M.; Cagcag Yolcu, O. A new CNN-based model for financial time series: TAIEX and FTSE stocks forecasting. Neural Process. Lett. 2022, 54, 3357–3374. [Google Scholar] [CrossRef]
  10. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. The performance of LSTM and BiLSTM in forecasting time series. In Proceedings of the 2019 IEEE International Conference on Big Data, Los Angeles, CA, USA, 9–12 December 2019. [Google Scholar]
  11. Zhou, F.; Huang, Z.; Zhang, C. Carbon price forecasting based on CEEMDAN and LSTM. Appl. Energy 2022, 311, 118601. [Google Scholar] [CrossRef]
  12. Zadeh, L.A. Fuzzy sets. Inf. Control 1965, 8, 338–353. [Google Scholar] [CrossRef]
  13. Takagi, T.; Sugeno, M. Fuzzy Identification of Systems and its Applications to Modeling and Control. IEEE Trans. Syst. Man Cybern. 1985, 15, 116–132. [Google Scholar] [CrossRef]
  14. Song, Q.; Chissom, B.S. Forecasting enrollments with fuzzy time series—Part I. Fuzzy Sets Syst. 1993, 54, 1–9. [Google Scholar] [CrossRef]
  15. Song, Q.; Chissom, B.S. Forecasting enrollments with fuzzy time series—Part II. Fuzzy Sets Syst. 1994, 62, 1–8. [Google Scholar] [CrossRef]
  16. Chen, S.M.; Wang, J.R. A Novel Fuzzy Time Series Forecasting Model Based on Optimal Partitioning of the Universe of Discourse. Fuzzy Sets Syst. 2000, 114, 159–175. [Google Scholar]
  17. Song, Q.; Chissom, B.S. Fuzzy time series and its models. Fuzzy Sets Syst. 1993, 54, 269–277. [Google Scholar] [CrossRef]
  18. Huarng, K. Effective Lengths of Intervals to Improve Forecasting in Fuzzy Time Series. Fuzzy Sets Syst. 2001, 123, 387–394. [Google Scholar] [CrossRef]
  19. Huarng, K.; Yu, H.K. Ratio-Based Lengths of Intervals to Improve Fuzzy Time Series Forecasting. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2006, 36, 328–340. [Google Scholar] [CrossRef]
  20. Chen, S.M.; Hsu, C.C. A New Method to Forecast Enrollments Using Fuzzy Time Series. Int. J. Appl. Sci. Eng. 2004, 2, 234–244. [Google Scholar]
  21. Chen, M.Y.; Chen, B.T. A Hybrid Fuzzy Time Series Model Based on Granular Computing for Stock Price Forecasting. Inf. Sci. 2015, 294, 227–241. [Google Scholar] [CrossRef]
  22. Wang, L.; Liu, X.; Pedrycz, W. Effective Intervals Determined by Information Granules to Improve Forecasting in Fuzzy Time Series. Expert Syst. Appl. 2013, 40, 5673–5679. [Google Scholar] [CrossRef]
  23. Wang, L.; Liu, X.; Pedrycz, W.; Shao, Y.Y. Determination of Temporal Information Granules to Improve Forecasting in Fuzzy Time Series. Expert Syst. Appl. 2014, 41, 3134–3142. [Google Scholar] [CrossRef]
  24. Yin, Y.; Sheng, Y.; Qin, J. Interval type-2 fuzzy C-means forecasting model for fuzzy time series. Appl. Soft Comput. 2022, 129, 109574. [Google Scholar] [CrossRef]
  25. Pedrycz, W.; Wang, X.M. Designing fuzzy sets with the use of the parametric principle of justifiable granularity. IEEE Trans. Fuzzy Syst. 2016, 24, 489–496. [Google Scholar] [CrossRef]
  26. Moreno, J.E.; Sanchez, M.A.; Mendoza, O.; Rodríguez-Díaz, A.; Castillo, O.; Melin, P.; Castro, J.R. Design of an interval Type-2 fuzzy model with justifiable uncertainty. Inf. Sci. 2020, 513, 206–221. [Google Scholar] [CrossRef]
  27. Zhang, B.; Pedrycz, W.; Wang, X.; Gacek, A. Design of Interval Type-2 Information Granules Based on the Principle of Justifiable Granularity. IEEE Trans. Fuzzy Syst. 2021, 29, 3456–3469. [Google Scholar] [CrossRef]
  28. Castillo, O.; Castro, J.R.; Melin, P. A Methodology for Building of Interval and General Type-2 Fuzzy Systems Based on the Principle of Justifiable Granularity. J. Mult. Valued Log. Soft Comput. 2023, 40, 253–284. [Google Scholar]
  29. Aladag, C.H.; Basaran, M.A.; Egrioglu, E.; Yolcu, U.; Uslu, V.R. Forecasting in High Order Fuzzy Times Series by Using Neural Networks to Define Fuzzy Relations. Expert Syst. Appl. 2009, 36, 4228–4231. [Google Scholar] [CrossRef]
  30. Aladag, C.H.; Yolcu, U.; Egrioglu, E. A High Order Fuzzy Time Series Forecasting Model Based on Adaptive Expectation and Artificial Neural Networks. Math. Comput. Simul. 2010, 81, 875–882. [Google Scholar] [CrossRef]
  31. Egrioglu, E.; Aladag, C.H.; Yolcu, U.; Uslu, V.R.; Basaran, M.A. Finding an Optimal Interval Length in High Order Fuzzy Time Series. Expert Syst. Appl. 2010, 37, 5052–5055. [Google Scholar] [CrossRef]
  32. Kuo, I.H.; Horng, S.J.; Chen, Y.H.; Run, R.S.; Kao, T.W.; Chen, R.J.; Lai, J.L.; Lin, T.L. Forecasting TAIFEX Based on Fuzzy Time Series and Particle Swarm Optimization. Expert Syst. Appl. 2010, 37, 1494–1502. [Google Scholar] [CrossRef]
  33. Huang, Y.L.; Horng, S.J.; He, M.; Fan, P.; Kao, T.W.; Khan, M.K.; Lai, J.L.; Kuo, I.H. A Hybrid Forecasting Model for Enrollments Based on Aggregated Fuzzy Time Series and Particle Swarm Optimization. Expert Syst. Appl. 2011, 38, 8014–8023. [Google Scholar] [CrossRef]
  34. Pant, M.; Kumar, S. Fuzzy time series forecasting based on hesitant fuzzy sets, particle swarm optimization and support vector machine-based hybrid method. Granul. Comput. 2022, 7, 861–879. [Google Scholar] [CrossRef]
  35. Didugu, G.; Gandhudi, M.; Alphonse, P.J.A.; Gangadharan, G.R. VWFTS-PSO: A novel method for time series forecasting using variational weighted fuzzy time series and particle swarm optimization. Int. J. Gen. Syst. 2024, 54, 540–559. [Google Scholar] [CrossRef]
  36. Xian, S.; Zhang, J.; Xiao, Y.; Pang, J. A novel fuzzy time series forecasting method based on the improved artificial fish swarm optimization algorithm. Soft Comput. 2018, 22, 3907–3917. [Google Scholar] [CrossRef]
  37. Bezdek, J.C.; Ehrlich, R.; Full, W. FCM: The fuzzy C-means clustering algorithm. Comput. Geosci. 1984, 10, 191–203. [Google Scholar] [CrossRef]
  38. Pedrycz, W.; Vukovich, G. Abstraction and specialization of information granules. IEEE Trans. Syst. Man Cybern. B Cybern. 2001, 31, 106–111. [Google Scholar] [CrossRef]
  39. Pedrycz, W.; Homenda, W. Building the fundamentals of granular computing: A principle of justifiable granularity. Appl. Soft Comput. 2013, 13, 4209–4218. [Google Scholar] [CrossRef]
  40. Geng, G.; He, Y.; Zhang, J.; Qin, T.; Yang, B. Short-Term Power Load Forecasting Based on PSO-Optimized VMD-TCN-Attention Mechanism. Energies 2023, 16, 4616. [Google Scholar] [CrossRef]
  41. Lu, W. Time Series Analysis and Modeling Method Research Based on Granular Computing; Dalian University of Technology: Dalian, China, 2015. (In Chinese) [Google Scholar]
  42. Shao, G.H. Modeling and Forecasting Based on Multivariate Granular Time Series; Dalian University of Technology: Dalian, China, 2017. (In Chinese) [Google Scholar]
  43. Zhou, W. Modeling Methods for Interval-Valued Time Series Based on Granular Computing; Dalian University of Technology: Dalian, China, 2019. (In Chinese) [Google Scholar]
  44. Sullivan, J.; Woodall, W.H. A Comparison of Fuzzy Forecasting and Markov Modeling. Fuzzy Sets Syst. 1994, 64, 279–293. [Google Scholar] [CrossRef]
  45. Chen, S.M. Forecasting Enrollments Based on Fuzzy Time Series. Fuzzy Sets Syst. 1996, 81, 311–319. [Google Scholar] [CrossRef]
  46. Huarng, K. Heuristic Models of Fuzzy Time Series for Forecasting. Fuzzy Sets Syst. 2001, 123, 369–386. [Google Scholar] [CrossRef]
  47. Yu, T. Weighted Fuzzy Time Series Model for TAIEX Forecasting. Physica A 2005, 349, 609–624. [Google Scholar] [CrossRef]
  48. Huarng, K.; Yu, T.H.K. The Application of Neural Networks to Forecast Fuzzy Time Series. Phys. A Stat. Mech. Appl. 2006, 363, 481–491. [Google Scholar] [CrossRef]
  49. Chen, S.M.; Wang, N.Y. Fuzzy Forecasting Based on Fuzzy-trend Logical Relationship Groups. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2010, 40, 10594–10605. [Google Scholar]
  50. Chen, S.M.; Chen, C.D. TAIEX Forecasting Based on Fuzzy Time Series and Fuzzy Variation Groups. IEEE Trans. Fuzzy Syst. 2011, 19, 1–12. [Google Scholar] [CrossRef]
  51. Wang, X. A hybrid forecasting model based on automatic clustering, axiomatic fuzzy set classification, and autoregressive integrated moving average (ARIMA) for stock market trends. Expert Syst. Appl. 2016, 55, 1–10. [Google Scholar] [CrossRef]
  52. Huarng, K.H.; Yu, T.H.K.; Hsu, Y.W. A Multivariate Heuristic Model for Fuzzy Time-Series Forecasting. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2007, 37, 836–846. [Google Scholar] [CrossRef]
  53. Yu, T.H.K.; Huarng, K.H. A Bivariate Fuzzy Time Series Model to Forecast the TAIEX. Expert Syst. Appl. 2008, 34, 2945–2952. [Google Scholar] [CrossRef]
  54. Chen, S.M.; Chang, Y.C. Multi-Variable Fuzzy Forecasting Based on Fuzzy Clustering and Fuzzy Rule Interpolation Techniques. Inf. Sci. 2010, 180, 4772–4783. [Google Scholar] [CrossRef]
Figure 1. Example of a triangular fuzzy information granule.
Figure 1. Example of a triangular fuzzy information granule.
Symmetry 17 00753 g001
Figure 2. The relationship between α and | Ω α | .
Figure 2. The relationship between α and | Ω α | .
Symmetry 17 00753 g002
Figure 3. The framework of fuzzy time series forecasting based on the novel partition of the universe of discourse.
Figure 3. The framework of fuzzy time series forecasting based on the novel partition of the universe of discourse.
Symmetry 17 00753 g003
Figure 4. TAIEX training set. (a) The original time series; (b) the change rate sequence.
Figure 4. TAIEX training set. (a) The original time series; (b) the change rate sequence.
Symmetry 17 00753 g004
Figure 5. Initial partitioning of the universe of discourse based on FCM.
Figure 5. Initial partitioning of the universe of discourse based on FCM.
Symmetry 17 00753 g005
Figure 6. Optimized partitioning of the universe of discourse based on PJG and PSO.
Figure 6. Optimized partitioning of the universe of discourse based on PJG and PSO.
Symmetry 17 00753 g006
Figure 7. Comparison of actual and predicted values of TAIEX.
Figure 7. Comparison of actual and predicted values of TAIEX.
Symmetry 17 00753 g007
Figure 8. Prediction results of TAIEX from 1991 to 2004.
Figure 8. Prediction results of TAIEX from 1991 to 2004.
Symmetry 17 00753 g008aSymmetry 17 00753 g008b
Figure 9. Average RMSE comparison of the prediction models for TAIEX from 1991 to 1999.
Figure 9. Average RMSE comparison of the prediction models for TAIEX from 1991 to 1999.
Symmetry 17 00753 g009
Figure 10. Average RMSE comparison of the prediction models for TAIEX from 2000 to 2004.
Figure 10. Average RMSE comparison of the prediction models for TAIEX from 2000 to 2004.
Symmetry 17 00753 g010
Figure 11. Prediction results of SHCI from 2011 to 2019.
Figure 11. Prediction results of SHCI from 2011 to 2019.
Symmetry 17 00753 g011aSymmetry 17 00753 g011b
Table 1. Ωα of different α.
Table 1. Ωα of different α.
αΩααΩα
0[−8,5]0.6[−2.3,2.1]
0.1[−8,5]0.7[−2.3,2.1]
0.2[−3.4,5]0.8[−2.3,1.2]
0.3[−3.4,5]0.9[−2.3,1.2]
0.4[−2.3,2.1]1[−1.6,1.2]
0.5[−2.3,2.1]
Table 2. The original time series and the change rate sequence of the TAIEX training set.
Table 2. The original time series and the change rate sequence of the TAIEX training set.
TimeThe Original DataThe Rate of Change (%)TimeThe Original DataThe Rate of Change (%)
03/01/19914258-02/07/19915613−2.69
04/01/199143672.5603/07/19915604−0.16
05/01/199144562.0404/07/199156070.05
07/01/19914191−5.9505/07/19915591−0.29
08/01/19913975−5.1506/07/19915412−3.20
25/06/19915872−3.5823/10/199141351.15
26/06/199160232.5724/10/199142532.85
27/06/19915931−1.5328/10/199143813.01
28/06/19915900−0.5229/10/19914364−0.39
29/06/19915768−2.2430/10/199143890.57
Table 3. Subintervals, fuzzy sets, and semantic values.
Table 3. Subintervals, fuzzy sets, and semantic values.
SubintervalFuzzy SetSemantic Value
[−7.00, −3.80)A1Sharp decrease
[−3.80, −2.24)A2Decrease
[−2.24, −0.57)A3Slight decrease
[−0.57, 1.17)A4No change
[1.17, 2.60)A5Slight increase
[2.60, 4.40)A6Increase
[4.40, 7.00]A7Sharp increase
Table 4. Fuzzy values of some data of the change rate sequence.
Table 4. Fuzzy values of some data of the change rate sequence.
TimeThe Rate of Change (%)Fuzzy ValueTimeThe Rate of Change (%)Fuzzy Value
03/01/1991--02/07/1991−2.69A2
04/01/19912.56A503/07/1991−0.16A4
05/01/19912.04A504/07/19910.05A4
07/01/1991−5.95A105/07/1991−0.29A4
08/01/1991−5.15A106/07/1991−3.20A2
25/06/1991−3.58A223/10/19911.15A4
26/06/19912.57A524/10/19912.85A6
27/06/1991−1.53A328/10/19913.01A6
28/06/1991−0.52A429/10/1991−0.39A4
29/06/1991−2.24A330/10/19910.57A4
Table 5. The first-order FLRs of some data of the rate change sequence.
Table 5. The first-order FLRs of some data of the rate change sequence.
TimeFuzzy ValueFirst-Order FLRTimeFuzzy ValueFirst-Order FLR
03/01/1991--02/07/1991A2A3A2
04/01/1991A5-03/07/1991A4A2A4
05/01/1991A5A5A504/07/1991A4A4A4
07/01/1991A1A5A105/07/1991A4A4A4
08/01/1991A1A1A106/07/1991A2A4A2
25/06/1991A2A3A223/10/1991A4A1A4
26/06/1991A5A2A524/10/1991A6A4A6
27/06/1991A3A5A328/10/1991A6A6A6
28/06/1991A4A3A429/10/1991A4A6A4
29/06/1991A3A4A330/10/1991A4A4A4
Table 6. Partial first-order FLRGs.
Table 6. Partial first-order FLRGs.
FLRGs
A 1 A 1 , A 4 A 4 A 2 , A 3 , A 4 , A 6 A 1 A 1 , A 4
A 2 A 4 , A 5 A 5 A 1 , A 3 , A 5 A 2 A 4 , A 5
A 3 A 2 , A 4 A 6 A 4 , A 6 A 3 A 2 , A 4
Table 7. Fuzzy logical relationship matrix of occurrence frequency.
Table 7. Fuzzy logical relationship matrix of occurrence frequency.
Ft−1Ft
A1A2A3A4A5A6A7
A13136311
A20049300
A3541116872
A45717271263
A5427131331
A60195221
A71122112
Table 8. Fuzzy logical relationship matrix of frequency.
Table 8. Fuzzy logical relationship matrix of frequency.
Ft−1Ft
A1A2A3A4A5A6A7
A10.170.050.170.340.170.050.05
A2000.250.560.1900
A30.090.080.210.300.150.130.04
A40.060.090.220.350.160.080.04
A50.090.050.160.300.300.080.02
A600.050.450.250.100.100.05
A70.100.100.200.200.100.100.20
Table 9. Information on training and test sets for TAIEX from 1991 to 2004.
Table 9. Information on training and test sets for TAIEX from 1991 to 2004.
YearSizeTraining SetSize of Training SetTest SetSize of Test Set
19912861/3~10/3023911/1~12/2847
19922841/4~10/3023811/2~12/2946
19932911/5~10/3024311/2~12/3148
19942861/5~10/2923611/1~12/3150
19952861/5~10/3023711/1~12/3049
19962881/4~10/3023611/1~12/3152
19972861/4~10/3023811/3~12/3148
19982711/3~10/3122611/2~12/3145
19992661/5~10/3022111/1~12/2845
20002711/4~10/3122411/1~12/3047
20012441/2~10/3120111/1~12/3143
20022481/2~10/3120511/1~12/3143
20032481/2~10/3120611/3~12/3142
20042501/2~10/2920511/1~12/3145
Table 10. The parameters for the PSO algorithm in Experiment A.
Table 10. The parameters for the PSO algorithm in Experiment A.
ParametersValue
Particle swarm size150
Number of iterations1000
Inertia weight coefficient0.8
Cognitive coefficient1.5
Social coefficient1.5
Table 11. RMSE comparison of the prediction models for TAIEX from 1991 to 1999.
Table 11. RMSE comparison of the prediction models for TAIEX from 1991 to 1999.
Models199119921993199419951996199719981999
AR(1) [44]87.195.8103.6111.790.386153.3149.2121.9
AR(2) [44] 59.276.9110.9111.169.262.9175.3137130.9
Chen [45]80601101127954148167149
Huarng [46]
based on average-based length intervals79.459.9105.2132.478.652.1148.8159.3159.1
based on distribution-based length intervals80.260.3110111.778.654.2148.0167.3148.7
Yu [47]
based on average-based length intervals61671051357054133151145
based on distribution-based length intervals67561051147052152154142
Huang and Yu [48]54.761.1117.988.764.152.1135.9136.2131.9
Chen and Wang [49]42.943.5103.489.852.252.8140.8116.9104.9
Chen and Chen [50]
using Dow Jones72.943.4103.278.666.759.8139.7124.4115.5
using NASDAQ66.149.6104.875.767.060.9140.9144.1119.3
using Dow Jones and NASDAQ74.943.8101.478.168.161.3139.3132.9116.6
Wang [51]
based on automatic clustering and axiomatic fuzzy set43.641.4102.489.055.049.4139.0118.2100.9
based on trend prediction and the autoregressive model42.544.0101.093.152.950.5145.1115.1101.3
based on fuzzy data mining43.543.3102.287.657.150.6139.5120.4102.9
The proposed method43.542.398.280.153.452.0132.5120.3101.2
Table 12. RMSE comparison of the prediction models for TAIEX from 2000 to 2004.
Table 12. RMSE comparison of the prediction models for TAIEX from 2000 to 2004.
Models20002001200220032004
Chen [45]176.3147.8101.274.584.3
Huarng, Yu and Hsu [52]
using Dow Jones165.8138.2593.7372.9573.49
using NASDAQ158.7136.4995.1565.5173.57
using Dow Jones and NASDAQ157.64131.9893.4865.5173.49
Yu and Huarng [53]
bivariate conventional regression model154120775485
bivariate neural network model274131695261
Chen and Chang [54]
using Dow Jones148.8113.779.864.0882.32
using NASDAQ131.1115.173.166.460.5
using Dow Jones and NASDAQ 130.1113.372.360.368.1
Chen and Chen [50]
using Dow Jones127.5122.074.766.058.9
using NASDAQ129.9123.171.065.161.9
using Dow Jones and NASDAQ 123.6123.971.958.157.7
Wang [51]
based on automatic clustering and axiomatic fuzzy set classification 138.0113.865.056.555.3
based on trend prediction and the autoregressive model132.0111.565.352.454.2
based on fuzzy data mining131.6113.668.559.356.7
LSTM [9]136101899270
The proposed method121.2112.765.557.655.2
Table 13. Information on training and test sets for TAIEX SHCI from 2010 to 2019.
Table 13. Information on training and test sets for TAIEX SHCI from 2010 to 2019.
YearSizeTraining SetSize of Training SetTest SetSize of Test Set
20112441/4~10/3120011/1~12/3044
20122431/4~10/3120011/1~12/3143
20132381/4~10/3119511/1~12/3143
20142451/2~10/3120211/3~12/3143
20152441/5~10/3020011/2~12/3144
20162441/4~10/3120011/1~12/3044
20172441/3~10/3120111/1~12/2943
20182431/2~10/3120111/1~12/2842
20192441/2~10/3120111/1~12/3143
Table 14. The parameters for the PSO algorithm in Experiment B.
Table 14. The parameters for the PSO algorithm in Experiment B.
ParametersValue
Particle swarm size150
Number of iterations1000
Inertia weight coefficient0.8
Cognitive coefficient1.5
Social coefficient1.5
Table 15. RMSE of the proposed model for SHCI from 2010 to 2019.
Table 15. RMSE of the proposed model for SHCI from 2010 to 2019.
Year201120122013201420152016201720182019
RMSE27.4024.1319.6952.0554.0022.0621.0526.7520.12
Table 16. LA of the proposed model for SHCI from 2010 to 2019.
Table 16. LA of the proposed model for SHCI from 2010 to 2019.
Year201120122013201420152016201720182019
LA(%)70.4558.1472.0960.4775.0075.0072.0997.6276.74
Table 17. RMSE of the proposed model with different number of partition intervals p for SHCI 2019.
Table 17. RMSE of the proposed model with different number of partition intervals p for SHCI 2019.
p345678910
RMSE37.2920.4320.2220.9620.1221.6921.6123.84
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, H.; Gao, X.; Wu, Q. An Enhanced Fuzzy Time Series Forecasting Model Integrating Fuzzy C-Means Clustering, the Principle of Justifiable Granularity, and Particle Swarm Optimization. Symmetry 2025, 17, 753. https://doi.org/10.3390/sym17050753

AMA Style

Chen H, Gao X, Wu Q. An Enhanced Fuzzy Time Series Forecasting Model Integrating Fuzzy C-Means Clustering, the Principle of Justifiable Granularity, and Particle Swarm Optimization. Symmetry. 2025; 17(5):753. https://doi.org/10.3390/sym17050753

Chicago/Turabian Style

Chen, Hailan, Xuedong Gao, and Qi Wu. 2025. "An Enhanced Fuzzy Time Series Forecasting Model Integrating Fuzzy C-Means Clustering, the Principle of Justifiable Granularity, and Particle Swarm Optimization" Symmetry 17, no. 5: 753. https://doi.org/10.3390/sym17050753

APA Style

Chen, H., Gao, X., & Wu, Q. (2025). An Enhanced Fuzzy Time Series Forecasting Model Integrating Fuzzy C-Means Clustering, the Principle of Justifiable Granularity, and Particle Swarm Optimization. Symmetry, 17(5), 753. https://doi.org/10.3390/sym17050753

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop