Next Article in Journal
A 280 GHz 30 GHz Bandwidth Cascaded Amplifier Using Flexible Interstage Matching Strategy in 130 nm SiGe Technology
Previous Article in Journal
Experimental Study, Simulation and Analysis of the Fracture Failure of the Drum Shaft of a Casting Bridge Crane
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Survey on Particle Swarm Optimization for Association Rule Mining

1
School of Computer and Information Engineering, Jiangxi Agriculture University, Nanchang 330045, China
2
Department of Computer, Mathematical and Physical Sciences, Sul Ross State University, Alpine, TX 79830, USA
3
College of Mathematics and Computer Science, Fuzhou University, Fuzhou 350116, China
4
Center for AI Research (CAIR), University of Agder (UiA), 4064 Grimstad, Norway
*
Authors to whom correspondence should be addressed.
Electronics 2022, 11(19), 3044; https://doi.org/10.3390/electronics11193044
Submission received: 23 August 2022 / Revised: 18 September 2022 / Accepted: 21 September 2022 / Published: 24 September 2022
(This article belongs to the Section Power Electronics)

Abstract

:
Association rule mining (ARM) is one of the core techniques of data mining to discover potentially valuable association relationships from mixed datasets. In the current research, various heuristic algorithms have been introduced into ARM to address the high computation time of traditional ARM. Although a more detailed review of the heuristic algorithms based on ARM is available, this paper differs from the existing reviews in that we expected it to provide a more comprehensive and multi-faceted survey of emerging research, which could provide a reference for researchers in the field to help them understand the state-of-the-art PSO-based ARM algorithms. In this paper, we review the existing research results. Heuristic algorithms for ARM were divided into three main groups, including biologically inspired, physically inspired, and other algorithms. Additionally, different types of ARM and their evaluation metrics are described in this paper, and the current status of the improvement in PSO algorithms is discussed in stages, including swarm initialization, algorithm parameter optimization, optimal particle update, and velocity and position updates. Furthermore, we discuss the applications of PSO-based ARM algorithms and propose further research directions by exploring the existing problems.

1. Introduction

Association rule mining (ARM), as one of the core techniques of data mining (DM), can discover potentially valuable association relationships between itemsets from complex and fuzzy mixed data. The Apriori algorithm is one of the classical algorithms for ARM, which was first proposed by Agrawal et al. [1] and has gained the attention and research of many scholars in the past two decades. Akbar et al. [2] reviewed the research on ARM algorithms based on evolutionary computation published before 2019 and classified the evolutionary ARM algorithms into four major categories: evolutionary-based, swarm intelligence-based, physics-inspired, and hybrid approaches. Minakshi et al. [3] systematically assessed numerical association rule mining (NARM) algorithms to help scholars select a suitable algorithm when performing NARM mining. Djenouri et al. [4] studied the application of various metaheuristics algorithms in frequent itemset mining (FIM) and high-utility itemset mining (HUIM), including genetic algorithms (GA), ant colony optimization (ACO) algorithms, artificial bee colony (ABC), and particle swarm optimization (PSO). Arani et al. [5] also used a metaheuristic-based clustering mechanism when analyzing cloud workloads. In addition, Gan et al. [6] combined the design of convolutional neural networks (CNN) to address the shortcomings of traditional mining models that cannot extract temporal data effectively by constructing a new temporal association rule mining model. Similar papers include [7,8]. Badhon et al. [9] developed a systematic and structured analysis of ARM algorithms in terms of different aspects, including chromosome representation, genetic operators, and initial swarms. In addition, Kheirollahi et al. [10] combined the clustering algorithm with the PSO algorithm, which can be applied to recognize effective, good clusters and extract strong rules among reservoir data. Datta et al. [11] introduced the concept of flexible dissociation to propose a significant ARM framework with high associativity. In response to the rapid growth of data, Sanjay et al. [12] proposed a distributed ARM algorithm called adaptive-miner, which can dynamically change the mining algorithm according to the characteristics of the dataset. Among them, the PSO algorithm has already received the attention of many scholars at home and abroad since it was proposed because of its few adjustable parameters, high efficiency, easy implementation, and fast convergence speed. The current research shows that ARM using the PSO algorithms has higher operational efficiency compared with other data mining algorithms and can generate more accurate and valuable rules [13,14,15,16,17,18,19,20,21].
However, recent papers reviewed only a small fraction of the PSO-based ARM algorithms; however, there have been many papers published in recent years. Therefore, we aimed to provide a more comprehensive and multi-faceted survey of emerging research. We expect this survey can provide a reference for researchers in the field to help them understand the state-of-the-art PSO-based ARM algorithms. We downloaded all the articles containing PSO or ARM keywords in recent years, removed the less cited and less relevant ones, and selected the remaining 117 papers for use in our survey.
To analyze these papers more systematically and comprehensively, we divided the ARM algorithms according to the different pattern types. They were fuzzy association rule mining, rare association rule mining, numerical association rule mining, quantitative association rule mining, binary association rule mining, high-utility association rule mining, and classification association rule mining, and they are shown in blue in the inner circle of Figure 1. In addition, we compared them from three perspectives: biologically inspired, physically inspired, and other algorithms, and listed the key information of each algorithm under the different perspectives, as shown in the outer circle of Figure 1. The specific description and summary of the characteristics of each algorithm can be found in Section 3 of the article.
The remaining paper is organized as follows: In Section 2, the concept of ARM is described in detail, and it is divided into three categories: classification association rule mining (CARM) [22], fuzzy association rule mining (FARM) [23,24] and numerical association rule mining (NARM) [15,25,26,27]; their algorithm evaluation metrics are described in three sections. In Section 3, different algorithms in ARM are compared and categorized into three groups: bio-inspired, physics-inspired, and other algorithms, and in each group, a paragraph describing the algorithms and a summary of their features are provided. The process of the PSO-based ARM algorithm is described in detail in Section 4, including the binary conversion of data and the encoding process. Section 5 discusses the current status of research on the ARM algorithms based on PSO algorithms from four perspectives: swarm initialization, algorithm parameter optimization, optimal particle update, and velocity and position update. The applications of the PSO algorithm in ARM are summarized in Section 6. Finally, the paper concludes with concluding statements and future aspects in Section 7.

2. Association Rule Mining

2.1. Definition of ARM

The definition of association rules was first introduced by Agrawal in 1993: I = { i 1 , i 2 , , i n } is called the itemsets, where i j { 0 , 1 } is called an item, D = { d 1 , d 2 , , d m } is called a database, and d k is called a transaction. A transaction is an itemset, i.e., a transaction is a subset d k I of a transaction, and each transaction has a unique ID. Each rule consists of two different itemsets, X and Y , where X is called the premise, and Y is called the conclusion. Rules are represented as X Y , and X , Y I .

2.2. Mode Types for ARM

2.2.1. CARM

CARM is a widely used method in the real world. Compared with traditional ARM, CARM has the characteristics of high classification accuracy and adaptability, etc. Its mining is mainly divided into the following processes: in the preprocessing stage, if the data are continuous values, they need to be discretized; then, class association rules that satisfy the minimum support and confidence are mined in the database, and the generated association rules are pruned. Finally, the classifier is used to predict the data objects of unknown classes. Currently, this model is widely used in several fields, such as medicine and biology.

2.2.2. FARM

For traditional ARM, its itemsets are often discretized by dividing the intervals into continuous values [28], but the method suffers from the drawback of dividing the intervals too hard. Therefore, Chan et al. [23] proposed the concept of fuzzy association rules to divide the interval into multiple fuzzy sets by softening the division boundary with fuzzy sets and obtaining the probability of an element on a fuzzy set based on the fuzzy affiliation function to provide a smooth transition between the members and nonmembers in the itemset.
Let the itemset set I = { i 1 , i 2 , , i n } , database D = { t 1 , t 2 , , t m } , and each element i j in the itemset correspond to a fuzzy set F i k = { f i k 1 , f i k 2 , , f i k l } . The fuzzy association rule for mining is that if T is A , then F is B . Where, T = { t 1 , t 2 , , t p } , F = { f 1 , f 2 , , f p } are the itemsets, A = { a 1 , a 2 , , a p } , B = { b 1 , b 2 , , b p } are the fuzzy sets corresponding to T and F, respectively. The rule premise X = T is A , and the conclusion Y = F is B . This is expressed as: when X is satisfied, it can be concluded that Y is also satisfied. For this purpose, we use the significance factor and the certainty factor to determine whether the rule is interesting or not.
Significance Factor: To generate a fuzzy association rule, it is first necessary to obtain the set of k items with a significance factor above a user-specified threshold, denoted as S < T , A > , which is calculated as follows:
S < T , A > = d k D t i T { α a j ( d k [ t i ] ) } T o t a l ( D )
α a j ( d k [ t j ] ) = { m a j A ( d k [ t j ] ) ,   i f   m a j > w 0 ,   o t h e r w i s e
where T o t a l ( D ) indicates the total number of transactions in database D, α a j ( d k [ t j ] ) indicates the membership grade of t j in the k th records.
Certainty Factor: After the first step of calculation to obtain a large set of K terms with significance factors above a given value < Z , C > , a deterministic factor is introduced to measure whether the rule is interesting or not, which is calculated as follows:
C < < T , A > , < F , B > > = d k D z i Z { α c i ( d k [ z i ] ) } d k D t j T { α a j ( d k [ t j ] ) }
α c i ( d k [ z i ] ) = { m c i c ( d k [ z i ] ) ,   i f   m c i > w 0 ,   o t h e r w i s e
Z = T F , C = A B

2.2.3. NARM

In the real world, data are often represented in the form of numerical values, and the process of finding the association rules in numerical data is called NARM. In solving the NARM problem, most of the existing studies are divided into three methods: optimization, discretization, and distribution. Among them, the optimization methods can be divided into several sub-methods, such as swarm intelligence-based and evolution-based; the discretization methods are for discrete-continuous values, which can be further divided into supervised and unsupervised, static and dynamic, and other sub-methods; and the distribution methods include the mean, median, and variance. The division methods are shown in Figure 2.

2.3. Evaluation Metrics for ARM

2.3.1. Basic Evaluation Indicators

To select the valuable rules from the possible frequent itemsets, some metrics are needed to filter and sift them. In most studies, the strength of a rule is described by support and confidence.
Support: Support indicates the frequency of the itemsets in the database. The support of the association rule X Y is the ratio of the number of transactions in the database that also contain the itemsets X and Y to the total number of transactions.
S u p p o r t ( X Y ) = P ( X Y ) = X Y a l l
Confidence: The confidence level is used to measure the confidence level of a rule. The confidence level of the association rule X Y is the ratio of the number of transactions in the database that contain both X and Y and the total number of transactions that contain X .
C o n f i d e n c e ( X Y ) = P ( X | Y ) = P ( X Y ) P ( X )
When S u p p o r t ( X Y ) and C o n f i d e n c e ( X Y ) are both greater than the minimum support threshold and the minimum confidence threshold, if L i f t ( X Y ) > 1 , then X Y is a strong association rule, and vice versa.

2.3.2. Quantitative Evaluation Indicators

Most researchers use support and confidence to evaluate rules. The inherent drawback of this is that rare data items with less than the minimum support cannot be studied, and the support–confidence-based evaluation system generates a huge number of invalid rules; thus, other evaluation metrics need to be introduced to further mine the rules of interest to users [29].
Correlativity: Correlation reflects the degree of interdependence between X and Y. If C o r r e l a t i v i t y ( X Y ) < 1 , then it means that X Y is an invalid strong correlation rule, and if C o r r e l a t i v i t y ( X Y ) = 1 , then X and Y are independent of each other.
C o r r e l a t i v i t y ( X Y ) = P ( X | Y ) P ( Y ) = C o n f i d e n c e ( X Y ) S u p p o r t ( Y )
Effect: The degree of influence reflects the degree of correlation between X and Y . It can filter invalid association rules and generate meaningful positive and negative association rules. | E f f e c t ( X Y ) | is larger, indicating a high degree of X Y correlation.
E f f e c t ( X Y ) = C o n f i d e n c e ( X Y ) S u p p o r t ( Y ) M a x { C o n f i d e n c e ( X Y ) , S u p p o r t ( Y ) }
Time: The timeliness reflects the value of time between X and Y and can help decision makers remove redundant or already outdated information. The temporal validity of the rules decreases exponentially with increasing time intervals. The temporal validity of support and confidence is calculated as shown in Equations (10) and (11), respectively.
T i m e _ S u p p o r t ( X Y ) = i = 1 n S u p p o r t ( X i ) e α | t i t 0 |
T i m e _ C o n f i d e n c e ( X Y ) = i = 1 n C o n f i d e n c e ( X i ) e α | t i t 0 |
where X i denotes the i th transaction in the X events and t i denotes the time when the i th transaction in the X events occurred.

2.3.3. Qualitative Evaluation Indicators

Utility: Utility reflects the potential usefulness of some attributes in the rule and their importance, and U t i l i t y [ 0 , 1 ] . When U t i l i t y = 1 , it means the attribute is the most important, and vice versa.
Concision: Simplicity is used to measure the comprehensibility of the association rule and is related to the arithmetic average of the number of items and the abstraction level of the antecedent items of the rule, and the weighted arithmetic average of the precedent items, whose weights depend on the importance of the antecedent and precedent items. The higher the number of items in a rule, the lower its comprehensibility, and the higher the level of abstraction contained in a rule, the lower its comprehensibility. The greater the simplicity of a rule, the more interesting it is to the user.

3. Comparison of Different ARM Algorithms

Due to the features of association rules in high-dimensional spaces, many scholars obtain better solutions through iterative heuristic processes. In this paper, we divided the ARM methods based on the heuristic algorithms into three main groups: bio-inspired, physics-inspired, and other algorithms, and provided a paragraph describing the algorithms and a summary of their features in each group.

3.1. Bio-Inspiration

3.1.1. Evolution-Based Algorithms

Some evolution-based algorithms are shown in Table 1.
Evolutionary algorithms are based on biological evolutionary mechanisms and are mostly inspired by the biological evolution of nature, including natural selection and genetics, which have higher robustness and wide adaptability, and efficiently solve complex problems that are difficult to solve in traditional optimization algorithms.

3.1.2. Swarm Intelligence-Based Algorithm

Some algorithms based on swarm intelligence are shown in Table 2.
Inspired by the behavior of insect clusters, herds of animals, flocks of birds, etc., the swarm intelligence algorithm can change the search direction through collaboration and information sharing among individuals of the swarm; it has a faster convergence rate and can obtain a better solution than evolutionary algorithms; the poor performance of individuals does not affect the solution of the problem by the whole swarm.

3.2. Physics-Inspired

A physics-inspired algorithm is a swarm optimization algorithm that models certain physical objective laws. Individuals of the swarm collaborate and exchange information with each other in the search space through rules inspired by some physical principles. Some physically inspired algorithms are shown in Table 3.

3.3. Other Algorithms

Some of the algorithms are shown in Table 4.
Other heuristic algorithms are inspired by natural phenomena, mathematical principles, man-made activities, etc. They include the fireworks algorithm, brainstorming algorithm, backtracking search algorithm, and sine and cosine algorithms, etc.

4. Particle Swarm Optimization Algorithms

4.1. Standard PSO Algorithms

The particle swarm optimization (PSO) algorithm [65] is an intelligent optimization algorithm studied by Kenney and Eberhart in 1995 for the swarm motion behavior of bird and fish flocks, and the basic idea is to seek the global optimal solution through mutual collaboration and information sharing among individuals. In this model, an individual is regarded as a particle and a flock of birds as a particle swarm. Suppose there are m particles in a D-dimensional target search space, where the position of the i th particle ( i = 1 , 2 , 3 , , m ) is denoted as X i = ( x i 1 , x i 2 , , x i D ) , i.e., the position of the i th particle in the D-dimensional target search space is X i D . Substituting X i into the objective function calculates its fitness value and measures its merit. The best position experienced by each particle is denoted as P i = ( p i 1 , p i 2 , , p i D ) , and the best position experienced by the whole particle swarm is denoted as P g = ( p g 1 , p g 2 , , p g D ) . The velocity of particle i is denoted as V i = ( v i 1 , v i 2 , , v i D ) . Particle i updates itself in each iteration by Equations (12) and (13) for P i , P g .
v i d + 1 = w v i d + c 1 r 1 ( p i x i d ) + c 2 r 2 ( p g x i d )
x i d + 1 = x i d + α v i d
where i = 1 , 2 , 3 , , m , d = 1 , 2 , 3 , D , w is the inertia weight, v i d is the velocity of the i th particle at the d th iteration, c 1 , c 2 is a nonnegative constant, r 1 , r 2 is an internally uniformly distributed random number, and α is a controlling factor that controls the velocity weight. The pseudocode is shown in Algorithm 1.
Algorithm 1 Procedure of standard PSO algorithm
1:for each particle i = 1 , 2 , , N  do
2: (a) Initialize the particle’s position and velocity.
3: (b) Evaluate the particle’s fitness value;
4: (c) Update the particle’s Pbest;
5: (d) Update the swarm’s Gbest;
6:while termination criteria is not met do
7:for each particle i = 1 , 2 , , N  do
8:  (a) Update particle’s velocity using Equation (12);
9:  (b) Update particle’s position using Equation (13);
10:  (c) Evaluate particle’s fitness value;
11:  if (x[i]) < f(Pbest[i]) then
12:   Update the best known position of particle i: Pbest[i]=x[i];
13:   if (Pbest[i]) < f(Gbest) then
14:    Update the swarm’s best known position: Gbest=Pbest[i];
15:  (d) t = t + 1;
16:return Gbest
Among them, we illustrate some of the more important parameters.
Maximum velocity V max : In each dimension, the velocity of the particle v i ( V max , V max ) reduces the possibility of the particle leaving the search space. Where V max = k x max ( 0.1 < k < 1 ) , the purpose is to limit the maximum distance the particle can move.
Inertia weight w: The larger the w, the greater the global searchability and the easier it is for the particle to escape the local extremes; however, it also reduces search efficiency and makes converging more difficult, and vice versa. w usually varies linearly ( 0 < w < 1 ), and the variation equation is shown in Equation (14).
w i = w max w max w min i t e r max i t e r
where i t e r max is the maximum number of iterations and i t e r is the number of current iterations.

4.2. Improved PSO Based on Algorithm Design

4.2.1. Swarm Initialization

The main function of swarm initialization is to generate several particles randomly, which has a significant impact on the diversity and convergence of the algorithm. Wang et al. [66] used the roulette wheel selection method and non-expectation coding vector construction strategy to initialize the swarm and select the initial optimized value of the next-generation swarm by a certain probability. Equation (15) represents the probability of itemset I being selected, Equation (16) represents the hetero set of the coding vector and Equations (17)–(19) represent the velocity update in the generated coding vector:
S i = f i t n e s s i k = 1 | S H U I | f i t n e s s k
B i t D i f f ( V e c i , V e c j ) = { n u m | 1 n u m l e n , b n u m ( V e c i ) b n u m ( V e c j ) = 1 }
v i = v i 1 + v i 2 + v i 3
v i 2 = [ | B i t D i f f ( p i , p b e s t i ) | r 1 ]
v i 3 = [ | B i t D i f f ( p i , p b e s t i ) | r 2 ]
where |SHUI| denotes the number of elements in the set SHUI, F i t n e s s i denotes the adaptation value of the i th itemset, b n u m ( V e c i ) denotes the value of V e c i in position num, and ⊕ denotes the iso-or operation. v i 1 defaults to 1.
Kalka et al. [67] introduced the chemical reaction optimization (CRO) algorithm to quickly generate the optimal set of solutions as the initial solution of the PSO algorithm and use the selection operator to improve the quality of the optimal solution and enhance the overall performance of the algorithm. Equation (20) is used to determine whether the generated latest solution is accepted after colliding through two chemical reactions:
f ( R j ) = φ 1 M S R + φ 2 E C R j + φ 3 ε
where φ 1 + φ 2 + φ 3 = 1 .
Hematpour et al. [68] introduced a nonlinear dynamical chaotic system to initialize the swarm positions. Equation (21) represents the first class of Chebyshev polynomials, α ( 0 , 2 ) , and Equation (22) represents the second class of Chebyshev polynomials, which can be simplified to a logistic mapping, α ( 1 2 , ) , and α ( 1 3 , 3 ) in Equation (23):
Φ 2 ( 1 ) = α 2 ( 2 x 1 ) 2 4 x ( 1 x ) + α 2 ( 2 x 1 ) 2
Φ 2 ( 2 ) = 4 α 2 x ( 1 x ) 1 + 4 ( α 2 1 ) x ( 1 x )
Φ 2 ( 1 , 2 ) = α 2 x ( 4 x 3 ) 2 α 2 x ( 4 x 3 ) 2 + ( 1 x ) ( 4 x 1 ) 2
Li et al. [69] used an S-type transfer function to initialize the swarm position by converting the inertia weights based on feature weighting into probability values. Let z be the variable, Equation (24) represents the inertia weight normalization process, Equation (25) represents the conversion of z to z by a standard logistic function, Equation (26) represents the conversion of z to reduce the upper limit of z , and p represents the initial swarm position:
z = 2 σ ( w w min ) w max w min σ
z = 1 1 + e z
p = λ z
where p ( 0 , λ ) , λ ( 0.5 , 1 ) , z [ σ , σ ] , σ R + .
Wang et al. [70] used the principal component analysis (PCA) technique to map high-dimensional data to low-dimensional data, retaining variance information and ranking it by priority, and initializing the mixture by selecting significant features. Equation (27) represents the normalization process of the PCA mean and variance, Equation (28) represents the PCA mapping the original data into new coordinates to maximize the variance, and Equation (29) represents the swarm position obtained after eigendecomposition of the covariance matrix:
{ u = 1 m i = 1 m x ( i ) σ j 2 = 1 m i = 1 m x j ( i )
1 m i = 1 m ( x ( i ) T u ) 2 = 1 m i = 1 m ( u T x ( i ) x ( i ) T ) 2 = u T ( 1 m i = 1 m x ( i ) x ( i ) T ) u
y ( i ) = [ u 1 T x ( i ) u 2 T x ( i ) u k T x ( i ) ]
where x ( i ) denotes the position of the i th particle in the swarm, j denotes the j th dimension of x ( i ) , u denotes the mean of the i th particle, σ j denotes the standard deviation of the j th dimension, and y ( i ) denotes the new position generated by the i th particle.
Bangyal et al. [71] introduced the Well technique with the help of low deviation sequences to make it applicable to high-dimensional optimization problems. Sobia et al. [72] reviewed a variety of swarm initialization strategies for the PSO algorithms and analyzed and compared each strategy in detail by the QRS method, including the performance of SO-PSO, H-PSO, WE-PSO, etc. in 15 benchmark tests. It was found that KE-PSO performs better in high-dimensional problems.

4.2.2. Algorithm Parameter Optimization

In the PSO algorithms, algorithm parameters play a great role in balancing the global search ability and local searchability, which can effectively improve swarm diversity and speed up algorithm convergence. Common parameter optimization includes inertia weights w, learning factors c1 and c2, the introduction of other factors, etc. Tian et al. [73] adaptively controlled the inertia weights through the fitness function, and the interaction of the two helps to control the swarm diversity. Equation (30) is used to assess swarm diversity, and Equation (31) is used to calculate inertia weights:
f D i v ( t ) = log 10 ( 1 m j = t m + 1 t ( f a v g ( j ) f A s v g ( t ) ) 2 )
w ( t ) = { w min + ( w max w min ) f A a v g ( t ) ,   f D i v ( t ) < α w min + ( w max w min ) ( T t T ) 2 ,   f D i v ( t ) α
where f a v g ( t ) denotes the mean value of adaptation in iteration t .
Li et al. [74] divided the swarm into three subsets by the fitness value and assigned different inertia weights, and Equation (32) represents the process of dynamically updating the inertia weights by introducing the cloud variation method when the particle fitness value is lower than the average fitness value of the swarm:
w = 0.9 0.7 e ( f i E x ) 2 2 ( E n ) 2
Zhang et al. [75] introduced a multi-stage strategy to improve the inertia weights through the cosine function. Equations (33)–(35) show the improvement process:
w cos ( t ) = ( w i n i + w f i n ) 2 + ( w i n i w f i n ) 2 cos ( I ( t ) π t max )
I ( t + 1 ) = I ( t ) + a , I ( 1 ) = 0
a = { a 1 ,   I ( t ) 6 t max a 2 ,   t max 6 < I ( t ) 5     t max 6 a 3 ,   5     t max 6 < I ( t ) t max ,   w h e r e   1 6 a 1 + 1 3 a 2 + 1 6 a 3 = 1
Ankit et al. [76] introduced a mapping function based on binomial probability and adaptively updated the inertia weights. Equations (36) and (37) represent the updating process:
w = w min + ( w max w min ) k = 0 X ( N k ) p k q N k
X = X ( i , t ) = i = 1 N I ( i , t )
where I ( i , t ) denotes the i th particle at the time t position function, X denotes the total number of particles passing through the binomial probability improvement position, p denotes the improvement success probability, q denotes the failure probability, the particle swarm is independent of each other, and p = q = 0.5 .
Komarudin et al. [77] combined the multi-dimensional direct fuzzy signature, parametric feedback success counter, and CBPE methods to transform the fuzzy process into aggregation equations for the adaptive updating of inertia weights. Equation (38) represents the structural hierarchy of the inertia weights, Equation (39) represents the best particle to reach the minimization goal, Equation (40) represents the best particle percentage, Equation (41) represents the particle distribution in the search space, and Equation (42) represents the NCBPE normalization process for measuring the best fitness value of the solution:
w = [ w max [ w min [ P S [ d i v e r s i t y d i v max ] N C B P E ] ] ]
S C i ( t ) = { 1 ,   f ( x i d ( t ) ) < f ( P i d ( t ) ) 0 ,   e l s e
P S ( t ) = 1 N i = 1 N S C i ( t )
d i v e r s i t y = 1 N i = 1 N d = 1 D ( x i d g d ) 2
N C B P E = C B P E C B P E min C B P E max C B P E min
Fan et al. [78] analyzed a variety of improved inertia weights and proposed a time-varying adaptive inertia weighting strategy, and Equation (43) represents the update strategy:
w ( t ) = ( w max w min ) f ( P n b e s t ) a v g f ( P t b e s t ) a v g + w max t M a x I t e r
where f ( P n b e s t ) a v g and f ( P t b e s t ) a v g denote the average fitness values of the global best position and the local best position in iteration t .
Chrouta et al. [79] proposed a factor selection strategy algorithm, FMSPSO, based on particle swarm number selection to adaptively adjust different parameter selection strategies during a swarm search to improve the algorithm’s ability to find global search and avoid premature convergence, which is based on the problem that MSPSO fails to obtain the optimal solution due to premature convergence during the search process. Equation (44) represents the adaptive inertia weight updating strategy, and Equation (45) represents the learning factor updating strategy:
w s = ( w i n i w e n d ) ar tanh ( δ ( 1 S S max ) m ) + w i n i
c = ( c i n i c e n d ) [ 1 ( k k max ) s m e s m ] + c i n i
where w i n i = 0.4 , w e n d = 0.9 , δ and m are control factors, and s is the number of current iterations.
Jeremiah et al. [80] introduced a shrinkage factor in the search process to ensure a smooth convergence of the algorithm to the global optimal solution to resolve the problem of premature convergence of the algorithm. They introduced a gravitational search algorithm to improve the global search ability of the algorithm and enhanced the stagnant particle activity through an adaptive response strategy. Equation (46) represents the shrinkage factor update strategy based on the cosine function, and Equation (47) represents the inertia weight update strategy. Equation (48) represents the relationship between the parameters:
k ( t ) = cos ( π G max T + 2.5 ) 4
{ w = w e n d + ( W s t a r t w e n d ) ( 1 T G max ) ,   i f   P g d x i d w = w e n d ,   i f   P g d = x i d
w c 1 + c 2 2 1 k ( t )

4.2.3. Optimal Particle Update

The self-renewal of particles in the PSO algorithm varies with the local optimal position P i and the global optimal position P g , which has a greater ability to guide the swarm evolution. Li et al. [81] combined the gray wolf optimization (GWO) algorithm with the worst particle for the evolution strategy and the optimal particle perturbation strategy to improve the swarm diversity, and the elite swarm guidance strategy to guide the particle swarm for enveloping the search and enhance the global searchability. The perturbation strategy is shown in Equations (49) and (50):
N b e s t ( t ) = s i g n ( r a n d 0.5 ) e ( a v g _ r ) G b e s t ( t ) + G b e s t ( t )
G b e s t ( t ) = { N b e s t ( t ) ,   i f ( f i t ( N b e s t ( t ) ) < f i t ( G b e s t ( t ) ) ) G b e s t ( t ) ,   o t h e r w i s e
where a v g _ r denotes the average distance of all particles to the center position. N b e s t ( t ) denotes the position after perturbation.
Keshavamurthy et al. [82] used a stochastic inertia weight strategy to balance the global and local search and adjusted the inertia weights by simulating the annealing probability, and the algorithm gradually approached the optimal solution as the temperature decreased. Equations (51)–(53) represent the inertia weight update strategy:
T t = f a v g t f b e s t t 1 = 1 m i = 1 m f i t f b e s t t 1
p = { 1 ,   i f   min f i t k 1 i m min f i t 1 i m e ( min f i t k 1 i m     min f i t 1 i m T t ) ,   i f   min f i t k 1 i m > min f i t 1 i m
w = { α 1 + r 2 ,   p r α 2 + r 2 ,   p < r
where Tt denotes the cooling temperature, p denotes the simulated annealing probability, k is an integer multiple of t , r is a random number within [0,1], α1, α2 are constants within [0,1], and α 1 > α 2 .
Han et al. [83] used the Karush–Kuhn–Tucker proximity measure to determine the contribution of optimally ranked particles to swarm convergence and introduced a gravitational search algorithm and an adaptive reset velocity strategy to enhance swarm diversity and avoid falling into local extremes. Equation (54) calculates the mass of the particle, and Equation (55) calculates the force of the j th particle on the i th particle to calculate the value of the particle’s contribution to the swarm convergence in Equation (56) to decide whether to guide other particles as the optimal particle. Equations (57) and (58) represent the stalled particle generation process:
m i t = 1 0.9 f i t i t w o r s t t
f i j d ( t ) = G ( t ) m i t + m j t R i j + ε ( x j d ( t ) x i d ( t ) )
a i d ( t ) = j k b e s t , j 1 r a n d j f i j d ( t ) m i t
p b e s t i , d = Φ + s d a m p v m e a n
v m e a n = 1 N i = 1 N 1 D j = 1 D ( v i , d ) 2
where f i t i t denotes the fitness value of the i th particle in the t th iteration, w o r s t t denotes the fitness worst value, R i j denotes the Euclidean distance between particles i and j , G ( t ) denotes the gravitational coefficient, and r a n d j is a random number uniformly distributed between [0,1].
Miao et al. [84] sorted each particle in the current swarm by the nonDS function to obtain the kth iteration of the non-dominated particle set and optimally combined it with the kth iteration of the particle set to obtain the current non-dominated particle set A r c k , where the particles choose their personal best position, P b e s t , in A r c k according to the Euclidean distance. This effectively improves the quality of P b e s t and prevents the algorithm from falling into local extremes. Equation (59) calculates the degree of influence of the optimal particles, and Equation (60) calculates Att, the size of which is proportional to the degree of influence:
d i s t ( x , y ) = { A t t ,   i f   x = y A t t ,   o t h e r w i s e
A t t = A t t max ( A t t max A t t min ) ( 1 g e n G max ) 5
where g e n denotes the current iteration number and G max denotes the maximum iteration number.
Zhu et al. [85] show that, based on the elastic collision principle and the law of momentum conservation, the worst particle of the swarm fitness value collides with the best particle of the fitness, and the worst particle obtains the best particle velocity. Equations (61) and (62) represent the calculation of the worst particle velocity, Equation (63) represents the particle velocity update strategy, and Equation (64) represents the particle position update strategy:
v 1 = Z e ( x w o r s t x r a n d ) 2
Z = 1 2 ( ( 1 2 t T ) 3 + 1 )
{ m 1 v 1 = m 1 v 1 + m 2 v 2 1 2 m 1 v 1 2 = 1 2 m 1 v 1 2 + 1 2 m 2 v 2 2 { v 1 = m 1     m 2 m 1   +   m 2 v 1 v 2 = 2     m 1 m 1   +   m 2 v 2
{ X w o r s t = X w o r s t + v 1 X b e s t = X b e s t + v 2
where X w o r s t denotes the particle with the worst adaptation value, X r a n d denotes the random particle position, T denotes the nonlinear factor, m 1 denotes the adaptation value of the worst particle, and m 2 denotes the adaptation value of the best particle.
Fu et al. [86] introduced the RBF model to generate the global optimum and individual optimum through MATLAB’s fmincon toolbox. While the existence of a large number of uncertain particles in the search space easily leads to local convergence of the algorithm, the authors also introduced the MaxEI method for pre-screening to measure the particle potential and uncertainty; thus, the global optimum has both maximum potential and maximum uncertainty.

4.2.4. Speed and Position Update

In the particle swarm optimization algorithms, velocity and position are properties specific to particles. Song et al. [87] combined successive higher Bessel curves and adaptive fractional-order velocities with particle evolution states to apply perturbations to the swarm to enhance its ability to jump out of local constraints. Equation (65) denotes the average distance of the i th particle from g b e s t in the k iterations, Equation (66) denotes the kth iteration and the swarm evolution state, Equation (67) denotes the swarm evolution adaptive adjustment factor, and Equations (68) and (69) denote the particle velocity and position update formulas:
d i k = 1 S 1 j = 1 , j i S k = 1 D ( x i k x j k ) 2
E f k = d g b k d min k d max k d min k
α = 0.9 1 1 + e E f k k k max
v i k + 1 = α v i k + 1 2 α ( 1 α ) v i k 1 + 1 6 α ( 1 α ) ( 2 α ) v i k 2 + 1 24 α ( 1 α ) ( 2 α ) ( 3 α ) v i k 3 + c 1 r 1 ( x i b k x i k ) + c 2 r 2 ( x g b k x i k )
x i k + 1 = x i k + v i k + 1
where d max k , d min k denotes the maximum and minimum values of the distance from a particle to other particles in the swarm, and d g b k denotes the average distance of the g b e s t .
Tian et al. [73] introduced the particle with the best comprehensive evaluation of the swarm fitness value and diversity in the current iteration as the latest guiding particle to guide the swarm evolution, together with the local and global optimum. Equation (70) represents the comprehensive evaluation of the i th particle, and Equation (71) represents the velocity update under the latest guiding particle:
f C o m ( i ) = d 1 f F i t ( i ) + d 2 f D i s ( i )
v i d ( t + 1 ) = w v i d ( t + 1 ) + c 1 r 1 ( p i d x i d ( t ) ) + c 2 r 2 ( g d x i d ( t ) ) + c 3 r 3 ( f C o m d ( i ) x i d ( t ) )
where d 1 = d 2 = 0.5 and d 1 + d 2 = 1 , f F i t ( i ) denotes the i th particle fitness value, and f D i s ( i ) denotes the i th particle diversity evaluation, i.e., the distance to the central particle c 1 = c 2 = c 3 = 1.49445 .
Mohamed et al. [88] and Suman et al. [89] introduced the GWO algorithm to improve the global search ability of the algorithm. Equation (72) represents α , β , δ the distance between the three wolves and other individuals, Equation (73) represents the current position of the three wolves, and Equation (74) represents the current gray wolf position:
{ D α = | C 1 X α X | D β = | C 2 X β X | D δ = | C 3 X δ X | }
{ X 1 = | X α A 1 ( D α ) | X 2 = | X β A 2 ( D β ) | X 3 = | X δ A 3 ( D δ ) | }
X ( t + 1 ) = X 1 + X 2 + X 3 3
where A = 2 a r 1 a , C = 2 r 2 , and a is a convergence factor that decreases from 2 to 0 with iteration and r 1 , r 2 are random numbers within [0,1].
Zhao et al. [84] used Euclidean distance to quantify the leading effect of the current optimal particle on evolution and to update the particle velocity and position. Equations (75) and (76) represent the velocity update process, and Equation (77) represents the particle position update probability:
V i ( p , q ) k + 1 = W V i ( p , q ) k + C 1 r 1 d i s t ( P b e s t i k ( p ) , q ) + C 2 r 2 d i s t ( G b e s t k ( p ) , q )
V i ( p , q ) k + 1 = { V max V i ( p , q ) k + 1 | V i ( p , q ) k + 1 | ,   i f | V i ( p , q ) k + 1 | V max V i ( p , q ) k + 1 ,   o t h e r w i s e
P k + 1 ( x i , p k + 1 = q ) = s i g m o i d ( V i ( p , q ) k + 1 ) l = 1 M s i g m o i d ( V i ( p , l ) k + 1 )
where dist() denotes the distance between the global optimum and the individual optimum for evaluating the particle’s importance.
Xu et al. [90] introduced variation, crossover, and selection operations of differential evolution to enhance the information exchange among swarms: Equation (78) for the variation factor, Equation (79) for the crossover operation, Equation (80) for the variation operation, and Equation (81) for the particle velocity update:
x i s o n = x r a + 0.5 2 e 1 i t max i t max + 1 n ( x r b x r c ) ( i = 1 , 2 , , N )
u i , d s o n = { x i , d s o n ,   i f ( r a n d < C R )   o r   ( d = d r a n d ) x i , d ,   o t h e r w i s e
x i s o n = { u i s o n ,   i f ( f ( u i s o n ) < f ( x i ) ) x i ,   o t h e r w i s e
v i d n + 1 = w v i d n + c 1 r 1 ( P i d n x i d n ) + c 2 r 2 ( P g d n x i d n ) + c 3 r 3 ( P c d n x i d n )
where ra, rb, rc, and d r a n d denote the random numbers within [ 0 , N ] , CR denotes the crossover rate and P c d n denotes the best communication location obtained after the variation, crossover, and selection.
Su et al. [91] introduced a time factor to overcome the problem that causes the traditional FOPSO algorithm to easily fall into local extremes, which reduces the overdependence of the algorithm on the fractional-order α. The time factor, T, can change adaptively with the swarm parameters, and when T is large, the algorithm has a strong global search capability and helps to obtain the global optimum, and vice versa. Equation (82) represents the time factor change process, and Equation (83) represents the position update process of the improved algorithm:
T = 1 2 2 e a t 1 a t ,   w h e r e   a t = min ( f g d , f t ) max ( f g d , f t )
x i d ( t + 1 ) = x i d ( t ) + v i d ( t ) T
where a t denotes the swarm diversity, f g d denotes the global optimum, and f t denotes the average fitness value in iteration t .

5. Association Rule Mining Based on PSO

5.1. Algorithm Principle

5.1.1. ARM Process

The particle swarm optimization (PSO) algorithm for ARM mainly consists of two stages: preprocessing and rule mining. Its basic flow is shown in Figure 3.
The preprocessing stage mainly completes the binary data conversion. The rule mining stage is the process of an iterative search and the continuous updating of particle positions and velocities in the search space by the PSO algorithm after data encoding and initialization of the swarm to finally obtain the best rules.

5.1.2. Binary Conversion

The standard PSO algorithm is mainly used for continuous real-valued optimization problems, while in real life, it is mostly used for discrete or combinatorial optimization problems. To address this issue, Kennedy and Eberhart proposed the discrete binary particle swarm optimization algorithms (BPSO) in 1997, which use the sigmoid() function to limit the position and velocity to the {0,1} interval, and where the particle position is calculated iteratively, as shown in Equation (84).
x i d + 1 = { 1 ,   r < s i g m o i d ( v i d + 1 ) ; 0 ,   r s i g m o i d ( v i d + 1 ) ;
where r is a random number between (0,1) and s i g m o i d ( v i d + 1 ) = 1 1 + e v i d + 1 .
The process of binary conversion of transaction data by the above means is shown in Figure 4.
Among them, the original database has five transaction records, which are T 1 , T 2 , , T 5 , and these five transaction records are converted into binary form and stored. Take T 5 as an example, where the transaction record contains I 1 , I 3 , I 4 , so its corresponding units are all 1, and the other units are 0. The converted transaction record is 10110.

5.1.3. Encoding

There are two main ways to encode data: the Michigan approach and the Pittsburgh approach. Among them, the Pittsburgh approach corresponds to a set of rules per chromosome and is suitable for solving classification problems. The Michigan approach corresponds to a single rule per chromosome and is more straightforward than the former, with a shorter rule syntax. Let us take the Michigan method as an example: Let there be N data items in the dataset, and each data item has two parts, and each part has two possible values that take the value of 1 or 0, as shown in Figure 5.
The dimension of each particle is 2N, and the data item X refers to whether the item is included in the rule. Y refers to whether the item is included in the result of the rule. For example, if the value I i is 11, then the i th item appears in the rule and is included in the result of the rule. The position representation of the rule is shown in Figure 6.

5.1.4. Search for the Best Particle

First, we need to define a fitness function for evaluating the importance of each particle. By using Equation (84) as the objective function:
F i t n e s s ( i ) = C o n f i d e n c e ( i ) + S u p p o r t ( i )
where C o n f i d e n c e ( i ) and S u p p o r t ( i ) denote the confidence and support of the ith particle, respectively. The objective of this function is to maximize the confidence and support of the particle; the stronger this support, the stronger the strength of the association, implying that this is an important association rule.
Second, we also need to generate an initial population, and the particles in the population are called initial particles. Finally, we need to set boundary conditions to constrain the range of motion of the particles and obtain the global best particle when G b e s t and P b e s t are iterated until the termination condition is reached. The boundary conditions include calculating the Euclidean distance of the particles, etc. The termination conditions can be set by setting a fixed number of iterations.

5.2. Application

In dealing with complex, realistic optimization problems, heuristic algorithms are inspired by nature and modeled by simulating the behavior of individuals or swarms, which have the advantage of not relying on specific problems, while such algorithms are not constrained by the form of optimization functions in the search process and are highly adaptable, efficient, and easy to implement. Compared with other algorithms, heuristic algorithms show obvious advantages in solving ARM problems. Among the heuristic algorithms, the genetic algorithm (GA) has a long convergence time due to the lack of individual memory and blind search without direction, and it shows inefficiency and difficulty in convergence in high-dimensional data. The ant colony optimization (ACO) algorithm uses pheromone to decide the search direction, which consumes a lot of time for a solution construction and also has the problem of a long convergence time and is prone to stagnation in high-dimensional data. The particle swarm optimization (PSO) algorithm is efficient and simple to implement, converges quickly, and can adjust the search direction by iteration, which is significantly better than other heuristics in ARM problems [92]. PSO algorithms are widely used in ARM, mainly in the fields of privacy protection, recommender systems, risk assessment, and medicine. The correlation is shown in Figure 7.

5.2.1. Privacy Protection

The rapid growth of the Web is accompanied by a dramatic increase in data, and more and more companies or organizations are sharing information by sharing data to gain more profits. To protect the sensitive information implicit in the data from leakage, users need to modify the individual numerical information of the database while sharing the data without compromising the validity of the data. Kalyani et al. [93] proposed an IFHARM algorithm based on PPARM, which includes identifying SAR, clustering SAR by similarity metric, identifying sensitive transactions, removing and rescanning the converted database, and performing error rule verification. Among them, the authors use the result of each item deletion on non-sensitive data as its impact factor to identify SAR to minimize the loss of non-sensitive association rules and ensure the high utility of the data. Jangra et al. [94] proposed the VIDPSO algorithm to solve the rapid performance degradation of evolutionary algorithms in dense datasets, which prioritizes the presence of more sensitive patterns in data and identifies SAR based on its sensitive patterns to avoid the failure of sensitive information hiding due to the transformation of sensitive frequent itemsets into infrequent itemsets by removing too much data. Suma et al. [95] developed a new method by integrating fractional calculus (FC) and the salp swarm algorithm (SSA) for data sanitization using privacy-preserved data. Krisgnamoorthy et al. [96] used different methods to perturb the dataset to generate a large number of solutions without affecting the validity of the data, which were then sequentially optimized by the PSO algorithm until the support was lower than the minimum support, at which point, the loop was closed, effectively reducing the number of modifications and avoiding the generation of spurious rules.

5.2.2. Recommender Systems

Analyzing consumers’ previous behaviors can provide users with personalized products or valuable information, and recommendation systems attempt to recommend goods to potential customers by applying data mining techniques. Among them, the shopping basket problem is a typical example of ARM, and Guo et al. [97] proposed an improved PSO algorithm based on the gray model for dynamic ARM of supermarket shopping data. Furthermore, since the introduction of the background values of the gray model tends to lead to a decrease in prediction accuracy, the authors introduced a secondary search mechanism based on the PSO algorithm to optimally solve the background values of the gray model at different moments, which significantly improves the local search ability of the algorithm and avoids premature particle convergence. To obtain rich product manufacturing knowledge from enterprise historical data to reflect the mapping relationship between a product’s design and manufacturing, Kou [98] proposed an ARM method based on a binary PSO algorithm, which encodes association rules and generates a certain number of particle swarms, obtains N rules through multiple iterations, integrates multiple indicators, such as reliability, relevance, and comprehension, and adds a similarity function to compare rules two by two, while eliminating the poor rules to achieve efficient rule mining.
Web data often implies a large amount of web search information. Kaur et al. [99] proposed a hybrid algorithm, HPSO-TS-ARM, to obtain the potential user information among them, which obtains web data through the PSO algorithm, arranges data information by the TAO algorithm, and then performs ARM. Its computation time and efficiency are better than other algorithms, and it has been validated in several datasets. Gangurde et al. [100] used the Weblog features related to Web mining to construct a Web recommendation model, introduced a PSO algorithm, and used a feedforward counter model to obtain association rules using a single data iteration technique, which was used as an adaptation value to execute the recommendation model, and the execution time was significantly reduced. Tyagi et al. [41] solved the low accuracy and data sparsity problems, considering support and confidence as the objectives for proposing the MOPSO algorithm. The algorithm can effectively filter out rules with significant association relationships, and its accuracy and precision are also better than other algorithms when the data is sparse. Jiang et al. [101] proposed an association rule method for generating sentiment designs based on online customer reviews for a large number of online user reviews on major websites by mining the sentiment dimension opinion of user reviews. This method can mine more reliable and interesting rules than the traditional PSO algorithm by performing sentiment analysis on customer reviews to achieve accurate user sentiment recommendations.

5.2.3. Risk Assessment

Credit risk management is one of the most important concerns of financial companies, including the collection of relevant information about borrowers, which is used to decide whether to approve their credit applications. Lanzarini et al. [102] proposed the LVQ-PSO algorithm by combining competitive neural networks with the PSO algorithm, which mines association rules by preprocessing the original data with the LVQ method and binary PSO algorithm. The simplified rules can assist the decision-maker in deciding whether to grant a loan or not. Priya et al. [103] proposed a novel PSO-based fuzzy associative classifier (PaSOFuAC) to detect phishing websites and avoid financial losses caused by users not recognizing phishing websites.
As the most convenient means of transportation, the flight process of aircraft is highly susceptible to environmental, equipment, and other complex factors, and the number of hazards increases. Li et al. [104] designed a weighted multiple swarm PSO algorithm for hazard source analyses based on the background of civil aviation safety risk assessments. The preprocessing stage assigns weights to items in the hazard source database and defines the candidate association rules for hazard sources; the analysis stage introduces the swarm interaction mechanism and uses a parallel search mode to generate the association rules for hazard sources and their generation causes, which provides important theoretical support for hazard source identification and prediction. The multiple swarms’ parallel search mode provides an interaction mechanism between the swarms to avoid swarms falling into local extremes due to the lack of information exchanged. Inspired by an online sequence learning machine, She et al. [105] used a single hidden layer feedforward neural network to train hazard source data and update the output weights of the learning machine in real-time to realize the online identification of hazard sources. The paper discards the traditional ARM algorithm and uses the ACO-ARM algorithm to mine the most frequent term and the PSO algorithm to initialize the pheromone concentration, which effectively avoids the algorithm’s initialization error and improves the system’s response speed.

5.2.4. Medical Field

In the medical field, a pathological diagnosis is mostly derived from physicians’ treatment experience due to the lack of quantitative criteria. ARM can extract features from multiple factors, classify and identify them, and extract the hidden information behind them from medical records to provide physicians with accurate scientific opinions. To improve the applicability and classification accuracy of the algorithm in processing medical data, Ripon et al. [106] combined the PSO algorithm with an artificial neural network to automatically analyze the dataset to achieve an accurate classification of tuberous sclerosis by ARM. To overcome the problem of excessive data dimensionality, Choubey et al. [107] introduced the PCA-PSO algorithm for a feature reduction of raw data to significantly reduce the computation time while ensuring prediction classification accuracy.
In terms of disease prediction, Karsidani et al. [108] proposed ANFIS-PSO to predict whether a patient has coronary artery disease effectively. To deal with the continuous and rapid growth of data volume problems, Mangat et al. [109] proposed a dynamic particle swarm-based rule mining classifier to achieve adaptive prediction of the expected life expectancy of patients with thoracic diseases by dynamically creating swarms and updating the algorithm parameters. Raja et al. [110] proposed a data mining strategy-based prediction model for type II diabetes mellitus by combining the PSO algorithm and fuzzy clustering mean with medical metrics, such as sensitivity, specificity, and accuracy, which are widely used in medicine to assess the performance of the system, and found that the prototype was 8.26% more accurate than traditional methods.
To address the existing problems of inefficient and time-consuming diagnoses of heart diseases, Alkeshuosh et al. [111] used a standard PSO algorithm to obtain heart disease rules from the original dataset and optimized them to generate the best rules. Mao Jie et al. [112] proposed the PSO-SVM algorithm based on support vector machines (SVM) for heart disease predisposing factor detection, which uses ARM to extract the disease features and train the feature dataset. Each PSO-SVM classifier performs the corresponding fitness function, formulated by a classified/total, to iteratively update the velocity and position of each particle until it obtains the global best value. The algorithm is tested on the UCI Cleveland dataset using confidence levels as an indicator and has better classification performance compared to traditional PSO algorithms.

6. Conclusions and Challenges

As an emerging field in ARM, PSO algorithms are becoming more mature through continuous exploration and research by many scholars and are widely used in privacy protection, recommender systems, risk assessment, and medicine. This section shows a comparative analysis of the different PSO algorithms, and the objective analysis of existing methods is shown in Table 5.
The advantages of the particle swarm optimization (PSO) algorithm over other algorithms are: (1) It has uncertainty, which is significantly better than the deterministic algorithms in solving certain aspects of a problem and is more advantageous in finding the global optimal solution; (2) it has better adaptability—the PSO algorithm is a bionic optimization algorithm based on multiple intelligences and is highly robust; (3) it is evolutionary and can preserve the optimal solution information in each iteration round.
The above research shows that the PSO algorithm still has great development space and application value in ARM, but with the expansion of data scale, the difficulty of mining effective association rules from large-scale data using the PSO algorithm gradually increases, and this field will still face great challenges in future research.

7. PSO Future Prospects

The future research directions of particle swarm optimization algorithms may be as follows:
(1)
PSO algorithms converge prematurely and are prone to falling into local extremes, which mainly lie in the imbalance between global search and local search capabilities. Nevertheless, many improvement algorithms have been proposed, such as the QPSO algorithm based on the cloud model, the adaptive multi-objective PSO algorithm based on Gaussian mixed variance and elite learning, etc. However, the two are relatively contradictory, and how to dynamically maintain the balance between them in the search process to obtain the optimal solution based on the actual search results or how to measure that the two have reached the best balance needs further study.
(2)
Due to the exponential growth of high-dimensional data search space and low data relevance, search efficiency decreases rapidly. In addition to the existing “divide and conquer” method and the introduction of other algorithms, the use of feature extraction methods to remove redundant data and natural computational methods, such as nonlinear dimensionality reduction, can also be introduced to reduce redundancy using the maximum linear irrelevance group. Although significant progress has been made in handling large amounts of data, there is still much room for exploration in the ARM models for handling high-speed data.
(3)
Inertia weights and learning factors are important parameters in PSO algorithms. Improvements to inertia weights include linearly decreasing inertia weights, fuzzy inertia weights, random inertia weights, etc. Improvements in learning factors include shrinkage factors, synchronous learning factors, asynchronous learning factors, etc. In the future, we can continue to optimize these two parameters and consider whether the two affect each other and their mutual influence weights.
(4)
Many scholars focus more on the swarm initialization problem, including the M-class random method, fixed-value setting method, two-step method, hybrid method, etc. However, the method is too simple in how to divide the swarm size and lacks a scientific and reasonable effective division, which will limit the speed of particle movement when the swarm is too small and, thus, lead to early convergence. When the swarm is too large, it leads to a too large search space and reduces the performance and efficiency of the algorithm.
(5)
When there are missing or wrong datasets, there is difficulty in guaranteeing the PSO algorithm’s accuracy and performance without artificial preprocessing, i.e., having certain fault tolerance.

Author Contributions

Conceptualization, G.L. and T.W.; methodology, G.L., T.W. and P.S.; validation, G.L., Q.C. and N.X.; writing—original draft preparation, T.W. and G.L.; writing—review and editing, G.L., T.W., P.S., Q.C. and A.V.; supervision, N.X. and A.V.; funding acquisition, G.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (#62041702); the Ministry of Education, Humanities and Social Sciences Project (#20YJA870010); the Jiangxi Provincial Social Science Planning Project (#19TQ05); the Key project of Education Science planning in Jiangxi Province (#19ZD024); the Jiangxi University Humanities and Social Science Planning Project (#TQ20105); the Basic Education Research Project of Jiangxi Province (#SZUNDZH2021-1143); and the Jiangxi Province Degree and Postgraduate Education and Teaching Reform Research Project (#JXYJG-2020-075).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Agrawal, R.; Tomasz, I.; Arun, S. Mining association rules between sets of items in large databases. In Proceedings of the 1993 ACM SIGMOD International Conference on Management of Data, Washington, DC, USA, 25–28 May 1993; pp. 207–216. [Google Scholar]
  2. Telikani, A.; Amir, H.G.; Asadollah, S. A survey of evolutionary computation for association rule mining. Inf. Sci. 2020, 524, 318–352. [Google Scholar] [CrossRef]
  3. Kaushik, M.; Sharma, R.; Peious, S.A.; Shahin, M.; Yahia, S.B.; Draheim, D. A systematic assessment of numerical association rule mining methods. SN Comput. Sci. 2021, 2, 1–13. [Google Scholar] [CrossRef]
  4. Djenouri, Y.; Fournier, V.P.; Belhadi, A.; Chun, W.L. Metaheuristics for frequent and high-utility itemset mining. In High-Utility Pattern Mining; Springer: Berlin/Heidelberg, Germany, 2019; pp. 261–278. [Google Scholar]
  5. Ghobaei, A.M.; Shahidinejad, A. An efficient resource provisioning approach for analyzing cloud workloads: A metaheuristic-based clustering approach. J. Supercomput. 2021, 77, 711–750. [Google Scholar] [CrossRef]
  6. Gan, X.; Gong, D.; Gao, X.; Zhang, Y.; Lin, L.; Lan, T. Research on Construction of Mining Model of Association Rules for Temporal Data based on CNN. In Proceedings of the 2021 IEEE International Conference on Electronic Communications, Internet of Things and Big Data (ICEIB), Yilan County, Taiwan, 10–12 December 2021; pp. 233–235. [Google Scholar]
  7. Chen, S.; Xi, J.; Chen, Y.; Zhao, J. Association mining of near misses in hydropower engineering construction based on convolutional neural network text classification. Comput. Intell. Neurosci. 2022, 2022, 1–6. [Google Scholar] [CrossRef]
  8. He, B.; Zhang, J. An Association Rule Mining Method Based on Named Entity Recognition and Text Classification. Arab. J. Sci. Eng. 2022, 6, 1–9. [Google Scholar] [CrossRef]
  9. Badhon, B.; Kabir, M.M.; Xu, S.; Kabir, M. A survey on association rule mining based on evolutionary algorithms. Int. J. Comput. Appl. 2021, 43, 775–785. [Google Scholar] [CrossRef]
  10. Kheirollahi, H.; Chahardowli, M.; Simjoo, M. A new method of well clustering and association rule mining. J. Pet. Sci. Eng. 2022, 214, 110479. [Google Scholar] [CrossRef]
  11. Datta, S.; Mali, K. Significant association rule mining with high associability. In Proceedings of the IEEE 2021 5th International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 6–8 May 2021; pp. 1159–1164. [Google Scholar]
  12. Rathee, S.; Kashyap, A. Adaptive-Miner: An efficient distributed association rule mining algorithm on Spark. J. Big Data 2018, 5, 1–17. [Google Scholar] [CrossRef]
  13. Zhong, Q.Y.; Qian, Q.; Feng, Y.; Fu, Y. BPSO Algorithm with Opposition-Based Learning Method for Association Rule Mining. In Advancements in Mechatronics and Intelligent Robotics; Springer: Berlin/Heidelberg, Germany, 2021; Volume 1220, pp. 351–358. [Google Scholar]
  14. Baró, G.B.; Martínez, T.J.F.; Rosas, R.M.V.; Ochoa, J.A.C.; González, A.Y.R.; Cortés, M.S.L. A PSO-based algorithm for mining association rules using a guided exploration strategy. Pattern Recognit. Lett. 2020, 138, 8–15. [Google Scholar] [CrossRef]
  15. Fister, J.I.; Fister, I. A Brief Overview of Swarm Intelligence-Based Algorithms for Numerical Association Rule Mining. Appl. Optim. Swarm Intell. 2021, 8, 47–59. [Google Scholar]
  16. Wang, C.; Bian, W.; Wang, R.; Chen, H.; Ye, Z.; Yan, L. Association rules mining in parallel conditional tree based on grid computing inspired partition algorithm. Int. J. Web Grid Serv. 2020, 16, 321–339. [Google Scholar] [CrossRef]
  17. Yuan, J. An anomaly data mining method for mass sensor networks using improved PSO algorithm based on spark parallel framework. J. Grid Comput. 2020, 18, 251–261. [Google Scholar] [CrossRef]
  18. Sukanya, N.S.; Thangaiah, P.R.J. Customized Particle Swarm Optimization Algorithm for Frequent Itemset Mining. In Proceedings of the IEEE 2020 International Conference on Computer Communication and Informatics (ICCCI), Coimbatore, India, 22–24 January 2020; pp. 1–4. [Google Scholar]
  19. Dubey, A.K.; Kumar, A.; Agrawal, R. An efficient ACO-PSO-based framework for data classification and preprocessing in big data. Evol. Intell. 2021, 14, 909–922. [Google Scholar] [CrossRef]
  20. Tofighy, S.; Rahmanian, A.A.; Ghobaei, A.M. An ensemble CPU load prediction algorithm using a Bayesian information criterion and smooth filters in a cloud computing environment. Softw. Pract. Exp. 2018, 48, 2257–2277. [Google Scholar] [CrossRef]
  21. Gad, A.G. Particle Swarm Optimization Algorithm and Its Applications: A Systematic Review. Arch. Comput. Methods Eng. 2022, 1–31. [Google Scholar]
  22. Nasr, M.; Hamdy, M.; Hegazy, D.; Bahnasy, K. An efficient algorithm for unique class association rule mining. Expert Syst. Appl. 2021, 164, 113978. [Google Scholar] [CrossRef]
  23. Kuok, C.M.; Fu, A.; Wong, M.H. Mining fuzzy association rules in databases. ACM SIGMOD Rec. 1998, 27, 41–46. [Google Scholar] [CrossRef]
  24. Ting, C.K.; Liaw, R.T.; Wang, T.C.; Hong, T.P. Mining fuzzy association rules using a memetic algorithm based on structure representation. Memetic Comput. 2018, 10, 15–28. [Google Scholar] [CrossRef]
  25. Varol, A.E.; Alatas, B. Computing, Performance analysis of multi-objective artificial intelligence optimization algorithms in numerical association rule mining. J. Ambient. Intell. Humaniz. Comput. 2020, 11, 3449–3469. [Google Scholar] [CrossRef]
  26. Altay, E.V.; Alatas, B. Intelligent optimization algorithms for the problem of mining numerical association rules. Phys. A Stat. Mech. Appl. 2020, 540, 123142. [Google Scholar] [CrossRef]
  27. Heraguemi, K.E.; Kamel, N.; Drias, H. Multi-objective bat algorithm for mining numerical association rules. Int. J. Bio-Inspired Comput. 2018, 11, 239–248. [Google Scholar] [CrossRef]
  28. Srikant, R.; Agrawal, R. Mining quantitative association rules in large relational tables. In Proceedings of the 1996 ACM SIGMOD International Conference on Management of Data, Montreal, QC, Canada, 4–6 June 1996; pp. 1–12. [Google Scholar]
  29. Song, X.D.; Zhai, K.; Gao, W. The Research of Evaulation index in Association Rules. Microcomput. Inf. 2007, 23, 174–176. [Google Scholar]
  30. Wang, C.; Liu, Y.; Zhang, Q.; Guo, H.; Liang, X.; Chen, Y.; Wei, Y. Association rule mining based parameter adaptive strategy for differential evolution algorithms. Expert Syst. Appl. 2019, 123, 54–69. [Google Scholar] [CrossRef]
  31. Altay, E.V.; Alatas, B. Differential evolution and sine cosine algorithm based novel hybrid multi-objective approaches for numerical association rule mining. Inf. Sci. 2021, 554, 198–221. [Google Scholar] [CrossRef]
  32. Guan, B.; Zhao, Y.; Yin, Y.; Li, Y. A differential evolution based feature combination selection algorithm for high-dimensional data. Inf. Sci. 2021, 547, 870–886. [Google Scholar] [CrossRef]
  33. Menaga, D.; Saravanan, S. GA-PPARM: Constraint-based objective function and genetic algorithm for privacy preserved association rule mining. Evol. Intell. 2021, 15, 1487–1498. [Google Scholar] [CrossRef]
  34. Lin, J.C.W.; Djenouri, Y.; Srivastava, G.; Yun, U.; Fournier-Viger, P. A predictive GA-based model for closed high-utility itemset mining. Appl. Soft Comput. 2021, 108, 107422. [Google Scholar] [CrossRef]
  35. Neysiani, B.S.; Soltani, N.; Mofidi, R.; NadimiShahraki, M.H. Improve performance of association rule-based collaborative filtering recommendation systems using genetic algorithm. Int. J. Inf. Technol. Comput. Sci. 2019, 11, 48–55. [Google Scholar] [CrossRef]
  36. Giri, P.K.; De, S.S.; Dehuri, S.; Cho, S.B. Biogeography based optimization for mining rules to assess credit risk. Intell. Syst. Account. Financ. Manag. 2021, 28, 35–51. [Google Scholar] [CrossRef]
  37. Ghobaei, A.M. A workload clustering based resource provisioning mechanism using Biogeography based optimization technique in the cloud based systems. Soft Comput. 2021, 25, 3813–3830. [Google Scholar] [CrossRef]
  38. Mo, H.; Xu, L. Immune clone algorithm for mining association rules on dynamic databases. In Proceedings of the 17th IEEE International Conference on Tools with Artificial Intelligence (ICTAI’05), Hong Kong, China, 14–16 November 2005; Volume 5, pp. 202–206. [Google Scholar]
  39. Husain, M.S. Exploiting Artificial Immune System to Optimize Association Rules for Word Sense Disambiguation. Int. J. Intell. Syst. Appl. Eng. 2021, 9, 184–190. [Google Scholar] [CrossRef]
  40. Cunha, D.S.; Castro, L.N. Evolutionary and immune algorithms applied to association rule mining in static and stream data. In Proceedings of the 2018 IEEE Congress on Evolutionary Computation (CEC), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–8. [Google Scholar]
  41. Tyagi, S.; Bharadwaj, K.K. Enhancing collaborative filtering recommendations by utilizing multi-objective particle swarm optimization embedded association rule mining. Swarm Evol. Comput. 2013, 13, 1–12. [Google Scholar] [CrossRef]
  42. Kuo, R.J.; Gosumolo, M.; Zulvia, F.E. Multi-objective particle swarm optimization algorithm using adaptive archive grid for numerical association rule mining. Neural Comput. Appl. 2019, 31, 3559–3572. [Google Scholar] [CrossRef]
  43. Agarwal, A.; Nanavati, N. Association rule mining using hybrid GA-PSO for multi-objective optimisation. In Proceedings of the 2016 IEEE International Conference on Computational Intelligence and Computing Research (ICCIC), Chennai, India, 15–17 December 2016; pp. 1–7. [Google Scholar]
  44. Moslehi, F.; Haeri, A.; Martínez-Álvarez, F. A novel hybrid GA–PSO framework for mining quantitative association rules. Soft Comput. 2020, 24, 4645–4666. [Google Scholar] [CrossRef]
  45. Roopa Devi, E.M.; Suganthe, R.C. Enhanced transductive support vector machine classification with grey wolf optimizer cuckoo search optimization for intrusion detection system. Concurr. Comput. Pract. Exp. 2020, 32, e4999. [Google Scholar] [CrossRef]
  46. Chiclana, F.; Kumar, R.; Mittal, M.; Khari, M.; Chatterjee, J.M.; Baik, S.W. ARM–AMO: An efficient association rule mining algorithm based on animal migration optimization. Knowl. Based Syst. 2018, 154, 68–80. [Google Scholar]
  47. Pradeep, G.; Ravi, V.; Krishna, G.J. Hybrid Evolutionary Computing-based Association Rule Mining. In Soft Computing in Interdisciplinary Sciences; Springer: Singapore, 2022; pp. 223–243. [Google Scholar]
  48. Pazhaniraja, N.; Sountharrajan, S.; Sathis, K.B. High utility itemset mining: A Boolean operators-based modified grey wolf optimization algorithm. Soft Comput. 2020, 24, 16691–16704. [Google Scholar] [CrossRef]
  49. Yildirim, G.; Alatas, B. New adaptive intelligent grey wolf optimizer based multi-objective quantitative classification rules mining approaches. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 9611–9635. [Google Scholar] [CrossRef]
  50. Chantar, H.; Mafarja, M.; Alsawalqah, H.; Heidari, A.A.; Aljarah, I.; Faris, H. Feature selection using binary grey wolf optimizer with elite-based crossover for Arabic text classification. Neural Comput. Appl. 2020, 32, 12201–12220. [Google Scholar] [CrossRef]
  51. Telikani, A.; Gandomi, A.H.; Shahbahrami, A.; Dehkordi, M.N. Privacy-preserving in association rule mining using an improved discrete binary artificial bee colony. Expert Syst. Appl. 2020, 144, 113097. [Google Scholar] [CrossRef]
  52. Turabieh, H.; Azwari, S.A.; Rokaya, M.; Alosaimi, W.; Alharbi, A.; Alhakami, W.; Alnfiai, M. Enhanced Harris Hawks optimization as a feature selection for the prediction of student performance. Computing 2021, 103, 1417–1438. [Google Scholar] [CrossRef]
  53. Dong, D.; Ye, Z.; Cao, Y.; Xie, S.; Wang, F.; Ming, W. An improved association rule mining algorithm based on ant lion optimizer algorithm and FP-growth. In Proceedings of the 2019 10th IEEE International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications (IDAACS), Metz, France, 18–21 September 2019; Volume 1, pp. 458–463. [Google Scholar]
  54. Nawaz, M.S.; Fournier, V.P.; Yun, U.; Wu, Y.; Song, W. Mining high utility itemsets with hill climbing and simulated annealing. ACM Trans. Manag. Inf. Syst. (TMIS) 2021, 13, 1–22. [Google Scholar] [CrossRef]
  55. Ospina, M.H.; Quintana, J.L.A.; Lopez, V.F.J.; Berrio, G.S.; Barrero, L.H.; Sana, S.S. Extraction of decision rules using genetic algorithms and simulated annealing for prediction of severity of traffic accidents by motorcyclists. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 10051–10072. [Google Scholar] [CrossRef]
  56. Chou, Y.H.; Kuo, S.Y.; Chen, C.Y.; Chao, H.C. A rule-based dynamic decision-making stock trading system based on quantum-inspired tabu search algorithm. IEEE Access 2014, 2, 883–896. [Google Scholar] [CrossRef]
  57. Derouiche, A.; Layeb, A.; Habbas, Z. Frequent Itemsets Mining with Chemical Reaction Optimization Metaheuristic. In Proceedings of the IEEE 2018 3rd International Conference on Pattern Analysis and Intelligent Systems (PAIS), Tebessa, Algeria, 24–25 October 2018; pp. 1–6. [Google Scholar]
  58. Taradeh, M.; Mafarja, M.; Heidari, A.A.; Faris, H.; Aljarah, I.; Mirjalili, S.; Fujita, H. An evolutionary gravitational search-based feature selection. Inf. Sci. 2019, 497, 219–239. [Google Scholar] [CrossRef]
  59. Tuba, E.; Jovanovic, R.; Hrosik, R.C.; Alihodzic, A.; Tuba, M. Web intelligence data clustering by bare bone fireworks algorithm combined with k-means. In Proceedings of the 8th International Conference on Web Intelligence, Mining and Semantics, Novi Sad, Serbia, 25–27 June 2018; pp. 1–8. [Google Scholar]
  60. Abualigah, L.; Dulaimi, A.J. A novel feature selection method for data mining tasks using hybrid sine cosine algorithm and genetic algorithm. Clust. Comput. 2021, 24, 2161–2176. [Google Scholar] [CrossRef]
  61. Djenouri, Y.; Drias, H.; Chemchem, A. A hybrid bees swarm optimization and tabu search algorithm for association rule mining. In Proceedings of the IEEE 2013 World Congress on Nature and Biologically Inspired Computing, Fargo, ND, USA, 12–14 August 2013; pp. 120–125. [Google Scholar]
  62. Ma, L.; Zhang, T.; Wang, R.; Yang, G.; Zhang, Y. Pbar: Parallelized brain storm optimization for association rule mining. In Proceedings of the 2019 IEEE Congress on Evolutionary Computation (CEC), Wellington, New Zealand, 10–13 June 2019; pp. 1148–1156. [Google Scholar]
  63. Djenouri, Y.; Djenouri, D.; Belhadi, A.; Fournier, V.P.; Lin, J.C.W.; Bendjoudi, A. Exploiting GPU parallelism in improving bees swarm optimization for mining big transactional databases. Inf. Sci. 2019, 496, 326–342. [Google Scholar] [CrossRef]
  64. Malik, M.M.; Haouassi, H. Efficient sequential covering strategy for classification rules mining using a discrete equilibrium optimization algorithm. J. King Saud Univ. Comput. Inf. Sci. 2021, in press. [Google Scholar] [CrossRef]
  65. Eberhart, R.; Kennedy, J. A new optimizer using particle swarm theory. In Proceedings of the IEEE Sixth International Symposium on Micro Machine and Human Science, MHS’95, Nagoya, Japan, 4–6 October 1995; Volume 2, pp. 39–43. [Google Scholar]
  66. Wang, C.W.; Yin, S.L.; Liu, W.Y.; Wei, X.M.; Zheng, H.J.; Yang, J.P. High utility itemset mining algorithm based on improved particle swarm optimization. J. Chin. Comput. Syst. 2020, 41, 1084–1090. [Google Scholar]
  67. Dubey, K.; Sharma, S.C. A novel multi-objective CR-PSO task scheduling algorithm with deadline constraint in cloud computing. Sustain. Comput. Inform. Syst. 2021, 32, 100605. [Google Scholar] [CrossRef]
  68. Hematpour, N.; Ahadpour, S. Execution examination of chaotic S-box dependent on improved PSO algorithm. Neural Comput. Appl. 2021, 33, 5111–5133. [Google Scholar] [CrossRef]
  69. Li, A.D.; Xue, B.; Zhang, M. Improved binary particle swarm optimization for feature selection with new initialization and search space reduction strategies. Appl. Soft Comput. 2021, 106, 107302. [Google Scholar] [CrossRef]
  70. Wang, S.; Yu, X.; Jia, W. A New Population Initialization of Particle Swarm Optimization Method Based on PCA for Feature Selection. J. Big Data 2021, 3, 1–9. [Google Scholar] [CrossRef]
  71. Bangyal, W.H.; Hameed, A.; Alosaimi, W.; Alyami, H. A new initialization approach in particle swarm optimization for global optimization problems. Comput. Intell. Neurosci. 2021, 2021, 1–17. [Google Scholar] [CrossRef]
  72. Pervaiz, S.; Haider, B.W.; Ashraf, A.; Nisar, K.; Haque, M.R.; Ibrahim, A.; Rawat, D.B. Comparative Research Directions of Population Initialization Techniques using PSO Algorithm. Intell. Autom. Soft Comput. 2022, 32, 1427–1444. [Google Scholar] [CrossRef]
  73. Tian, S.; Li, Y.; Kang, Y.; Xia, J. Multi-robot path planning in wireless sensor networks based on jump mechanism PSO and safety gap obstacle avoidance. Future Gener. Comput. Syst. 2021, 118, 37–47. [Google Scholar] [CrossRef]
  74. Xiang, L.I.; Chen, J. A modified PSO Algorithm based on Cloud Theory for optimizing the Fuzzy PID controller. J. Phys. Conf. Ser. 2022, 2183, 012014. [Google Scholar]
  75. Zhang, J.; Sheng, J.; Lu, J.; Shen, L. UCPSO: A uniform initialized particle swarm optimization algorithm with cosine inertia weight. Comput. Intell. Neurosci. 2021, 2021, 41–59. [Google Scholar] [CrossRef]
  76. Agrawal, A.; Tripathi, S. Particle swarm optimization with adaptive inertia weight based on cumulative binomial probability. Evol. Intell. 2021, 14, 305–313. [Google Scholar] [CrossRef]
  77. Komarudin, A.; Setyawan, N.; Kamajaya, L.; Achmadiah, M.N.; Zulfatman, Z. Signature PSO: A novel inertia weight adjustment using fuzzy signature for LQR tuning. Bull. Electr. Eng. Inform. 2021, 10, 308–318. [Google Scholar] [CrossRef]
  78. Fan, M.; Akhter, Y. A time-varying adaptive inertia weight based modified PSO algorithm for UAV path planning. In Proceedings of the IEEE 2021 2nd International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST), Dhaka, Bangladesh, 5–7 January 2021; pp. 573–576. [Google Scholar]
  79. Chrouta, J.; Farhani, F.; Zaafouri, A. A modified multi swarm particle swarm optimization algorithm using an adaptive factor selection strategy. Trans. Inst. Meas. Control. 2021, 3, 01423312211029509. [Google Scholar] [CrossRef]
  80. Osei, K.J.; Han, F.; Amponsah, A.A.; Ling, Q.; Abeo, T.A. A hybrid optimization method by incorporating adaptive response strategy for Feedforward neural network. Connect. Sci. 2022, 34, 578–607. [Google Scholar] [CrossRef]
  81. Li, Z.; Wang, F.; Wang, R.J. An Improved Particle Swarm Optimization Algorithm. Available online: http://kns.cnki.net/kcms/detail/11.4762.TP.20210831.0841.008.html (accessed on 31 August 2021).
  82. Keshavamurthy, B.N. Improved PSO for task scheduling in cloud computing. In Evolution in Computational Intelligence; Springer: Singapore, 2021; pp. 467–474. [Google Scholar]
  83. Amponsah, A.A.; Han, F.; Osei, K.J.; Bonah, E.; Ling, Q.H. An improved multi-leader comprehensive learning particle swarm optimisation based on gravitational search algorithm. Connect. Sci. 2021, 33, 803–834. [Google Scholar] [CrossRef]
  84. Miao, Z.; Yong, P.; Mei, Y.; Quanjun, Y.; Xu, X. A discrete PSO-based static load balancing algorithm for distributed simulations in a cloud environment. Future Gener. Comput. Syst. 2021, 115, 497–516. [Google Scholar] [CrossRef]
  85. Zhu, D.; Huang, Z.; Xie, L.; Zhou, C. Improved Particle Swarm Based on Elastic Collision for DNA Coding Optimization Design. IEEE Access 2022, 10, 63592–63605. [Google Scholar] [CrossRef]
  86. Fu, K.; Cai, X.; Yuan, B.; Yang, Y.; Yao, X. An Efficient Surrogate Assisted Particle Swarm Optimization for Antenna Synthesis. IEEE Trans. Antennas Propag. 2022, 70, 4977–4984. [Google Scholar] [CrossRef]
  87. Song, B.; Wang, Z.; Zou, L. An improved PSO algorithm for smooth path planning of mobile robots using continuous high-degree Bezier curve. Appl. Soft Comput. 2021, 100, 106960. [Google Scholar] [CrossRef]
  88. Shaheen, M.A.M.; Hasanien, H.M.; Alkuhayli, A. A novel hybrid GWO-PSO optimization technique for optimal reactive power dispatch problem solution. Ain Shams Eng. J. 2021, 12, 621–630. [Google Scholar] [CrossRef]
  89. Suman, G.K.; Guerrero, J.M.; Roy, O.P. Optimisation of solar/wind/bio-generator/diesel/battery based microgrids for rural areas: A PSO-GWO approach. Sustain. Cities Soc. 2021, 67, 102723. [Google Scholar] [CrossRef]
  90. Xu, L.; Huang, C.; Li, C.; Wang, J.; Liu, H.; Wang, X. Estimation of tool wear and optimization of cutting parameters based on novel ANFIS-PSO method toward intelligent machining. J. Intell. Manuf. 2021, 32, 77–90. [Google Scholar] [CrossRef]
  91. Su, S.; Zhai, Z.; Wang, C.; Ding, K. Improved fractional-order PSO for PID tuning. Int. J. Pattern Recognit. Artif. Intell. 2021, 35, 2159016. [Google Scholar] [CrossRef]
  92. Zhong, Q.Y.; Qian, Q.; Fu, Y.F.; Yong, F.E.N.G. Survey of particle swarm optimization algorithm for association rule mining. J. Front. Comput. Sci. Technol. 2021, 15, 777–793. [Google Scholar]
  93. Kalyani, G.; Chandra, S.R.M.V.P.; Janakiramaiah, B. Particle swarm intelligence and impact factor-based privacy preserving association rule mining for balancing data utility and knowledge privacy. Arab. J. Sci. Eng. 2018, 43, 4161–4178. [Google Scholar] [CrossRef]
  94. Jangra, S.; Toshniwal, D. VIDPSO: Victim item deletion based PSO inspired sensitive pattern hiding algorithm for dense datasets. Inf. Process. Manag. 2020, 57, 102255. [Google Scholar] [CrossRef]
  95. Suma, B.; Shobha, G. Fractional salp swarm algorithm: An association rule based privacy-preserving strategy for data sanitization. J. Inf. Secur. Appl. 2022, 68, 103224. [Google Scholar]
  96. Krishnamoorthy, S.; Sadasivam, G.S.; Rajalakshmi, M.; Kowsalyaa, K.; Dhivya, M. Privacy preserving fuzzy association rule mining in data clusters using particle swarm optimization. Int. J. Intell. Inf. Technol. (IJIIT) 2017, 13, 1–20. [Google Scholar] [CrossRef]
  97. Guo, S.W.; Meng, Y.Y.; Chen, S.L. Application of improved PSOGM algorithm in dynamic association rule mining. Comput. Eng. Appl. 2018, 54, 160–165. [Google Scholar]
  98. Kou, Z.C. Binary particle swarm optimization-based association rule mining for machine capabilities and part features. J. Univ. Jinan Sci. Technol. 2019, 33, 381–388. [Google Scholar]
  99. Kaur, S.; Goyal, M. Fast and robust hybrid particle swarm optimization tabu search association rule mining (HPSO-ARM) algorithm for web data association rule mining (WDARM). Int. J. Adv. Res. Comput. Sci. Manag. Stud. 2014, 2, 448–451. [Google Scholar]
  100. Gangurde, R.A.; Kumar, B. Next Web Page Prediction using Genetic Algorithm and Feed Forward Association Rule based on Web-Log Features. Int. J. Perform. Eng. 2020, 16, 10–18. [Google Scholar] [CrossRef]
  101. Jiang, H.; Kwong, C.K.; Park, W.Y.; Yu, K.M. A multi-objective PSO approach of mining association rules for affective design based on online customer reviews. J. Eng. Des. 2018, 29, 381–403. [Google Scholar] [CrossRef]
  102. Lanzarini, L.; Villa, M.A.; Aquino, G.; De, G.A. Obtaining classification rules using lvqPSO. In Proceedings of the ICSI 2015: Advances in Swarm and Computational Intelligence, Beijing, China, 25–28 June 2015; Springer: Cham, Switzerland, 2015; pp. 183–193. [Google Scholar]
  103. Priya, S.; Selvakumar, S. PaSOFuAC: Particle Swarm Optimization Based Fuzzy Associative Classifier for Detecting Phishing Websites. Wirel. Pers. Commun. 2022, 125, 755–784. [Google Scholar] [CrossRef]
  104. Li, S.Y.; Zhou, L.; Liu, H. Hazard ldentification Algorithm Based on Deep Extreme Learning Machine. Comput. Sci. 2017, 44, 89–94. [Google Scholar]
  105. She, Y.L.; Zhou, L. Hazard ldentification Algorithm Based on Improved Online Sequential Extreme Learning Machine. Comput. Technol. Dev. 2018, 28, 72–77. [Google Scholar]
  106. Ripon, S.; Sarowar, G.; Qasim, F.; Cynthia, S.T. An Efficient Classification of Tuberous Sclerosis Disease Using Nature Inspired PSO and ACO Based Optimized Neural Network. In Nature Inspired Computing for Data Science; Springer: Cham, Switzerland, 2020; pp. 1–28. [Google Scholar]
  107. Choubey, D.K.; Kumar, P.; Tripathi, S.; Kumar, S. Performance evaluation of classification methods with PCA and PSO for diabetes. Netw. Model. Anal. Health Inform. Bioinform. 2020, 9, 1–30. [Google Scholar] [CrossRef]
  108. Karsidani, S.D.; Farhadian, M.; Mahjub, H.; Mozayanimonfared, A. Prediction of Major Adverse Cardiovascular Events (MACCE) Following Percutaneous Coronary Intervention Using ANFIS-PSO Model. BMC Cardiovasc. Disord. 2022, 22, 389. [Google Scholar]
  109. Mangat, V.; Vig, R. Novel associative classifier based on dynamic adaptive PSO: Application to determining candidates for thoracic surgery. Expert Syst. Appl. 2014, 41, 8234–8244. [Google Scholar] [CrossRef]
  110. Raja, J.B.; Pandian, S.C. PSO-FCM based data mining model to predict diabetic disease. Comput. Methods Programs Biomed. 2020, 196, 105659. [Google Scholar] [CrossRef]
  111. Alkeshuosh, A.H.; Moghadam, M.Z.; Al, M.I.; Abdar, M. Using PSO algorithm for producing best rules in diagnosis of heart disease. In Proceedings of the IEEE 2017 International Conference on Computer and Applications (ICCA), Doha, United Arab Emirates, 6–7 September 2017; pp. 306–311. [Google Scholar]
  112. Mao, J.; Wang, H.Y. Association rule optimization for detection algorithm of factors inducing heart disease. Control. Eng. China 2017, 24, 1286–1290. [Google Scholar]
  113. Shao, P.; Wu, Z.; Peng, H.; Wang, Y.; Li, G. An Adaptive Particle Swarm Optimization Using Hybrid Strategy. In Proceedings of the International Symposium on Intelligence Computation and Applications, Guangzhou, China, 20–21 November 2021; Springer: Singapore, 2017; pp. 26–39. [Google Scholar]
  114. Liang, B.; Zhao, Y.; Li, Y. A hybrid particle swarm optimization with crisscross learning strategy. Eng. Appl. Artif. Intell. 2021, 105, 104418. [Google Scholar] [CrossRef]
  115. Rostami, M.; Forouzandeh, S.; Berahmand, K.; Soltani, M. Integration of multi-objective PSO based feature selection and node centrality for medical datasets. Genomics 2020, 112, 4370–4384. [Google Scholar] [CrossRef] [PubMed]
  116. Zhong, Q.Y.; Qian, Q.; Fu, Y.F. Association rule mining based on multi-strategy BPSO algorithm. Bull. Sci. Technol. 2021, 37, 40–46. [Google Scholar]
  117. Dai, Y.B. Multi-Objective Particle Swarm Algorithm Based on Quasi-Circular Mapping. Appl. Res. Comput. 2021, 38, 3673–3677. [Google Scholar]
Figure 1. The research process of ARM.
Figure 1. The research process of ARM.
Electronics 11 03044 g001
Figure 2. Solution of NARM.
Figure 2. Solution of NARM.
Electronics 11 03044 g002
Figure 3. Association rule mining process.
Figure 3. Association rule mining process.
Electronics 11 03044 g003
Figure 4. Binary conversion of data.
Figure 4. Binary conversion of data.
Electronics 11 03044 g004
Figure 5. Data item representation of the Michigan method.
Figure 5. Data item representation of the Michigan method.
Electronics 11 03044 g005
Figure 6. Location representation of the Michigan method rule.
Figure 6. Location representation of the Michigan method rule.
Electronics 11 03044 g006
Figure 7. Application of particle swarm optimization algorithms.
Figure 7. Application of particle swarm optimization algorithms.
Electronics 11 03044 g007
Table 1. ARM, based on evolutionary algorithm.
Table 1. ARM, based on evolutionary algorithm.
AlgorithmsAuthorsTechnologyAdvantagesDisadvantages
DEWang et al. [30]Adaptive adjustment F and CrEnhanced algorithmic global search capabilityHigh memory overhead
Altay et al. [31]DE-SCA, Multi-target,Performs well on datasets with few attributes and many instancesHigh algorithmic complexity
Guan et al. [32]HGDEIncreased swarm diversity and high stabilityUnable to handle high-dimensional data items
GAMenaga et al. [33]GAPPARMHigh practicalityEasy to cause data loss
Lin et al. [34]Clustering the dataHigh accuracy and performanceSingle objective
Neysiani et al. [35]Identify similarity between itemsets by association rulesHigh-quality rulesLong execution time
BBOGiri et al. [36]LGBBO-RuleMinerHigh accuracy and a simple algorithmOnly for single target problems
Arani et al. [37]K-means, Bayesian learningReduce the delay, SLA violation ratio, cost, and energy consumptionHigh CPU usage
CSAMo et al. [38]DMARICAShort execution time, Highly scalableLess integrity of generation rules
AISHusain et al. [39]CLONALGHigh accuracyAccuracy is positively correlated with the number of iterations
Danilo et al. [40]CLONALG-GAPerforms better in sparse datasetsNot suitable for dense datasets
Table 2. ARM, based on swarm intelligence algorithm.
Table 2. ARM, based on swarm intelligence algorithm.
AlgorithmsAuthorsTechnologyAdvantagesDisadvantages
PSOTyagi et al. [41]Multi-target, MOPSO-ARMPerforms well on sparse dataCalculated overload, low efficiency
Kuo et al. [42]Adaptive Archive Grid, MOPSOAutomatically finds the best interval between datasets; no data preprocessing is required.Fewer targets to consider
Baro et al. [14]Guided search strategyHigh-quality rules and short calculation timesLower average fitness values
Agarwal et al. [43]Multi-objectiveBalancing global and local search capabilitiesNot applicable to quantitative association rules
Moslehi et al. [44]GA-PSONo predefined minimum support and confidence levels are requiredGenerate redundant rules
CSODevi et al. [45]HGWCSO-ETSVM, Min-MaxHigh accuracy and precisionReduced overall system performance
AMOSon et al. [46]ARM-AMOReduces the time and memory required for frequent itemset generationLower quality rules
BATKamel et al. [27]MSBARM, loop strategyRule quality is higher than other ways algorithmUnable to handle large databases
FAPradeep et al. [47]BFFO-TA, Feature SelectionNo redundant rules are createdNot applicable to multi-objective problems
GWOPazhaniraa et al. [48]BGWO-HUILow time complexityHigh memory overhead
Yildirim et al. [49]Adaptive multi-target intelligent search, MOGWOFor discrete, quantitative, and mixed datasets with high comprehensibilityNot suitable for distributed datasets
Chantar et al. [50]SVM, Elite Cross, BGWOHigh accuracyHigh time complexity
ABCAkbar et al. [51]ABC4ARHHighly efficient and stableLess scalability and higher complexity
HHOTurabieh et al. [52]HHO-KNNEnhanced ability to think outside the local optimumLess stability
ALODong et al. [53]ALO-ARMThe optimized search process has high efficiencyNot suitable for large datasets
Table 3. ARM, based on the physics-inspired algorithm.
Table 3. ARM, based on the physics-inspired algorithm.
AlgorithmsAuthorsTechnologyAdvantagesDisadvantages
SANawaz et al. [54]HUIM-SAHUIM-SA varies linearly with the number of iterationsLack of mutation mechanisms
Holman et al. [55]GA-SA, Confusion MatrixHigh accuracyAccuracy depends on dataset size
TSChou et al. [56]QTSObtain more rulesAlgorithm performance degrades when the dataset is too small
CROAbir et al. [57]CROLow algorithm complexity; no need to specify minimum supportLess stability, calculated overload
GSATaradeh et al. [58]HGSA, Crossover, variationFast convergence and high-quality rulesNot applicable to mixed datasets
Table 4. ARM, based on other heuristic algorithms.
Table 4. ARM, based on other heuristic algorithms.
AlgorithmsAuthorsTechnologyAdvantagesDisadvantages
FAEva et al. [59]Combining k-means for web data clusteringFaster convergence; better performance at the minimum distance; shorter computation timeThe placement strategy of the initial center of mass needs to be further optimized
MATing et al. [24]Optimizing the affiliation function in FARMImprove the search capability of the algorithm and obtain high fuzzy supportLess scalability
SCALaith et al. [60]SCAGAMaximum classification accuracy with minimal attributesThe imbalance between global search capability and local search capability
BSODjenouris et al. [61]HBSO-TSShort execution timePoor results in large datasets
Ma et al. [62]PLBSOParallelized processing with low computational costsLess performance when dealing with quantitative rules
Youcef et al. [63]GBSO-MinerAdaptable to large text and graphical databasesThread divergence exists
EOMalik et al. [64]DEOA-CRMHigh-quality rules and highly interpretable algorithmsPoor performance in some datasets
Table 5. An objective analysis of the PSO algorithm.
Table 5. An objective analysis of the PSO algorithm.
ObjectiveThesis
Convergence[66,67,73,75,79,80,82,83,86,87,91,113]
Versatility[66,68,73,81,83,84,87,90,113,114,115,116,117]
High-dimensional data[70,71,72,115,116,117]
Result accuracy[46,73,81,84,116,117]
Balance exploration and utilization[67,69,74,77,78,80,82,85,86,114]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, G.; Wang, T.; Chen, Q.; Shao, P.; Xiong, N.; Vasilakos, A. A Survey on Particle Swarm Optimization for Association Rule Mining. Electronics 2022, 11, 3044. https://doi.org/10.3390/electronics11193044

AMA Style

Li G, Wang T, Chen Q, Shao P, Xiong N, Vasilakos A. A Survey on Particle Swarm Optimization for Association Rule Mining. Electronics. 2022; 11(19):3044. https://doi.org/10.3390/electronics11193044

Chicago/Turabian Style

Li, Guangquan, Ting Wang, Qi Chen, Peng Shao, Naixue Xiong, and Athanasios Vasilakos. 2022. "A Survey on Particle Swarm Optimization for Association Rule Mining" Electronics 11, no. 19: 3044. https://doi.org/10.3390/electronics11193044

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop