You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

15 October 2022

An Improved Gray Wolf Optimization Algorithm with a Novel Initialization Method for Community Detection

,
,
,
,
and
1
Software College, Yunnan University, Kunming 650504, China
2
School of Artificial Intelligence, Hohai University, Nanjing 211100, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Optimisation Algorithms and Their Applications

Abstract

Community discovery (CD) under complex networks is a hot discussion issue in network science research. Recently, many evolutionary methods have been introduced to detect communities of networks. However, evolutionary optimization-based community discovery still suffers from two problems. First, the initialization population quality of the current evolutionary algorithm is not good, resulting in slow convergence speed, and the final performance needs to be further improved. Another important issue is that current methods of CD have inconsistent network detection performance at different scales, showing a dramatic drop as the network scale increases. To address such issues, this paper proposes an algorithm based on the novel initial method and improved gray wolf optimization (NIGWO) to tackle the above two problems at the same time. In this paper, a novel initialization strategy is proposed to generate a high-quality initial population and greatly accelerate the convergence speed of population evolution. The strategy effectively fused the elite substructure of the community and different features based on the dependency and other features among nodes. Moreover, an improved GWO is presented with two new search strategies. An improved hunting prey stage is proposed to retain the excellent substructures of populations and quickly improve the community structure. Furthermore, new mutation strategies from node level to community level are designed in an improved encircling prey stage. Specifically, boundary nodes are mutated according to a proposed function to improve the search efficiency and save the computation assumption. Numerous experiments have proven our method obtains more excellent performance in most networks compared with 11 state-of-the-art algorithms.

1. Introduction

Community discovery (CD) is to find closely related nodes in a complex network and attribute these nodes to different communities. The community is highly dense within a complex network and highly sparse between communities. “Highly dense” means that the ratio of the number of edges to the number of distant nodes is relatively high. A simple explanation for this is the ratio of edges to nodes. However, this is not universal. A more general indicator of community density is the intralink density [1]. The higher the intralink density, the higher the density of the community, which indicates that the community is good. One of the tasks of community discovery is to find such good communities. The CD has excellent applications in criminology, such as fraud detection, crime recognition, criminal activity detection, and robot detection [2]. For example, Qi et al. applied the CD to the evolution of the GitHub software ecosystem and derived the evolution process of the survival and extinction of some popular software or programming languages under the software ecosystem in the GitHub software ecosystem [3]. Tao et al. proposed a Chinese medicine CD algorithm to discover the community structure of the Chinese medicine network structure; it effectively provides technical support to traditional Chinese medicine in the process of diagnosis and treatment [4]. Hence, finding important community structures in large-scale social networks is meaningful work.
In the past years, many methods for the CD have been proposed, such as node importance-based (NIB) methods [5,6,7], good initial substructure-based (GISB) methods [8,9,10], non-negative matrix factorization-based (NMFB) methods [11,12,13,14], deep learning-based (DLB) methods [15,16,17,18], and population intelligence-based (PIB) methods [1,14,19,20]. The methods of NIB, which find the existing dependency and apply mathematical operations between nodes in the graph, or the methods of GISB, which find good initial substructure, can divide the network faster and more efficiently, so they are chosen by more researchers [21,22].
The methods of NIB give the degree of association between nodes by ranking [7], since the importance of the node’s neighbor nodes is not the same for the node. This method provides a good basis for network division to a certain extent, but it ignores many small, tightly connected communities in the network. In addition, there are many small good initial substructures in complex networks. Utilizing these large numbers of substructures can divide closely related nodes into the same community [23]. However, there is currently little work combining node importance and good initial substructure for the CD. Therefore, we further attempt to fuse the characteristics of the node importance and the good initial substructure to solve the CD more efficiently. In addition, since complex networks with complicated and diversified structures often have various nonlinear features, the authors propose a non-negative matrix factorization-based method to deal with nonlinear features [11]. The other one to solve nonlinear problems is to employ swarm intelligence algorithms [24]. Moreover, it is worth mentioning that swarm intelligence algorithms are also widely adopted to deal with complex optimization problems. Until now, swarm intelligence-based CD is still a hot topic, and many methods are constantly presented in high-prestige journals and conferences. For example, Siwei et al. proposed a particle swarm optimization method (PSO) which is for process monitoring and the selection of Non-Zero Loadings [25]. John et al. used the intelligent search strategy for disjoint principal component analysis to solve the interpretation of the vector subspace when reducing the dimension in multivariate analysis [26]. Taiyong et al. combined pyramid PSO (PPSO) with novel competitive and cooperative strategies to achieve superior performance [27]. One of the latest directions of swarm intelligence algorithms is to combine neural networks to solve problems in the field of complex artificial intelligence. For example, Eren et al. proposed an architecture based on particle swarm optimization combined with the recurrent neural network to deal with the prediction problem, which solves the problem of gradient explosion or disappearance in the recurrent neural network [28]. Since the gray wolf optimization (GWO) has a good global optimization ability, we employ the GWO to handle the CD.
Different from the existing swarm intelligence-based methods, we propose a hybrid algorithm named NIGWO by integrating improved GWO and novel initialization strategies to solve the above problems. The new initialization method is proposed to combine excellent initial substructure and node importance. New hunting operators and mutation strategies are leveraged in GWO to improve the quality of the populations. In this paper, the contributions can be summarized as follows:
  • We propose a new initialization method that combines the importance of node and elite community structures to generate an initial elite population quickly, which effectively improves the quality of solutions. The elite substructures are obtain by utilizing motifs and their intersections. The importance of node is generated according to matrices depending on dependency and the other features of nodes.
  • The GWO is adopted in discrete spaces of the CD. We design an improved hunting prey behavior of GWO to maintain the good community substructure and evolve the population to improve the performance of the CD.
  • A multi-scales mutation operator is designed to improve the encircling prey behavior of GWO from node level to community level. The multi-scales mutation includes boundary node-level mutation and community-level mutation. Specifically, the boundary nodes and small communities are mutated on the basis of the proposed formula to effectively improve the search efficiency and reduce the computational cost.
  • The effectiveness of the algorithm on small and large datasets is demonstrated by extensive experiments on artificial networks and real-world networks.
The rest of this paper is arranged as follows: in Section 2, the related work and methodology are introduced. NIGWO is proposed in Section 3. The effectiveness of the NIGWO is demonstrated by comparison with 11 other algorithms using 11 real-world networks and two goups of synthetic networks in Section 4. There is a summary of this paper in Section 5.

3. Proposed NIGWO

3.1. Problem Definition

For complex network datasets, the relationship between entities can be abstracted into a graph structure optimization problem. The data structure of a graph can be expressed as G = V , E , where V and E are a collection of entities and a collection of edges, respectively. The topology information of the graph is composed by the node i and its neighbor node vector. The community structure can be represented as a set of communities C = C 1 , C 2 , , C k , and each community C j C is represented as a network subgraph G s = V s , E s . For clarity, we summarize all frequently used variables in Table 1.
Table 1. Symbolic representation of the CD.
The CD with graph structure can be modeled as an optimization problem as follows:
m a x y = F ( X ) , X = [ X 1 , X 2 , , X p o p ] , X j = [ x 1 , x 2 , , x n ] , j = 1 , 2 , , p o p
where F ( · ) is the objective function. Modularity is utilized to measure the strength of the community structure, that is, “high cohesion, low coupling”. Thence, to find a good community structure, the CD can be modeled as a modularity optimization problem as follows:
Q ( G , i n d i v i d u a l ) = c = 1 k l c m ( d c 2 m ) 2
where i n d i v i d u a l denotes a representation with network division. We can obtain the current network division according to i n d i v i d u a l . It is worth noting that the degree of the node here is not in a community but in the entire network.

3.2. Solutions Representation

The solution encoding method is generally encoded as an individual that uses the LAR method [1,20]. Each solution is represented as an n-dimensional integer vector X j . The i-th dimensional value x i of the solution X j is denoted as the neighbor node of the node n i . Such an encoding scheme often results in a low-quality initialization, which slows down the convergence of the population evolution. Therefore, we propose a promising solution representation to improve the problems encountered above in Figure 1. As shown in Figure 1, node n i and node n j will belong to the same community if x i and x j have the same value. Nodes in different colors indicate that the nodes do not belong to the same communities. The i-th dimensional value x i of the solution X j is denoted as the label of the community in which node n i is located.
Figure 1. Two examples of solution representation.

3.3. Gray Wolf Optimization with a Novel Initialization and Two Improved Stages (NIGWO)

3.3.1. The Framework of NIGWO

In the GWO, the behavior of the elite gray wolves increases the global search ability and the diversity of the population. However, the GWO still suffers from the problems of being prone to local optima and slow convergence. Hence, the proposed NIGWO makes three significant improvements to the standard GWO algorithm, the framework of which is shown in Figure 2. The first improvement is a fast elite population initialization strategy to greatly accelerate the convergence speed of population evolution. Secondly, an improved hunting prey stage to share the excellent structure of the first three optimal solutions α , β , δ of the contemporary population to improve other gray wolf individuals ω is presented. Thirdly, a multi-scales mutation strategy is proposed in the encircling prey stage to further improve search capability and increase the diversity of the gray wolf population. The specific NIGWO pseudocode is shown in Algorithm 1.
Algorithm 1: NIGWO
Input:
G , p o p , p m u , M a x I t e r
Output:
The optimal individuals
 1:
Initialize the gray wolf population X = [ X 1 , X 2 , , X p o p ] according to the fast elite population Initialization
 2:
Calculate the fitness of each gray wolf
 3:
t 0
 4:
while t < M a x I t e r do
 5:
    Get X = [ X α , X β , X δ , X ω ] by sorting the population X
 6:
    Get e l i t e = [ X α , X β , X δ ] by X
 7:
    while  2 < i < p o p  do
 8:
       Hunting operation by e l i t e for gray wolf individual ω
 9:
    end while
10:
   Execute a multi-scales mutation strategy
11:
   tt + 1
12:
end while
13:
Return X α
Figure 2. The framework of the NIGWO.

3.3.2. Fast Elite Population Initialization

Two strategies are utilized to improve the quality and diversity of the initial population.
The first solutions are generated as the adjacent node-based initialization method [37]. In this initialization method, individual nodes are encoded using the neighboring node, which is selected randomly from the current node; then, the current node and its neighboring node are aggregated in the same community by decoding.
The second solutions are composed of solutions obtained by a fast elite initialization method. First, in the elite structure phase, a motif-based strategy is presented to find the elite structure of the community. Moreover, in the fast sorting phase, the remaining nodes are quickly sorted and labeled according to the existing substructures.
In the elite structure phase, our methods obtain sub-stable structures to decrease the search space by merging motifs structures. An example of the elite structure phase is presented in Figure 3. The sub-stable structures are the different set of motifs that is the stable structure in the community. Their definitions are given as follows:
Figure 3. The elite structure phase. (a) A graph structure of the network. (b) Generate motifs according to Defition 1. (c) Calculate sub-stable structure by Defition 2. (d) Nodes that have not yet been aggregated. (e) Preliminary division of the network.
Definition 1
(Motifs). Network motifs are dense and interconnected subgraphs occurring in complex networks at significantly higher numbers than those in randomized networks [23]. In this paper, only the three-node motifs are calculated, since the computation complexity of the three-node motifs is relatively low. In our work, a motif is defined as G s o l i d . Let G s o l i d be a triple ( x , y , z ) where x , y , z are three nodes of G and x , y G . a d j [ z ] , then G s o l i d is a motif. It is the basis of the sub-stable community.
Definition 2
(Sub-stable community). A sub-stable community G s e c _ s o l i d is a community composed of two motifs, and the intersection between two motifs is greater than 1. It can be formalized as G s e c _ s o l i d = G s o l i d 1 G s o l i d 2 , s . t . G s o l i d 1 G s o l i d 2 > 1 . G s o l i d 1 and G s o l i d 2 are represented as the motifs.
In Figure 3, the initial graph network first generates three motifs and merges them based on Definition 2. The network is initially divided by the above process. In the next phase, these nodes n 0 , n 4 and n 9 of the undivided community are divided into communities.
In the fast sorting phase, the remaining nodes are sorted and divided into communities according to the obtained elite structure. Inspired by the fast clustering algorithm (FCA) [21], a series of operations are defined on the graph of the community, and we extracted several features from it. The nodes are ordered according to a set of features by utilizing mathematical operations on the graph. Set the number of remaining nodes is q. The effect matrix E = { E i i = 0 , 1 , , q } . We calculate the priority of these nodes by E i as follows:
E i = S i j = 1 n K j × S j K j = 0 , i f m d ( i , j ) = 0 1 , i f m d ( i , j ) 0
where the sum matrix S is an n × 1 matrix, and the value of each row represents the sum S i of the degree of the adjacent nodes of the node i [ 0 , n ) , m d denotes the neighborhood degree matrix. The effect matrix E is an n × 1 matrix. It is important to note that the largest E i is for a node that has not yet been aggregated and the clustering process of the remaining nodes starts with this node. Let n i and n j denote two nodes in the graph so that an edge between n i and n j indicates the existing dependency between them. We cluster nodes that have not yet been aggregated by performing some operations on the dependence matrix m d and extracting several features from it. The node importance array i n d e x is obtained in descending order according to the effect matrix E, and the aggregation of the remaining nodes is completed by i n d e x . If each node is adjacent to the previous elite sub-community, it will be merged into this community; otherwise, it will be assigned by the label of the neighbor node. As shown in Figure 4, the nodes are sorted by E i to obtain the processing order, and then, their quadratic neighboring nodes are found to obtain corresponding communities C o .
Figure 4. The fast sorting phase. Ne (Ne1) represents the neighbor nodes of the previous nodes. Co represents the neighbor nodes shared by the previous nodes.
Finally, the uninitialized nodes will be divided into communities based on the elite substructure, and the C o values obtained from the above two phases are shown in Figure 5. Nodes n 4 , n 0 , and  n 9 will be divided into the communities where nodes n 8 , n 3 , and  n 7 are located, respectively. The pseudocode of the initialization method is shown in Algorithm 2.
Algorithm 2: Fast elite population Initialization
Input:
G , p o p
Output:
Population X
 1:
Find motifs of G by Definition 1
 2:
Get sub-stable communities by Definition 2
 3:
Get effect matrix E by Equation (4)
 4:
Get node sorted array i n d e x according to the effect matrix E
 5:
i 0
 6:
while i < p o p / 2 do
 7:
   Select sub-stable communities to update node label of X i
 8:
    z nodes without labels according to the order of i n d e x .
 9:
    j 0
10:
  while j is less than the length of z do
11:
     Get the corresponding node y by z [ j ]
12:
      k 0
13:
     while k is less than the length of N s ( y )  do
14:
         s e t = N s ( N s ( y ) [ k ] ) N s ( y )
15:
        if  s e t not n u l l  then
16:
            Update the label of node z to the label of s e t
17:
        else
18:
            Update the label of node z by the label of the neighbor node
19:
        end if
20:
        k k + 1
21:
     end while
22:
     j j + 1
23:
   end while
24:
   i i + 1
25:
end while
26:
i i p o p / 2
27:
while i i < p o p do
28:
   Initialize the gray wolf individual by the adjacent node-based method
29:
end while
30:
Return population X
Figure 5. Population initialization based on the elite structure phase and the fast sorting phase. (a) Preliminary division of the network. (b) Initial partition of the network.

3.3.3. Improved Hunting Stage

Based on the characteristics of the CD problem, we design a new hunting stage to improve the hunting behavior of the gray wolf. In the classical GWO, each individual explores prey locations through elite wolfs (the leaders α , β , and δ ), and the position of the prey is calculated according to Equation (2). However, this strategy cannot be directly applied to the CD, since CD is a discrete space search optimization problem. We leverage the elite substructure among the population to reduce the search space and improve search efficiency. Therefore, we take the intersection of α , β , and δ as the prey position, and then the rest of the individuals update their positions according to the prey position. In this way, a good substructure among the elite solutions is obtained. Specifically, the intersection of α , β , and δ is shown as follows:
s u b c = { c s | c s ( C α C β C δ ) }
Note that C α , C β , and C δ are the communities set of α , β , and δ , respectively. The c s represents the common subset of C α , C β , and C δ . We perform the intersection operation on the communities of C α , C β , and C δ . If the intersection of α , β , and δ is empty, we select another gray wolf to replace ω until their intersection is not empty.
The new solution is obtained by updating the old solution with s u b c . Specifically, by randomly selecting a node i from s u b c , the other nodes in s u b c are assigned the same community as node i. The updating process is given as follows:
x j t + 1 = x i t , i f ( i , j ) s u b c x j t , e l s e
In this way, the elite substructure is fused into the community structure of the old solution and is effectively used to improve the search efficiency.
To further illustrate the improved hunting stage, we use an example to show the main idea of the improved hunting stage in Figure 6. The s u b c generation process is given in Figure 6a. The α , β , and δ are decoded separately to obtain the community sets. There, the high-quality network subsets are obtained as s u b c among the three elite individuals. In Figure 6a, subc is {1, 3} and {6, 7, 8} as it the intersect of the C α , C β , and C δ . Then, in Figure 6b, the good substructure of s u b c has remained in the new individual. Thus, the search space is greatly reduced by moving non-elite individuals closer to elite individuals and maintaining quality subsets of the community in the improved hunting phase.
Figure 6. The improved hunting stage. (a) is the generation process of s u b c , and (b) is the process by which ω wolf hunt for prey based on s u b c .

3.3.4. Improved Encircling Stage

Encircling prey behavior is essentially a random method-based mutation behavior from fast to slow. Therefore, we propose multi-scales mutation strategies to improve the quality of solutions and increase the diversity of the wolf population in the encircling stage. Boundary nodes and sub-community are utilized to mutate the individual on different scales.
Firstly, the boundary nodes are selected to mutate as Liu et al. proposed the DECS algorithm to utilize the number of edges that boundary nodes to different communities to mutate and got a better performance [22]. However, the boundary node mutation process in DECS will not necessarily improve the performance of the CD when the two communities connected to the boundary node are complex. For example, the boundary node mutation process is shown in Figure 7. It shows that after the mutation of boundary node 5, the modularity value varies from 0.2149 to 0.3926. This improves the performance of the CD to a certain extent. However, in another case, as shown in Figure 7b, the initial value of modularity is 0.1667, and the value of modularity after boundary node mutation is 0.1235.
Figure 7. Boundary node mutation in DECS. (a) The performance of Ind1 has been improved. (b) The performance of Ind2 has been degraded.
The detailed definition of a boundary node can be found in [22]. We have improved the judgment conditions for mutation in the community mutation strategy. We adopt the Q calculation method of EP-WOCD [20], which defines the Q value from the perspective of the community. Different from his study, we propose a Δ Q method for boundary node mutation and improve the calculation method of the change value of Q. The derivation of the simplified formula for the change value of Q is as shown in Equation (5).
Δ Q = Q l a t t e r Q f o r m e r   = c = 1 k [ l c l m ( d c l 2 m ) 2 ] c = 1 k [ l c f m ( d c f 2 m ) 2 ]
where l c l and l c f represent the total number of community edge connections after and before the boundary nodes mutation, respectively. Since l c and d c are not changed except for the two involved communities in the process of mutation, the calculation of Δ Q can be updated as:
Δ Q = ( Δ l 1 + Δ l 2 ) / m ( Δ d 1 2 + Δ d 2 2 ) / 4 m 2
where Δ l 1 ( Δ l 2 ) and Δ d 1 2 ( Δ d 2 2 ) represent the changes in the number of edges and squares of the sum of nodes degree in the community, respectively. The change of the Q value is affected by the number of edges of two communities. Some large-scale networks have hundreds or thousands of communities. The time calculation cost can be greatly reduced by Δ Q > 0 since Δ Q > 0 only needs to calculate two communities. When Δ Q > 0 , the boundary node mutation is executed. Otherwise, it will remain unchanged. In the above example, we only need to calculate the change value of l k and the change value of d k square of the two communities. Therefore, we effectively improve the search efficiency and save the computation assumption by mutating boundary nodes through Δ Q .
Secondly, the community mutation operator is shown in Figure 8, in which the label of one community is changed to the label of its neighbor community. It can be observed that there exist a lot of small communities after the initialization. These small communities can be effectively integrated by the community mutation operator. In this way, the community mutation can be effectively utilized to improve the performance of the individuals.
Figure 8. Community mutation.
Obviously, these nodes in {9, 1, 0} in Figure 8 belongs to the same community. In the process of community mutation, these three nodes will find the community closest to them for community mutation and assimilate into the same community.

3.4. Time Complexity Analysis

As shown in Algorithm 1, the gray wolf population initialization is executed p o p times in step 1. The network has n nodes, so the total time is O ( n × p o p ) in population initialization. During steps 7–8, the hunting operation is performed for all gray wolf individuals except α , β , δ , since the encoding time of each individual is O ( n ) . Therefore, the total time cost of the improved encircling stage is O ( M a x I t e r × ( n × ( p o p 3 ) ) ) . In the improved encircling stage, which contains boundary node variation and community variation, the time for boundary node variation is mainly spent on finding boundary nodes, so the time complexity is O ( n ) . For community variation, the time complexity is O ( k + n + p o p ) because the number of communities in the network is k and the number of nodes in each community needs to be counted. The time complexity of the improved encircling stage is O ( M a x I t e r × ( k + 2 n + p o p ) ) in step 10. Therefore, this total time complexity is O ( p o p × n + M a x I t e r × ( n ( p o p 1 ) + k + p o p ) ) .

4. Experiment and Evaluation

4.1. Datasets and Comparison Algorithms

To evaluate the effectiveness of NIGWO in community discovery, we test and experiment on 11 different scales and different fields of real datasets and two groups of artificial networks. The artificial network used in the experiment is the LFR benchmark network [44]. The LFR network can generate the network to be tested in the investigation according to any configuration parameters. Eleven real data sets are used for experiments from Network Data (http://www-personal.umich.edu/~mejn/netdata/, accessed on 21 July 2022), SNAP [45], LINQS (https://linqs.soe.ucsc.edu/data/, accessed on 11 June 2022). There are six datasets: Karate, Football, Dolphins, Polbooks, Lesmis, and Netscience from Network Data. For the two networks, CA-GrQc and CA-HepTh in SNAP, it is used to evaluate the performance of the large-scale complex networks under different algorithms. The network datasets in LINQS are Cora, Citeseer, and Pebmud, respectively. The details of the eleven real data are in Table 2, where d m a x represents the maximum degree of nodes in the dataset, and d a v e represents the average degree of nodes, which takes an approximate integer value by rounding down.
Table 2. The details of 11 networks.
The performance of NIGWO is compared with four popular EA-based CD algorithms for complex networks, namely, Meta-LPAM+ [19], DMFWA [14], EP-WOCD [20], RMOEA [1]; and five popular NMF-based CD algorithms namely NSED [46], BigClam [47], MNMF [13], DANMF [12], and NMFGAAE [11]; and other CD algorithms, namely, DeepWalk [48], and LPA [34]. NMFGAAE is also a method based on NMF, which is a combination of the graph neural network and NMF-based method. The details of the eleven algorithms are in Table 3.
Table 3. The details of 11 algorithms.

4.2. Parameters Setting and Evaluation Metrics

The value of parameters has a great influence on the performance of the algorithm. There are three parameters that need to be specified before the NIGWO is executed, namely, the initial population size p o p , the number of iterations M a x I t e r , and the mutation probability p m u . In order to facilitate the comparison and analysis of the experiments, the parameters of the evolutionary algorithms used in the experiments are set uniformly and compared. The parameter settings are shown in Table 4. Significantly, our algorithm does not need to adjust the parameters, and there is no randomness. We set the population size of DMFWA, EP-WOCD, RMOEA, and NIGWO to 100 and the number of iterations to 40 uniformly. The other parameters are set by default in the original paper.
Table 4. Parameters setting of the compared evolutionary algorithms.
For real-world networks and synthetic networks, we use modularity and NMI as evaluation metrics, respectively. Modularity is defined in Equation (3). The NMI is defined as follows.
N M I ( A , B ) = 2 i = 1 k A j = 1 k B N i j log n · N i j N i N j i = 1 k A N i log N i n + j = 1 k B N j log N j n
where A is the network division obtained through CD, and B is the real division of the network. k A and k B represent the number of communities of A and B, respectively. n represents the dimension of the complex network, that is, the number of nodes. N i j represents the number of nodes that co-occur in community i and j. N i and N j are the number of nodes in community i and j, respectively.

4.3. Experiments on Real-World Networks

For real-world network datasets, some datasets have labels and some datasets have no labels. We make the modularity comparison of the twelve algorithms listed above for real-world networks, and the result is shown in Table 5.
Table 5. Modularity values comparison between algorithms.
In Table 5, it can seem that our algorithm has more obvious advantages than other algorithms. Especially on the Netscience network, NIGWO has better modularity than other algorithms. In these experiments, an interesting phenomenon is that the LPA algorithm performs poorly on small data. However, as the network size increases, LPA will instead perform well. For example, our method shows good competitiveness in Netscience, Citeseer, Ca-HepTh, and Pubmed networks. One of the main problems in CD is the inability of the existing algorithms to maintain consistently good performance as the network scales. However, the results of experiments clearly show that NIGWO can maintain a good community detection performance on datasets of different scales.
To evaluate the convergence of the experiments, we perform the population evolution comparison on two algorithms (RMOEA and NIGWO) on six networks. Figure 9 shows the changes in the modularity of the six datasets under the population evolution of the two algorithms. Obviously, NIGWO has better performance and faster convergence in most networks, such as Karate, Dolphins, Polbooks, Netscience, and Lesmis networks.
Figure 9. Modularity based on population evolution.
To verify the validity of our proposed multi-scales mutation, we performed an experimental comparison of different p m u values in Figure 10. It can be clearly observed that the performance decreases a lot when no mutation strategy is implemented. After implementing the mutation strategy, the performance is improved to a great extent. At the same time, we can find that the performance increases with a higher mutation probability, and the convergence is also incremental. The convergence and community detection performance reach the optimum when p m u is at 1. However, due to the time cost, the performance of NIGWO is finally close to the optimal value at 0.4. Therefore, we set the probability value of p m u at 0.4 to balance time and performance. In the last citeseer data, the value of modularity decreases when the variation probability becomes 1. This is because of the inability to jump out of the local optimum due to excessive variation, which is the reason why we finally set the variation probability to 0.4.
Figure 10. Comparison of Q in different p m u with four real-world networks.
To verify the validity of our initialization, we performed a comparison of the improved LAR initialization method of EP-WOCD (i.e., LAR Initialization here for convenience) in Figure 11. Since the time difference between small and large data is large, the comparison with small and large datasets separately is executed. We perform a comparison of Q and time consumption for four small datasets. The result is shown in Figure 11a,b. For the four small datasets Karate, Football, Dolphins, and Lesmis, our initialization method has significantly better performance on the small datasets and is only a small difference in time cost from LAR initialization. Similarly, on the large datasets Citeseer, Ca-GrQc, Ca-HeppTh, and Pubmed, although the performance of the LAR-based initialization method is improved, the improvement is limited. As Figure 11c,d show, our initialization method achieves better performance at about a similar time cost. In addition, higher quality initialized populations tend to allow the evolutionary algorithm to reach the optimal solution faster, speeding up the convergence of community detection.
Figure 11. Comparison of different initialization methods. (a) Comparison of Q for four small datasets. (b) Comparison of time for four small datasets. (c) Comparison of Q for four big datasets. (d) Comparison of time for four big datasets.

4.4. Experiments on LFR Networks

In our work, two groups of LFR networks are used to further evaluate NIGWO. The first group consists of networks with 500 nodes and the mixing parameter μ from 0.1 to 0.8, μ is a fraction that each node shares with the other nodes of the network. In addition, 1- μ is that each node shares a fraction of its links with the other nodes of its community [44]. The average degree d a v e = 5 , the maximum degree d m a x = 15 , the minimum community size C m i n = 20 , the maximum community size C m a x = 50 , the exponents of the power law distribution of node degrees τ 1 = 2 , and the power law exponent community sizes distribution τ 2 = 1 ; the second group is these networks where the number of nodes is from 1000 to 5000, and the μ value is fixed to 0.3, the minimum community C m i n = 50 , the maximum community size C m a x = 100 , the exponents of the power law distribution of node degrees τ 1 = 2 , the power law exponent community sizes distribution τ 2 = 1 , the average degree d a v e = 4 , and the maximum degree d m a x = 15 . A comparative experiment is completed for the second group with NMI, and the experimental results are shown in Table 6.
Table 6. NMI comparison for LFR with different nodes.
In Table 6, we can clearly see that our algorithm outperforms other algorithms at any time. In the worst case, our algorithm also improves the performance by 5.1%. Overall, the performance of NIGWO has been greatly improved compared to other algorithms.
Then, we compare the accuracy of six algorithms in the first group of the LFR which have different μ , and the experimental results are shown in Figure 12a. As shown in Figure 12a, the NMI value shows a decreasing trend with the increase of μ . The increase of μ will make the community structure more complex. In addition, the results of NIGWO have well NMI when compared with the other algorithms. As the value of μ increases, the NMI of RMOEA and DMFWA decreases sharply, while our proposed NIGWO method can be clearly seen to be decreasing slowly. When μ > 0.4, our method decreases smoothly and obviously has a more stable performance than Meta-LPAm+, DMFWA, and RMOEA.
Figure 12. Comparison on LFR networks. (a) Comparison of NMI at different μ . (b) Comparison of time at different number of nodes.
We further compare our experiments with other excellent experiments in terms of time. The results of experiments are obviously much faster than Meta-LPAm+, DMFWA, EP-WOCD, and DANMF in terms of operating efficiency, and they are only slower than RMOEA and NMFGAAE. RMOEA algorithmically treats a good community structure as a node, which leads to a large-scale reduction of large-scale complex networks, thereby greatly improving operating efficiency, so it has a very fast operating speed. This sacrifices some performance in exchange for time efficiency. In the end, our algorithm achieves a balance in terms of time complexity and performance.
As can be seen in Figure 12b, for the second group of the LFR, our algorithm is slightly longer in time complexity than the current excellent algorithm RMOEA. However, with the expansion of the network size, our algorithm requires more time, because the more complex the network, the more information is hidden, which requires a longer time for mining. We sacrifice time on complex networks, which brings a great improvement in modularity and NMI on complex networks.
As analyzed above, the advantage of our algorithm’s performance is that it can reach the optimal NMI and modularity on most networks, and it can achieve a high NMI and modularity. Our method also performs well on both small and large datasets.

4.5. Visualization of NIGWO

We apply our proposed method to divide the three datasets with real labels into communities and visualize them. The three datasets are the Karate network, Dolphins network, Football network.
The visualization is shown in Figure 13. In this figure, the communities detected by our algorithm are not consistent with the original communities, because some of the accuracies are sacrificed while improving the robustness of the network. For example, in the Karate dataset, the Q value of Karate is only 0.3715, but the detection performance is improved to 0.4198 by our algorithm. Similarly, in the Football network, the Q value of the real community is only 0.554, but the Q value reaches 0.6046 by NIGWO. In the Dolphins network, the Q value under the real community is only 0.3735, while the Q value can reach 0.5285 after NIGWO.
Figure 13. Community structure of three networks. (a) Ground-truth communities of Karate network. (b) Communities detected by NIGWO of Karate network. (c) Ground-truth communities of Dolphins network. (d) Communities detected by NIGWO of Dolphins network. (e) Ground-truth communities of Football network. (f) Communities detected by NIGWO of Football network.

5. Conclusions

In the paper, we proposed a new algorithm NIGWO to solve CD problems. In NIGWO, we designed a novel initialization method and two improved operators of GWO. The initialization method greatly enhances the quality of the initial population and improves the convergence speed by defining sub-stable communities from the network structure and diffusing the neighborhood according to various features of nodes. During this evolution, the improved hunting stage and improved encircling stage are further designed to enhance the efficiency of the algorithm and ensure the quality of the results on small and large datasets. The two operators have improved the search performance of GWO by utilizing the characteristics of the good substructure and boundary node.
Experiments show that our algorithm obtains higher-quality solutions for detecting communities in large-scale complex networks. Therefore, the increasingly complex network can be solved efficiently, and communities in social networks, biological networks, protein molecular networks, etc. can be detected effectively. A good initialization method is crucial to the swarm intelligence algorithm. We will design an effective initialization strategy to further improve the quality of the initialization population. At the same time, the features of the local community and paralleled framework can be investigated to reduce the dimension of complex networks and reduce time consumption in large-scale network problems.

Author Contributions

Conceptualization, Y.K. and Z.X.; methodology, Y.K. and Z.X.; software, Y.K., Z.X. and H.W.; validation, Y.K. and H.W.; formal analysis, Z.X., H.W. and Y.Y.; investigation, Z.X. and Y.Y.; resources, Y.K., Z.X. and H.W.; data curation, Z.X., K.P. and H.W.; writing—original draft preparation, Z.X.; writing—review and editing, Y.K., X.Y. and H.W.; visualization, Z.X., H.W. and X.Y.; supervision, Y.K., K.P. and H.W.; project administration, Y.K., Z.X. and H.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [National Natural Science Foundation of China] grant number [61762092], [Open Foundation of the Key Laboratory in Software Engineering of Yunnan Province] grant number [2020SE303], [Major Science and Technology Project of Precious Metal Materials Genome Engineering in Yunnan Province] grant number [2019ZE001-1, 202002AB080001-6], [Yunnan Provincial major science and technology: Research and Application of Key Technologies for Resource Sharing and Collaboration Toward Smart Tourism] grant number [202002AD080047], Postgraduate Scientific Research Innovation Project of Hunan Province.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: [http://www-personal.umich.edu/mejn/netdata/, https://linqs.soe.ucsc.edu/data, http://snap.stanford.edu/index.html] (accessed on 21 July 2022).

Acknowledgments

The authors are grateful to the reviewers for their insightful suggestions and comments, which helped us improve the presentation and content of the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, X.; Zhou, K.; Pan, H.; Zhang, L.; Zeng, X.; Jin, Y. A network reduction-based multiobjective evolutionary algorithm for community detection in large-scale complex networks. IEEE Trans. Cybern. 2018, 50, 703–716. [Google Scholar] [CrossRef] [PubMed]
  2. Karataş, A.; Şahin, S. Application areas of community detection: A review. In Proceedings of the 2018 International Congress on Big Data, Deep Learning and Fighting Cyber Terrorism (IBIGDELFT), Ankara, Turkey, 3–4 December 2018; pp. 65–70. [Google Scholar]
  3. Qing, Q.; Jian, C.; Yancen, L. The Evolution of Software Ecosystem in GitHub. J. Comput. Res. Dev. 2020, 57, 513. [Google Scholar]
  4. Yang, T.; Xie, J.; Zhao, J.; Dong, H.; Hu, K. Design and Application of Chinese Medicine Association Discovery Algorithm Based on Association Network and Hierarchical Clustering. Mod. Tradit. Chin. Med. Mater. Med.—Sci. Technol. 2020, 22, 1962–1968. [Google Scholar]
  5. Lu, D.D. Leader-Based Community Detection Algorithm in Attributed Networks. IEEE Access 2021, 9, 119666–119674. [Google Scholar] [CrossRef]
  6. Liu, Q.; Su, Y.; Peng, Q.; Chen, K.; Lu, Y. An Overlapping Community Detection Algorithm for Label Propagation Based on Node Influence. In Proceedings of the 2021 3rd International Conference on Advances in Computer Technology, Information Science and Communication (CTISC), Shanghai, China, 23–25 April 2021; pp. 183–187. [Google Scholar] [CrossRef]
  7. Roghani, H.; Bouyer, A. A Fast Local Balanced Label Diffusion Algorithm for Community Detection in Social Networks. IEEE Trans. Knowl. Data Eng. 2022. [Google Scholar] [CrossRef]
  8. Huang, J.; Hou, Y.; Li, Y. Efficient community detection algorithm based on higher-order structures in complex networks. Chaos Interdiscip. J. Nonlinear Sci. 2020, 30, 023114. [Google Scholar] [CrossRef]
  9. Li, C.; Tang, Y.; Tang, Z.; Cao, J.; Zhang, Y. Motif-based embedding label propagation algorithm for community detection. Int. J. Intell. Syst. 2022, 37, 1880–1902. [Google Scholar] [CrossRef]
  10. Li, P.Z.; Huang, L.; Wang, C.D.; Lai, J.H.; Huang, D. Community detection by motif-aware label propagation. ACM Trans. Knowl. Discov. Data (TKDD) 2020, 14, 1–19. [Google Scholar] [CrossRef]
  11. He, C.; Zheng, Y.; Fei, X.; Li, H.; Hu, Z.; Tang, Y. Boosting nonnegative matrix factorization based community detection with graph attention auto-encoder. IEEE Trans. Big Data 2021, 8, 968–981. [Google Scholar] [CrossRef]
  12. Ye, F.; Chen, C.; Zheng, Z. Deep autoencoder-like nonnegative matrix factorization for community detection. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management, Torino, Italy, 22–26 October 2018; pp. 1393–1402. [Google Scholar]
  13. Wang, X.; Cui, P.; Wang, J.; Pei, J.; Zhu, W.; Yang, S. Community preserving network embedding. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  14. Guendouz, M.; Amine, A.; Hamou, R.M. A discrete modified fireworks algorithm for community detection in complex networks. Appl. Intell. 2017, 46, 373–385. [Google Scholar] [CrossRef]
  15. Liu, F.; Xue, S.; Wu, J.; Zhou, C.; Hu, W.; Paris, C.; Nepal, S.; Yang, J.; Yu, P.S. Deep learning for community detection: Progress, challenges and opportunities. arXiv 2020, arXiv:2005.08225. [Google Scholar]
  16. Wang, P.; Kong, B.; Bao, C.; Zhou, L.; Wang, C. Community Detection Based On Graph Neural Network. In Proceedings of the 2021 6th International Conference on Intelligent Computing and Signal Processing (ICSP), Xi’an, China, 9–11 April 2021; pp. 89–93. [Google Scholar] [CrossRef]
  17. Su, X.; Xue, S.; Liu, F.; Wu, J.; Yang, J.; Zhou, C.; Hu, W.; Paris, C.; Nepal, S.; Jin, D.; et al. A Comprehensive Survey on Community Detection With Deep Learning. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–21. [Google Scholar] [CrossRef] [PubMed]
  18. Kang, Y.; Pu, B.; Kou, Y.; Yang, Y.; Chen, J.; Muhammad, K.; Yang, P.; Xu, L.; Hijji, M. A deep graph network with multiple similarity for user clustering in human–computer interaction. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM), 2022; just accepted. [Google Scholar] [CrossRef]
  19. Le, B.D.; Shen, H.; Nguyen, H.; Falkner, N. Improved network community detection using meta-heuristic based label propagation. Appl. Intell. 2019, 49, 1451–1466. [Google Scholar] [CrossRef]
  20. Feng, Y.; Chen, H.; Li, T.; Luo, C. A novel community detection method based on whale optimization algorithm with evolutionary population. Appl. Intell. 2020, 50, 2503–2522. [Google Scholar] [CrossRef]
  21. Teymourian, N.; Izadkhah, H.; Isazadeh, A. A fast clustering algorithm for modularization of large-scale software systems. IEEE Trans. Softw. Eng. 2020, 48, 1451–1462. [Google Scholar] [CrossRef]
  22. Liu, F.; Wu, J.; Xue, S.; Zhou, C.; Yang, J.; Sheng, Q. Detecting the evolving community structure in dynamic social networks. World Wide Web 2020, 23, 715–733. [Google Scholar] [CrossRef]
  23. Milo, R.; Shen-Orr, S.; Itzkovitz, S.; Kashtan, N.; Chklovskii, D.; Alon, U. Network motifs: Simple building blocks of complex networks. Science 2002, 298, 824–827. [Google Scholar] [CrossRef]
  24. Tang, J.; Liu, G.; Pan, Q. A review on representative swarm intelligence algorithms for solving optimization problems: Applications and trends. IEEE/CAA J. Autom. Sin. 2021, 8, 1627–1643. [Google Scholar] [CrossRef]
  25. Lou, S.; Wu, P.; Guo, L.; Duan, Y.; Zhang, X.; Gao, J. Sparse principal component analysis using particle swarm optimization. J. Chem. Eng. Jpn. 2020, 53, 327–336. [Google Scholar] [CrossRef]
  26. Ramirez-Figueroa, J.A.; Martin-Barreiro, C.; Nieto-Librero, A.B.; Leiva, V.; Galindo-Villardón, M.P. A new principal component analysis by particle swarm optimization with an environmental application for data science. Stoch. Environ. Res. Risk Assess. 2021, 35, 1969–1984. [Google Scholar] [CrossRef]
  27. Li, T.; Shi, J.; Deng, W.; Hu, Z. Pyramid particle swarm optimization with novel strategies of competition and cooperation. Appl. Soft Comput. 2022, 121, 108731. [Google Scholar] [CrossRef]
  28. Bas, E.; Egrioglu, E.; Kolemen, E. Training simple recurrent deep artificial neural network for forecasting using particle swarm optimization. Granul. Comput. 2022, 7, 411–420. [Google Scholar] [CrossRef]
  29. Cai, Q.; Gong, M.; Shen, B.; Ma, L.; Jiao, L. Discrete particle swarm optimization for identifying community structures in signed social networks. Neural Netw. 2014, 58, 4–13. [Google Scholar] [CrossRef] [PubMed]
  30. Ahmed, K.; Hafez, A.I.; Hassanien, A.E. A discrete krill herd optimization algorithm for community detection. In Proceedings of the 2015 11th International Computer Engineering Conference (ICENCO), Cairo, Egypt, 29–30 December 2015; pp. 297–302. [Google Scholar]
  31. Ma, T.; Xia, Z. A community detection algorithm based on local double rings and fireworks algorithm. In Proceedings of the International Conference on Intelligent Data Engineering and Automated Learning, Guilin, China, 30 October–1 November 2017; Springer: Cham, Switzerland, 2017; pp. 129–135. [Google Scholar]
  32. Zhang, Y.; Liu, Y.; Li, J.; Zhu, J.; Yang, C.; Yang, W.; Wen, C. WOCDA: A whale optimization based community detection algorithm. Phys. A Stat. Mech. Its Appl. 2020, 539, 122937. [Google Scholar] [CrossRef]
  33. Ma, T.; Xia, Z.; Yang, F. An ant colony random walk algorithm for overlapping community detection. In Proceedings of the International Conference on Intelligent Data Engineering and Automated Learning, Guilin, China, 30 October–1 November 2017; Springer: Cham, Switzerland, 2017; pp. 20–26. [Google Scholar]
  34. Raghavan, U.N.; Albert, R.; Kumara, S. Near linear time algorithm to detect community structures in large-scale networks. Phys. Rev. E 2007, 76, 036106. [Google Scholar] [CrossRef]
  35. Zeng, X.; Wang, W.; Chen, C.; Yen, G.G. A consensus community-based particle swarm optimization for dynamic community detection. IEEE Trans. Cybern. 2019, 50, 2502–2513. [Google Scholar] [CrossRef]
  36. Handl, J.; Knowles, J. An evolutionary approach to multiobjective clustering. IEEE Trans. Evol. Comput. 2007, 11, 56–76. [Google Scholar] [CrossRef]
  37. Wang, Y.J.; Song, J.X.; Sun, P. Research on Dynamic Community Detection Method Based on an Improved Pity Beetle Algorithm. IEEE Access 2022, 10, 43914–43933. [Google Scholar] [CrossRef]
  38. Li, H.; Zhang, R.; Zhao, Z.; Liu, X. LPA-MNI: An Improved Label Propagation Algorithm Based on Modularity and Node Importance for Community Detection. Entropy 2021, 23, 497. [Google Scholar] [CrossRef]
  39. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  40. Gupta, S.; Deep, K. Hybrid Grey Wolf Optimizer with Mutation Operator. In Soft Computing for Problem Solving; Advances in Intelligent Systems and Computing; Bansal, J., Das, K., Nagar, A., Deep, K., Ojha, A., Eds.; Springer: Singapore, 2019; Volume 817, pp. 961–968. [Google Scholar] [CrossRef]
  41. Gupta, S.; Deep, K. Enhanced leadership-inspired grey wolf optimizer for global optimization problems. Eng. Comput. 2020, 36, 1777–1800. [Google Scholar] [CrossRef]
  42. Liu, Y.; Sun, J.; Yu, H.; Wang, Y.; Zhou, X. An Improved Grey Wolf Optimizer Based on Differential Evolution and OTSU Algorithm. Appl. Sci. 2020, 10, 6343. [Google Scholar] [CrossRef]
  43. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man, Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  44. Lancichinetti, A.; Fortunato, S. Benchmarks for testing community detection algorithms on directed and weighted graphs with overlapping communities. Phys. Rev. E 2009, 80, 016118. [Google Scholar] [CrossRef]
  45. Leskovec, J.; Krevl, A. SNAP Datasets: Stanford Large Network Dataset Collection. 2014. Available online: http://snap.stanford.edu/data (accessed on 7 May 2022).
  46. Sun, B.J.; Shen, H.; Gao, J.; Ouyang, W.; Cheng, X. A non-negative symmetric encoder-decoder approach for community detection. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 597–606. [Google Scholar]
  47. Yang, J.; Leskovec, J. Overlapping community detection at scale: A nonnegative matrix factorization approach. In Proceedings of the Sixth ACM International Conference on Web Search and Data Mining, Rome, Italy, 4–8 February 2013; pp. 587–596. [Google Scholar]
  48. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.