Next Article in Journal
Layered HARQ Design for LDPC-Based Multi-Level Coded Modulation
Previous Article in Journal
An Efficient Algorithmic Way to Construct Boltzmann Machine Representations for Arbitrary Stabilizer Code
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Extreme Value Statistics of Community Detection in Complex Networks with Reduced Network Extremal Ensemble Learning (RenEEL)

by
Tania Ghosh
1,2,
Royce K. P. Zia
1,3 and
Kevin E. Bassler
1,2,4,*
1
Department of Physics, University of Houston, Houston, TX 77204, USA
2
Texas Center for Superconductivity, University of Houston, Houston, TX 77204, USA
3
Center for Soft Matter and Biological Physics, Department of Physics, Virginia Tech, Blacksburg, VA 24061, USA
4
Department of Mathematics, University of Houston, Houston, TX 77204, USA
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(6), 628; https://doi.org/10.3390/e27060628
Submission received: 6 November 2024 / Revised: 28 May 2025 / Accepted: 11 June 2025 / Published: 13 June 2025
(This article belongs to the Section Complexity)

Abstract

Arguably, the most fundamental problem in Network Science is finding structure within a complex network. Often, this is achieved by partitioning the network’s nodes into communities in a way that maximizes an objective function. However, finding the maximizing partition is generally a computationally difficult NP-complete problem. Recently, a machine learning algorithmic scheme was introduced that uses information within a set of partitions to find a new partition that better maximizes an objective function. The scheme, known as RenEEL, uses Extremal Ensemble Learning. Starting with an ensemble of K partitions, it updates the ensemble by considering replacing its worst member with the best of L partitions found by analyzing a reduced network formed by collapsing nodes, which all the ensemble partitions agree should be grouped together, into super-nodes. The updating continues until consensus is achieved within the ensemble about what the best partition is. The original K ensemble partitions and each of the L partitions used for an update are found using a simple “base” partitioning algorithm. We perform an empirical study of how the effectiveness of RenEEL depends on the values of K and L and relate the results to the extreme value statistics of record-breaking. We find that increasing K is generally more effective than increasing L for finding the best partition.

1. Introduction

Finding the community structure within a complex network that relates to its function or behavior is a fundamental challenge in Network Science [1,2,3,4,5]. It is a highly non-trivial problem, as even defining what one means by “structure” must be specified [6]. A variety of approaches can be used to partition the nodes of a network into communities [7,8]. Each approach will divide the nodes according to a different definition of the structure and, in general, will find a different partition [9,10]. Often, the goal is to find a partition that maximizes an objective function. However, finding such a partition can be a computationally difficult NP-complete problem [11,12]. Finding a guaranteed exact solution for a large network is therefore generally infeasible. Thus, for practical applications, it is important to have an approximate algorithm that has polynomial-time complexity, i.e., is fast, and that finds near-exact best solutions for large networks, i.e., is accurate.
Recently, an algorithmic scheme has been introduced that uses information in an ensemble of partitions to produce a better, more accurate partition when seeking to maximize an objective function. The scheme, Reduced Network Extremal Ensemble Learning (RenEEL) [13], uses a machine learning paradigm for graph partitioning, Extremal Ensemble Learning (EEL). EEL begins with an ensemble P K of k K unique partitions. It then iteratively updates the ensemble using extremal criteria until consensus is reached within the ensemble about what the “best” partition is, i.e., the one with the largest value of the objective function. Each update considers adding a new partition P to P K as follows: If P P K , then the “worst” partition in P K , P worst , is removed from P K , reducing the size of P K by one k k 1 , and the update is complete. If P P K and P is worse than P worst , then again P worst is removed from P K , reducing the size of P K by one k k 1 , and the update is complete. If P P K and P is better than P worst , then if k = K , P worst is replaced by P in P K and the update is complete, or if k < K , P is added to P K , increasing the size of P K by one k k + 1 , and the update is complete. Iterative updates continue until k = 1 . The remaining partition is the consensus choice for the best partition.
To ensure fast convergence to a consensus choice in EEL updates, RenEEL conserves the consensus within P K that exists up to that point each time it finds a new partition to be used in an update. It achieves this by partitioning a reduced network rather than the original network. The reduced network is constructed by collapsing the nodes that every partition in P K agrees should be in the same community into “super” nodes. Reduced networks are smaller than the original network and can be analyzed faster, focusing effort only on improving partitioning where there is disagreement within P K . The consensus within P K increases monotonically, and the size of the reduced networks decreases monotonically as the EEL updates are made. RenEEL creates an ensemble P L consisting of L partitions found by analyzing the reduced network to decide the partition to use in each EEL update. The best partition in P L is then used in the update.
There is wide flexibility within the RenEEL scheme. A base algorithm is used to find the partitions that initially form the P K ensemble and those that form each P L ensemble. The base algorithm can be any algorithm that finds a partition that maximizes an objective function. Multiple base algorithms can even be used. There is also freedom in choosing the values of K and L, which represent the maximum size of P K and the size of each P L , respectively. The best base algorithm choice and values of K and L depend on the network being analyzed, desired accuracy, and available computational resources.
This paper investigates the effect of varying K and L on the performance of RenEEL. Larger values of K and L will typically lead to a final, consensus best partition with larger values of the objective function [13]. But how does the value of the objective function of the consensus partition reached typically depend on K and L? How quickly is the value found expected to approach its true maximum value? Given only limited computational resources, is it better to increase K or L? We empirically study these questions when seeking the partition of three well-known real-world networks that maximizes the objective function Modularity.

2. Results

A commonly used approach to find structure in a complex network is to partition the nodes into communities that are more densely connected than expected in a random network. In this approach, the community structure corresponds to the partition that maximizes an objective function called Modularity [2,14]. For a given nodal partition C { c 1 , c 2 , } , Modularity q is defined as
q ( C ) = 1 2 m i j A i j k i k j 2 m δ c i c j ,
where the sum is over all pairs of nodes i j , c i is the community of the ith node, and m is the total number of links present in the network. k i and A i j are, respectively, the degree of the ith node and the i j th element of the adjacency matrix. Thus, Modularity is the difference between the fraction of links inside the partition’s communities, the first term in Equation (1), and what the expected fraction would be if all links of the network were randomly placed, the second term in Equation (1). The task is to find the partition C that maximizes q. We denote the maximum value of q as Q, the value of which is called “the Modularity” of the network.
A number of algorithms with polynomial-time complexity have been developed to find a partition that maximizes Modularity. They range from very fast but not-so-accurate algorithms, such as the Louvain method [15] or randomized greedy agglomerative hierarchical clustering [16], to more accurate but slower algorithms [17], such as one that combines both agglomeration and division steps [18,19]. The accuracy of all of the algorithms tends to decrease as the size of the network increases.
All of the fast Modularity-maximizing algorithms are stochastic because at intermediate steps of their execution, there are seemingly equally good choices to make that are randomly made. In the end, those choices can be consequential because different runs of an algorithm with different sets of random intermediate choices can result in different solutions. Because of this, multiple runs of an algorithm are often made, say 100, to analyze a network, producing an ensemble of approximate partitions. The partition in the ensemble with the largest Modularity is then taken as the network’s community structure, while all other partitions in the ensemble are discarded. RenEEL instead uses the information within the ensemble to find a more accurate partition.
Here we use a RenEEL algorithm that has a randomized greedy base algorithm [16] to find the community structure by maximizing Modularity in real-world networks A, B, and C. Network A is the As-22july06 Network [20]. It is a snapshot in time of the structure of the Internet at the level of autonomous systems. It has 22,963 nodes, which represent autonomous systems, and 48,436 links of data connection. Network B is the PGP Network [20]. It is a snapshot in time of the giant component of the Pretty-Good-Privacy (PGP) algorithm user network. It has 10,680 nodes, which are the users of the PGP algorithm, and 24,316 links indicating the interactions among them. Lastly, Network C is the Astro-ph network. It is a coauthorship network of scientists in Astrophysics consisting of 16,706 nodes representing scientists and 12,1251 links representing coauthorship in preprints on the Astrophysics arXiv database [21].
For each of the three networks, 300 different runs of RenEEL were made for independent values of K and L of 10, 20, 40, 80, 160, and 320, respectively. The compute time required to find the consensus partition was measured for each run. The mean and standard errors of the compute times for the runs at a given value of K and L were then calculated. The full results are listed in Table A1, Table A4 and Table A7 in Appendix A. For a fixed value of L or K, we find that the mean compute times t increase asymptotically as a power of the other ensemble size,
t K α K L   fixed and t L α L K   fixed .
For example, Figure 1 shows this power-law behavior for Network A when L and K have fixed values of 80.
Two-parameter, nonlinear least-squares fits to data for ensemble sizes greater than 10 were then used to determine the proportionality constant and α . Table 1a,b show the values for α K and α L , respectively, that result from fits at different fixed values of L and K for each of the three networks. All statistical errors reported in this paper are ± 2 σ .
The values of the exponents α K ( L ) and α L ( K ) weakly vary with the value of L and K, respectively. The standard errors of α K ( L ) and α L ( K ) tend to remain consistent between smaller and larger ensemble sizes. The distribution of computed time, as depicted in Figure A1 for Network A, does not follow a normal distribution. Consequently, increasing the ensemble size does not lead to a decrease in the standard error. For each network, however, the value of α K is significantly larger than α L . Thus, the expected compute time increases faster with K than L. Given that larger values of K and L typically lead to a better result, i.e., a consensus partition with a larger Q, one might naively conclude that it is better to increase L rather than K. But, to determine if that conclusion is, in fact, correct, the way that Q increases with K and L must be taken into account.
To this end, we begin by noting that for any finite-size network, there is only a finite number of possible partitions. Many modularity-maximizing algorithms will consistently find the actual best partition for very small networks. As the network size and number of possible partitions grow, the task becomes harder; algorithms start to fail to find the exact solution and only provide estimates of the actual, or exact, best partition. RenEEL appears to perform very well at finding the actual best partition of networks of sizes of up to a few thousand nodes [13]. Still, even RenEEL can only find estimates of the exact best partition of larger networks, such as the three we analyze in this paper. As values of K and L increase, the estimates improve, and the value of Q of the consensus partition approaches Q max , the Modularity of the exact best partition. To explore how the values of Q of RenEEL’s consensus partitions approach Q max as a function of K and L, the mean and standard errors of Q found in the runs that were made on each network were calculated as a function of K and L. The results are listed in Table A2, Table A3, Table A5, Table A6, Table A8 and Table A9 in Appendix A. For a fixed value of L or K, we find that Q approaches a maximum value, Q max , as a power-law of the other ensemble size,
Q Q m a x A K K γ K L   fixed and Q Q m a x A L L γ L K   fixed ,
where the As are constants. Figure 2 shows this behavior for Network A when L and K have fixed values of 80. The exact value of Q max is unknown for Networks A, B, and C. Three-parameter, nonlinear least-squares fits were used to determine the values of Q max , A, and γ . Table 2a,b list the values of Q max and γ K , respectively, that result from fits at fixed values of L for each of the three networks. Similarly, Table 3a,b list the values of Q max and γ L that result from fits at fixed values of K for each of the three networks.
The fitted values of Q max increase systematically with increasing L and K and converge to statistically equivalent values at the largest ensemble sizes studied (320), regardless of whether L or K is increased. However, the values of Q max are generally larger when fixing L rather than K when comparing results from when they are fixed at the same size. This fact implies that the maximum value of Q is approached faster by fixing L and increasing K rather than the opposite. The rate of convergence to Q max is quantified by the γ exponents. The values of the exponents α K ( L ) and α L ( K ) depend on the network but only weakly vary with the values of L and K, respectively. For each network, however, the value of α K is significantly larger than that of α L . Thus, the expected compute time increases faster with K than it does with L.
To understand these results, recognize that finding the best partition is an extremal process that, when repeated, is akin to the process of record-breaking. Let us recall some of the theory of the extreme value statistics of record-breaking [22]. Consider a sequence of independent and identically distributed random numbers
x 1 , x 2 , x 3 , , x t ,
chosen from a probability distribution of the form
p ( x ) = μ B μ ( B x ) μ 1 , 0 x B ,
where B is the maximum possible value of x, and define the record R ( t ) as the maximum value of x in the first t numbers in the sequence:
R ( t ) max x 1 , x 2 , x 3 , , x t .
Then, in the limit of large t, the mean record will approach B as
R ( t ) = B 1 Γ ( 1 / μ )   t 1 / μ ,
i.e., as a power-law function with an exponent of 1 / μ . From this, we see that μ = 1 is a borderline case; from Equation (5), it is the case of a uniform distribution of x. If μ < 1 , then p ( x ) is maximal at x = B , and if μ > 1 , then p ( x ) vanishes as x B .
While the analogy with this simple, analytically tractable model of record-breaking is not perfect, Equation (3) can be compared with Equation (7) by identifying Q max with B and γ with 1 / μ . Then, the fact that empirically γ K > 1 and γ L < 1 suggests that the distributions of the Q of the consensus partitions found by increasing K and L correspond to different cases. Namely, as K is increased, the consensus partition Q is likely to be near Q max as it is for μ < 1 , while as L is increased, it is more likely to have a smaller value as it is for μ > 1 . To confirm this, we made 800 runs of RenEEL analyzing Network A with L = 80 and K = 320 and with L = 320 and K = 80 . Figure 3 shows the consensus values of Q found in those runs. As expected, the values found with L = 80 and K = 320 (red bars) are much more likely to be near the maximum value than those found with L = 320 and K = 80 (blue bars).
We can now answer the central question of this paper: Given only limited computational resources, is it better to increase K or L? We have found that the average compute time grows faster with K than with L, but also that the consensus Q approaches Q max faster with K than with L. Does the consensus Q approach Q max as a function of compute time faster by increasing K or L? To answer this, we invert Equation (2) and combine it with Equation (3) to obtain
Q m a x Q t γ / α
So, the larger γ / α is, the faster Q approaches Q m a x as a function of average compute time. Table 4a shows the values of γ K / α K at different fixed L for the three networks. Similarly, Table 4b shows the values of γ L / α L at different fixed L for the three networks.
From these results, it can be clearly concluded that increasing K rather than L will cause Q to approach Q m a x faster. With limited computational resources, it is therefore better to increase K rather than L. Although we have shown that this is true only for three example networks and only when maximizing Modularity, we speculate that these networks are not special and maximizing modularity, rather than a different objective function, is also not special. Therefore, the conclusion that it is more computationally efficient to increase K rather than L in RenEEL should generally be true. However, it would be interesting to explore this question when maximizing other objective functions with RenEEL.

Author Contributions

T.G., R.K.P.Z. and K.E.B. conceived of the presented idea. T.G. performed the numerical simulations and data analysis. K.E.B. verified the analytical methods and supervised the project. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The three networks used in this study are openly available at https://sites.cc.gatech.edu/dimacs10/archive/clustering.shtml (accessed on 19 April 2023). The raw data, which are used for analysis, are provided in Appendix A.

Acknowledgments

The authors would like to thank the Sabine cluster at the University of Houston, which was utilized for performing the simulations presented in this work.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
RenEELReduced Network Extremal Ensemble Learning
EELExtremal Ensemble Learning

Appendix A

Table A1. Average computed time for Network A.
Table A1. Average computed time for Network A.
K
L 10 20 40 80 160 320
1050.6(0.3)123(1)298(4)755(5)2065(57)5811(104)
2059.7(0.3)161.2(0.9)391(2)1094(6)2732(24)7241(42)
4072.0(0.3)202(1)564(3)1617(9)3972(21)10,839(58)
8093(1)326(2)960(5)2427(13)6857(41)18,568(212)
160145.1(0.9)501(3)1559(10)4641(29)12,123(109)34,007(245)
320215(1)919(5)2908(16)8385(56)22,964(220)58,568(388)
Table A2. Modularity values for Network A.
Table A2. Modularity values for Network A.
K
L 10 20 40
100.67691112(300)0.67829673(134)0.67877639(779)
200.677076101(1973)0.67844664(446)0.67881534(751)
400.67715786(897)0.67853550(1947)0.67887617(823)
800.67717013(1549)0.67851858(941)0.67890690(564)
1600.67730312(899)0.67863411(271)0.67890899(716)
3200.67732834(2819)0.67865861(980)0.67893038(621)
Table A3. Modularity values for Network A.
Table A3. Modularity values for Network A.
K
L 80 160 320
100.67906248(512)0.67921816(584)0.67928542(149)
200.67910093(401)0.67925767(629)0.67932669(188)
400.67912582(465)0.67928322(294)0.67933913(177)
800.67915278(412)0.67929836(261)0.67935888(196)
1600.67917852(401)0.67931103(397)0.67936292(202)
3200.67919068(391)0.67932548(421)0.67936962(97)
Table A4. Average computed time for Network B.
Table A4. Average computed time for Network B.
K
L 10 20 40 80 160 320
1013.84(0.07)30.7(0.2)63.1(0.1)143.1(0.6)293(1)625(3)
2016.37(0.09)39.7(0.2)88.5(0.3)196(1)426(2)833(4)
4020.2(0.1)60.5(0.3)130(1)293(1)649(3)1340(7)
8031.1(.2)102(0.9)256(2)514(3)1199(8)2604(22)
16046.8(0.3)170(1)431(3)969(6)2036(13)4363(26)
32073.5(0.6)282(2)877(6)1955(14)4016(29)8214(51)
Table A5. Modularity values for Network B.
Table A5. Modularity values for Network B.
K
L 10 20 40
100.88642816(536)0.88674884(440)0.88671202(790)
200.88647127(1895)0.88674307(594)0.88675001(918)
400.88641007(295)0.88677498(371)0.8867800(1021)
800.88649938(1104)0.88674605(430)0.88678501(865)
1600.88655162(1003)0.88676799(246)0.88680202(672)
3200.88655147(675)0.88676975(315)0.88681905(495)
Table A6. Modularity values for Network B.
Table A6. Modularity values for Network B.
K
L 80 160 320
100.88673690(1014)0.88674209(745)0.88675501(832)
200.88677493(799)0.88678016(634)0.88678290(405)
400.88680491(805)0.88680996(678)0.886801492(418)
800.88680099(697)0.88681788(395)0.88681484(302)
1600.88682706(707)0.88683401(471)0.88683493 (247)
3200.88683201(381)0.88683687(382)0.88683980(201)
Table A7. Average computed time for Network C.
Table A7. Average computed time for Network C.
K
L 10 20 40 80 160 320
10139.5(0.7)295(2)640(5)1466(10)3483(23)8485(70))
20181(1)415(2)1010(9)2389(16)5806(35)14274(108)
40295(2)716(6)1754(10)4350(31)10,846(82)25,722(161)
80484(4)1251(13)3203(30)8033(68)20,680(162)45,012(402)
160909(9)2340(28)6096(52)15,600(123)41,308(344)87,912(801)
3201790(19)4600(53)11,936(106)29,605(249)77,968(620)170,610(1001)
Table A8. Modularity values for Network C.
Table A8. Modularity values for Network C.
K
L 10 20 40
100.73982575(12141)0.74042051(10865)0.74119763(9309)
200.73946114(16171)0.74082625(12067)0.74199647(6703)
400.73847614(7023)0.74163987(6899)0.74230803(7021)
800.73959559(15273)0.74192572(9061)0.74221369(6149)
1600.73966281(10312)0.7411455(13235)0.7424959(6301)
3200.74011840(13382)0.74154016(12306)0.74209565(8293)
Table A9. Modularity values for Network C.
Table A9. Modularity values for Network C.
K
L 80 160 320
100.74155587(8022)0.74188600(8036)0.74189623(8136)
200.74220959(7080)0.74202369(8062)0.74203165(8518)
400.74230761(6930)0.74229803(8301)0.74216325(8772)
800.74239582(7500)0.74251785(6805)0.74251525(6181)
1600.7425311(7104)0.7426539(7409)0.7426924(6417)
3200.74273495(6921)0.74270134(4921)0.74271412(2012)

Appendix B

Figure A1. Distribution of required computed time to find the consensus partition for Network A. Blue bars correspond to the data for L = 320 and K = 80 , and red bars to L = 80 and K = 320 .
Figure A1. Distribution of required computed time to find the consensus partition for Network A. Blue bars correspond to the data for L = 320 and K = 80 , and red bars to L = 80 and K = 320 .
Entropy 27 00628 g0a1

References

  1. Fortunato, S. Community detection in graphs. Phys. Rep. 2010, 486, 75–174. [Google Scholar] [CrossRef]
  2. Newman, M.E.J.; Girvan, M. Finding and evaluating community structure in networks. Phys. Rev. E 2004, 69, 026113. [Google Scholar] [CrossRef] [PubMed]
  3. Newman, M.E.J. Networks: An Introduction; Oxford University Press: Oxford, UK, 2010. [Google Scholar]
  4. Peixoto, T.P. Descriptive vs. Inferential Community Detection in Networks: Pitfalls, Myths and Half-Truths; Elements in the Structure and Dynamics of Complex Networks; Cambridge University Press: Cambridge, MA, USA, 2023. [Google Scholar]
  5. Guimerà, R.; Sales-Pardo, M.; Amaral, L.A.N. Modularity from fluctuations in random graphs and complex networks. Phys. Rev. E 2004, 70, 025101. [Google Scholar] [CrossRef] [PubMed]
  6. Schaub, M.T.; Delvenne, J.C.; Rosvall, M.; Lambiotte, R. The many facets of community detection in complex networks. Appl. Netw. Sci. 2017, 2, 4. [Google Scholar] [CrossRef] [PubMed]
  7. Newman, M.E.J. The Structure and Function of Complex Networks. SIAM Rev. 2003, 45, 167–256. [Google Scholar] [CrossRef]
  8. Ng, A.; Jordan, M.; Weiss, Y. On Spectral Clustering: Analysis and an algorithm. In Proceedings of the Advances in Neural Information Processing Systems, NIPS 2001, Vancouver, BC, Canada, 3–8 December 2001; Dietterich, T., Becker, S., Ghahramani, Z., Eds.; MIT Press: Cambridge, MA, USA, 2001; Volume 14. [Google Scholar]
  9. Peel, L.; Larremore, D.B.; Clauset, A. The ground truth about metadata and community detection in networks. Sci. Adv. 2017, 3, e1602548. [Google Scholar] [CrossRef] [PubMed]
  10. Wolpert, D.; Macready, W. Coevolutionary free lunches. IEEE Trans. Evol. Comput. 2005, 9, 721–735. [Google Scholar] [CrossRef]
  11. Brandes, U.; Delling, D.; Gaertler, M.; Gorke, R.; Hoefer, M.; Nikoloski, Z.; Wagner, D. On Modularity Clustering. IEEE Trans. Knowl. Data Eng. 2008, 20, 172–188. [Google Scholar] [CrossRef]
  12. Brandes, U.; Delling, D.; Gaertler, M.; Görke, R.; Hoefer, M.; Nikolski, Z.; Wagner, D. On Modularity—NP-Completeness and Beyond; Technical Report 19; Universität Karlsruhe (TH): Karlsruhe, Germany, 2006. [Google Scholar] [CrossRef]
  13. Guo, J.; Singh, P.; Bassler, K.E. Reduced network extremal ensemble learning scheme for community detection in complex networks. Sci. Rep. 2019, 9, 14234. [Google Scholar] [CrossRef] [PubMed]
  14. Newman, M.E.J. Modularity and community structure in networks. Proc. Natl. Acad. Sci. USA 2006, 103, 8577–8582. [Google Scholar] [CrossRef] [PubMed]
  15. Blondel, V.D.; Guillaume, J.L.; Lambiotte, R.; Lefebvre, E. Fast unfolding of communities in large networks. J. Stat. Mech. Theory Exp. 2008, 2008, P10008. [Google Scholar] [CrossRef]
  16. Ovelgönne, M.; Geyer-Schulz, A. Cluster Cores and Modularity Maximization. In Proceedings of the 2010 IEEE International Conference on Data Mining Workshops, Sydney, Australia, 13 December 2010; pp. 1204–1213. [Google Scholar]
  17. Sun, Y.; Danila, B.; Josić, K.; Bassler, K.E. Improved community structure detection using a modified fine-tuning strategy. Europhys. Lett. 2009, 86, 28004. [Google Scholar] [CrossRef]
  18. Treviño, S.; Nyberg, A.; Genio, C.I.D.; Bassler, K.E. Fast and accurate determination of modularity and its effect size. J. Stat. Mech. Theory Exp. 2015, 2015, P02003. [Google Scholar] [CrossRef]
  19. Liu, Z.; Ma, Y. A divide and agglomerate algorithm for community detection in social networks. Inf. Sci. 2019, 482, 321–333. [Google Scholar] [CrossRef]
  20. Newman, M. Network Data. Available online: http://www-personal.umich.edu/~mejn/netdata/ (accessed on 19 April 2025).
  21. Newman, M.E.J. The structure of scientific collaboration networks. Proc. Natl. Acad. Sci. USA 2001, 98, 404–409. [Google Scholar] [CrossRef] [PubMed]
  22. Schmittmann, B.; Zia, R.K.P. “Weather” records: Musings on cold days after a long hot Indian summer. Am. J. Phys. 1999, 67, 1269–1276. [Google Scholar] [CrossRef]
Figure 1. The compute time for RenEEL to complete when analyzing Network A. Data are measured in seconds. Blue circles are results as a function of K for fixed L = 80 , and red squares are results as a function of L for fixed K = 80 . The straight lines are power-law fits to the data of ensemble sizes greater than 10. The slope of the fit to the blue circles is α K = 1.489 ( 4 ) , and to the red squares is α L = 0.810 ( 3 ) .
Figure 1. The compute time for RenEEL to complete when analyzing Network A. Data are measured in seconds. Blue circles are results as a function of K for fixed L = 80 , and red squares are results as a function of L for fixed K = 80 . The straight lines are power-law fits to the data of ensemble sizes greater than 10. The slope of the fit to the blue circles is α K = 1.489 ( 4 ) , and to the red squares is α L = 0.810 ( 3 ) .
Entropy 27 00628 g001
Figure 2. Modularity Q of RenEEL consensus partitions for Network A (a) as a function of K for fixed L = 80 (blue circles), and (b) as a function of L for fixed K = 80 (red squares). The black solid lines are fits to the data, representing a power-law approach to the estimated maximum value Q max indicated by the green dashed lines. The value of Q max is 0.679368(6) in (a) and 0.679229(40) in (b). The insets show the data on log-log axes. The slope of the fit to the blue circles in the inset of (a) is γ K = 1.34 ( 2 ) and to the red squares in the inset of (b) is γ L = 0.36 ( 18 ) .
Figure 2. Modularity Q of RenEEL consensus partitions for Network A (a) as a function of K for fixed L = 80 (blue circles), and (b) as a function of L for fixed K = 80 (red squares). The black solid lines are fits to the data, representing a power-law approach to the estimated maximum value Q max indicated by the green dashed lines. The value of Q max is 0.679368(6) in (a) and 0.679229(40) in (b). The insets show the data on log-log axes. The slope of the fit to the blue circles in the inset of (a) is γ K = 1.34 ( 2 ) and to the red squares in the inset of (b) is γ L = 0.36 ( 18 ) .
Entropy 27 00628 g002
Figure 3. Distribution of Q obtained for Network A. Blue bars correspond to the data for L = 320 and K = 80 , and red bars to L = 80 and K = 320 .
Figure 3. Distribution of Q obtained for Network A. Blue bars correspond to the data for L = 320 and K = 80 , and red bars to L = 80 and K = 320 .
Entropy 27 00628 g003
Table 1. (a) Values of scaling exponent α K of the compute time divergence at fixed values of L. (b) Values of scaling exponent α L of the compute time divergence at fixed values of K.
Table 1. (a) Values of scaling exponent α K of the compute time divergence at fixed values of L. (b) Values of scaling exponent α L of the compute time divergence at fixed values of K.
(a)
L Network ANetwork BNetwork C
201.382(2)1.262(2)1.331(4)
401.431(2)1.212(2)1.300(4)
801.489(4)1.240(2)1.300(2)
1601.483(2)1.164(2)1.320(4)
3201.469(2)1.194(4)1.311(6)
(b)
K Network ANetwork BNetwork C
200.732(3)0.800(6)0.821(4)
400.752(2)0.820(4)0.872(4)
800.810(3)0.851(2)0.900(8)
1600.810(4)0.800(2)0.924(4)
3200.792(3)0.879(4)0.898(6)
Table 2. (a) Values of asymptotic maximum Modularity Q m a x found at fixed values of L. (b) Values of scaling exponent γ K of the Modularity convergence to Q max at fixed values of L.
Table 2. (a) Values of asymptotic maximum Modularity Q m a x found at fixed values of L. (b) Values of scaling exponent γ K of the Modularity convergence to Q max at fixed values of L.
(a)
L Network ANetwork BNetwork C
100.679265(4)0.886761(2)0.742303(12)
200.679309(6)0.886784(6)0.742589(18)
400.679320(6)0.886814(2)0.742829(16)
800.679368(6)0.886825(4)0.742956(18)
1600.679370(8)0.886832(2)0.743013(16)
3200.679377(4)0.886834(2)0.743168(20)
(b)
L Network ANetwork BNetwork C
101.21(2)2.16(4)1.18(18)
201.29(2)2.15(6)1.31(20)
401.38(2)2.01(8)1.28(20)
801.34(2)2.04(4)1.71(16)
1601.35(4)2.29(14)1.22(18)
3201.25(2)2.05(6)1.29(20)
Table 3. (a) Values of asymptotic maximum Modularity Q m a x found at fixed values of K. (b) Values of scaling exponent γ L of the Modularity convergence to Q max at fixed values of K.
Table 3. (a) Values of asymptotic maximum Modularity Q m a x found at fixed values of K. (b) Values of scaling exponent γ L of the Modularity convergence to Q max at fixed values of K.
(a)
K Network ANetwork BNetwork C
100.677480(34)0.886521(12)0.740259(170)
200.678725(24)0.886763(10)0.741572(174)
400.678963(36)0.886830(12)0.743070(172)
800.679229(40)0.886832(10)0.743072(170)
1600.679246(26)0.886832(6)0.743077(172)
3200.679370(20)0.886833(8)0.743088(168)
(b)
K Network ANetwork BNetwork C
100.37(6)0.86(10)0.29(26)
200.49(8)0.88(10)0.29(22)
400.50(12)0.85(16)0.22(24)
800.36(18)0.85(12)0.20(22)
1600.47(20)0.69(12)0.25(20)
3200.67(10)0.86(16)0.25(20)
Table 4. (a) Values of scaling exponent γ α at fixed values of L. (b) Values of scaling exponent γ α at fixed values of K.
Table 4. (a) Values of scaling exponent γ α at fixed values of L. (b) Values of scaling exponent γ α at fixed values of K.
(a)
L Network ANetwork BNetwork C
200.93(1)1.70(5)0.98(15)
400.96(1)1.65(6)0.98(15)
800.90(1)1.64(3)1.31(12)
1600.91(2)1.96(12)0.93(13)
3200.89(1)1.71(5)0.98(12)
(b)
K Network ANetwork BNetwork C
200.66(11)1.09(13)0.35(26)
400.66(16)1.03(20)0.25(26)
800.44(22)0.99(14)0.22(24)
1600.57(22)0.88(15)0.27(21)
3200.68(12)0.97(18)0.28(22)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ghosh, T.; Zia, R.K.P.; Bassler, K.E. Extreme Value Statistics of Community Detection in Complex Networks with Reduced Network Extremal Ensemble Learning (RenEEL). Entropy 2025, 27, 628. https://doi.org/10.3390/e27060628

AMA Style

Ghosh T, Zia RKP, Bassler KE. Extreme Value Statistics of Community Detection in Complex Networks with Reduced Network Extremal Ensemble Learning (RenEEL). Entropy. 2025; 27(6):628. https://doi.org/10.3390/e27060628

Chicago/Turabian Style

Ghosh, Tania, Royce K. P. Zia, and Kevin E. Bassler. 2025. "Extreme Value Statistics of Community Detection in Complex Networks with Reduced Network Extremal Ensemble Learning (RenEEL)" Entropy 27, no. 6: 628. https://doi.org/10.3390/e27060628

APA Style

Ghosh, T., Zia, R. K. P., & Bassler, K. E. (2025). Extreme Value Statistics of Community Detection in Complex Networks with Reduced Network Extremal Ensemble Learning (RenEEL). Entropy, 27(6), 628. https://doi.org/10.3390/e27060628

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop