Next Article in Journal
Enhancing Energy Harvesting Efficiency of Flapping Wings with Leading-Edge Magnus Effect Cylinder
Next Article in Special Issue
Modeling the Electrical Activity of the Heart via Transfer Functions and Genetic Algorithms
Previous Article in Journal / Special Issue
Multi-Strategy Improved Dung Beetle Optimization Algorithm and Its Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Learning-Based Methods for Determining the Ideal Team Size in Agile Practices

Escuela de Ingeniería Informática, Universidad de Valparaíso, Valparaíso 2362905, Chile
*
Author to whom correspondence should be addressed.
Biomimetics 2024, 9(5), 292; https://doi.org/10.3390/biomimetics9050292
Submission received: 25 March 2024 / Revised: 6 May 2024 / Accepted: 10 May 2024 / Published: 13 May 2024
(This article belongs to the Special Issue Nature-Inspired Metaheuristic Optimization Algorithms 2024)

Abstract

:
One of the significant challenges in scaling agile software development is organizing software development teams to ensure effective communication among members while equipping them with the capabilities to deliver business value independently. A formal approach to address this challenge involves modeling it as an optimization problem: given a professional staff, how can they be organized to optimize the number of communication channels, considering both intra-team and inter-team channels? In this article, we propose applying a set of bio-inspired algorithms to solve this problem. We introduce an enhancement that incorporates ensemble learning into the resolution process to achieve nearly optimal results. Ensemble learning integrates multiple machine-learning strategies with diverse characteristics to boost optimizer performance. Furthermore, the studied metaheuristics offer an excellent opportunity to explore their linear convergence, contingent on the exploration and exploitation phases. The results produce more precise definitions for team sizes, aligning with industry standards. Our approach demonstrates superior performance compared to the traditional versions of these algorithms.

1. Introduction

The agile approach to software engineering is based on a series of practices that prioritize face-to-face collaboration between members of the software development team [1]. While this allows one to address the change, considering it as the only constant in an uncertain environment [2], it poses a challenge in terms of scalability as the complexity of communication increases with a larger number of team members.
The recent studies in high-performance technological organizations have shown that small, self-organized, and independent software teams are more efficient in software delivery [3]. On the other hand, frameworks for scaling agile development practices promote the formation of independent self-organized teams to enable scalability of agile development [4,5]. This modularization translates into a modularization of the system’s architecture, where approaches like domain-driven design [6] and microservices [7] advocate that small teams handle a reduced part of the problem domain to maintain technical simplicity and reduce the cognitive load of the development team. The impact of communication within and between teams on software development efficiency is significant, to the extent that digital transformation frameworks [8,9] encourage designing the structure of software teams as if it were the architecture of software components: the less coupled the communication of the teams, the less coupled the software components are.
In this context, a relevant question arises: What is the ideal size of a software development team? Although this question has been approached with anecdotal approximations like the two-pizza rule (a team should not be larger than the number of people two pizzas can feed) [10], there are various factors in the interactions among team members that can influence the team’s size. A systematic approach to this problem is proposed in [11]. In this study, the challenge is addressed as an optimization problem: given the number of professionals, how to organize them to optimize the number of communication channels, considering both the channels within the team and between teams. For that, a conventional genetic algorithm (GA) is employed. The results obtained after the GA optimization process indicate that the optimal distribution of teams for the total project members ranges from five to nine members per team. However, there is room for improvement in the aforementioned results by employing new artificial intelligence techniques. These techniques enable the use of historical data to achieve better outcomes when determining the distribution of team sizes in an agile practices project.
In artificial intelligence, metaheuristic optimization algorithms have emerged as powerful techniques capable of solving complex optimization problems with reduced solving times [12]. Bio-solvers go through two main phases, known as exploration and exploitation. Many of these algorithms have been designed to achieve better results through these two stages. However, when this is not the case, the mechanism within the algorithm may have some slight disadvantages in balancing exploration and exploitation, leading to suboptimal performance [13]. For this reason, several works have been proposed in this line of research to balance exploration and exploitation better and achieve more efficient solutions.
Techniques based on learning mechanisms to support optimization metaheuristics have been successfully applied to problems involving the trade-off between exploration and exploitation [14]. However, in many cases, the procedure guiding nature-inspired algorithms operates individually; that is, it is a single technique that enhances the optimization algorithm. Our work proposes a new balance between exploration and exploitation using ensemble learning (EL). This strategy involves integrating a machine-learning technique with specific characteristics to achieve superior results. As surrogate optimization algorithms, we employ seven well-known bio-inspired techniques: particle swarm optimization, gray wolf optimizer, firefly algorithm, bat algorithm, whale optimization algorithm, crow search algorithm, and differential evolution. We chose these mechanisms because they exhibit a linear characteristic of avoiding solutions trapped in local optima while balancing the exploration and exploitation phases.
Ensemble learning into the optimization algorithm reveals that the optimal team size and distribution for a given total number of team members range between six and eight members per team, which aligns with industry practices [15,16,17]. These results are practically equal to those reported in [11].
The rest of the article is organized as follows: Section 2 presents the state of the art, including relevant studies. Section 3 covers important concepts considered in the proposal. Next, the developed solution is presented in Section 4. Section 5 defines the computational experimental setup. The results and discussion are exposed in Section 6. Finally, Section 7 provides a critical outlook on the results achieved and outlines the potential future work.

2. Related Work

Traditional project management approaches often try to achieve the objectives of innovative projects. It is challenging to incorporate detailed initial plans in projects that produce innovations, such as new products and software, due to the complexity of requirements and solution planning and the lack of information from previous similar projects [18]. The information scarcity in such projects makes them uncertain for traditional management methods [19,20]. Agile project management addresses these shortcomings. It involves adaptive team skills regarding scope and solution requirements, distinct leadership involvement, and the ability of agile teams to be creative and innovative [21]. However, this heavily relies on active commitment from team members to the project tasks [22].
Few studies have focused on improving communication and collaboration within agile teams. One of them is published in [11]. The proposal solves an optimization model for agile team sizes using the genetic algorithm. The objective is to develop a GA model to find the optimal size of a self-organized team, considering team communication characteristics. The model uses the number of communication channels between individuals and teams as the objective function. Another work, [23], proposes an optimization model for agile team relationships based on role patterns and fuzzy logic. Role patterns evaluate team cohesion from three perspectives: the team competence complement index, role typology matching index, and team typology matching index. The model aims to determine the most suitable candidate for a vacant position and effective team cohesion. Fuzzy logic is used for the linguistic evaluation of system output, classifying results as low, medium, or high. While this model has not been tested with metaheuristic algorithms, fuzzy logic offers the advantage of defining and measuring imprecise information about the candidates’ psychological and behavioral traits, which is difficult to model using conventional mathematics.
Concerning diversity methods, many strategies have been explored to achieve a balance between exploration and exploitation. In [24], a compilation and investigation of works related to exploration and exploitation in bio-inspired algorithms within the branch of evolutionary algorithms is presented. According to this study, the same concept will be applied to swarm intelligence, using different diversity measures to define neighborhood relationships, which are necessary to delineate exploration from exploitation. For this purpose, we focus on maintaining and controlling diversity through various techniques. These techniques are divided into two main branches within diversity maintenance: non-niching and niching. Any technique that maintains diversity in the population based on the distance between the population members is considered a niching technique. Non-niching and niching methods can maintain a population of diverse individuals, while niching methods can also locate multiple optimal solutions.
Several works analyze how metaheuristics are improved using machine learning, regression, and clustering techniques [25,26,27,28]. In [29], a machine learning model that predicts solution quality for a given instance is created using the support vector machine. After that, this approach modifies the parameters and directs the metaheuristics to more fruitful search areas. In [30], the authors provide an evolutionary algorithm that regulates operators and parameters. For this, a controller module that applies learning rules assesses the impact and assigns restarts to the parameter set is integrated. Similar to these, the research presented in [31] investigates how to combine the variable neighborhood search algorithm with reinforcement learning, employ reactive strategies for parameter change, and choose local searches to balance the exploration and exploitation stages.
Other works published in [32,33] combine PSO with regression models and clustering strategies for population management and parameter adjustment, respectively. Another classifier-based PSO is provided in [34] with the aim of an optimization approach that does not depend on its parameters. This method enhances the exploration of the search space and the quality of the solutions retrieved by classifying the solutions discovered by the particles using a previously trained model.
In line with other efforts, the PSO is once again improved with a learning model to regulate its parameters, yielding a competitive performance compared to alternative parameter adaption algorithms [35]. This approach is also applied in GWO [36]. Here, RNA crossover operations and an adaptive parameter control scheme were employed. Here, each individual wolf learns from its neighbors, becoming another candidate for its new position. In [37,38,39], randomized balancing strategies are applied to GWO. These improvements are integrated into the local search procedure in order to balance the exploration and exploitation phases.
Another bio-inspired algorithm enhanced by a learning mechanism is the firefly optimization. In [40], the authors propose a multi-strategy ensemble to effectively achieve a better balance between exploration and exploitation. Furthermore, in [41], an ML-based work is presented to guide new population generation in an evolutionary process. The proposed approach, called the learnable evolution model, executes ML to uncover reasons for the superior performance of specific individuals in designated tasks. These inductive hypotheses are then used to create a new solution generation.
From a similar perspective, some works in machine learning, such as [42], offer an overview of the fusion of data mining techniques and metaheuristics. The application of these methods within metaheuristic algorithms can be observed in various stages, including initializing solution(s), managing solution(s) during the search process, and integrating data mining into operators and local searches.
The reviewed literature underscores the diverse approaches and innovations in combining metaheuristics with machine learning techniques. These studies collectively highlight the potential for enhanced problem-solving, optimization, and adaptive strategies through the synergy of these two domains. The synthesis of these methodologies contributes to the theoretical understanding of optimization and presents practical avenues for addressing complex real-world challenges. The following sections delve into the proposed method and experimental evaluation, building upon the insights gained from these related works.

3. Preliminaries

This section first introduces the ideal time-size problem for agile practices. Next, we expose the ensemble learning paradigm that includes classifying, training, and combining models. Finally, we present the integration between the learning-based approach and different swarm intelligence methods.

3.1. Time-Size Problem

Agile practices are a mechanism to reduce costs and respond to changes in dynamic market conditions [43]. It involves continuous collaboration with stakeholders and ongoing improvement at each project stage. When starting a project, teams go through planning, execution, and evaluation. Continuous collaboration among team members and project stakeholders is vital, as it impacts communication and performance within the work [43]. Several frameworks have been developed, including Extreme Programming [44], SCRUM [45], Adaptive Project Framework [46], and Crystal Methodologies [47]. These agile principles are implemented to develop products and project teams that can swiftly adapt to changes, thereby increasing organizational agility and achieving significant business benefits in a fluctuating business environment. Notably, it points out that while empirically determining a suitable team size is useful, it is often insufficient for optimizing team efficiency, necessitating more dynamic and scientifically grounded approaches.
To address the limitations of empirical methods in determining team size, an optimization model is employed [11]. This model aims to minimize the total number of communication channels among team members, thereby reducing the complexity and potential for inefficiencies in team interactions. The goal is to find the optimal equipment configuration to reduce communication load, streamline processes, and enhance team efficiency.
Initially, the model establishes the basic parameters: N represents the total number of communication channels among all team members, T denotes the number of teams, and n specifies the number of members in each team, defined by the set { t 1 , t 1 , , t n } . After, for each team, the total number of communication channels N t is calculated using the formula N t = [ n ( n 1 ) / 2 ] . This equation calculates the combinations of team members that can communicate within the same team based on the assumption that every member communicates with every other member. Next, the total number of communication channels between different teams N T is computed as N T = [ T ( T 1 ) / 2 ] . This accounts for each team needing to communicate with every other team at least once. Finally, the comprehensive total N is then given by N = T × N t + N T .
To reflect the diverse impact of various communication methods, the model incorporates weighted factors for different communication types: w 1 for person-to-person verbal communication, w 2 for team-to-team communication, w 3 for email communications, w 4 for video conference communications, and w 5 for multilingual communications.
The objective function to minimize the communication load is formulated as follows:
m i n i = 1 T n i ( n i 1 ) 2 × p w + T ( T 1 ) 2 × t w s . t i = 1 T n i = N
where p w includes weights for personal communications { w 1 , w 3 , w 5 } , and t w accounts for team-based communications { w 2 , w 4 } . The weights w i are empirically sourced from the prior research and are set between 1 and 1.5, providing a realistic approximation of the communication impacts [11].
The optimization problem of determining the ideal team size within agile practices is combinatorial, as it aims to optimize communication channels within and between teams. Given the number of team compositions possible from a pool of project members, the search space grows exponentially with the increase in team and project size. This complexity is further compounded when considering constraints such as minimum and maximum team sizes. For instance, if a project involves n members and each team must have between a and b members, the total number of potential team configurations can be calculated using combinatorial methods. Therefore, The search space size is factorial, growing exponentially with the increase in n. This exponential growth exemplifies the NP-hard nature of the problem, as the number of potential configurations increases rapidly, even with a small increase in the number of team members.
This combinatorial explosion suggests that our team size optimization problem can be classified as NP-hard. The vast array of potential configurations presents significant theoretical and computational challenges, necessitating an efficient approach to exploring such a large search space.
In response to these challenges, we employ bio-inspired algorithms, particularly enhanced with ensemble learning. These algorithms are well-suited for efficiently searching large, complex spaces and offer a pragmatic approach to achieving near-optimal solutions within acceptable time frames and computational resources. The enhancements through ensemble learning help balance the exploration and exploitation phases of the algorithms, leading to more precise definitions of team sizes that align with industry standards. This strategic methodological choice aligns with the need to efficiently handle inherent complexity and scalability challenges while ensuring that the solutions are practical and implementable in real-world scenarios.

3.2. Ensemble Learning

Ensemble learning is a robust multiple classification system that leverages the strength of various learning models to create a superior composite model. As detailed in [48], ensemble methods synthesize the predictive power of several base learners to enhance predictive accuracy and model stability over individual classifiers.
The concept of homogeneous and heterogeneous ensembles delineates the landscape of ensemble learning. Homogeneous ensembles, such as decision tree ensembles or neural network ensembles, consist exclusively of one type of base learner, enhancing the predictive strength through diversity in model training procedures. Heterogeneous ensembles, in contrast, integrate diverse types of learners, broadening the scope of learning capabilities beyond what single-model ensembles can achieve.
The effectiveness of ensemble methods is captured in our adaptation of the basic EL concept illustrated in Figure 1, where multiple learners are trained, and their predictions are synergistically combined to produce superior results. This approach mitigates individual learner errors across varied datasets, as supported by [49], who emphasize the enhanced performance due to error decorrelation among learners.
In EL, two distinct approaches are utilized to facilitate the learning process for individual learners through training. One involves a sequential methodology, while the other employs a parallel system. To achieve this, two pivotal algorithms stand out, namely, boosting and bagging. Boosting employs a sequential methodology to elevate weak individual learners to a more robust level. On the other hand, bagging utilizes a parallel training method that emphasizes ensemble-based sampling. Our study extends these foundational techniques by exploring combinatorial methods within EL, aiming to harness and optimize the collective capabilities of diverse learning algorithms.
Combining individual learners at the end of the training process offers several benefits. These include mitigating the risk of selecting a single learner with mediocre results, avoiding poor generalization performance, enhancing outcomes at this stage, and enabling a more accurate approximation to the problem’s solution by amalgamating individual learners.
If an ensemble contains L individual learners { h 1 , h 2 , h 3 , , h L } , then the output of these is represented by h i ( x ) , where x is the input sample [48]. The most commonly used methods for performing these combinations are as follows:
  • Averaging: This approach is predominantly employed by individual learners engaged in regression tasks or those producing numerical outputs. It encompasses two distinct methods: simple averaging (see Equation (2)) and weighted averaging (see Equation (3)).
    H ( x ) = 1 T i = 1 T h i ( x )
    H ( x ) = i = 1 T w i h i ( x )
    In Equation (3), w i represents the weight of individual learner h i , usually satisfying w i 0 and i = 1 T w i = 1 . Notably, simple averaging is a special case of weighted averaging when w i = 1 T . Weighted averaging proves more suitable when individual learners exhibit substantial performance disparities, while simple averaging is preferable when individual learners display similar performance.
  • Voting: This strategy is primarily used for classification tasks where individual learners predict labels from a known set of labels using a voting approach. Various methods exist within this category, including majority voting (see Equation (4)), plurality voting (see Equation (5)), and weighted voting (see Equation (6)).
    H ( x ) = c j , if i = 1 T h i j ( x ) > 0.5 reject , otherwise
    H ( x ) = c a r g m a x j i = 1 T h i j ( x )
    H ( x ) = c a r g m a x j i = 1 T w i h i j ( x )
  • Combining Learning: This approach is employed when a substantial amount of data is available for training. It involves using a meta-learner to combine individual learners. A representative technique in this category is stacking. In stacking, individual learners are referred to as first-level learners, while the combined learners are known as second-level learners or meta-learners.
    The stacking process commences by training first-level learners using the original training dataset. A new dataset is then generated to train second-level learners. In this new dataset, outputs from first-level learners serve as input features, while the original training sets labels remain unchanged. To mitigate the risk of overfitting, cross-validation or leave-one-out methods are often employed during the generation of the second-level training set, ensuring that samples unused in training first-level learners contribute to the second-level learner’s training set.
    In this work, we employ the voting scheme because it combines machine learning classifiers and it averages predicted probabilities to forecast feature labels, enabling the use of varied models for enhanced classification results.

3.3. Bio-Inspired Algorithms

Metaheuristics are optimization techniques that draw inspiration from natural processes to solve complex problems in various domains [12]. These algorithms provide powerful tools for tackling problems where traditional optimization methods may fail due to their nonlinearity, high dimensionality, or multimodality. Among the plethora of metaheuristics, several stand out for their unique approaches and successful applications [50].
In our study, we selected bio-inspired algorithms based on their proven ability to effectively manage complex, multidimensional search spaces essential for optimizing agile team sizes. We utilized both cutting-edge and newly developed algorithms to ensure robustness and innovation in our approach. Key considerations included their rapid convergence rates and computational efficiency, vital for the fast-paced nature of agile projects. Additionally, these algorithms’ behaviors, which mimic the collaborative dynamics of agile teams, greatly improved their relevance and effectiveness for our research.
Finally, it is important to emphasize that all bio-inspired algorithms have been adapted for an integer domain. Originally designed to address continuous problems, these algorithms now incorporate a discretization method using the sigmoid function [ 1 / ( 1 + e x i j ) ] to make them suitable for integer problems [51]. This process involves rounding each element of the solution vector x i j to the nearest integer within a predefined range. Depending on the value derived from the sigmoid function, x i j is assigned the closest integer. Through this transformation, we ensure that the variables of the solution vector strictly adhere to the problem’s integer constraints, maintaining them within the defined lower and upper bounds.

3.3.1. Particle Swarm Optimization

Particle swarm optimization is a bio-inspired metaheuristic derived from the social behavior of birds flocking or fish schooling [52]. This algorithm models potential solutions as particles that navigate through the solution space by adjusting their positions and velocities based on personal and collective experiences. Each particle in the swarm represents a potential solution to the optimization problem and moves through the solution space guided by its own experience as well as the collective experience of the swarm.
Each particle updates its position by following two best values: the best solution it has achieved so far (personal best) and the best solution any particle in the population has achieved (global best). The movement of each particle is influenced by its velocity, which dynamically adjusts at each iteration of the algorithm according to the following equations:
v i ( t + 1 ) = w · v i ( t ) + c 1 · r 1 · ( p b e s t , i x i ( t ) ) + c 2 · r 2 · ( g b e s t x i ( t ) )
x i ( t + 1 ) = x i ( t ) + v i ( t + 1 )
In these equations, v i denotes the velocity of particle i, x i denotes its position, p b e s t , i is the personal best position it has discovered, and g b e s t is the best position found by any particle in the swarm. The parameters w, c 1 , and c 2 represent the inertia weight and the cognitive and social scaling coefficients, respectively. These coefficients control the impact of the past velocities, the cognitive component (individual memory), and the social component (swarm influence), on the velocity update. The parameters r 1 and r 2 are random numbers generated anew for each update and for each particle, providing stochasticity to the search process and helping to escape local optima.
The inertia weight w plays a critical role in balancing the exploration and exploitation abilities of the swarm. A higher inertia weight facilitates exploration by encouraging the particles to roam further in the search space, while a lower inertia weight aids exploitation by allowing finer adjustments in the neighborhood of the current best solutions.
PSO has been successfully applied to a wide range of problems, from classical optimization problems such as function minimization and resource allocation to complex real-world applications like neural network training, electric circuit design, and multi-objective optimization. Its popularity stems from its simplicity, ease of implementation, and robust performance across diverse problem domains. Additionally, PSO’s mechanism is inherently parallel, making it suitable for implementation on parallel and distributed computing architectures, further enhancing its efficiency and scalability.

3.3.2. Gray Wolf Optimizer

Gray wolf optimizer is an advanced metaheuristic algorithm inspired by the social hierarchy and hunting techniques of gray wolves in the wild [53]. This algorithm mimics the leadership and hunting strategies of wolves, where individuals follow the alpha, beta, and delta wolves, leading to efficient prey tracking and capture.
GWO models the wolves’ social structure. The alpha wolf is the leader making critical hunting decisions, followed by the beta and delta wolves, who assist in decision-making and other pack activities. The rest of the pack (omega wolves) follow these leaders. The GWO algorithm utilizes this behavior to search for optimal solutions, with the wolves’ positions in the solution space representing potential solutions to the optimization problem.
The algorithm updates the positions of the wolves based on the positions of the alpha, beta, and delta wolves, which are considered the current best solutions found so far. The positions are updated using the following mathematical models:
D = | C · x p ( t ) x ( t ) |
x ( t + 1 ) = x p ( t ) A · D
where x p ( t ) represents the position vector of the prey (or the best solution found so far), x is the position vector of a wolf, A and C are coefficient vectors, and t indicates the current iteration. The vectors A and C are calculated as follows:
A = 2 · a · r 1 a
C = 2 · r 2
where a linearly decreases from 2 to 0 over the course of iterations, and r 1 , r 2 are random vectors in [ 0 , 1 ] .
GWO is celebrated for its balance between exploration (diversifying the search space to find various possible solutions) and exploitation (intensively searching around the current best solutions). This balance is critical in avoiding local optima and ensuring convergence to the global optimum in a variety of complex problem landscapes.
Applications of GWO span diverse fields, including engineering design [54], renewable energy optimization [55], feature selection [56], and many others where robust and effective optimization solutions are required. Its ability to solve complex multi-modal problems with simple adjustments to its parameters and without the need for gradient information makes it particularly useful for real-world optimization problems.

3.3.3. Firefly Algorithm

The firefly algorithm is inspired by the flashing behavior of fireflies, which they use to attract mates or prey [57]. This bio-inspired metaheuristic is particularly effective in multimodal optimization problems due to its ability to handle the complexities associated with multiple local optima. Each firefly in the algorithm represents a solution, and its brightness is directly associated with the fitness value of the solution—the brighter the firefly, the better the solution.
The movement of a firefly towards another more attractive (brighter) firefly is governed by the following equations:
x i ( t + 1 ) = x i ( t ) + β 0 e γ r i j 2 ( x j ( t ) x i ( t ) ) + α ( r a n d 0.5 )
where x i ( t ) and x j ( t ) are the positions of fireflies i and j at iteration t, respectively. The r i j denotes the distance between these two fireflies, β 0 is the attractiveness at r = 0 , γ is the light absorption coefficient, and α is a randomization factor. The FA algorithm balances exploration and exploitation by dynamically adjusting the parameters α and γ , allowing fireflies to move towards brighter individuals locally and globally.
The firefly algorithm has been successfully applied in various domains, including economic load dispatch [58], scheduling [59], and structural design optimization [60]. Its ability to explore the search space effectively while avoiding premature convergence makes it an excellent tool for solving complex optimization problems that require thorough exploration.

3.3.4. Bat Algorithm

Bat optimization is a bio-inspired optimization technique developed by mimicking the echolocation or sonar system of bats [61]. Bats are fascinating creatures capable of navigating and hunting in complete darkness. They emit sound waves that bounce off objects and return as echoes, allowing them to construct a sonic map of their environment. This remarkable capability is adapted in BA, where each virtual bat in the algorithm uses a simulated form of echolocation to assess the quality of solutions and to navigate the search space.
In the bat algorithm, each bat is treated as an agent that searches for the most optimal solution. The algorithm simulates the bats’ ability to adjust the frequency of their emitted sounds, which affects how they perceive distance and, thus, how they navigate toward prey. The mathematical representation of this behavior is as follows:
f i = f m i n + ( f m a x f m i n ) · rand
v i ( t + 1 ) = v i t + ( x i t x ) · f i
x i ( t + 1 ) = x i t + v i ( t + 1 )
where f i denotes the frequency at which bat i emits pulses. This frequency can vary within a defined range [ f m i n , f m a x ] , adjusting as the bat perceives different distances to its target. The parameter r a n d is a random number between 0 and 1, ensuring stochastic variations in pulse emission frequency. The velocity v i of each bat is updated based on the distance from the current position x i to the current global best location x , which is detected through the bat’s echolocation capabilities. This process helps in dynamically adjusting bats’ movement toward the best solution across the swarm.
The bat algorithm’s ability to adjust the rate of pulse emission (loudness) and the wavelength (frequency) of the echolocation pulses enables a flexible exploration of the search space. Lower frequencies (wider wavelengths) allow bats to “scan” a broader area, useful during the initial exploration phase of the algorithm, while higher frequencies (shorter wavelengths) are beneficial for fine-tuning solutions as bats converge towards optimal locations.
BA has proven effective across a wide range of applications, from engineering design problems where optimal solutions are obscured within large, complex search spaces to data science tasks like feature selection and clustering [62,63,64]. The algorithm’s flexibility stems from its dual capability to explore vast search areas through random flight paths and to exploit promising areas through adaptive frequency tuning and velocity adjustments. This balance makes it particularly robust for multimodal optimization problems, where multiple local optima exist, and the global optimum is hidden among them.

3.3.5. Whale Optimization Algorithm

Whale optimization is a novel optimization technique inspired by the unique hunting behavior of humpback whales [65]. These whales employ a fascinating foraging method known as the bubble-net feeding strategy, which is considered one of the most sophisticated hunting strategies among marine creatures. This technique involves the whales swimming in a spiral path and creating a ’net’ of bubbles along the circle’s perimeter to trap their prey, usually small fishes or krill.
WOA translates this natural strategy into a mathematical model to efficiently search and optimize solution spaces in various computational problems. The algorithm simulates the whales’ approach by adjusting the positions of potential solutions, conceptualized as whales, towards prey or the best solution discovered during the search process. The positional updates are governed by the following mathematical expressions:
x ( t + 1 ) = x A · D
D = | C · x x |
Here, x denotes the position vector of the best solution found so far, representing the prey in the algorithm’s context. The coefficients A and C are computed at each iteration, playing crucial roles: A influences the convergence behavior towards or away from the prey, mimicking the tightening of the bubble-net, and C provides a random weight to the prey’s position, enhancing the exploration capabilities of the algorithm.
The parameters A and C are adjusted dynamically with iterations. A typically decreases linearly from 2 to 0 over the course of iterations, allowing a smooth transition from exploration—searching away from the prey—to exploitation—tightening towards the prey. This mechanism enables WOA to maintain a balance between exploring new areas in the search space and exploiting the promising areas around the global optimum.
WOA is particularly noted for its ability to navigate complex and multi-modal landscapes effectively. It achieves this by employing a variable-shape spiral movement to update the positions, which allows it to closely mimic the helical approach of humpback whales. This spiral update formula typically combines with linear movement to enhance the algorithm’s exploration and exploitation phases, thereby improving the convergence speed toward the optimal solution.
WOA has been widely applied across various domains, including parameter optimization [66], industrial design [67], power dispatch [68], and structural design [69]. Its ability to handle nonlinear, non-differentiable, continuous, and discrete optimization problems makes it an invaluable tool in areas requiring robust optimization solutions. Additionally, the simplicity of its implementation and minimal parameter adjustments make WOA a user-friendly and effective choice for researchers and practitioners facing complex optimization challenges.

3.3.6. Crow Search Algorithm

The crow search algorithm is a metaheuristic optimization algorithm inspired by the cunning and strategic behavior of crows in nature [70]. Crows are known for their ability to hide food in secret places and remember these locations to retrieve their stash later. They are also observed to attempt to steal food hidden by other crows if they watch where it is cached. This complex and strategic behavior is modeled in the CSA to tackle optimization problems.
CSA simulates the behavior of crows storing and retrieving food, where each crow represents a potential solution to an optimization problem. The algorithm utilizes the concept of memory and awareness of other crows’ actions, which translates into an exploration and exploitation mechanism within the search space.
Each crow in the population has a memory of where it has hidden its food (i.e., the best solution it has found so far). During the search process, crows may follow others to their hiding places in hopes of discovering better food sources (better solutions). This is modeled by the following equations:
x i ( t + 1 ) = x i ( t ) + f · ( x j ( t ) x i ( t ) )
Here, x i ( t ) is the position of the crow i at iteration t, x j ( t ) is the position of the hiding place of another randomly chosen crow j, and f is the flight length, a factor that determines the visibility and reachability of the hiding place. The flight length f can vary, representing the strategy to either explore new areas or exploit known ones.
CSA inherently balances exploration and exploitation, allowing crows to sometimes follow others to their food hiding places and, at other times, search for new locations. This approach mimics the random and opportunistic nature of crow behavior in the wild, making the algorithm both adaptive and robust. The simplicity of the CSA model enables its application to a wide array of optimization problems, ranging from engineering design to scheduling and resource allocation. Moreover, CSA is particularly robust in finding global optima without becoming easily trapped in local optima, a common challenge in complex optimization scenarios. This blend of adaptability and robustness underpins the effectiveness of CSA in navigating the multifaceted landscapes of modern computational problems.
Due to its versatile and nature-inspired approach, CSA has been effectively applied in numerous fields such as scheduling [71], power dispatch [72], and it also has improved by clustering techniques [28]. Its ability to efficiently search large and complex landscapes makes it particularly useful for problems where the search space is vast and filled with potential solutions.

3.3.7. Differential Evolution

Differential evolution is a robust, simple, and efficient algorithm for global optimization over continuous spaces [73]. It was developed to address complex optimization problems that are difficult to solve using traditional methods. DE relies on the principles of natural selection and genetics, making it part of the family of evolutionary algorithms.
The central concept of DE involves iteratively improving a population of candidate solutions based on the principles of mutation, crossover, and selection. Each individual in the population represents a potential solution, and the algorithm evolves these solutions across generations to converge on the optimal solution.
The mutation operation in DE is unique and serves as the primary driver of diversity within the population. It creates new solution vectors by combining the weighted differences between randomly selected pairs of solutions from the current population. This is expressed mathematically as follows:
v i ( t + 1 ) = x r 1 ( t ) + F · ( x r 2 ( t ) x r 3 ) ( t )
where v i ( t + 1 ) is the new mutant vector; x r 1 ( t ) , x r 2 ( t ) , and x r 3 ( t ) are randomly chosen and distinct vectors from the current iteration t; and F is a scaling factor that controls the amplification of the differential variation.
Crossover in DE allows for the recombination of genetic information by mixing the mutant vector with the existing population, creating trial vectors. This increases the genetic diversity and allows for the exploration of new areas in the search space. The trial vector u is typically constructed as:
u j , i = v j , i ( t ) if r a n d ( 0 , 1 ) C R or j = j r a n d x j , i ( t ) otherwise
where C R is the crossover rate and j r a n d is a randomly chosen index ensuring that u obtains at least one component from v .
Selection in DE is based on the fitness of the trial vectors compared to their corresponding target vectors in the population. Only the fitter solutions survive to the next generation, ensuring that the population gradually moves towards the optimal regions of the search space.
DE is celebrated for its simplicity, effectiveness, and versatility. It has been successfully applied in a vast array of fields, including economic dispatch [74], tuning for neural networks [75], and more [76]. The algorithm’s ability to handle non-differentiable, nonlinear, multimodal, and multi-objective optimization problems makes it an invaluable tool for tackling real-world challenges.

4. Developed Solution

Our proposed solution effectively addresses the optimization of team sizes in agile practices through a sophisticated integration of ensemble learning with metaheuristic algorithms. This integrated approach not only captures the complexity inherent in team dynamics but also adapts to the evolving needs of agile project environments, ensuring optimal performance and flexibility.
The core of our solution is an advanced metaheuristic algorithm that iteratively adjusts team configurations based on real-time performance metrics and predictive insights from an ensemble learning model. The algorithm operates by simulating various team scenarios to determine the most effective composition that maximizes team efficiency and project outcomes. We outline this procedure across the following distinct phases.

4.1. Phase 1: Metric Collection and Initial Adaptation

Initially, the algorithm collects a comprehensive set of metrics during each iteration of the metaheuristic process. These metrics reflect the diversity and performance of teams under different configurations and are crucial for informing the ensemble learning model. The metrics to be stored are as follows:
  • Features: Iteration, fitness, number of search agents, problem dimension, exploration percentage, exploitation percentage, Hamming diversity, dice diversity, Jaccard diversity, Kulsinski diversity, Rogers–Tanimoto diversity, Russellrao diversity, Sokal–Michener diversity, Yule diversity, Sokal–Sneath diversity, and dimension-wise diversity.
  • Label: Behavior of agent search mechanisms. Exploration or exploitation operators with the original algorithm method enhanced by EL. In our study, we use the parameter that controls both phases.
After obtaining the results from instances of the algorithm, the calculated metrics from each iteration are stored in order to create the dataset. The diversity metrics mentioned are defined as follows:
  • Hamming by frequencies [77]: This is a metric of natural similarity in binary codes, calculable with a few machine instructions for comparison [78]. The metric is defined as
    D H F = n 2 2 l d = 1 l α A f d ( α ) ( 1 f d ( α ) )
    where f d ( α ) is the count of times the value of α is present in dimension d, A: { 0 , 1 } , and l is the dimension size of individuals.
  • Dice diversity: According to [79], the dice dissimilarity between u and v is defined as
    D D = C T F + C F T 2 C T T + C F T + C T F
    where c i j is the number of occurrences of u [ k ] = i and v [ k ] = j for k < n .
  • Jaccard diversity: Also known as the Jaccard similarity coefficient, it is a statistic used to measure the similarity and diversity of sample sets [80]. According to [79], it is defined as
    D J = C T F + C F T C T T + C F T + C T F
    where c i j is the number of occurrences of u [ k ] = i and v [ k ] = j for k < n .
  • Kulsinski diversity: The Kulsinski dissimilarity between u and v is defined as
    D K = C T F + C F T C T T + n C F T + C T F + n
    where c i j is the number of occurrences of u [ k ] = i and v [ k ] = j for k < n .
  • Rogers–Tanimoto diversity: The Rogers–Tanimoto dissimilarity between u and v is defined as:
    D R T = R C T T + C F F + R
    where c i j is the number of occurrences of u [ k ] = i and v [ k ] = j for k < n and R = 2 ( C T F + C F T ) .
  • Russellrao diversity: The Russellrao dissimilarity between u and v is defined as
    D R = n C T T n
    where c i j is the number of occurrences of u [ k ] = i and v [ k ] = j for k < n .
  • Sokal–Michener diversity: The Sokal–Michener dissimilarity between u and v is defined as
    D S M = R S + R
    where c i j is the number of occurrences of u [ k ] = i and v [ k ] = j for k < n , R = 2 ( C T F + C F T ) , and S = C F F + C T T .
  • Yule diversity: The Yule dissimilarity is defined as
    D Y = R C T T × C F F + R 2
    where c i j is the number of occurrences of u [ k ] = i and v [ k ] = j for k < n and R = 2.0 × C T F × C F T .
  • Sokal–Sneath diversity: The Sokal–Sneath dissimilarity between u and v is defined as
    D S S = R C T T + R
    where c i j is the number of occurrences of u [ k ] = i and v [ k ] = j for k < n and R = 2 ( C T F + C F T ) .
  • Dimension-wise diversity [13]: This metric calculates the increase and decrease in distance between search agents. Under this method, population diversity is defined as follows:
    D D w = 1 m n i = 1 m j = 1 n | m e d i a n ( x j ) x i j |
    where m e d i a n ( x j ) represents the median of dimension j across the population, x i j is dimension j of search agent i, n is the number of agents in the population, and m symbolizes the number of design variables in the optimization problem.
    Diversity in each dimension is defined as the distance between the dimension j of each search agent and the median of that dimension, averaged. The diversity of the entire population is then calculated by averaging each over each dimension. Both values are calculated in each iteration.
    The comprehensive balance response is characterized by the percentage of exploration and exploitation invested by a given metaheuristic scheme. These values are calculated in each iteration using the following models:
    X P L % = D D w m a x ( D D w ) × 100
    X P T % = | D D w m a x ( D D w ) | m a x ( D D w ) × 100
    where m a x ( D D w ) represents the maximum diversity value encountered throughout the optimization process. The exploration percentage XPL% represents the level of exploration as the ratio of diversity in each iteration to the maximum diversity achieved. The exploitation percentage XPT% corresponds to the level of exploitation. It is calculated as the complementary percentage to XPL% because the difference between the maximum diversity and the current diversity of an iteration is a consequence of the concentration of search agents [13].
    According to [13], achieving a good balance involves focusing on the following points in the result analysis:
    -
    Algorithms with a balance of the dimension-wise diversity metric percentage [81] above 90% exploitation and less than 10% exploration yield better results. For our problem, a percentage of over 70% exploitation and less than 30% exploration is used, as the optimization problem at hand does not reach such high exploitation values in algorithms due to the nature of the problem’s constraints.
    -
    The combination of competitive search mechanisms and an appropriate balance response is essential. Effective results are achieved through operators generating promising solutions that utilize the observed diversity conditions in the balance response.
The machine learning algorithms or individual learners utilize the dataset generated from the stored metrics to create the enhanced EL model.
In this work, we use supervised machine learning algorithms because they are pre-classified into two classes by the label. These classes indicate whether an exploration process is needed, activating the mechanism based on selection/crossover/mutation, or if exploitation is required using the original strategy of the algorithm. This approach aims to be applied throughout iterations of the population of wolves to modify their diversity.
For performing these classifications through a predictive model, the ML methods to be used are
  • DecisionTreeClassifier (DTC): One of the most well-known methods for data classification. DTC’s key feature is its capability to transform complex decision-making problems into simple processes, resulting in an understandable and more interpretable solution [82].
  • Support vector machines (SVM): This ML method attempts to find the best hyperplane that can be used to classify two different classes of data. SVM’s main objective is to find a hyperplane that maximizes the distance from itself to the nearest data points. This distance is called the margin. The larger the margin, the higher the likelihood of achieving lower errors in generalization [83].
  • Gaussian Naive Bayes (GaussianNB): A probabilistic classification mechanism rooted in Bayes’ Theorem. From a classification perspective, the primary goal is to find the best mapping between new data and a set of classifications within a specific problem domain [84].
  • Random forest (RF): Classification results of new data are based on the scores formed by the vote of many classification trees. Its essence is an enhancement of the decision tree algorithm with multiple decision trees combined. The construction of each tree depends on independently extracted samples [85,86].
The provided input data are used to train these algorithms, generating predictive models that are ultimately combined through ensemble learning methods. This produces an improved prediction model with fewer errors and better accuracy.

4.2. Phase 2: Ensemble Learning Model Training

An ensemble learning model is trained to predict the most effective team configurations using the collected metrics. This model utilizes several machine learning algorithms, including decision tree classifier, support vector machines, and random forest, to analyze patterns and derive insights from the historical data. The model assesses whether a team configuration needs more diversity (exploration) or should intensify focus on its current strengths (exploitation).
The combinatorial technique of weighted voting, also known as a voting classifier, is used with soft voting. As previously mentioned, the essence of this technique is to combine conceptually different machine learning classifiers and use averaged predicted probabilities to predict feature labels (see Figure 2). This EL method is useful as it allows the employment of heterogeneous or different classification models, leading to greater variety and better results when classifying [48].

4.3. Phase 3: Dynamic Adjustment and Real-Time Optimization

The metaheuristic algorithm dynamically adjusts team sizes and compositions by incorporating the predictions from the ensemble learning model. For that, the improved EL model is incorporated into optimization algorithms. Figure 3 depicts the proposal’s operation. In this algorithm instance, the EL model is assigned. The simulation of search agent behavior then begins, and the EL model, through iteration metrics, the strategy taken, and the fitness, helps in selecting a strategy or mechanism aimed at enhancing the balance of exploration and exploitation throughout the algorithm iterations. This process updates the position of each search agent. Finally, the results of the obtained metrics in the EL-executed algorithm instance are stored in another file (optional) for use in further model enhancements using the EL process.

4.4. Integration of Metaheuristics and Ensemble Learning in Agile Team Configuration

The integration of metaheuristic optimization with ensemble learning for optimizing team sizes in agile environments is elaborately structured in Algorithms 1 and 2. Algorithm 1 starts by setting the necessary input parameters specifically aimed at resolving the challenge of optimal team configuration, focusing particularly on minimizing communication channels within and between teams. This algorithm initiates by generating an initial population within predefined constraints, setting a foundation for optimal team interactions and communication efficiencies.
Algorithm 1: Pseudocode for Metaheuristic Optimization Enhanced by Ensemble Learning Solving the Team-Size Problem.
Biomimetics 09 00292 i001
As the process unfolds, Algorithm 1 systematically evaluates each potential team configuration within the loop, which continues until the maximum number of iterations (maxIter) is achieved. Each team’s configuration or agent is assessed for compliance with operational boundaries and effectiveness in communication minimization. The fitness of each team setup is evaluated based on its ability to reduce intra-team and inter-team communication channels. When a team’s configuration surpasses the previously established best in terms of reduced communication overhead, it is earmarked as the new optimal solution.
Algorithm 2: Pseudocode for Training the Ensemble Learning Model.
Biomimetics 09 00292 i002
Simultaneously, diversity metrics are collected for each team’s configuration to measure the variety and effectiveness of the communication strategies employed within teams. These metrics are crucial for the ensemble learning model, which uses them to determine the current optimization phase—whether to explore new team configurations or to exploit existing successful setups. Based on the model’s prediction, the algorithm adjusts each team’s structure using genetic algorithm techniques such as crossover and mutation for exploration or refines successful configurations for exploitation, tailoring adjustments to the specific dimensions of team size and communication pathways.
Algorithm 2 outlines how the ensemble learning model, used in Algorithm 1, is developed or refined. The process starts by loading data that capture the historical performance of various team setups along with their communication metrics. If an existing model is available, it is reused; otherwise, a new model is constructed using a blend of machine learning techniques—decision tree classifier, support vector machines, Gaussian Naive Bayes, and random forest—configured into a soft voting ensemble. This model undergoes meticulous training and calibration to accurately predict the most effective team sizes and configurations that minimize communication overhead. Once optimized, the model is saved for operational deployment. By leveraging a comprehensive array of machine learning techniques, this method ensures that the ensemble learning model is finely tuned to dynamically adapt team configurations in response to validated historical insights and current performance metrics.
The synergy between these algorithms not only ensures a dynamic optimization of team sizes, tailored to minimize communication channels within and between teams but also leads to the development of an optimal team configuration for each project instance. This configuration is supported by comprehensive performance and communication metrics, which are crucial for ongoing analysis and iterative improvement, thereby enhancing project management agility and efficiency.

5. Experimental Setup

This section outlines the test cases targeting the communication and collaboration challenges in agile practices and the algorithms’ implementation.
Before initiating the experimentation with the algorithms, the initial step involved defining the instances to be executed to capture the diversity metric results. As discussed in earlier sections, this process aims to construct the ensemble learning model using these outcomes. The instances to be defined are outlined in Table 1. Notably, all instances are conducted within the scope of teams ranging from a minimum of 3 to a maximum of 17 members. The selection of the experimental team size range of 3 to 17 members, despite the widely recognized ideal agile team size of 6 to 8, was strategically chosen to explore the flexibility and adaptability of our proposed optimization methods across a broader spectrum of team configurations. This extended range allows us to assess the performance and efficacy of our bio-inspired algorithms and ensemble learning enhancements under extreme and non-ideal conditions, often encountered in larger or smaller projects due to specific organizational needs, project scopes, or constraints. Furthermore, the broader range of team sizes allows for a detailed analysis of how communication dynamics shift with team scale, enhancing our understanding of the method’s scalability and robustness. By exploring these extremes, we pinpoint the boundaries of our algorithms’ applicability and identify necessary adjustments.
The parameters for these instances are set as follows: popSize represents the overall project size (to be executed for each case in the section), nAgents denotes the number of virtual agents to operate, maxIter describes the number of iterations, and mechanism refers to the update method for the bio-inspired algorithm. In the latter, the original value indicates the execution of the algorithm’s original exploitation-oriented behavior (m1), while m2 signifies the instantiation of the algorithm with the behavior, encompassing selection, crossover, and mutation. To generate the dataset, we employ all bio-inspired algorithms. The results will be stored and consumed posteriorly.
The next step includes the ML implementation and, consequently, the EL strategy. This is accomplished by assembling the classifiers within the VotingClassifier voting process, as depicted in Figure 4. This process generates a unified model from the metrics previously acquired. With the module prepared, it is executed, and the ensemble learning model is stored. This model offers improved predictions to guide strategy selection between exploration and exploitation, as indicated in [87].
The third step involves testing evaluation scenarios. The proposal was tested on seven state-of-the-art algorithms: PSO, GWO, FA, BA, WOA, CSA, and DE. We called PSOEL, GWOEL, FAEL, BAEL, WOAEL, CSAEL, and DEEL to each of them, respectively (codes can be downloaded in [88]). Table 2 describes the test instances that are carried out. These tests are executed 30 times to obtain a result that can be analyzed later. All the methods were implemented in Python and executed on macOS 14.2.1 Darwin Kernel version 23 with an Ultra M2 chip and 64 GB of RAM. The codes are available from [89].

6. Results and Discussion

The results of the experiments pertaining to the optimization problem are presented in Table 3. The findings reveal that the optimal team size and composition, considering a fixed total number of team members, generally ranges between six and eight members per team. This outcome is consistent with the established management practices documented in various studies [15,16,17] and closely mirrors the results reported in [11].
In all six scenarios, the algorithms consistently showcased superior performance in discovering optimal solutions: PSOEL, PSO, GWOEL, and GWO excelled compared to their counterparts. These algorithms diligently evaluated and minimized team communication channels, achieving remarkable outcomes that significantly enhance the effectiveness of communications within teams. This optimization aligns with the principle that effective communication within agile teams deteriorates as the number of communication channels per person increases. Our model aims to mitigate this problem by optimizing team configurations to reduce these channels.
The ensemble learning-enhanced algorithms under study were tested for 1500 iterations, and each was run 30 times to ensure the robustness of results, as detailed in Table 4. Our objective function, designed to minimize the total number of communication channels among teams ( f ( x ) or N T ), showcased how effectively these algorithms could reduce unnecessary communication overhead, thereby streamlining interactions and improving project efficiency.
These insightful findings provide a clear direction for organizations that might otherwise rely on assumptions or emulate practices from other companies. Traditional methodologies often struggle with team sizes above 40 members, whereas agile methodologies like Scrum suggest practical solutions by subdividing larger teams into smaller, more manageable units. Our results endorse this approach, indicating that even in larger project settings, teams can be optimally divided into groups of 5 to 9 members to maintain communication efficacy and individual productivity.
Finally, to underscore the robustness of our findings regarding optimal team size configurations, we applied the Kruskal–Wallis H test. This non-parametric method was selected due to the non-normal distribution of our data. The test was conducted to assess if there were statistically significant differences in the performance medians of the various bio-inspired algorithms evaluated. Each algorithm was treated as a separate group, focusing on its performance metrics as reported in Table 5. The Kruskal–Wallis H test confirmed significant variations in algorithm performance ( p < 0.05 ), indicating that not all algorithms perform equally in optimizing team sizes within agile environments.
These variations highlight the need for selecting optimization algorithms that are specifically tailored to project specifics and team dynamics in agile settings. The enhanced algorithms consistently outperformed the standard versions, demonstrating that incorporating advanced features such as ensemble learning can significantly improve the optimization of team sizes. This adaptability is crucial for agile environments, suggesting that strategic algorithm selection, aligned with project needs and team dynamics, is vital for enhancing agility and efficiency in project management.
The mentioned algorithms’ exploration and exploitation behaviors were analyzed, and it was observed that integration with ensemble learning slightly enhances the attainment of an optimal value. These behaviors and their impact on the search process are further illustrated in Figure 5 and Figure 6, which show the exploration and exploitation activities within the search space of each algorithm.

7. Conclusions

The agile approach to software engineering, which emphasizes collaboration and adaptability to change, presents the challenge of scalability in terms of communication within development teams. The previous studies have demonstrated that small, self-organizing, and independent teams are more efficient in software delivery. To address this challenge, frameworks and approaches like domain-driven design and microservices have been proposed, promoting team modularization and system architecture.
In this context, the question of the ideal size of a software development team arises. This study identified the opportunity to enhance the results of previously conducted research by employing artificial intelligence techniques, specifically ensemble learning. Integrating EL into optimization algorithms was proposed to predict and determine how the search agent’s position should be updated, potentially leading to more efficient solutions in team size distribution for agile projects.
The results obtained using the EL technique applied to optimization algorithms showed that the optimal team size and distribution range between six and eight members per team, which aligns with industry practices. The PSOEL, PSO, GWOEL, and GWO algorithms demonstrated better performance in finding optimal solutions in terms of minimizing communication channels. Although this research successfully addressed an optimization problem, there is still a need to develop more optimization models that accurately represent the communication and collaboration dynamics in practical agile settings within organizations. This necessity is underscored by the limited number of studies on this topic, as discussed in previous sections. While the algorithms chosen for this study showed optimal performance in preliminary evaluations, the future research could explore additional algorithms with similar characteristics. This exploration would not only potentially enhance the outcomes but also validate our findings, ensuring broader applicability and robustness of the optimization models in real-world agile environments.
With this in mind, the future research should explore incorporating elements related to project-involved resources, required time, workspace or organization’s meeting spaces, official communication channels within the organization, and potential mood states during intervals demanding communication and collaboration with agile practice teams from the multifactorial optimization perspective.
In our forthcoming research, we will delve deeper into applying advanced machine learning techniques, such as deep learning and reinforcement learning models, to analyze and optimize team dynamics in agile environments. Specifically, we plan to utilize deep Q-learning and proximal policy optimization to dynamically model and refine the decision-making processes involved in team size configuration. These models will leverage historical data from previous projects to recognize complex collaboration, communication, and team performance patterns. This approach allows for real-time adjustments in team configurations to maximize project efficiency. By integrating these sophisticated techniques, we aim to develop models that provide more accurate, personalized recommendations for team formation, thereby enhancing the precision, adaptability, and effectiveness of our recommendations, ultimately leading to improved project outcomes and enhanced team performance.

Author Contributions

Formal analysis, R.O., R.N. and R.M.; investigation, R.O., R.N, S.M.G., D.M. and R.M.; methodology, R.O., R.N. and R.M.; resources, R.M.; software, R.O., S.M.G. and. R.M.; validation, R.O., R.N, S.M.G., D.M. and R.M.; writing—original draft, R.O., R.N. and R.M.; writing—review and editing, R.O., R.N, S.M.G., D.M. and R.M. All the authors of this paper hold responsibility for every part of this manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

R.O. is supported by grant ANID/FONDECYT/INICIACION/11231016. R.M. is supported by grant ANID/FONDECYT/REGULAR/1211905. S.M.G. is supported by Programa de Magíster en Informática Aplicada—Universidad de Valparaíso. D.M. is supported by ANID/BECA/DOCTORADO NACIONAL/21231737.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data and codes are available from [89].

Conflicts of Interest

The authors declare no conflicts of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Beck, K.; Beedle, M.; Van Bennekum, A.; Cockburn, A.; Cunningham, W.; Fowler; Thomas, D. The Agile Manifesto; Agile Alliance: Dallas, TX, USA, 2001. [Google Scholar]
  2. Beck, K. Extreme Programming Explained: Embrace Change; Addison-Wesley Professional: Boston, MA, USA, 2000. [Google Scholar]
  3. Forsgren, N.; Humbpotifle, J.; Kim, G. Accelerate: The Science of Lean Software and DevOps Building and Scaling High Performing Technology Organizations; IT Revolution Press: Portland, OR, USA, 2018. [Google Scholar]
  4. Scaled Agile, Inc. SAFe 5 for Lean Enterprises. Available online: https://www.scaledagileframework.com/ (accessed on 4 October 2021).
  5. Larman, C.; Vodde, B. Large-Scale Scrum: More with LeSS; Addison-Wesley Professional: Boston, MA, USA, 2016. [Google Scholar]
  6. Evans, E.; Evans, E.J. Domain-Driven Design: Tackling Complexity in the Heart of Software; Addison-Wesley Professional: Boston, MA, USA, 2004. [Google Scholar]
  7. Lewis, J.; Fowler, M. Microservices: A Definition of This New Architectural Term. 2014. Available online: https://martinfowler.com/articles/microservices.html (accessed on 5 May 2024).
  8. Highsmith, J.; Luu, L.; Robinson, D. EDGE: Value-Driven Digital Transformation; Addison-Wesley Professional: Boston, MA, USA, 2019. [Google Scholar]
  9. Skelton, M.; Pais, M. Team Topologies: Organizing Business and Technology Teams for Fast Flow; It Revolution: Portland, OR, USA, 2019. [Google Scholar]
  10. Service, A.W. Two-Pizza Teams. 2023. Available online: https://docs.aws.amazon.com/whitepapers/latest/introduction-devops-aws/two-pizza-teams.html (accessed on 25 July 2023).
  11. Almadhoun, W.; Hamdan, M. Optimizing the Self-Organizing Team Size Using a Genetic Algorithm in Agile Practices. J. Intell. Syst. 2018, 29, 1151–1165. [Google Scholar] [CrossRef]
  12. Talbi, E.G. Metaheuristics: From Design to Implementation; John Wiley and Sons, Inc.: Hoboken, NY, USA, 2009. [Google Scholar]
  13. Morales-Castañeda, B.; Zaldívar, D.; Cuevas, E.; Fausto, F.; Rodríguez, A. A better balance in metaheuristic algorithms: Does it exist? Swarm Evol. Comput. 2020, 54, 100671. [Google Scholar] [CrossRef]
  14. Calvet, L.; de Armas, J.; Masip, D.; Juan, A.A. Learnheuristics: Hybridizing metaheuristics with machine learning for optimization with dynamic inputs. Open Math. 2017, 15, 261–280. [Google Scholar] [CrossRef]
  15. Rodríguez, D.; Sicilia, M.; García, E.; Harrison, R. Empirical findings on team size and productivity in software development. J. Syst. Softw. 2012, 85, 562–570. [Google Scholar] [CrossRef]
  16. Heričko, M.; Živkovič, A.; Rozman, I. An approach to optimizing software development team size. Inf. Process. Lett. 2008, 108, 101–106. [Google Scholar] [CrossRef]
  17. Pendharkar, P.C.; Rodger, J.A. The relationship between software development team size and software development cost. Commun. Acm 2009, 52, 141–144. [Google Scholar] [CrossRef]
  18. Cristóbal, J.R.S.; Carral, L.; Diaz, E.; Fraguela, J.A.; Iglesias, G. Complexity and Project Management: A General Overview. Complexity 2018, 2018, 1–10. [Google Scholar] [CrossRef]
  19. Apaolaza, U.; Lizarralde, A.; Oyarbide-Zubillaga, A. Modern Project Management Approaches in Uncertainty Environments: A Comparative Study Based on Action Research. Sustainability 2020, 12, 10542. [Google Scholar] [CrossRef]
  20. Hallegatte, S. Strategies to adapt to an uncertain climate change. Glob. Environ. Change 2009, 19, 240–247. [Google Scholar] [CrossRef]
  21. Fernandez, D.J.; Fernandez, J.D. Agile Project Management -Agilism versus Traditional Approaches. J. Comput. Inf. Syst. 2008, 49, 10–17. [Google Scholar]
  22. Malik, M.; Sarwar, S.; Orr, S. Agile practices and performance: Examining the role of psychological empowerment. Int. J. Proj. Manag. 2021, 39, 10–20. [Google Scholar] [CrossRef]
  23. Bach-Dabrowska, I.; Pawlewski, P. Optimization model of agile team’s cohesion: Knowledge-based and intelligent information and engineering systems. In Proceedings of the 18th Annual Conference, KES—2014, Gdynia, Poland, 15–17 September 2014; Volume 35, pp. 1577–1585. [Google Scholar] [CrossRef]
  24. Črepinšek, M.; Liu, S.H.; Mernik, M. Exploration and Exploitation in Evolutionary Algorithms: A Survey. ACM Comput. Surv. 2013, 45, 1–33. [Google Scholar] [CrossRef]
  25. Gómez-Rubio, Á.; Soto, R.; Crawford, B.; Jaramillo, A.; Mancilla, D.; Castro, C.; Olivares, R. Applying Parallel and Distributed Models on Bio-Inspired Algorithms via a Clustering Method. Mathematics 2022, 10, 274. [Google Scholar] [CrossRef]
  26. Caselli, N.; Soto, R.; Crawford, B.; Valdivia, S.; Olivares, R. A Self-Adaptive Cuckoo Search Algorithm Using a Machine Learning Technique. Mathematics 2021, 9, 1840. [Google Scholar] [CrossRef]
  27. Soto, R.; Crawford, B.; Molina, F.G.; Olivares, R. Human Behaviour Based Optimization Supported with Self-Organizing Maps for Solving the S-Box Design Problem. IEEE Access 2021, 9, 84605–84618. [Google Scholar] [CrossRef]
  28. Valdivia, S.; Soto, R.; Crawford, B.; Caselli, N.; Paredes, F.; Castro, C.; Olivares, R. Clustering-Based Binarization Methods Applied to the Crow Search Algorithm for 0/1 Combinatorial Problems. Mathematics 2020, 8, 1070. [Google Scholar] [CrossRef]
  29. Zennaki, M.; Ech-Cherif, A. A new machine learning based approach for tuning metaheuristics for the solution of hard combinatorial optimization problems. J. Appl. Sci. 2010, 10, 1991–2000. [Google Scholar] [CrossRef]
  30. Maturana, J.; Lardeux, F.; Saubion, F. Autonomous operator management for evolutionary algorithms. J. Heuristics 2010, 16, 881–909. [Google Scholar] [CrossRef]
  31. dos Santos, J.P.Q.; de Melo, J.D.; Neto, A.D.D.; Aloise, D. Reactive search strategies using reinforcement learning, local search algorithms and variable neighborhood search. Expert Syst. Appl. 2014, 41, 4939–4949. [Google Scholar] [CrossRef]
  32. Lessmann, S.; Caserta, M.; Arango, I.M. Tuning metaheuristics: A data mining based approach for particle swarm optimization. Expert Syst. Appl. 2011, 38, 12826–12838. [Google Scholar] [CrossRef]
  33. Liang, X.; Li, W.; Zhang, Y.; Zhou, M. An adaptive particle swarm optimization method based on clustering. Soft Comput. 2015, 19, 431–448. [Google Scholar] [CrossRef]
  34. Harrison, K.R.; Ombuki-Berman, B.M.; Engelbrecht, A.P. A parameter-free particle swarm optimization algorithm using performance classifiers. Inf. Sci. 2019, 503, 381–400. [Google Scholar] [CrossRef]
  35. Dong, W.; Zhou, M. A supervised learning and control method to improve particle swarm optimization algorithms. IEEE Trans. Syst. Man Cybern. Syst. 2016, 47, 1135–1148. [Google Scholar] [CrossRef]
  36. Liu, X.; Wang, N. A novel gray wolf optimizer with RNA crossover operation for tackling the non-parametric modeling problem of FCC process. Knowl.-Based Syst. 2021, 216, 106751. [Google Scholar] [CrossRef]
  37. Adhikary, J.; Acharyya, S. Randomized Balanced Grey Wolf Optimizer (RBGWO) for solving real life optimization problems. Appl. Soft Comput. 2022, 117, 108429. [Google Scholar] [CrossRef]
  38. Preeti; Deep, K. A random walk Grey wolf optimizer based on dispersion factor for feature selection on chronic disease prediction. Expert Syst. Appl. 2022, 206, 117864. [Google Scholar] [CrossRef]
  39. Nadimi-Shahraki, M.H.; Taghian, S.; Mirjalili, S. An improved grey wolf optimizer for solving engineering problems. Expert Syst. Appl. 2021, 166, 113917. [Google Scholar] [CrossRef]
  40. Peng, H.; Xiao, W.; Han, Y.; Jiang, A.; Xu, Z.; Li, M.; Wu, Z. Multi-strategy firefly algorithm with selective ensemble for complex engineering optimization problems. Appl. Soft Comput. 2022, 120, 108634. [Google Scholar] [CrossRef]
  41. Michalski, R.S. Learnable Evolution Model: Evolutionary Processes Guided by Machine Learning. Mach. Learn. 2000, 38, 9–40. [Google Scholar] [CrossRef]
  42. Jourdan, L.; Dhaenens, C.; Talbi, E.G. Using datamining techniques to help metaheuristics: A short survey. In Proceedings of the Hybrid Metaheuristics, Gran Canaria, France, 13–14 October 2006; Volume 4030, pp. 57–69. [Google Scholar]
  43. Pikkarainen, M.; Haikara, J.; Salo, O.; Abrahamsson, P.; Still, J. The impact of agile practices on communication in software development. Empir. Softw. Eng. 2008, 13, 303–337. [Google Scholar] [CrossRef]
  44. Beck, K. Embracing change with extreme programming. Computer 1999, 32, 70–77. [Google Scholar] [CrossRef]
  45. Schwaber, K.; Beedle, M. Agile Software Development with Scrum, 1st ed.; Prentice Hall PTR: Hoboken, NJ, USA, 2001. [Google Scholar]
  46. Wysocki, R.K. Adaptive Project Framework; Addison-Wesley Educational: Boston, MA, USA, 2010. [Google Scholar]
  47. Cockburn, A. Crystal Clear; The Agile Software Development Series; Addison Wesley: Boston, MA, USA, 2004. [Google Scholar]
  48. Zhou, Z.H. Machine Learning; Springer: Singapore, 2021. [Google Scholar] [CrossRef]
  49. Yang, Y. Chapter 4—Ensemble learning. In Temporal Data Mining via Unsupervised Ensemble Learning; Yang, Y., Ed.; Elsevier: Amsterdam, The Netherlands, 2017; pp. 35–56. [Google Scholar] [CrossRef]
  50. Ma, Z.; Wu, G.; Suganthan, P.N.; Song, A.; Luo, Q. Performance assessment and exhaustive listing of 500+ nature-inspired metaheuristic algorithms. Swarm Evol. Comput. 2023, 77, 101248. [Google Scholar] [CrossRef]
  51. Crawford, B.; Soto, R.; Astorga, G.; García, J.; Castro, C.; Paredes, F. Putting Continuous Metaheuristics to Work in Binary Search Spaces. Complexity 2017, 2017, 1–19. [Google Scholar] [CrossRef]
  52. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar] [CrossRef]
  53. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  54. Varan, M.; Erduman, A.; Menevşeoğlu, F. A Grey Wolf Optimization Algorithm-Based Optimal Reactive Power Dispatch with Wind-Integrated Power Systems. Energies 2023, 16, 5021. [Google Scholar] [CrossRef]
  55. Silaa, M.Y.; Barambones, O.; Bencherif, A.; Rahmani, A. A New MPPT-Based Extended Grey Wolf Optimizer for Stand-Alone PV System: A Performance Evaluation versus Four Smart MPPT Techniques in Diverse Scenarios. Inventions 2023, 8, 142. [Google Scholar] [CrossRef]
  56. Too, J.; Abdullah, A.; Mohd Saad, N.; Mohd Ali, N.; Tee, W. A New Competitive Binary Grey Wolf Optimizer to Solve the Feature Selection Problem in EMG Signals Classification. Computers 2018, 7, 58. [Google Scholar] [CrossRef]
  57. Yang, X.S. Firefly algorithms for multimodal optimization. In Stochastic Algorithms: Foundations and Applications; Springer: Berlin/Heidelberg, Germany, 2009; pp. 169–178. [Google Scholar] [CrossRef]
  58. Tariq, F.; Alelyani, S.; Abbas, G.; Qahmash, A.; Hussain, M.R. Solving Renewables-Integrated Economic Load Dispatch Problem by Variant of Metaheuristic Bat-Inspired Algorithm. Energies 2020, 13, 6225. [Google Scholar] [CrossRef]
  59. Yousif, A.; Alqhtani, S.M.; Bashir, M.B.; Ali, A.; Hamza, R.; Hassan, A.; Tawfeeg, T.M. Greedy Firefly Algorithm for Optimizing Job Scheduling in IoT Grid Computing. Sensors 2022, 22, 850. [Google Scholar] [CrossRef]
  60. Sánchez-Olivares, G.; Tomás, A.; García-Ayllón, S. A Minimum Cost Design Approach for Steel Frames Based on a Parallelized Firefly Algorithm and Parameter Control. Appl. Sci. 2023, 13, 11801. [Google Scholar] [CrossRef]
  61. Yang, X.-S. A new metaheuristic bat-inspired algorithm. In Nature Inspired Cooperative Strategies for Optimization (NICSO 2010); Springer: Berlin/Heidelberg, Germany, 2010; pp. 65–74. [Google Scholar] [CrossRef]
  62. Saleem, N.; Zafar, K.; Sabzwari, A. Enhanced Feature Subset Selection Using Niche Based Bat Algorithm. Computation 2019, 7, 49. [Google Scholar] [CrossRef]
  63. Nguyen, T.T.; Pan, J.S.; Dao, T.K. A Compact Bat Algorithm for Unequal Clustering in Wireless Sensor Networks. Appl. Sci. 2019, 9, 1973. [Google Scholar] [CrossRef]
  64. Kumar Mohapatra, P.; Kumar Rout, S.; Kishoro Bisoy, S.; Kautish, S.; Hamzah, M.; Jasser, M.B.; Mohamed, A.W. Application of Bat Algorithm and Its Modified Form Trained with ANN in Channel Equalization. Symmetry 2022, 14, 2078. [Google Scholar] [CrossRef]
  65. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  66. Sun, G.; Shang, Y.; Yuan, K.; Gao, H. An Improved Whale Optimization Algorithm Based on Nonlinear Parameters and Feedback Mechanism. Int. J. Comput. Intell. Syst. 2022, 15, 38. [Google Scholar] [CrossRef]
  67. Yildiz, A.R. A novel hybrid whale–Nelder–Mead algorithm for optimization of design and manufacturing problems. Int. J. Adv. Manuf. Technol. 2019, 105, 5091–5104. [Google Scholar] [CrossRef]
  68. Zhang, J.; Zhang, T.; Zhang, G.; Wang, D.; Kong, M. Using the Whale Optimization Algorithm to Solve the Optimal Reactive Power Dispatch Problem. Processes 2023, 11, 1513. [Google Scholar] [CrossRef]
  69. Chen, Z.; Zhang, K.; Chan, T.H.T.; Li, X.; Zhao, S. A Novel Hybrid Whale-Chimp Optimization Algorithm for Structural Damage Detection. Appl. Sci. 2022, 12, 9036. [Google Scholar] [CrossRef]
  70. Askarzadeh, A. A novel metaheuristic method for solving constrained engineering optimization problems: Crow search algorithm. Comput. Struct. 2016, 169, 1–12. [Google Scholar] [CrossRef]
  71. Huang, K.W.; Girsang, A.; Wu, Z.X.; Chuang, Y.W. A Hybrid Crow Search Algorithm for Solving Permutation Flow Shop Scheduling Problems. Appl. Sci. 2019, 9, 1353. [Google Scholar] [CrossRef]
  72. Meddeb, A.; Amor, N.; Abbes, M.; Chebbi, S. A Novel Approach Based on Crow Search Algorithm for Solving Reactive Power Dispatch Problem. Energies 2018, 11, 3321. [Google Scholar] [CrossRef]
  73. Storn, R.; Price, K. Differential Evolution - A Simple and Efficient Heuristic for global Optimization over Continuous Spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  74. Naila; Haroon, S.; Hassan, S.; Amin, S.; Sajjad, I.; Waqar, A.; Aamir, M.; Yaqoob, M.; Alam, I. Multiple Fuel Machines Power Economic Dispatch Using Stud Differential Evolution. Energies 2018, 11, 1393. [Google Scholar] [CrossRef]
  75. Baioletti, M.; Di Bari, G.; Milani, A.; Poggioni, V. Differential Evolution for Neural Networks Optimization. Mathematics 2020, 8, 69. [Google Scholar] [CrossRef]
  76. Eltaeib, T.; Mahmood, A. Differential Evolution: A Survey and Analysis. Appl. Sci. 2018, 8, 1945. [Google Scholar] [CrossRef]
  77. Mattiussi, C.; Waibel, M.; Floreano, D. Measures of Diversity for Populations and Distances Between Individuals with Highly Reorganizable Genomes. Evol. Comput. 2004, 12, 495–515. [Google Scholar] [CrossRef] [PubMed]
  78. Norouzi, M.; Fleet, D.J.; Salakhutdinov, R.R. Hamming distance metric learning. In Advances in Neural Information Processing Systems; Pereira, F., Burges, C., Bottou, L., Weinberger, K., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2012; Volume 25. [Google Scholar]
  79. Scipy. Distance Computations, 2008–2022. Available online: https://docs.scipy.org/doc/scipy/reference/spatial.distance.html (accessed on 5 May 2024).
  80. Wikipedia. Jaccard Index 2022. Available online: https://en.wikipedia.org/wiki/Jaccard_index (accessed on 5 May 2024).
  81. Salleh, M.N.M.; Hussain, K.; Cheng, S.; Shi, Y.; Muhammad, A.; Ullah, G.; Naseem, R. Exploration and exploitation measurement in swarm-based metaheuristic algorithms: An empirical analysis. In Proceedings of the Recent Advances on Soft Computing and Data Mining, Cham, Switzerland, 16–18 June 2018; Ghazali, R., Deris, M.M., Nawi, N.M., Abawajy, J.H., Eds.; pp. 24–32. [Google Scholar]
  82. Priyanka; Kumar, D. Decision tree classifier: A detailed survey. Int. J. Inf. Decis. Sci. 2020, 12, 246–269. [Google Scholar] [CrossRef]
  83. Rustam, Z.; Ariantari, N.P.A.A. Support Vector Machines for Classifying Policyholders Satisfactorily in Automobile Insurance. J. Phys. Conf. Ser. 2018, 1028, 012005. [Google Scholar] [CrossRef]
  84. Yang, F.J. An implementation of naive bayes classifier. In Proceedings of the 2018 International Conference on Computational Science and Computational Intelligence (CSCI), Las Vegas, NV, USA, 12–14 December 2018; pp. 301–306. [Google Scholar] [CrossRef]
  85. Zhang, L.; Liu, K.; Wang, Y.; Omariba, Z.B. Ice Detection Model of Wind Turbine Blades Based on Random Forest Classifier. Energies 2018, 11, 2548. [Google Scholar] [CrossRef]
  86. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  87. Sagi, O.; Rokach, L. Ensemble learning: A survey. Wires Data Min. Knowl. Discov. 2018, 8, e1249. [Google Scholar] [CrossRef]
  88. Olivares, R.; Noel, R.; Medina, S.; Miranda, D.; Munoz, R. Metaheuristics with Ensemble Learning to Solve the Team-Size Problem in Agile Practices. Electronics 2023, 13, 178. [Google Scholar] [CrossRef]
  89. Ravelo, C.; Olivares, R. Biomimetic Orca Predator Algorithm improved by Deep Reinforcement Learning for Feature Selection. Mathematics 2024, 12, 1249. [Google Scholar] [CrossRef]
Figure 1. General workflow of an EL inspired by [49].
Figure 1. General workflow of an EL inspired by [49].
Biomimetics 09 00292 g001
Figure 2. Phase 1: Creation of the enhanced ensemble learning model in the proposed solution.
Figure 2. Phase 1: Creation of the enhanced ensemble learning model in the proposed solution.
Biomimetics 09 00292 g002
Figure 3. Phase 2: Execution of an algorithm instance with enhanced ensemble learning.
Figure 3. Phase 2: Execution of an algorithm instance with enhanced ensemble learning.
Biomimetics 09 00292 g003
Figure 4. Creation of the ensemble learning model.
Figure 4. Creation of the ensemble learning model.
Biomimetics 09 00292 g004
Figure 5. Convergenceand boxplot charts displaying results of all algorithms involved in all experiments.
Figure 5. Convergenceand boxplot charts displaying results of all algorithms involved in all experiments.
Biomimetics 09 00292 g005
Figure 6. Exploration and exploitation percentages of PSO, PSOEL, GWO, GWOEL, FA, FAEL, BA, BAEL, WOA, WOAEL, CSA, CSAEL, DE, and DEEL involved in Exp6.
Figure 6. Exploration and exploitation percentages of PSO, PSOEL, GWO, GWOEL, FA, FAEL, BA, BAEL, WOA, WOAEL, CSA, CSAEL, DE, and DEEL involved in Exp6.
Biomimetics 09 00292 g006
Table 1. Test instances for dataset generation.
Table 1. Test instances for dataset generation.
PopSize nAgents MaxIter Mechanism
50, 100, 150, 2006250m1
50, 100, 150, 2006250m2
50, 100, 150, 20012250m1
50, 100, 150, 20012250m2
50, 100, 150, 2006500m1
50, 100, 150, 2006500m2
50, 100, 150, 20012500m1
50, 100, 150, 20012500m2
Table 2. Experimental instances.
Table 2. Experimental instances.
IDSize of the ProjectMaxIternAgents
Exp15015006
Exp27015006
Exp310015006
Exp415015006
Exp517015006
Exp620015006
Table 3. Outcome of the top-performing individuals in the best algorithms for each experiment.
Table 3. Outcome of the top-performing individuals in the best algorithms for each experiment.
IDSize of the ProjectProposalFitnessSolution Vector
Exp150GWO159.504[6, 5, 5, 5, 5, 6, 5, 6, 5]
Exp270PSOEL259.92[6, 5, 5, 6, 6, 6, 6, 5, 5, 6, 6, 6]
Exp3100PSOEL431.97[6, 6, 7, 6, 6, 7, 7, 7, 6, 6, 6, 7, 7, 7, 7]
Exp4150PSOEL763.89[8, 7, 8, 8, 7, 8, 7, 8, 7, 8, 9, 8, 7, 8, 8, 8, 8, 8, 8]
Exp5170PSO910.94[8, 8, 9, 8, 8, 9, 8, 8, 8, 7, 10, 8, 8, 8, 8, 8, 7, 8, 7, 7, 8]
Exp6200PSOEL1142.86[8, 8, 10, 9, 8, 9, 8, 8, 7, 8, 8, 9, 8, 8, 8, 9, 8, 7, 8, 10, 7, 9, 8, 8]
Table 4. Comparison of best, worst, mean, median, and standard deviation values for GWO, BA, DE, FA, PSO, WOA, and CSA against the proposed method applied to them, namely, GWOEL, BAEL, FAEL, DEEL, PSOEL, WOAEL, and CSAEL metaheuristic algorithms using the problem size team over 30 runs, 1500 iterations, and six search agents.
Table 4. Comparison of best, worst, mean, median, and standard deviation values for GWO, BA, DE, FA, PSO, WOA, and CSA against the proposed method applied to them, namely, GWOEL, BAEL, FAEL, DEEL, PSOEL, WOAEL, and CSAEL metaheuristic algorithms using the problem size team over 30 runs, 1500 iterations, and six search agents.
GWOGWOELFAFAELPSOPSOELBABAELWOAWOAELCSACSAELDEDEEL
best159.504160.76167.13166.42160.30160.14183.28179.00161.15160.88177.59175.41166.05165.88
worst230.80230.44227.97230.71230.52232.28207.87215.24230.51223.71213.77211.34217.33217.33
mean161.76161.99167.32166.71161.74161.94183.35179.12162.54162.60179.22177.51166.26166.10
med160.55161.39167.13166.42160.37160.49183.28179.00161.42161.58178.58176.82166.05165.88
sd3.382.892.012.183.593.690.761.473.683.381.812.382.222.19
best262.26263.65276.82276.61261.13259.92303.59297.60263.80264.23293.23282.06270.16272.58
worst347.76347.76347.90349.45358.69350.74327.86337.15349.86348.81325.27327.46331.86336.70
mean265.81266.18277.11276.88264.20264.32303.74297.76267.07268.31296.34286.02270.47272.95
med263.96265.05276.82276.61261.41261.54303.62297.63265.17266.49296.05284.32270.16272.58
sd5.244.172.602.605.965.580.861.435.716.202.775.182.793.00
best438.26443.36471.56469.59434.58431.97497.34488.20442.27441.49491.01467.87449.33451.20
worst544.62539.36552.78543.38549.77548.74516.35528.71542.30549.52524.90531.49533.45536.29
mean444.81446.76471.89469.95440.49440.12497.62488.41447.13449.30496.75473.22450.04451.96
med441.19445.01471.56469.59435.88434.97497.53488.23443.85444.53494.05470.05449.33451.20
sd9.325.543.143.189.289.570.821.488.9811.475.599.524.644.75
best782.79789.54850.82847.54769.76763.89869.50867.52777.52780.08853.06821.74789.24791.78
worst911.38917.40917.61914.85913.95909.36888.79906.04909.52909.18903.43900.72902.20908.39
mean794.70795.92851.18848.02780.50780.40869.83868.27788.26792.89865.36832.24790.48793.19
med788.41792.65850.82847.54772.71772.88869.53867.98781.53786.22861.98827.47789.24791.78
sd15.349.893.053.1815.4914.630.861.6416.5817.4010.9415.036.847.39
best932.61942.731014.521019.88910.94918.321025.241018.93929.88931.101017.99957.54938.94941.23
worst1081.511077.371073.771069.051070.651067.411046.721053.581062.271070.401057.701051.731054.991062.59
mean947.71950.121014.791020.19929.02930.461025.661019.73943.49945.541027.23973.24940.41942.82
med937.12945.671014.521019.88920.02922.451025.461019.25936.78937.671021.76964.95938.94941.23
sd19.9410.552.762.2216.6117.201.071.6418.0019.089.4422.367.868.35
best1175.581186.541280.261289.471151.931142.861277.781273.041164.971165.401262.371259.071183.341175.52
worst1311.641316.901309.931317.351319.931315.141300.081304.631306.931313.741311.611311.241317.371301.26
mean1193.741194.291280.431289.551166.581164.431278.341273.741183.291182.991274.041273.971185.401177.52
med1185.541192.531280.261289.471156.711153.321277.901273.301177.141175.111270.481272.301183.341175.52
sd22.9110.591.400.9020.0020.781.121.6821.2122.1711.1511.199.549.07
Table 5. Kruskal–Wallis H Test Results for Standard vs. Enhanced Algorithms.
Table 5. Kruskal–Wallis H Test Results for Standard vs. Enhanced Algorithms.
AlgorithmMedian Scorep-Value
Standard Gray Wolf Optimizer1500.038
Enhanced Gray Wolf Optimizer180
Standard Firefly Algorithm1520.050
Enhanced Firefly Algorithm178
Standard Particle Swarm Optimization1600.045
Enhanced Particle Swarm Optimization195
Standard Bat Algorithm1480.025
Enhanced Bat Algorithm175
Standard Whale Optimization Algorithm1550.033
Enhanced Whale Optimization Algorithm188
Standard Crow Search Algorithm1600.042
Enhanced Crow Search Algorithm190
Standard Differential Evolution1580.020
Enhanced Differential Evolution185
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Olivares, R.; Noel, R.; Guzmán, S.M.; Miranda, D.; Munoz, R. Intelligent Learning-Based Methods for Determining the Ideal Team Size in Agile Practices. Biomimetics 2024, 9, 292. https://doi.org/10.3390/biomimetics9050292

AMA Style

Olivares R, Noel R, Guzmán SM, Miranda D, Munoz R. Intelligent Learning-Based Methods for Determining the Ideal Team Size in Agile Practices. Biomimetics. 2024; 9(5):292. https://doi.org/10.3390/biomimetics9050292

Chicago/Turabian Style

Olivares, Rodrigo, Rene Noel, Sebastián M. Guzmán, Diego Miranda, and Roberto Munoz. 2024. "Intelligent Learning-Based Methods for Determining the Ideal Team Size in Agile Practices" Biomimetics 9, no. 5: 292. https://doi.org/10.3390/biomimetics9050292

Article Metrics

Back to TopTop