You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

8 March 2023

HWACOA Scheduler: Hybrid Weighted Ant Colony Optimization Algorithm for Task Scheduling in Cloud Computing

,
,
,
and
1
School of Computer Science and Engineering, Vellore Institute of Technology (VIT), Chennai 600127, India
2
Center for Cyber Physical Systems, School of Computer Science and Engineering, Vellore Institute of Technology (VIT), Chennai 600127, India
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Cyber–Physical Systems in Real-Time and Edge Computing for Smart Grids

Abstract

With the advancement of technology and time, people have always sought to solve problems in the most efficient and quickest way possible. Since the introduction of the cloud computing environment along with many different sub-substructures such as task schedulers, resource allocators, resource monitors, and others, various algorithms have been proposed to improve the performance of the individual unit or structure used in the cloud environment. The cloud is a vast virtual environment with the capability to solve any task provided by the user. Therefore, new algorithms are introduced with the aim to improve the process and consume less time to evaluate the process. One of the most important sections of cloud computing is that of the task scheduler, which is responsible for scheduling tasks to each of the virtual machines in such a way that the time taken to execute the process is less and the efficiency of the execution is high. Thus, this paper plans to propose an ideal and optimal task scheduling algorithm that is tested and compared with other existing algorithms in terms of efficiency, makespan, and cost parameters, that is, this paper tries to explain and solves the scheduling problem using an improved meta-heuristic algorithm called the Hybrid Weighted Ant Colony Optimization (HWACO) algorithm, which is an advanced form of the already present Ant Colony Optimization Algorithm. The outcomes found by using the proposed HWACO has more benefits, that is, the objective for reaching the convergence in a short period of time was accomplished; thus, the projected model outdid the other orthodox algorithms such as Ant Colony Optimization (ACO), Quantum-Based Avian Navigation Optimizer Algorithm (QANA), Modified-Transfer-Function-Based Binary Particle Swarm Optimization (MTF-BPSO), MIN-MIN Algorithm (MM), and First-Come-First-Serve (FCFS), making the proposed algorithm an optimal task scheduling algorithm.

1. Introduction

Since the time that cloud computing was introduced, almost all the data and tasks were handled by the cloud and, to assign each task to a particular virtual machine, a processor called the task scheduler was introduced, which, up to today, new algorithms have been proposed in order to enhance the performance of the task scheduler and also enhance the parameters involved. The main reason why we focus on the task scheduling component compared to other components is that task scheduling is one of the most essential and crucial parts, as it is responsible for assigning each of the virtual machines present in the cloud with the task requested by the user and returning the completed task back to the user. As mentioned before, various parameters and hyperparameters have been considered to establish a more enhanced task scheduler, some of which include the total cost of task execution, completion time, tolerance of faults, resource usage, and energy consumption. The procedure of scheduling tasks appears to be an NP complete problem, as the difficulty of the problem depends on the expanse of time taken to determine the solution, which may differ in different situations [1,2,3].
The main objective of the algorithm is to reduce the cost and time taken to execute the task while, at the same time, improving the performance of the task schedulers. The algorithm used for scheduling the task plays an important role in cloud computing, as it proficiently searches and makes decisions actively to determine the exact and apt virtual machine to execute each task. In order to arrive at an ideal and fine algorithm, various hyperparameters and parameters are used, such as makespan, time of response, utilization of the system, and network-based measures, which include the expenditure of the network communication, round trip, traffic volume, and many more. The method of finding the tasks available in the cloud can be characterized as a hybrid, heuristic, and meta-heuristic scheduling of task methods. There are algorithms where hybrid scheduling is used, and then there are some methods where a combination of heuristic and meta-heuristics approaches is used. In the algorithm where heuristic task scheduling is used, assigning the task is easy and often the quickest way to return the required outcome, but the ideal result is not guaranteed. When only taking into mind the time required for computation, algorithms using the meta-heuristic approach are better, as they attain a huge probing space to find the required result, thus decreasing the time required for computation. The area of concern now is the task scheduler, which is an important part, as it impacts the operation of cloud computing. Users such as creators and programmers, use common possessions due to the help of the cloud and, depending on the user’s needs and specifications, several different dedicated algorithms are recorded to aid the users in attaining the essential outcome [4,5].
Thus, a robust and efficient task scheduler algorithm is required that offers an ideal solution set. There is no problem seen while executing it on a simple system, but when executing it on a complex system, it should be able to handle enormous data and be able to complete the process. Nevertheless, during this process, there may be some problems encountered. Therefore, an efficient algorithm will be required to handle and process the data to obtain the required results.
Moreover, enumeration is needed to build a scheduler that is optimal, and, for this to happen, a process of building all task scheduling systems is required, that is, it mainly focuses on building a task structure from the already available set of tasks to get the related result set and cross check with each other to get an ideal set of solutions that can be used in the cloud environment for assigning each task to the correct virtual machine to improve the performance by minimizing the time consumed for handling a huge number of tasks. An algorithm tries to solve the scheduling-of-tasks problem by utilizing either a meta-heuristic or heuristic-based algorithm. The optimal result set that is considered ideal can be obtained by making use of a heuristic-based approach, as these approaches utilize some already defined rules, and results obtained by these programs vary based on the size of the problem and the underlying rules.
In addition to this, the meta-heuristic approach is used to solve extremely complex problems, as this approach generally makes use of a set of solutions to obtain the required solution space, which is completely different in terms of operation when compared to the functions used in the mathematical approach or the heuristic method, where only one candidate result space is used. Therefore, meta-heuristic-based algorithms perform better in comparison to the mathematical function and heuristic-based methods. Some of the famous algorithms used to solve the scheduling of task problem are the Quantum-Based Avian Navigation Optimizer Algorithm (QANA), the Modified-Transfer-Function-Based Binary Particle Swarm Optimization (MTF-BPSO), the Ant Colony Optimization (ACO), the MIN-MIN Optimization Algorithm (MM), and the First-Come-First-Serve Algorithm (FCFS).
In this paper, a new algorithm called the Hybrid Weighted Ant Colony Optimization Algorithm (HWACOA) is introduced that optimizes the process of task scheduling. The main objective of this program or method is to reduce the time taken for execution and the cost of operation. To summarize this paper:
o
In order to understand the problems faced by other task scheduling techniques and algorithms and to analyze how this affects the performance of the cloud, many various works have been reviewed.
o
Based on the understanding from the various works, this paper introduces a new innovative hybrid algorithm called the Hybrid Weighted Ant Colony Optimization Algorithm (HWACOA), which enhances the shortcomings and problems encountered by various different algorithms and, thus, improves the performance of scheduling the task.
o
Last but not the least, comparison between the results obtained by the introduced algorithm with many other algorithms has been conducted, and this was done by comparing extensive experimental simulation results obtained by various different algorithms using paired t-tests.

3. Problem with Solution Framework

The service providers for the cloud have designed the cloud environment in such a way that the cloud comprises virtual machines (VMs) and physical machines (PMs), and it also provides an interface for the public to access it. The users who want to execute some tasks submit their respective task to the cloud through the interface, after which the resource manager will aggregate and effectively manage all the tasks as it monitors and keeps all the updates of the available records of the resources present in the cloud, which include CPU, memory, and storage. Following this, the scheduler effectively does task scheduling such that the resultant fitness function is minimized, and the tasks are given to the respective virtual machine in accordance with the performance of the process involved in scheduling. Scheduling of a task is done only after the scheduler receives all the data from both the request manager and resource monitor. Thus, upon receiving all the required data, a decision concerning the assignment of the tasks to the apt virtual machine is done. The process of assigning the task to the respective virtual machine can be improved after knowing the exact location details of the available virtual machines along with the list of the requested tasks. This data will help reduce a lot of parameters, such as migration cost, total time, load utilization, and cost.
Let T = {t1, t2, …, tm} represent a list of tasks that must be executed and V = {v1, v2, …, vn} represents a list of VMs (virtual machines), where each virtual machine is capable of processing and executing the task assigned to it. Thus, the process of task scheduling can be represented by F: T→VM. The main objective is to build a new task scheduler algorithm which has a low makespan while, at the same time, having a low cost.
O b j e c t i v e   f u n c t i o n = i = 1 m T i α   · M a k e s p a n +   β   ·   C o s t  
From the above equation, variable Ti represents a task selected from the list T and the value of i ranges from [1–m] (m represents the number of available tasks); α represents the makespan constant whose value is between the range [0–1]; and β represents a cost constant whose value is also between the range [0–1].
Makespan Formula:
M i j = S i j + W t i j + E C t i j
M a k e s p a n = M i n j = 1 n i = 1 m M i j
where, Sij is the time taken for submission of the ith task to the jth virtual machine, Wtij denotes the time of waiting for the ith task on the virtual machine, and ECtij represents the time of execution for the ith task by the jth virtual machine.
Cost calculation Formula:
c o s t = d a t a c e n t r e h o s t . c o s t P e r S t o r a g e v m . s i z e + d a t a c e n t r e h o s t . c o s t P e r R a m v m . r a m + d a t a c e n t r e h o s t . c o s t P e r B w v m . b w + d a t a c e n t r e h o s t . c o s t P e r M i p s v m . m i p s v m . n u m b e r O f P e s .
In the above-mentioned formula, the variable d a t a c e n t r e h o s t represents the capacity or the number of hosts present, c o s t P e r S t o r a g e is the time taken to access the right storage unit, v m . s i z e refers to the size of the virtual machine, c o s t P e r R a m is the cost of using the ram, v m . r a m represents the ram capacity present in the virtual machine, c o s t P e r B w represents the bandwidth size, v m . b w is the bandwidth of the virtual machine, c o s t P e r M i p s is the virtual machine’s capacity to process a million instructions per second, and v m . n u m b e r O f P e s represents the number of prefetch evaluation systems available in each of the virtual machines.
As discussed before, in a cloud computing environment, when any users send a particular task to the cloud and expects some result, certain processes and methods are followed so that the job performed by the cloud is done in an effective and efficient manner. There is a division of work seen in the cloud environment to handle different task so that the performance and through output are maximized. Whenever the users (end users) send their respective task to the cloud environment, the task manager is responsible for monitoring and storing all the tasks received from different users, after which the task scheduler algorithm is applied to schedule the task. Here, the introduced new algorithm named HWACO algorithm is applied and certain parameters, such as cost and makespan, are considered for the evaluation of the performance of the algorithm. After this, the process/control is passed to the cloud service provider, which contains different multiple processes, such as resource management, resource monitoring, and other processes. After this, each task will be sent to the data center, which contains a set of virtual machines (VMs), where each of the tasks will be executed by the correct virtual machine that was assigned by the task scheduler. Thus, the cloud will be able to receive different tasks from the users and assign correct VMs to each of the tasks using the introduced task scheduler. The architectural diagram of the cloud using the HWACO task scheduling algorithm is depicted in the Figure 1 below.
Figure 1. Architectural diagram of the task scheduler in cloud environment.

3.1. Task Scheduling Process Using HWACO Algorithm

The ACO (Ant Colony Optimization) algorithm, as the term suggests, tries to copy the behavior and movements of the ant toward a particular food source and back to its colony [19]. The algorithm comprises two major steps: the Construct_ANT solution and the Update_Pheromone function. In the Construct_ANT solution function, a set of artificial ants are constructed in order to find results from the finite set of available answer components, and, in the Update_Pheromone function, the pheromone level related to the solution set is maximized in order to arrive at an ideal solution-set from the available possible solution sets, that is, by maximizing the pheromone level related to a particular solution set, more importance is given to that particular solution set compared to other solution sets; therefore, selection of the ideal solution set occurs. Efficient discovery of the shortest path is important. At the same time, optimization is also one of the most vital factors in order to obtain and enhance the functioning of the task scheduler in the cloud environment. The Ant Colony Optimization algorithm and some other similar type of algorithms provide a near ideal solution with some drawbacks in certain cases; thus, an enhanced version of the ACO has been introduced, that is, the HWACO algorithm. The HWACO is equipped with improved construct ANT solution and Update Pheromone methods that make use of the weight concept, which provides the required consistency among the capabilities, such as exploitation and exploration. In this work, the HWACO algorithm was defined and executed in the task scheduling process using the following notation:
  • S is the exploration space with a finite set of distinct choice variables generated for each variable Ti;
  • Ω represents the set of restraints;
  • f:S→R is the main function which must be reduced.
The definition of the exploration space S is as follows. A distinct variable set X i , where the value of i ranges from [1, m] (where m is the number of available tasks) with the variable’s value present in the set, is defined as v i j D i , and D i = v 1 j , ,   v | D i | j , where i represents a particular selected variable in the set S , and D i stands for the search variable for finding the correct virtual machine; therefore, X i represents the set of the available virtual machine for executing the required tasks. Each element of the set S is defined properly, that is, every element variable X i has a value v i j allocated from its field S . A new set S Ω is defined, which contains the achievable results that satisfy all the constraints specified on the set Ω and set X i . The constraints present in the set   Ω are usually pre-fed beforehand to include various rules and constraints, such as checking the value of the elements from the different sets such as S ,   T   and   V , in order to make sure that these values are inside the acceptable range of values, and no negative numbers are present. The rule for forming various sets such as X i ,   D i , and others will eventually lead to finding the required objective solution set by setting up the fitness function and the selection of values based on the fitness function, and certain rules, such the maximum number of the available virtual machines to be considered, that is, the j value, the acceptable range of values for the constants such as α , β , and many more depend on the user’s needs.
A result set S* that belongs to the set SΩ is known as a global optimal solution set if it satisfies the below equation:
f S f S Ω   ,   S Ω S
In order to obtain the best optimal path with the shortest distance, the following processes are followed:
o
Construct_ANT solution.
o
Update_Pheromone.

3.2. Construct ANT Solution Process

A group of m artificial ants try to create the solution set by combining elements from the available finite set of existing solution mechanisms S , and the new set C , which is a subset of X i union Ω (here set C and S Ω are the same), which is represented by the set C c i j , where j = 1, …, | D i | and i = 1, …, m. For an appropriate optimization process, the function begins with a null partial result, which is represented by S p = . After this, all of the present partial solution S p is then protracted at every construction step-by-step by adding a feasible solution component from the set of feasible neighbours N ( S p )     C , that is, in order to get to the solution component, the set of feasible neighbours N ( S p ) is used, and this set is calculated from the previously calculated solution component S p , where the values of the ith task and the selected jth virtual machine are left, and, for assigning the (i+1)th task, the range of virtual machines to look for will be designed in such a way that it gives less priority to the already assigned virtual machines and gives more priority to the available virtual machines. The process of developing solutions can be thought of as a path on the construction graph G C V , E . The solution construction mechanism, which defines the set N ( S p )   with respect to a partial solution S p , implicitly defines the allowed paths in the G C . At each construction step, a remedy element from N ( S p )   is chosen probabilistically. Thus, for assigning one task to the correct virtual machine, multiple comparison will be done based on the range of virtual machines, when the value of the element c i j obtained (ith task assigned to the jth virtual machine) matches with other elements from the S p set, that is, c i j = c i l and so on. Then, the Equation (5a) is used to calculate the best possible pair, but when the set has different values, Equation (5b) is used to calculate the best possible pair by using the weight concept. The main reason to use weights in the second case is because, when the value of the elements in set is different, this formula helps to determine the best pair of the task-to-virtual machine match, but, most of the time, it just acts like a cross-checker to be sure that the selection is optimal, and the selection eventually results in the required solution set that is optimal. The calculation of weights is done based on the first and second best values of the elements from the S p set, where c i j refers to the first highest value, and c i l refers to the second highest value. The precise rules for the stochastic selection of solution components differ based on the cases mentioned below.
Case 1when c i j = c i l
p c i j k S p = T i j α η i j β C i l N s p T i l α η i l β c i j c i l N S ρ
Case 2when c i j   c i l
p c i j k S p = max T i j α η i j β C i l N s p T i l α η i l β w 1   , T i j α η i j β C i l N s p T i l α η i l β 1 w 1 c i j c i l N S ρ  
The weights are calculated using the below-mentioned formula:
w 1 = p ( j | l )   η i j p ( j | l )   η i j + p ( l | j )   η i l
where the current ith task can be assigned to a virtual machine from the range [j, k]. Here, the value of j initially starts from one, and, as the number of iterations increases, the value of j also increases, and the value of k is equivalent to the value of D i for that particular task; thus, p c i j k S p represents the probabilistic value of selecting each virtual machine from the above range for processing the ith task. Variables, η i j and τ i j are correspondingly the heuristic value and pheromone value, respectively, connected with the component c i j . The variable η i j is a constant value for any value of i, j, which is fixed beforehand, and the variable τ i j is calculated using the Update_Pheromone update function. The constant variables such as α and β are the makespan and cost constant variables, respectively, whose values are similar to the makespan and cost constants values used in the objective function in Equation (1).

3.3. Update_Pheromone Process

The pheromones update step aims to maximizes the pheromone values related to the result set while decreasing those related with corrupt result sets. Typically, this is accomplished by (i) reducing all pheromone values via evaporation of the pheromones and (ii) maximizing the levels of pheromones related with a selected set of good results S*:
τ i j = 1 p τ i j + Δ T i j
Δ T i j = p s ϵ S u p d   |   c i j   ϵ   S   F s
where S* represents the solution set that is used in the updating process, and p is based on the previously calculated value of the term p c i j k S p . This constraint is called the evaporation rate. F is defined as the function of the the equation that follows:
f s < f s = > F s F s ,   s s S
Here, f s represents the fitness function of an element (s) from the set such that, when compared with other elements (suppose s’), the fitness function of f s is low compared to other elements present in the set, and, at the same time, the F s value should be high. Here, F s stands for the performance in of that element in comparison to other elements in the set. The performance test will be calculated based on the fitness function f s and f s , and the average mean to get the performance in terms of percentage value and this value should be higher. The reason to select the average mean method is because this formula will be used multiple times, and conducting a dedicated parameterized test, such as a t-test, will increase the makespan of the algorithm. Thus, a simple test is selected to measure the performance metrics of the selected task-to-virtual machine pairs and obtain an optimal solution set.
Pheromone evaporation is a beneficial way of overlooking that encourages the search for new ranges in the exploration space. Various ACO programs and methods, such as the Ant Colony System (ACS), Ant Colony Optimization (ACO), and many others try to increase the performance of the cloud environment by using this method, but they have certain drawbacks, due to which the objective of improving the performance is not seen in such algorithms, yet the performance of the introduced HWACO algorithm is theoretically high, due to the concept of weight and other functionalities used in the algorithm. Current knowledge, including substantive findings of the above-mentioned update rule, are obtained through various conditions of the Supd, which is often considered a subset of Siter ∪ {Sbs}, where Siter is the solution set constructed in the current iteration, and Sbs is the best-so-far solution, that is, the best answer discovered since the first algorithm iteration is another mechanism that boosts the performance of the introduced algorithm.
S *   S i t e r   U   S b s  

3.4. Algorithm

From the below HWACO Algorithm 1, it is clear that the introduced algorithm makes use of weight, and this is reflected in Equations (5a) and (5b) based on the conditions, whereas the normal ACO algorithm will use just one equation that is similar to Equation (5a). Thus, the search space will be more compared to the HWACO, and the updating process is more static in nature, as it assumes the p evaporation rate to be constant, thus making the process more static in nature. In contrast, in the HWACO, the evaporation rate is determined based on the past values, thus making it more dynamic in nature. Therefore, all the formulas and explanation point to the fact that the introduced algorithm is better compared to the ACO algorithm.
Algorithm 1: HWACO Scheduler.
Input:
Inputs considered:
Number of tasks— T 1 ,   T 2 ,   ,   T m (m refers to number of tasks)
Number of virtual machines— v 1 ,   v 2 ,   ,   v n (n refers to number of virtual machine)
Maximum iteration— M a x t
Number of host machines— k
Output:
Mapping (Ti -> Vj) for all the available tasks for the set of virtual machines, which may be present in the same host machine or distributed over multiple virtual machines.
Proposed HWACO Scheduler:
BEGIN
Initialize the population and number of generations of ants and other parameters
WHILE t < Maxt
Compute the path of the ant by Equation (5);
Perform the Update_Pheromone update process by using Equation (6);
IF (The fitness function is evaluated using the Equation (7))
Evaluate the best solution of that iteration using Equations (2) and (3);
Compute the objective function using Equation (1);
IF (objective_function_value < S* check the condition in Equation (4))
update S* using Equation (8)
END IF
END IF
Increment the value of t by 1;
END WHILE
Return the S*, which contains the best solution
END

4. Results and Discussion

All the finds and experiments were conducted in a simulated environment with initial setup details mentioned in Table 2 and Table 3 using the Java (JDK 1.6)-based Cloudsim tool. All the experiments were conducted by varying the task inputs from 100–400 units [20,24]. The performance of the introduced Hybrid Weighted Ant Colony Optimized algorithm was compared with various other algorithms, such as the MTF-BPSO [14],QANA [25], ACO [28], MIN-MIN [29], FCFS [30], with respect to their time taken for execution (makespan) and cost parameters. The main reason to select the ACO algorithm is due to the fact that the introduced algorithm is based on ACO, and, thus, the first algorithm used for comparison was the ACO algorithm, while the other algorithm used for comparison could have been any recently proposed or a traditional one, but out of all the algorithms, FCFS and MIN-MIN were selected due to various reasons. Both of them are the most basic and traditional algorithms which were initially utilized in almost all cloud computing and the fact that most of the algorithms that we know of today are related to these algorithms in some or the other way. These were mainly the task scheduling algorithms which consider makespan and cost as their parameters for evaluation of their performance. Thus, these two algorithms acted as a stepping stone to decide whether the introduced algorithm would be able to optimize the result and achieve the objective goals. The selection for QANA is based on the fact that QANA is one of the state-of-the-art algorithms introduced for solving global optimization problems, that is, QANA is better in terms of performances for returning ideal optimization solutions when compared to other modern algorithms like the MWOA (Modified Whale Optimization Algorithm), EEGWO (Exploration-Enhanced Grey Wolf Optimizer), and many more. Thus, the introduced algorithm was compared with ideal algorithms, such as FCFS and MIN-MIN. At the same time, it was also compared with modern algorithms such as QANA and MTF-BPSO in terms of performance. Finally, MTF-BPSO was selected because MTF-BPSO is one of the recently proposed algorithms which aims to improve the performance task scheduling and quality of service in cloud computing and for the fact that MTF-BPSO outperforms many task scheduling algorithms such as the Sailfish Optimization Algorithm, Opposition-Based Sunflower Optimization Algorithm, and many other algorithms.
There were numerous parameters utilized for the input, which are as follows:
Table 2. Simulation setup.
Table 2. Simulation setup.
S/NoEntity TypeParameterValue
1UserNumber of users50
2TaskSize400
Number of tasks100–400
3HostHost memory (RAM)2048BM
Host bandwidth10,000
Host storage100,000
4Virtual Machine (VM)Type of policyTime-Shared
Number of VMs40
VM RAM512BM
VMMXen
OSWindows
Number of CPUs1 on each
5DatacentreNumber of datasets4
Number of hosts2
The values used for initializing the parameters and hyperparameters in all the six algorithms are present in Table 3 and Table 4. Parameters are generally values which are calculated during the process of finding the solution set, that is, they are variables that help determine the final optimal solution set. On the other hand, hyperparameters are fixed values which help determine the parametric values during each iteration.
Table 3. Parameter initialization.
Table 3. Parameter initialization.
S/No.Scheduling AlgorithmsParametersValues
1HWACONumber of ants in colony1000
Makespan constant (α)0.3
Cost constant (β)0.6
Heuristic constant (n)0.498
2ACONumber of ants in colony1000
Evaporation factor (p)0.398
Pheromone tracking weight (α)0.3
Heuristic information weight (β)0.6
Pheromone updating constant (q)10
3MIN-MINMakespan constant (α)0.3
Cost constant (β)0.6
Population size1000
4FCFSMakespan constant (α)0.3
Cost constant (β)0.6
Population size1000
5QANAThe number of flocks (k) 10
Memory size (K’) 8
Short-term memory with memory size (Ҡ”)48
6MTF-BPSODecision variable £1 or 0 depending on task size
Social learning factor c10.5
Personal learning factor c22.4
Inertia factor0.7

4.1. Makespan Valuation

In this unit, the performance of the presented algorithm in terms of the time taken for the execution of the task was assessed. To evaluate performance, 100 to 400 tasks were considered, and these tasks were executed with onset rates ranging between 10 and 40. Here the onset rate represents the time duration (in microseconds) after which each task is given to the algorithm, that is, the algorithm will become overwhelmed with handling a large number of tasks when assigned immediately. Thus, to prevent this from happening and maintain a low bandwidth in the communication bus, the onset rate was used with different values. The results of the proposed method were compared among those of the HWACO, QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS. Figure 1 and Figure 2 show the results of the simulation.
Figure 2. Execution time for the tasks with onset rate of 10.

4.1.1. Makespan, Onset Rate = 10

In this unit, we measured the time taken to execute the tasks with 10 as the onset rate. The time taken for time, in microseconds, for various algorithms such as HWACO, QANA, MTF-BPSO, ACO, MIN-MIN and FCFS was 13.78, 15.01, 16.21, 18.8, 15.73, and 32.32, respectively. When the number of tasks was changed to 200, the time taken for the execution for the HWACO, QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS was 25.9, 27.23, 31.09, 34.61, 29.05, and 67, respectively. When the number of tasks was changed to 300, the time taken for the execution for HWACO, QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS was 36.56, 38.31, 46.31, 50.31, 40.02, and 75.5, respectively. Finally, when the number of tasks was changed to 400, the time taken for the execution for HWACO, QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS was 50.4, 52.3, 58.42, 66, 53.1, and 124, respectively. Figure 2 shows the execution time for the tasks by different algorithms when the onset rate = 10.

4.1.2. Makespan, Onset Rate = 40

In this unit, we considered the time taken to execute the tasks with 40 as the onset rate. The time taken for execution for HWACO, QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS was 16.98, 17.01, 19.67, 22.68, 18.786, and 38.784, respectively. When the number of tasks was increased to 200, the time taken for the execution for HWACO, QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS was 30.76, 32.36, 38.12, 41.53, 34.86, and 80.4, respectively. When the number of tasks was changed to 300, the time taken for the execution for HWACO, QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS was 45.13, 46.12, 55.41, 60.37, 48.02, and 90.6, respectively. Finally, when the number of tasks was changed to 400, the time taken for the execution for HWACO, QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS was 60.8, 62.3, 71.01, 79.2, 63.98, and 148.8, respectively. Figure 3 shows the execution time for the tasks by different algorithms when the onset rate = 40.
Figure 3. Execution time for the tasks with onset rate of 40.
Figure 4 shows the convergence plot when the makespan of 200 tasks was considered. The results of the proposed method were compared among the HWACO, QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS.
Figure 4. The convergence plot for makespan of 200 tasks.
Figure 5 shows the convergence plot when the makespan of 400 tasks was considered. The results of the proposed method were compared among the HWACO, QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS.
Figure 5. The convergence plot for makespan of 400 tasks.

4.2. Cost Assessment

The cost assessment was evaluated in this section. In this virtual simulation process, 200 and 400 tasks with varying deadlines were considered. The iteration was between 10 to 50. Figure 6 and Figure 7 show the results that were obtained. Figure 6 considers 200 task cost values, while Figure 7 considers 400 task cost values. Cost assessment of the proposed algorithm HWACO was compared with other algorithms, including QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS. From the experimental results obtained, they suggest that the introduced algorithm HWACO achieved a low cost for both cases when number of tasks were 200 and 400.
Figure 6. The task cost values of 200.
Figure 7. The task cost values of 400.

4.3. Statistical Analysis Using t-Test

Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12 and Table 13 contain the outcomes of a paired t-test that was carried out to assess whether there are substantial changes between the makespan and cost obtained from the proposed algorithm—HWACO—and other algorithms, namely, QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS, using comparable stopping standards for all task instances. The satisfactory p-value must be less than alpha α < 0.5. As shown in the tables below, almost all of the datasets had p-values that were less than this alpha value, which shows a substantial enhancement in the proposed HWACO algorithm compared to the other algorithms used for comparison. Thus, it can be inferred that the HWACO outperforms other task scheduling algorithm in terms of makespan and cost.
Table 4. t-Test for makespan between HWACO and QANA.
Table 4. t-Test for makespan between HWACO and QANA.
MakespanTask SizeHWACOQANAImprovementt-Test Valuep-Value
Onset rate 1010013.7815.018.1945369753.45347.430 × 10−3
20025.927.234.8843187662.45334.405 × 10−2
30036.5638.314.5679979123.30936.351 × 10−2
40050.452.33.6328871892.45349.984 × 10−2
Onset rate 4010016.9817.010.1763668431.90931.308 × 10−2
20030.7632.264.6497210172.45324.304 × 10−4
30045.1346.122.1465741542.09846.094 × 10−2
40060.862.32.4077046552.32215.534 × 10−3
Table 5. t-Test for makespan between HWACO and MFT-BPSO.
Table 6. t-Test for makespan between HWACO and ACO.
Table 7. t-Test for makespan between HWACO and MIN-MIN.
Table 8. t-Test for makespan between HWACO and FCFS.
Table 9. t-Test for cost between HWACO and QANA.
Table 10. t-Test for cost between HWACO and MTF-BPSO.
Table 11. t-Test for cost between HWACO and ACO.
Table 12. t-Test for cost between HWACO and MIN-MIN.
Table 13. t-Test for cost between HWACO and FCFS.

5. Conclusions

Given that the scheduling of the task process is one of the most vital issues encountered by the cloud environment, this paper proposed a new algorithm called the HWACO algorithm that improves the basic functions of the ACO algorithm in order to efficiently search through the solution space and return the best optimal solution. This was achieved by using two main functions, which are the Construct_ANT solution function and Update_Pheromone function, respectively, that try to find the best and optimal solution, as was mentioned in Section 3. As mentioned earlier, the selection of the best solution is based on the values computed using the objective formula which combines both the makespan and cost parameters. Thus, the proposed algorithm outperformed other algorithms in terms of cost and makespan in such a way that the makespan and cost of the entire process were minimized, and, at the same time, the performance of the cloud computing environment was increased. To support this claim, an experimental simulation was conducted along with the statistical t-test, which suggest the fact that the HWACO was able to attain an average of 3.83%, 16.54%, 25.34%, 8.66% and 57.11% of improvement when compared to the QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS in terms of makespan, respectively, and 12.15%, 18.88%, 23.63%, 27.05% and 32.94% of improvement when compared to QANA, MTF-BPSO, ACO, MIN-MIN, and FCFS, respectively, in terms of cost. Therefore, it can be concluded that the introduced algorithm optimizes the performance of the task scheduler in terms of makespan and cost.

Author Contributions

Conceptualization, C.C. and P.K.; methodology, C.C.; software, C.C.; validation, C.C. and P.K.; formal analysis, V.K.P.; investigation, B.A.; resources, K.R.; data curation, K.R.; writing—original draft preparation, C.C.; writing—review and editing, P.K.; visualization, B.A.; supervision, K.R.; project administration, K.R.; funding acquisition, K.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not Applicable.

Data Availability Statement

Dataset with 100 tasks: GoCJ_Dataset_100.txt—Mendeley Data, dataset with 200 tasks: GoCJ_Dataset_200.txt—Mendeley Data, dataset with 300 tasks: GoCJ_Dataset_300.txt—Mendeley Data, dataset with 400 tasks: GoCJ_Dataset_400.txt—Mendeley Data. The data files are available in https://data.mendeley.com/datasets/b7bp6xhrcd/1 (accessed on 23 August 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
HWACOHybrid Weighted Ant Colony Optimization Algorithm
ACOAnt Colony Optimization Algorithm
QANAQuantum-based Avian Navigation Optimizer Algorithm
FCFSFirst-Come-First-Serve Algorithm
MMMIN-MIN Algorithm
MTF-BPSOModified-Transfer-Function-Based Binary Particle Swarm Optimization Algorithm
NP-problemNondeterministic Polynomial problem

References

  1. Zuo, X.; Zhang, G.; Tan, W. Self-Adaptive Learning PSO-Based Deadline Constrained Task Scheduling for Hybrid IaaS Cloud. IEEE Trans. Autom. Sci. Eng. 2013, 11, 564–573. [Google Scholar] [CrossRef]
  2. Krishnadoss, P.; Pradeep, N.; Ali, J.; Nanjappan, M.; Krishnamoorthy, P.; Kedalu Poornachary, V. CCSA: Hybrid cuckoo crow search algorithm for task scheduling in cloud computing. Int. J. Intell. Eng. Syst. 2021, 14, 241–250. [Google Scholar] [CrossRef]
  3. Pradeep, K.; Jacob, T.P. CGSA scheduler: A multi-objective-based hybrid approach for task scheduling in cloud environment. Inf. Secur. J. 2018, 27, 77–91. [Google Scholar] [CrossRef]
  4. Pradeep, K.; Ali, L.J.; Gobalakrishnan, N.; Raman, C.J.; Manikandan, N. CWOA: Hybrid Approach for Task Scheduling in Cloud Environment. Comput. J. 2021, 65, 1860–1873. [Google Scholar] [CrossRef]
  5. Krishnadoss, P.; Chandrashekar, C.; Poornachary, V.K. RCOA Scheduler: Rider Cuckoo Optimization Algorithm for Task Scheduling in Cloud Computing. Int. J. Intell. Eng. Syst. 2022, 15, 1–10. [Google Scholar]
  6. Nabi, S.; Aleem, M.; Ahmed, M.; Islam, M.A.; Iqbal, M.A. RADL: A resource and deadline-aware dynamic load-balancer for cloud tasks. J. Supercomput. 2022, 78, 14231–14265. [Google Scholar] [CrossRef]
  7. Konjaang, J.K.; Murphy, J.; Murphy, L. Energy-efficient virtual-machine mapping algorithm (EViMA) for workflow tasks with deadlines in a cloud environment. J. Netw. Comput. Appl. 2022, 203, 103400. [Google Scholar] [CrossRef]
  8. Kumar, M.; Suman, S. Scheduling in IaaS Cloud Computing Environment using Sailfish Optimization Algorithm. Trends Sci. 2022, 19, 4204. [Google Scholar] [CrossRef]
  9. Gupta, S.; Singh, R.S.; Vasant, U.D.; Saxena, V. User defined weight based budget and deadline constrained workflow scheduling in cloud. Concurr. Comput. Pract. Exp. 2021, 33, e6454. [Google Scholar] [CrossRef]
  10. Bacanin, N.; Zivkovic, M.; Bezdan, T.; Venkatachalam, K.; Abouhawwash, M. Modified firefly algorithm for workflow scheduling in cloud-edge environment. Neural Comput. Appl. 2022, 34, 9043–9068. [Google Scholar] [CrossRef]
  11. Chandrashekar, C.; Krishnadoss, P. Opposition based sunflower optimization algorithm using cloud computing environments. Mater. Today Proc. 2022, 62, 4896–4902. [Google Scholar] [CrossRef]
  12. Gokuldhev, M.; Singaravel, G. Local Pollination-Based Moth Search Algorithm for Task-Scheduling Heterogeneous Cloud Environment. Comput. J. 2022, 65, 382–395. [Google Scholar] [CrossRef]
  13. Belgacem, A.; Beghdad-Bey, K.; Nacer, H.; Bouznad, S. Efficient dynamic resource allocation method for cloud computing environment. Clust. Comput. 2020, 23, 2871–2889. [Google Scholar] [CrossRef]
  14. Kumar, M.; Sharma, S.C.; Goel, S.; Mishra, S.K.; Husain, A. Autonomic cloud resource provisioning and scheduling using meta-heuristic algorithm. Neural Comput. Appl. 2020, 32, 18285–18303. [Google Scholar] [CrossRef]
  15. Nanjappan, M.; Natesan, G.; Krishnadoss, P. An Adaptive Neuro-Fuzzy Inference System and Black Widow Optimization Approach for Optimal Resource Utilization and Task Scheduling in a Cloud Environment. Wirel. Pers. Commun. 2021, 121, 1891–1916. [Google Scholar] [CrossRef]
  16. Nadimi-Shahraki, M.H.; Zamani, H. DMDE: Diversity-maintained multi-trial vector differential evolution algorithm for non-decomposition large-scale global optimization. Expert Syst. Appl. 2022, 198, 116895. [Google Scholar] [CrossRef]
  17. Nadimi-Shahraki, M.H.; Zamani, H.; Mirjalili, S. Enhanced whale optimization algorithm for medical feature selection: A COVID-19 case study. Comput. Biol. Med. 2022, 148, 105858. [Google Scholar] [CrossRef]
  18. Zamani, H.; Nadimi-Shahraki, M.H.; Gandomi, A.H. Starling murmuration optimizer: A novel bio-inspired algorithm for global and engineering optimization. Comput. Methods Appl. Mech. Eng. 2022, 392, 114616. [Google Scholar] [CrossRef]
  19. Zuo, L.; Shu, L.; Dong, S.; Zhu, C.; Hara, T. A multi-objective optimization scheduling method based on the ant colony algorithm in cloud computing. IEEE Access 2015, 3, 2687–2699. [Google Scholar] [CrossRef]
  20. Batista, B.G.; Estrella, J.C.; Ferreira, C.H.G.; Filho, D.M.L.; Nakamura, L.H.V.; Reiff-Marganiec, S.; Santana, M.J.; Santana, R.H.C. Performance evaluation of resource management in cloud computing environments. PLoS ONE 2015, 10, e0141914. [Google Scholar] [CrossRef]
  21. Abdulhamid, S.I.M.; Abd Latiff, M.S.; Abdul-Salaam, G.; Hussain Madni, S.H. Secure scientific applications scheduling technique for cloud computing environment using global league championship algorithm. PLoS ONE 2016, 11, e0158102. [Google Scholar] [CrossRef] [PubMed]
  22. Abdullahi, M.; Ngadi, M.A. Symbiotic organism search optimization based task scheduling in cloud computing environment. Future Gener. Comput. Syst. 2016, 56, 640–650. [Google Scholar] [CrossRef]
  23. Ramezani, F.; Lu, J.; Hussain, F.K. Task-Based System Load Balancing in Cloud Computing Using Particle Swarm Optimization. Int. J. Parallel Program. 2014, 42, 739–754. [Google Scholar] [CrossRef]
  24. Domanal, S.G.; Guddeti, R.M.R.; Buyya, R. A Hybrid Bio-Inspired Algorithm for Scheduling and Resource Management in Cloud Environment. IEEE Trans. Serv. Comput. 2017, 13, 3–15. [Google Scholar] [CrossRef]
  25. Zamani, H.; Nadimi-Shahraki, M.H.; Gandomi, A.H. QANA: Quantum-based avian navigation optimizer algorithm. Eng. Appl. Artif. Intell. 2021, 104, 104314. [Google Scholar] [CrossRef]
  26. Sardaraz, M.; Tahir, M. A parallel multi-objective genetic algorithm for scheduling scientific workflows in cloud computing. Int. J. Distrib. Sens. Netw. 2020, 16, 1550147720949142. [Google Scholar] [CrossRef]
  27. Konjaang, J.K.; Xu, L. Multi-objective workflow optimization strategy (MOWOS) for cloud computing. J. Cloud Comput. 2021, 10, 11. [Google Scholar] [CrossRef]
  28. Yu, Q.; Ling, C.; Bin, L. Ant colony optimization applied to web service compositions in cloud computing. Comput. Electr. Eng. 2015, 41, 18–27. [Google Scholar] [CrossRef]
  29. Murad, S.S.; Badeel, R.O.Z.I.N.; Salih, N.; Alsandi, A.; Faraj, R.; Ahmed, A.R.; Alsandi, N. Optimized Min-Min task scheduling algorithm for scientific workflows in a cloud environment. J. Theor. Appl. Inf. Technol. 2022, 100, 480–506. [Google Scholar]
  30. Ramkumar, K.; Gunasekaran, G. Preserving security using crisscross AES and FCFS scheduling in cloud computing. Int. J. Adv. Intell. Paradig. 2019, 12, 77–85. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The st atements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.