Skip to main content

Optimizing scientific workflow scheduling in cloud computing: a multi-level approach using whale optimization algorithm

Abstract

Cloud computing has evolved into an indispensable tool for facilitating scientific research due to its ability to efficiently distribute and process workloads in a virtual environment. Scientific tasks that involve complicated task dependencies and user-defined constraints related to quality of service (QoS) and time constraints require the efficient use of cloud resources. Planning these scientific workflow tasks represents an NP-complete problem, prompting researchers to explore various solutions, including conventional planners and evolutionary optimization algorithms. In this study, we present a novel, multistage algorithm specifically designed to schedule scientific workflows in cloud computing contexts. This approach addresses the challenges of efficiently mapping complex workflows onto distributed cloud resources while considering factors like resource heterogeneity, dynamic workloads, and stringent performance requirements. The algorithm uses the whale optimization algorithm (WOA) with a two-phase approach to shorten execution time, minimize financial costs, and effectively maintain load balancing.

Introduction

Cloud computing has emerged as a rapidly expanding area of distributed computing, providing scalable services over the Internet through hardware and software virtualization. This business model allows customers to access and purchase services under service-level agreements (SLAs) and pay for what they use, similar to conventional utilities [1]. The main strengths of cloud computing lie in its adaptability and flexibility, enabling individuals to access resources and services customized to their requirements remotely [2]. Cloud providers typically offer two resource provisioning plans for different user needs [3]. The first plan is the on-demand plan, where users request resources whenever needed, accommodating fluctuating and unpredictable inquiries. The second plan is the long-term preservation plan, where users reserve resources in advance, providing predictability and stability for their resource allocation. Leading cloud providers like Amazon EC2 and GoGrid offer services with both on-demand and reservation plans [4, 5].

Heterogeneous distributed computing environments consist of various computers, machines, and processors interconnected by high-speed networks [6]. These systems can collectively utilize their resource capacities to tackle complicated issues. Previously, scientific workloads were often executed on distributed grid computing, also known as e-science [7]. However, as cloud computing has evolved, it has been increasingly adopted for e-business purposes and for addressing e-science challenges. Cloud computing’s widespread availability, cost-effectiveness, and flexibility, enabled by virtualization technology, have garnered significant interest from scientific and research communities. Scientific workflows, which are mathematical parallelizable processes, are implemented in various actual engineering tasks like FFT, GJ elimination, and LU decomposition. These workflows are commonly modeled as directed acyclic graphs (DAGs), with nodes reflecting application tasks and directed edges connecting data-dependent tasks [8]. The shapes and resource demands of such parallelizable applications vary widely. As users may not be able to expand existing infrastructure, distributed systems like cloud computing offer scalability, especially with the elasticity attribute, enabling the provision of services according to users’ varying resource demands [9]. In a cloud environment, multiple virtual machines (VMs) can execute independent tasks simultaneously. Turnaround time represents one of the most critical QoS indicators, measuring the overall elapsed time between the initiation of the first task in an application and the completion of the last. A user’s experience is directly affected by this time duration, known as makespan. Therefore, minimizing makespan is a significant objective function in this research [10].

The successful execution of scientific workflows relies on efficiently utilizing available resources. To achieve this, researchers focus on developing effective strategies to assign workflow tasks to computing resources, known as workflow scheduling. Workflow scheduling involves coordinating the execution of interdependent tasks, taking into account resource priority constraints [11]. Due to the NP completeness of this problem, researchers have focused on finding near-optimal solutions. To ensure workflows are well-defined and handled for future execution, an effective workflow management system (WMS) is essential. Figure 1 illustrates that a workflow scheduler (bridge) is required in a cloud environment to organize workflow tasks and allocate them to targeted resources. This workflow scheduler is crucial for efficient resource distribution and management of scientific workflows within cloud computing environments.

Fig. 1
figure 1

Scientific workflow execution model in cloud computing

The task scheduling process presents a significant issue within cloud computing, especially for complex and diverse scientific workflows. Scientific workflows can be sensitive to large data volumes, complex processing, and multiple criteria simultaneously [12]. This complexity has motivated researchers to come up with strategies that optimize the handling of scientific workflows, with a particular focus on finding a balance between two conflicting QoS variables: cost and time [13]. QoS is a measure of user satisfaction with a given service and is often evaluated based on criteria like reliability, computational cost, and computational time. Balancing the conflicting goals of minimizing processing time and reducing costs can be challenging. Faster processing often requires more powerful and expensive resources, while cheaper resources may result in slower completion times [14]. To address this contradiction, it is necessary to shorten processing time while reducing costs, all while adhering to deadlines and budgets. This strategy seeks to maintain an ideal balance between meeting performance requirements and optimizing resource utilization.

This study aims to optimize scientific workflow scheduling performance within cloud computing environments. This is achieved by focusing on three key objectives: enhancing execution time, reducing financial costs, and maintaining effective load balancing across resources. By utilizing the whale optimization algorithm (WOA) in a multi-level approach, the proposed method seeks to minimize the makespan, ensuring that workflows are completed in the shortest possible time. Additionally, the algorithm aims to minimize the monetary costs associated with resource utilization by optimizing the task distribution among VMs. Finally, the study emphasizes maintaining load balancing, ensuring that the computational load is equally dispersed across all available resources to prevent bottlenecks and enhance overall system performance.

Related work

This section reviews significant contributions to scientific workflow scheduling in cloud-based systems, highlighting various methodologies and their effectiveness. Table 1 lays out a detailed comparison of these approaches, highlighting objectives, evaluation methods, and key findings.

Table 1 Related work on workflow scheduling in cloud computing

Shi et al. [15] developed a resource allocation and task management framework in the cloud to handle scientific workflows in an elastic manner. This method is designed to efficiently execute critical workflow tasks within specified time and budget limitations. The approach encompasses four key phases: task preprocessing, task authorization, resource assignment on an elastic basis, and task scheduling. The performance assessment involves four types of real scientific workflow tasks with varying financial limitations. Additionally, uncertainties regarding task failures, processing delays, and estimated runtimes are considered in the assessment. The results demonstrate that in most scenarios, their mechanism outperforms other alternatives.

Aziza and Krichen [16] have introduced a novel model aimed at optimizing the execution time of interconnected tasks in the cloud while simultaneously reducing the workload and fulfilling strict demands and budget constraints. Their approach utilizes a hybrid process that incorporates a genetic algorithm to effectively model and improve cloud computing workflow scheduling. The heterogeneous earliest finish time (HEFT) generates the sample population. Through extensive experiments applied to actual scientific tasks, they have demonstrated the superior performance of their developed approach compared to existing HEFT and other approaches analyzed in their research. The experiments clearly demonstrate the high efficiency of their approach, making it a viable approach to cloud workflow planning. To implement their proposed model effectively, they have developed a GA-based module, which has been seamlessly adapted to the CloudSim-based WorkflowSim framework.

Workflow optimization heavily relies on efficient task scheduling, considered a widely recognized NP-hard issue. The genetic algorithm has been used in numerous strategies for cloud task scheduling, but the strategy puts forth by Iranmanesh and Naji [17] stands out as more efficient than others. This is due to the application of altered genetic operators and the integration of a load-balancing mechanism. Notably, their method employs a heuristic solution to generate one primary population chromosome and utilizes a heuristic method to generate the remaining primary population chromosomes. To ensure a comprehensive evaluation, an adaptive fitness function is employed, taking both cost and makespan into consideration. By introducing a load-balancing routine, their algorithm maximizes resource efficiency during execution. The results are evaluated against leading algorithms in this domain to gauge the performance of their algorithm. The findings reveal that the approach exhibits substantial advantages over other approaches, achieving the most efficient and cost-effective task scheduling.

Mohammadzadeh and Masdari [18] have introduced a mixed multi-objective optimization algorithm named HGSOA-GOA, combining the grasshopper optimization algorithm (GOA) and seagull optimization algorithm (SOA). The algorithm utilizes chaotic patterns to generate stochastic values, achieving a balance between exploiting and exploring, resulting in improved convergence rates. The HGSOA-GOA algorithm addresses scheduling challenges for scientific workflows in cloud settings, incorporating multiple factors like throughput, energy, cost, and makespan. It aims to optimize these objectives simultaneously, enabling better decision-making for task assignments in a multi-cloud scenario. In this approach, Pareto Solutions are chosen using the knee-point strategy, which helps identify an optimal compromise solution. This selected solution is then used to assign scientific workflow tasks in multi-cloud environments. To evaluate the performance of HGSOA-GOA, comprehensive analyses are performed via the CloudSim and WorkflowSim simulators, and the outcomes are evaluated against those obtained from the SPEA2 model. The findings demonstrate that the HGSOA-GOA algorithm outperforms alternative methods, as evidenced by indicators including IGD (inverted generational distance), coverage rate, and others.

Choudhary et al. [19] have implemented a task clustering and partial critical path algorithm to enhance the efficiency of scientific workflow execution. This algorithm groups fine-grained tasks into jobs and assigns sub-deadlines recursively to tasks situated on the partial critical path. This approach helps optimize the workflow and improves the overall performance. Additionally, they address the energy efficiency aspect by incorporating the dynamic voltage and frequency scaling (DVFS) method. This procedure constantly adjusts the voltage and frequency of computing nodes’ processors based on workload, allowing for energy-saving opportunities during execution. To validate their proposed framework, simulations are performed on various scientific applications, including LIGO inspiral analysis, SIPHT, CyberShake, and Montage. The results from these simulations demonstrate that the implemented task clustering and DVFS techniques effectively address the mentioned issues. The analysis of the results reveals considerable savings in transmission costs and energy utilization, indicating the efficiency of the suggested approach in achieving energy savings and improved performance in scientific workflow execution.

Khaleel [20] addressed the scientific workflow scheduling problem as a multi-objective optimization challenge, aiming to find an equilibrium between scheduling reliability and energy efficiency. They suggest a two-step approach to reach this compromise. In the first step, tasks requiring moderate computational demands are placed on fog resources, while tasks with higher computational demands are assigned to cloud resources. This strategy avoids high-failure-rate resources, thereby enhancing scheduling reliability. In the second step, they employ the reliability-sensitive stepwise performance-to-power ratio (PPR) optimization method to minimize energy usage significantly. This involves measuring the machine’s utilization rates using PPR. The PPR is determined by the ratio of transactions completed at certain times to the energy consumed. Through simulations in real-world and synthesized workflow applications, the proposed approach, called dis-RMEE, achieves significantly enhanced reliability while maintaining reduced energy constraints. It outperforms its competing workflow mapping methods in terms of both reliability and energy efficiency.

Methods

This section introduces a many-objective algorithm for task scheduling. It consists of two main components: preprocessing and a scheduling approach based on the WOA. Before the scheduling process begins, the workflow and its associated tasks are preprocessed to gather essential information. This step involves analyzing the workflow’s structure, identifying task dependencies, and assessing the resource requirements of each task. The algorithm searches for optimal solutions by iteratively updating the positions of candidate solutions, known as whales, in the search space. In the context of task scheduling, the WOA algorithm is adapted to find an optimal assignment of tasks to available computing resources. The positions of the whales represent different possible task-resource assignments, and the algorithm iteratively explores and updates these assignments to improve the overall performance of the workflow. By employing the WOA-based scheduling approach, our multi-objective hybrid algorithm aims to optimize multiple objectives, such as minimizing makespan (processing time), reducing resource utilization costs, and meeting deadline constraints.

Problem statement

In the context of workflow scheduling in cloud computing, workflows are expressed as DAGs, denoted by G = (V, E), where V corresponds to the set of points that signify individual workflow tasks and E indicates the set of edges representing task dependencies. In a DAG, each task has a specific execution order, and a parent task must be completed before any of its child tasks can start execution. Each workflow task in the DAG has various parameters that are crucial for scheduling and resource allocation. Execution time denotes the time required for a task to be executed on a particular resource. Different tasks may have different execution times depending on their computational complexity. Dependencies stand for the relationships between tasks, where a child’s task depends on the completion of its parent task(s). These dependencies need to be considered during scheduling to ensure correct execution order.

Workflow tasks can vary in their computational and data intensity. Some tasks may be computation intensive, involving complex computations, while others may be data intensive, requiring significant data handling and transfer. In some cases, tasks may be both data and computation intensive, making the scheduling process more challenging. The cloud structure is formed of multiple data centers, each containing a number of physical machines. These machines are equipped with data storage resources and bandwidth capabilities. In cloud computing, resources are depicted as VMs. VMs are virtualized instances that can be deployed and managed on the physical machines within data centers. Each VM has fixed attributes such as bandwidth, processing capabilities (measured by the number of processing elements or PEs), and storage cost per unit of time. The processing capacity of VMs is calculated by Eq. 1.

$${C}_{i}=({MIPS}_{i}\times PE)$$
(1)

The capacity of n VMs is calculated by Eq. 2.

$$C=\sum_{i=1}^{n}{C}_{i}$$
(2)

The VMs’ loads are computed by Eq. 3 as the percentage of the total number of tasks performed by VMs to their capacities.

$${L}_{{vm}_{i}}=\frac{TL}{{C}_{i}}$$
(3)

Equation 4 calculates the load of all VMs in the cloud system. The time taken to finish the ith task is the sum of the duration needed to retrieve the necessary data for execution (time to obtain data for task i) and the execution time of the task itself. This time of completion is a vital measurement to consider in task scheduling and load balancing algorithms, as it directly impacts the overall makespan of the workflow and the performance of the system. Equation 5 calculates the completion time of a workflow task i.

$$L=\sum_{i=1}^{n}{L}_{{vm}_{i}}$$
(4)
$$Time\left({t}_{i}\right)={Time}_{E}(\left({t}_{i},{VM}_{k}\right)+Time({Trans}_{{t}_{i},{t}_{j}})$$
(5)

Equation 5 defines \({Trans}_{{t}_{i},{t}_{j}}\) as the duration of data transmission between tasks i and j, while \({Time}_{E}\left({t}_{i},{VM}_{k}\right)\) represents the execution time of the ith task on \({VM}_{k}\). The transmission duration is estimated using Eq. 6, while Eq. 7 provides the ability to compute the execution time.

$$Trans\left({t}_{i},{t}_{j}\right)=\frac{sizeof({t}_{i},{t}_{j})}{{\beta (VM}_{k},{VM}_{m})}$$
(6)
$${T}_{E}=\frac{{l}_{i}}{{C}_{{m}_{j}}}$$
(7)

Within the aforementioned equations, the term \(sizeof({t}_{i},{t}_{j})\) corresponds to the data size transmitted between tasks i and j, and \({\beta (VM}_{k},{VM}_{m})\) denotes the bandwidth of data centers in which \({VM}_{k}\) and \({VM}_{m}\) are situated. When both VMs are located within the same data center, the transmission cost is effectively reduced to zero. Equation 7 incorporates the variables \({l}_{i}\) and \({C}_{{m}_{j}}\), where \({l}_{i}\) represents the length of task i and \({C}_{{m}_{j}}\) denotes the processing capacity of \({VM}_{j}\), as determined using Eq. 2. The makespan corresponds to the last task’s completion time within the workflow, calculated by Eq. 8.

$$Makespan={FT}_{i=1}^{n}[{task}_{i}time]$$
(8)

Equation 8 defines the makespan as the task’s finish time (FT). The monetary cost (MC) associated with a workflow encompasses both the data transfer and execution costs between workflow tasks. This cost can be computed using Eq. 9. The MC further comprises the total transfer cost (TTC) determined using Eq. 10 and the total execution cost (TEC) evaluated with Eq. 10.

$$MC=TTC+TEC$$
(9)
$${TTC}_{wi}=\sum_{i=1}^{n}\frac{size({t}_{i},{t}_{j})}{\beta cost}$$
(10)
$${TEC}_{wi}=\sum_{i=1}^{n}{vm}_{i}^{time} {vm}_{i}^{cost}$$
(11)

Equation 10 presents the relationship between the size, representing the data size and cost, which denotes the bandwidth cost for data transfer. Additionally, load balance is expressed as the variance of the load across all nodes, as depicted in Eq. 12. Smaller values of load balance indicate more effective load management.

$$L=\sqrt{\frac{\sum_{i=1}^{m}{({L}_{{vm}_{i}}-R)}^{2}}{m}}$$
(12)

In Eq. 12, \({L}_{{vm}_{i}}\) denotes the load of \({vm}_{i}\), R represents the mean load across all VMs, and m indicates the total number of VMs in the system.

Whale optimization algorithm

The WOA is a computational intelligence algorithm that draws inspiration from the foraging behavior of humpback whales, specifically their bubble-net hunting mechanism. Humpback whales are highly intelligent creatures with brain cells called spindle cells, similar to humans [21]. They exhibit complex behaviors, including communication through their own dialect, making them an intriguing inspiration for optimization algorithms. The WOA algorithm begins by randomly generating a population of candidate solutions across the search space. Each candidate solution’s fitness is calculated based on the objective function, and the best solution is identified. The WOA operates in three main phases: encircling prey, bubble-net hunting, and prey searching.

Once the humpback whales find the best search agent (the one with the highest fitness value), the rest of the whale population starts encircling that agent. This behavior is a part of the WOA’s exploration phase. Search agents update their positions during this process as they approach prey (the best search agent). The mathematical model for the current position update during the encircling behavior in the WOA can be described as follows:

$${WX}_{curr}\left(t+1\right)={WX}_{best}\left(t\right)-A\times {D}_{dis}$$
(13)
$$A=2a\times {r}_{1}-a$$
(14)
$${D}_{dis}=\left|C\times {WX}_{best}\left(t\right)-{WX}_{curr}(t)\right|$$
(15)
$$C=2\times {r}_{2}$$
(16)

\({D}_{dis}\) represents the distance between the current solution \({WX}_{curr}\left(t\right)\) and the best solution \({WX}_{best}\left(t\right)\) at iteration t. The values of \({r}_{1}\) and \({r}_{2}\) are random numbers within the range [0, 1]. The variable \(a\) is the shrinking decreased linearity value, which starts at two and gradually decreases to 0 as the iterations progress. If the WOA algorithm chooses the first method, i.e., the shrinking encircling mechanism, for updating the solution, it follows the same process as described earlier using the decreasing value of Eq. 18. This method involves calculating the encircling vector and updating the current solution’s position based on the shrinking decreased linearity value and the current iteration number. On the other hand, if the WOA algorithm opts for the second method, i.e., the spiral update based on random selection probability, the population of whales will navigate around the current optimal solution denoted as \({WX}_{best}\). The process for this method is defined as follows:

$${WX}_{curr}\left(t+1\right)={Dt{^{\prime}}}_{dis}\times {e}^{bk}\times \text{cos}\left(2\pi k\right)+{WX}_{best}(t)$$
(17)
$${D{^{\prime}}}_{dis}=\left|{WX}_{best}\left(t\right)-{WX}_{curr}\left(t\right)\right|$$
(18)

k varies randomly between [− 1, 1], and b controls the magnitude of the spiral movement. The algorithm switches between the shrinking and spiral paths for solution updates based on a probability distribution. The equation to determine whether to use the shrinking encircling mechanism or the spiral update method is as follows:

$${WX}_{curr}\left(t+1\right)=\left\{\begin{array}{c}{WX}_{best}\left(t\right)-A\times {D}_{dis} if {r}_{3}\ge 0.5\\ {D{^{\prime}}}_{dis}\times {e}^{bk}\times \text{cos}\left(2\pi k\right)+{WX}_{best}\left(t\right) if {r}_{3}<05\end{array}\right.$$
(19)

The whales engage in a random search for their optimal prey, denoted as \({WX}_{best}\), by exploring a random position (\({WX}_{rand}\)) instead of relying on the best solution (also \({WX}_{best}\)), which is determined using the following equations.

$${WX}_{curr}\left(t+1\right)={WX}_{rand}\left(t\right)-A\times {D}_{dis}$$
(20)
$${D}_{dis}=\left|{C.WX}_{rand}\left(t\right)-{WX}_{curr}\left(t\right)\right|$$
(21)

The WOA algorithm utilizes four parameters, namely a, A, C, and r3, to update solutions. If r3 is greater than 0.5, Eq. 17 is employed for solution updating. On the other hand, if r3 is less than 0.5, Eqs. 20 and 21 are used. Furthermore, depending on the value of |A|, the solution can also be updated using Eqs. 13 and 15. The WOA updates the position vector of the solution until the specified conditions are satisfied. Upon meeting the criteria for termination, WOA’s best solution is returned as WXbest. Figures 2 and 3 depict how solutions are updated in the WOA, demonstrating that solutions have the flexibility to move close to a solution in random positions or follow a spiral path.

Fig. 2
figure 2

Bubble-net search strategy

Fig. 3
figure 3

Position update in a spiral

Proposed algorithm

The proposed algorithm utilizes WOA to optimize both execution time (makespan) and monetary cost of workflows while ensuring a balanced load distribution across all nodes. WOA stands out among other meta-heuristics due to its speed and faster convergence, making it a favorable choice for this application. The algorithm finds its application in various cloud computing areas, including VM scheduling and placement, among others. To effectively execute the algorithm, the scheduler is assumed to possess knowledge of task dependencies within the workflow, and the execution times of individual tasks are predetermined. The main objective of the suggested algorithm is to efficiently allocate cloud resources to workflow tasks, thereby minimizing both execution time and monetary cost. Achieving this allocation is crucial for ensuring the effective utilization of cloud resources. In addition to this, the scheduler needs to take other relevant parameters into account while scheduling cloud resources, further enhancing the overall efficiency and effectiveness of the scheduling process.

Before applying the WOA, the proposed algorithm employs a preprocessing step to organize resources and tasks for optimization. Initially, the algorithm prioritizes tasks based on the number of descendants they have, giving priority to tasks with a large number of descendants. These tasks tend to become bottlenecks for cloud resources and consequently lead to longer execution times. Similarly, the algorithm also sorts cloud resources according to their processing power, distinguishing between low-processing power and high-processing power resources. It maintains two lists of resources for processing workflow tasks. Parent tasks, which require significant processing time due to their dependencies, are handled by high-processing nodes to expedite dependency elimination. Once the parent tasks are completed, the children’s tasks are addressed based on their place in the workflow graph. Leaf tasks are performed using low-processing nodes, while parent and intermediate tasks are assigned to available high-processing nodes. To ensure load balance across resources, the algorithm periodically switches tasks between two lists: the dependency list containing tasks with dependencies and the independent list containing tasks without dependencies. The WOA is then utilized for resource-to-task assignments for both lists. Each task begins execution with random whale positions and velocities, and each whale’s performance is assessed using the fitness function. Variables are adjusted iteratively through multiple iterations until a stopping criterion is reached. The fitness function, used to calculate the fitness of a whale, considers both the execution time and the cost, as represented in Eq. 22. Figure 4 depicts the flowchart of the proposed algorithm.

Fig. 4
figure 4

The preprocessing step in the proposed algorithm

$$F=w\frac{{max}^{ms}-ms}{{max}^{ms}{min}^{ms}}+w\frac{{max}^{mc}-mc}{{max}^{mc}-{min}^{mc}}+w\frac{{max}^{lb}-lb}{{max}^{lb}-{min}^{lb}}$$
(22)

Results and discussion

In this section, we provide an overview of performance measures, simulation settings, and the workflow design employed in the proposed algorithm. A collection of interconnected computational tasks characterizes a workflow. Various scientific workflows like Pan-STARRS, Epigenomics, Cybershake, Montage, and LIGO have been documented in the literature. For this paper, the Pan-STARRS scientific workflow is adopted as the framework for task execution. The Pan-STARRS project is dedicated to ongoing sky monitoring to identify movable or varying celestial objects. The PS1 telescopic instrument is employed for this purpose. The astronomy data generated by this project is managed by John Hopkins University and Microsoft, utilizing two distinct workflows: PSMerge and PSLoad. The PSMerge workflow focuses on database updates, while the PSLoad workflow collects data from the telescope and records them in the database. Detailed depictions of these workflows are illustrated in Figs. 5 and 6. Comprehensive characteristics of these workflows are outlined in Table 2.

Fig. 5
figure 5

PSLoad workflow

Fig. 6
figure 6

PSMerge workflow

Table 2 Workflow description

The simulation of the cloud environment and validation of the proposed scheduling model are carried out using the CloudSim simulator. Two distinct host types were deployed: HP ProLiant ML110 G4 and HP ProLiant ML110 G5. These host types exhibit variations in energy consumption rates, quantified in watts per second (Ws-1). Energy consumption rates for deployed hosts are specified. Also, a separate value is provided for the energy consumption associated with transferring 1 gigabyte (GB) of data. Four unique VM configurations were implemented in the simulation environment. These machines differ in terms of their processing power, measured in millions of instructions per second (MIPS), and their available random access memory (RAM) capacity, specified in megabytes (MB). Each VM configuration is broken down:

  • VM type 1: This VM offers 500 MIPS of processing power and is equipped with 613 MB of RAM.

  • VM type 2: This VM configuration provides 1000 MIPS of processing power and boasts 1740 MB of RAM.

  • VM type 3: The third VM configuration delivers 2000 MIPS of processing power while maintaining the same 1740 MB of RAM capacity as VM type 2.

  • VM type 4: This specialized VM configuration is designated for executing the scientific workflow. It offers 2500 MIPS of processing power but has a reduced RAM capacity of 870 MB.

The average VM startup time was established within the simulation to align with scientific workflow requirements. Furthermore, the average bandwidth between VMs was configured to approximate the bandwidth capabilities offered by a prominent cloud computing service provider, Amazon Web Services (AWS). A real-world scientific workflow, derived from the Pan-STARRS astronomical survey project, was chosen as the benchmark for this simulation. This workflow is categorized into three distinct groups based on the number of tasks it entails. A detailed breakdown of these task groupings is provided in Table 2.

The assessment of the proposed method is carried out through a comparative analysis against PESVMC [22] and EVMP [23], considering key performance metrics encompassing execution time (makespan), energy consumption, and resource utilization. The makespan, representing the time taken for the complete execution of the scientific workflow from its initial tasks to the final task, is evaluated using Eq. 23. Here, \({subTime}_{workflow}\) denotes the submission time of the workflow. Energy consumption quantifies the overall energy expended by the servers during the execution of the scientific workflow. This is computed utilizing Eq. 24, where \({ecr}_{ijk}\) signifies the energy consumption rate of VM j on host k and \({x}_{ijk}\) indicates the mapping of task i onto VM j at host k. The variable \({x}_{ijk}\) is set to 1 if task i is scheduled on VM j at host k for execution; otherwise, it is 0. The average resource utilization denotes the proportion of allocated computing resources for accomplishing the tasks of the scientific workflow relative to the total computing resources available on the server. This is calculated using Eq. 25.

$$makespan=\text{max}\left({ft}_{ijk}\right)-{subTime}_{workflow}$$
(23)
$$EC=\sum_{k=1}^{\left|{Host}_{a}\right|}\sum_{j=1}^{\left|{VM}_{k}\right|}\sum_{i=1}^{T}{x}_{ijk}.{ecr}_{ijk}.{et}_{ijk}$$
(24)
$$ARU=\sum_{k=1}^{\left|{Host}_{a}\right|}\sum_{j=1}^{\left|{VM}_{k}\right|}\sum_{i=1}^{T}{l}_{ijk}.{x}_{ijk}\div \sum_{k=1}^{\left|{Host}_{a}\right|}{c}_{k}.{at}_{k}$$
(25)

The comparison of makespan across different methods, including WOA, PESVMC, and EVMP, is presented for both PSMerge and PSLoad workflows, each with varying numbers of tasks. Figures 7 and 8 depict simulation outcomes concerning makespan. Results show that the WOA algorithm is superior to the others in terms of makespan efficiency in both workflows. On average, there is a reduction of 10% and 98% in makespan when compared to the EVMP and PESVMC algorithms, respectively. The inferior efficiency of the current algorithm results from its failure to consider child-parent relationships during task scheduling on VMs. This deficiency has had a noticeable impact on the workflow’s makespan.

Fig. 7
figure 7

Makespan comparison for PSMerge workflow

Fig. 8
figure 8

Makespan comparison for PSLoad workflow

The comparison of total energy consumption across different methods, including WOA, PESVMC, and EVMP, is conducted for both PSMerge and PSLoad scientific workflows with varying task counts. The test results concerning total energy consumption, expressed in kilowatts (kWs), are illustrated in Figs. 9 and 10. The results suggest that the WOA algorithm demonstrates an impressive reduction in energy usage. This achievement stems from the WOA algorithm’s ability to allocate resources in alignment with workflow task requirements, effectively curbing energy usage. Furthermore, the suggested algorithm takes into account child-parent relationships in task scheduling, leading to a reduction in data transfer-related energy consumption. On average, the WOA algorithm showcases a remarkable reduction of 20% and 42% in energy consumption when compared to the EVMP and PESVMC algorithms, respectively. These substantial energy savings highlight the potential of the WOA algorithm in optimizing resource utilization and energy efficiency in the context of workflow scheduling.

Fig. 9
figure 9

Energy consumption comparison for PSMerge workflow

Fig. 10
figure 10

Energy consumption comparison for PSLoad workflow

The assessment of average resource utilization across different methods, including WOA, PESVMC, and EVMP, is conducted for both PSMerge and PSLoad scientific workflows, each encompassing varying numbers of tasks. The experimental findings regarding average resource utilization are graphically presented in Figs. 11 and 12. The findings underscore the better resource utilization performance of the WOA algorithm when compared to EVMP and PESVMC. The dynamic nature of the WOA algorithm contributes to its enhanced performance. The proposed algorithm strategically creates new VMs when the currently deployed VMs prove inadequate to meet task deadlines. This dynamic provisioning ensures optimal resource utilization. Additionally, the utilization of a VM migration policy facilitates resource consolidation, leading to impressive gains in resource utilization. On average, the WOA algorithm achieves a noteworthy increase of 10% and 8.6% in resource utilization over the baseline algorithms. These substantial improvements validate the effectiveness of the WOA algorithm in dynamically adapting to resource demands and efficiently utilizing available computing resources.

Fig. 11
figure 11

Resource utilization comparison for PSMerge workflow

Fig. 12
figure 12

Resource utilization comparison for PSLoad workflow

Conclusion

In the ever-evolving environment of cloud computing, where the submission of scientific workflows by user groups is a prevalent scenario, the efficient scheduling of extensive task sets within these workflows has emerged as a pivotal challenge. Task scheduling optimization holds paramount significance in cloud computing, aiming to minimize workflow execution time, optimize cloud resource consumption, and reduce execution costs for users. To cope with these imperatives, this study introduced a novel WOA-based task scheduling technique. This algorithm achieved an optimal distribution of tasks across available computing resources. The algorithm capitalizes on the concept of “whales,” which symbolize potential task-resource assignments. Through an iterative process, the algorithm explores and refines these assignments to enhance the overall workflow performance, thereby striving to cut down execution time, resource usage, and associated budgets. The findings of comprehensive simulations provide compelling evidence for the better efficiency of the proposed WOA-based algorithm. Notably, in comparison to previous algorithms, the proposed approach excels in energy consumption, resource utilization, and makespan.

Availability of data and materials

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Abbreviations

AWS:

Amazon Web Services

DAG:

Directed acyclic graphs

DVFS:

Dynamic voltage and frequency scaling

FT:

Finish time

GOA:

Grasshopper optimization algorithm

HEFT:

Heterogeneous earliest finish time

IGD:

Inverted generational distance

MC:

Monetary cost

PPR:

Performance-to-power ratio

QoS:

Quality of service

SOA:

Seagull optimization algorithm

SLA:

Service-level agreement

TEC:

Total execution cost

TTC:

Total transfer cost

VM:

Virtual machine

WOA:

Whale optimization algorithm

WMS:

Workflow management system

References

  1. Hayyolalam V, Pourghebleh B, Kazem AAP, Ghaffari A (2019) Exploring the state-of-the-art service composition approaches in cloud manufacturing systems to enhance upcoming techniques. Int J Adv Manufact Technol 105(1–4):471–498

    Article  Google Scholar 

  2. Wang X, Sun Y, Sun Q, Lin W, Wang JZ, Li W (2023) HCIndex: a Hilbert-curve-based clustering index for efficient multi-dimensional queries for cloud storage systems. Clust Comput 26(3):2011–2025

    Article  Google Scholar 

  3. Hayyolalam V, Pourghebleh B, Chehrehzad MR, Pourhaji Kazem AA (2022) Single-objective service composition methods in cloud manufacturing systems: recent techniques, classification, and future trends. Concurr Comput 34(5):e6698

    Article  Google Scholar 

  4. Yakubu IZ, Murali M (2023) An efficient meta-heuristic resource allocation with load balancing in IoT-Fog-cloud computing environment. J Ambient Intell Humaniz Comput 14(3):2981–2992

    Article  Google Scholar 

  5. Sefati S, Mousavinasab M, Zareh Farkhady R (2022) Load balancing in cloud computing environment using the grey wolf optimization algorithm based on the reliability: performance evaluation. J Supercomputing 78(1):18–42

    Article  Google Scholar 

  6. Al-Jumaili AHA, Muniyandi RC, Hasan MK, Paw JKS, Singh MJ (2023) Big data analytics using cloud computing based frameworks for power management systems: status, constraints, and future recommendations. Sensors 23(6):2952

    Article  Google Scholar 

  7. He J (2022) Cloud computing load balancing mechanism taking into account load balancing ant colony optimization algorithm. Comput Intell Neurosci 2022:3120883

    Google Scholar 

  8. Mangalampalli S et al (2023) Prioritized task-scheduling algorithm in cloud computing using cat swarm optimization. Sensors 23(13):6155

    Article  Google Scholar 

  9. Praveenchandar J, Tamilarasi A (2021) Dynamic resource allocation with optimized task scheduling and improved power management in cloud computing. J Ambient Intell Humaniz Comput 12(3):4147–4159

    Article  Google Scholar 

  10. Dubey K, Sharma SC (2021) A novel multi-objective CR-PSO task scheduling algorithm with deadline constraint in cloud computing. Sustain Comput 32:100605

    Google Scholar 

  11. M. Hosseinzadeh, M. Y. Ghafour, H. K. Hama, B. Vo, and A. Khoshnevis, (2020) Multi-objective task and workflow scheduling approaches in cloud computing: a comprehensive review. J Grid Comput  18:1–30

  12. Kamanga CT, Bugingo E, Badibanga SN, Mukendi EM (2023) A multi-criteria decision making heuristic for workflow scheduling in cloud computing environment. J Supercomput 79(1):243–264

    Article  Google Scholar 

  13. Mikram H, El Kafhali S, Saadi Y (2024) HEPGA: a new effective hybrid algorithm for scientific workflow scheduling in cloud computing environment. Simul Model Pract Theory 130:102864

    Article  Google Scholar 

  14. Asghari Alaie Y, Hosseini Shirvani M, Rahmani AM (2023) A hybrid bi-objective scheduling algorithm for execution of scientific workflows on cloud platforms with execution time and reliability approach. J Supercomputing 79(2):1451–1503

    Article  Google Scholar 

  15. Shi J, Luo J, Dong F, Zhang J, Zhang J (2016) Elastic resource provisioning for scientific workflow scheduling in cloud under budget and deadline constraints. Clust Comput 19:167–182

    Article  Google Scholar 

  16. Aziza H, Krichen S (2020) A hybrid genetic algorithm for scientific workflow scheduling in cloud environment. Neural Comput Appl 32:15263–15278

    Article  Google Scholar 

  17. Iranmanesh A, Naji HR (2021) DCHG-TS: a deadline-constrained and cost-effective hybrid genetic algorithm for scientific workflow scheduling in cloud computing. Clust Comput 24:667–681

    Article  Google Scholar 

  18. A Mohammadzadeh, M Masdari (2021) Scientific workflow scheduling in multi-cloud computing using a hybrid multi-objective optimization algorithm. J Ambient Intellig Humanized Comput 14:3509–3529

  19. Choudhary A, Govil MC, Singh G, Awasthi LK, Pilli ES (2022) Energy-aware scientific workflow scheduling in cloud environment. Clust Comput 25(6):3845–3874

    Article  Google Scholar 

  20. Khaleel MI (2022) Multi-objective optimization for scientific workflow scheduling based on performance-to-power ratio in fog–cloud environments. Simul Model Pract Theory 119:102589

    Article  Google Scholar 

  21. Al-Moalmi A, Luo J, Salah A, Li K, Yin L (2021) A whale optimization system for energy-efficient container placement in data centers. Expert Syst Appl 164:113719

    Article  Google Scholar 

  22. Mohanapriya N, Kousalya G, Balakrishnan P, Pethuru Raj C (2018) Energy efficient workflow scheduling with virtual machine consolidation for green cloud computing. J Intell Fuzzy Syst 34(3):1561–1572

    Article  Google Scholar 

  23. N. Garg, M. Raj, I. Gupta, V. Kumar, and G. Sinha, “Energy-efficient scientific workflow scheduling algorithm in cloud environment,” Wireless Communications and Mobile Computing, vol. 2022, 2022.

Download references

Acknowledgements

Not applicable.

Funding

No funding.

Author information

Authors and Affiliations

Authors

Contributions

XZ contributed to writing—original draft preparation and conceptualization.

Corresponding author

Correspondence to Xiaowen Zhang.

Ethics declarations

Competing interests

The author declares no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, X. Optimizing scientific workflow scheduling in cloud computing: a multi-level approach using whale optimization algorithm. J. Eng. Appl. Sci. 71, 175 (2024). https://doi.org/10.1186/s44147-024-00512-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s44147-024-00512-9

Keywords