RP ML Workflow Scheduling
RP ML Workflow Scheduling
Victoria Huang1 , Chen Wang1 , Hui Ma2 , Gang Chen2 , and Kameron Christopher1
1
National Institute of Water and Atmospheric Research, Wellington, New Zealand
{victoria.huang, chen.wang, kameron.christopher}@niwa.co.nz
2
Victoria University of Wellington, Wellington, New Zealand
{hui.ma, aaron.chen}@ecs.vuw.ac.nz
1 Introduction
Large-scale and highly complex computational applications (e.g., weather fore-
casting and Tsunami prediction) are usually modeled as workflows in cloud [17,
⋆
This work is in part supported by the NZ Government’s Strategic Science Invest-
ment Fund (SSIF) and the New Zealand Marsden Fund with the contract number
(VUW1510), administrated by the Royal Society of New Zealand.
2 V. Huang et al.
determined and remains fixed [13, 20, 25]. (2) Their performance is sensitive to
hyper-parameter settings while hyper-parameter search is difficult.
The aim of this paper is to propose an effective approach for training newly
designed scheduling policies to handle a changing number of VMs and workflows
to address the cost-aware DMWS. Specifically, we design a new Deep Neural
Network (DNN) based scheduling policy to dynamically rent new VMs and allo-
cate simultaneously any tasks ready for execution to the rented VMs. We show
that the designed scheduling policies can be used to efficiently assign a priority
value to each candidate VM for a given task. The VM with the highest priority
will be selected for renting and/or task execution. In line with the new policy
network design, a new training approach called Evolutionary Strategy based
Reinforcement Learning (ES-RL) for DMWS is proposed. Our new training ap-
proach features the use of Evolutionary Strategy (ES), a deep neuroevolution
algorithm, to achieve stable and effective training of our policy network. Mean-
while, ES-RL is not sensitive to hyper-parameter settings and offers significant
performance gain and time reduction through its parallel training capabilities.
Specifically, the key contributions of this paper are listed as follows:
– A Priority-based Deep Neural Network (DNN) scheduling policy design is
proposed to flexibly adapt to a changing number of VMs and workflows.
– An evolutionary DRL approach called Evolutionary Strategy based Rein-
forcement Learning (ES-RL) is proposed to achieve robust and time-efficient
training of new policy networks that can solve the DMWS problem effec-
tively.
– Extensive experiments with real-world datasets have been conducted to show
that the scheduling policies trained by ES-RL can significantly reduce the
costs by more than 90% compared to the state-of-the-art WS approaches.
2 Related Work
3 Problem Formulation
In this paper we study the cost-aware Dynamic Multi-Workflow Scheduling
(DMWS) problem. This section presents a formal definition of the problem.
where I(π, T ) is the set of VMs being rented within the time period T .
where DAG(w) includes a set of tasks {T ask(w, i)|i ∈ {1, 2, ...}} and directed
edges connecting the tasks to enforce their execution order. Note that T ask(w, i)
is associated with execution time Ref T (T ask(w, i)) and can only be executed if
all its predecessor tasks P re(T ask(w, i)) are completed. T ask(w, i) is an entry
task if P re(T ask(w, i)) = ∅. Similarly, the successors of T ask(w, i) are denoted
as Suc(T ask(w, i)). A task with no successors is an exit task.
In this paper, we assume that the arrival time of any new workflows is not
known in advance. To avoid SLA violation and flexibly utilize the VM resources,
6 V. Huang et al.
scheduling decisions are made in real time, e.g., whenever a task is ready. In
particular, a task T ask(w, i) is defined as ready if it is either an entry task
of a workflow (i.e., P re(T ask(w, i)) = ∅) or a task with all its predecessors
P re(T ask(w, i)) completed. We define a set of candidate VMs as CV M (t) which
includes all leased VMs at time t and a set of VM options with all VM types
that can be created. Whenever T ask(w, i) is ready at time t, π selects a VM v
from CV M (t) for T ask(w, i) allocation.
Following π, the start time for T ask(w, i) is ST (T ask(w, i), π) and the com-
pletion time is
Ref T (T ask(w, i))
CT (T ask(w, i), π) = ST (T ask(w, i), π) +
Capa(v(π))
Thus, the completion time W CT of a workflow w is the maximum completion
time among all tasks:
W CT (w, π) = max {CT (T ask(w, i), π)}
T ask(w,i)∈DAG(w)
SLA Penalty: Following existing works [28, 31], the SLA violation penalty of
workflow w can be defined as follows:
(
0, if W CT (w, π) ≤ DL(w)
P enalty(w, π) =
ϵ + β(w) × (W CT (w, π) − DL(w)), otherwise
Scheduling
Schedulingpolicy ⇡ from
policy from SaaS provider
workflow broker VM
Dynamic workflows submitted by users
…
VM
…
…
O
…
…
…
f✓ VM
Queuing Executing
tasks tasks
Priority Mapping: Using the extracted state features z(v, t), the priority func-
tion fθ with trainable parameters θ calculates a priority value p(v, t) for every
VM candidate v:
VM Selection: Given the priorities, a VM a(t) with the highest priority value
is selected by Φ:
The goal of ES-RL is to find θ that can minimize the total cost defined in
Eq. (4) through minimizing the expected objective value over the population
Cost-Aware DMWS in Cloud Data Center Using Evolutionary RL 9
Cloud environment
…
VM
… …
… …
O Step 4: Scheduling
…
decision a(t)
VM
…
Step 2: Update Dynamic Queuing Executing
individuals ✓ + ✏ workflows tasks tasks
2
Step 1: Sample noise ✏ ⇠ N( , I)
Multi-variance Gaussian
2
N( , I)
distribution, i.e., Eϵ∼N (0,I) F (θ + σϵi ). To achieve this goal, ES updates θ using
the following estimator:
1
∇θ Eθ∼pψ F (θ) = ∇θ Eϵ∼N (0,I) F (θ + σϵ) = Eϵ∼N (0,I) F (θ + σϵ)ϵ
σ
N (5)
1 X
≈ F (θ + σϵi )ϵi
N σ i=1
10 V. Huang et al.
6 Performance Evaluation
Cloud Environment: We consider a cloud data center equipped with six differ-
ent VM types. Following [3,9], VMs are configured based on the general-purpose
on-demand VM group in the US East region provided by Amazon EC2. VM
details are summarized in Tab. 1 which were collected in September 2020 from
Amazon6 .
Workflows: Following existing studies [3, 9, 29], four classes of scientific work-
flows are used in our experiment which include CyberShake, Inspiral, Montage,
and Sipht. All workflows are computational intensive and their communication
time is trivial compared to the task execution time [9]. Detail analysis of these
workflows can be found in [4,12]. Note that workflows in each class share similar
structures but can differ from the numbers of their tasks (e.g., ranging from
25 to 50). During our training, we simulate small workflows from four classes
where each of them contains 25 to 30 tasks. Nevertheless, our trained scheduling
policy is generalized well and can be directly applied to large workflows with
50 tasks without retraining. Following [6, 15], we simulate the dynamic arrival
of workflow applications by randomly sampling 30 workflows from four classes
of workflows in each simulation. The arrival time of workflows follows a Poisson
distribution with λ = 0.01 [15]. The penalty rate β is set to be $0.24 per hour
for each workflow [32].
To compare the algorithm performance under different deadlines, a deadline
relaxation factor η is used. Similar to [3], we consider η changes from 1 to 2.25.
Given η, the deadline of a workflow w is set as below:
VM Utilization
VM Cost ($) 250 1.75 250 1.75 0.5 1.75
2.0 2.0 2.0
200 2.25 200 2.25 0.4 2.25
150 150 0.3
100 100 0.2
50 50 0.1
0 0 0.0
ProLiS GRP-HEFT ES-RL ProLiS GRP-HEFT ES-RL ProLiS GRP-HEFT ES-RL
Fig. 3: Comparison of GRP-HEFT [9], ProLiS [29], and ES-RL on different fac-
tors with respect to different η with small workflows.
Table 2: The average fitness values (i.e., total cost) tested over multiple runs
for GRP-HEFT [9] and ProLiS [29], and ES-RL with different η with small
workflows. (Note: a lower value is better)
η ProLiS GRP-HEFT ES-RL
1.00 395.5520 ± 11.633617 - 74.490828 ± 7.060981
1.25 108.5184 ± 15.815567 1775.9232 ± 161.077450 71.817811 ± 7.664415
1.50 91.7376 ± 12.662934 1171.8144 ± 75.565892 73.737249 ± 6.339560
1.75 83.4304 ± 10.828026 1258.9056 ± 90.980289 71.436684 ± 5.725663
2.00 76.9088 ± 10.752586 1164.7488 ± 76.973755 68.036960 ± 5.993622
2.25 68.3456 ± 10.332963 969.9840 ± 102.602041 65.590513 ± 7.967546
1.75 1.75
VM Cost ($)
2.0 2.0
200 2.25 200 2.25
100 100
0 0
ProLiS ES-RL ProLiS ES-RL
VM Utilization
Fig. 4: Comparison of ProLiS [29] and ES-RL on different factors with respect
to different η with large workflows.
7 Conclusions
In this paper, we proposed an effective ES based approach for the cost-aware
dynamic multi-workflow scheduling (DMWS) problem. In particular, we formu-
late a dynamic multi-workflow scheduling problem with the goal of minimizing
both the VM rental cost and SLA violation penalties. To effectively solve this
problem, we proposed a new scheduling policy design of a priority-based deep
neural network that can be used in a dynamic environment with a changing
number of VMs and workflows. Meanwhile, a new Evolutionary Strategy based
RL (ES-RL) algorithm for DMWS is proposed to efficiently train a generally ap-
plicable scheduling policy in parallel. Our experiments with real-world datasets
showed that the scheduling policies trained by ES-RL can effectively reduce the
overall costs compared to two state-of-the-art algorithms.
14 V. Huang et al.
References
1. Ahmad, S.G., Liew, C.S., Munir, E.U., Ang, T.F., Khan, S.U.: A hybrid genetic
algorithm for optimization of scheduling workflow applications in heterogeneous
computing systems. Journal of Parallel and Distributed Computing 87, 80–90
(2016)
2. Alsurdeh, R., Calheiros, R.N., Matawie, K.M., Javadi, B.: Hybrid workflow schedul-
ing on edge cloud computing systems. IEEE Access 9, 134783–134799 (2021)
3. Arabnejad, V., Bubendorfer, K., Ng, B.: Budget and deadline aware e-science work-
flow scheduling in clouds. IEEE Transactions on Parallel and Distributed Systems
30(1), 29–44 (2019)
4. Bharathi, S., Chervenak, A., Deelman, E., Mehta, G., Su, M., Vahi, K.: Character-
ization of scientific workflows. In: 2008 Third Workshop on Workflows in Support
of Large-Scale Science. pp. 1–10 (2008)
5. Byun, E.K., Kee, Y.S., Kim, J.S., Deelman, E., Maeng, S.: Bts: Resource capacity
estimate for time-targeted science workflows. Journal of Parallel and Distributed
Computing 71(6), 848–862 (2011)
6. Chen, H., Zhu, X., Liu, G., Pedrycz, W.: Uncertainty-aware online scheduling for
real-time workflows in cloud service environment. IEEE Transactions on Services
Computing pp. 1–1 (2018)
7. Dong, T., Xue, F., Xiao, C., Zhang, J.: Workflow scheduling based on deep rein-
forcement learning in the cloud environment. Journal of Ambient Intelligence and
Humanized Computing 12(12), 10823–10835 (2021)
8. Escott, K.R., Ma, H., Chen, G.: Genetic programming based hyper heuristic ap-
proach for dynamic workflow scheduling in the cloud. In: International Conference
on Database and Expert Systems Applications. pp. 76–90. Springer (2020)
9. Faragardi, H.R., Saleh Sedghpour, M.R., Fazliahmadi, S., Fahringer, T., Rasouli,
N.: GRP-HEFT: A Budget-Constrained Resource Provisioning Scheme for Work-
flow Scheduling in IaaS Clouds. IEEE Transactions on Parallel and Distributed
Systems 31(6), 1239–1254 (2020)
10. Genez, T.A.L., Bittencourt, L.F., Madeira, E.R.M.: Workflow scheduling for saas /
paas cloud providers considering two sla levels. In: 2012 IEEE Network Operations
and Management Symposium. pp. 906–912 (2012)
11. Hoseiny, F., Azizi, S., Shojafar, M., Tafazolli, R.: Joint qos-aware and cost-efficient
task scheduling for fog-cloud resources in a volunteer computing system. ACM
Transactions on Internet Technology (TOIT) 21(4), 1–21 (2021)
12. Juve, G., Chervenak, A., Deelman, E., Bharathi, S., Mehta, G., Vahi, K.: Charac-
terizing and profiling scientific workflows. Future Generation Computer Systems
29(3), 682–692 (2013)
13. Li, H., Huang, J., Wang, B., Fan, Y.: Weighted double deep q-network based rein-
forcement learning for bi-objective multi-workflow scheduling in the cloud. Cluster
Computing 25(2), 751–768 (2022)
14. Liessner, R., Schmitt, J., Dietermann, A., Bäker, B.: Hyperparameter optimization
for deep reinforcement learning in vehicle energy management. In: ICAART (2).
pp. 134–144 (2019)
15. Liu, J., Ren, J., Dai, W., Zhang, D., Zhou, P., Zhang, Y., Min, G., Najjari, N.: On-
line multi-workflow scheduling under uncertain task execution time in iaas clouds.
IEEE Transactions on Cloud Computing pp. 1–1 (2019)
16. Lopez-Garcia, P., Onieva, E., Osaba, E., Masegosa, A.D., Perallos, A.: Gace: A
meta-heuristic based in the hybridization of genetic algorithms and cross entropy
Cost-Aware DMWS in Cloud Data Center Using Evolutionary RL 15
methods for continuous optimization. Expert Systems with Applications 55, 508–
519 (2016)
17. Masdari, M., ValiKardan, S., Shahi, Z., Azar, S.I.: Towards workflow scheduling
in cloud computing: a comprehensive analysis. Journal of Network and Computer
Applications 66, 64–82 (2016)
18. Oliver, H., Shin, M., Matthews, D., Sanders, O., Bartholomew, S., Clark, A., Fitz-
patrick, B., van Haren, R., Hut, R., Drost, N.: Workflow automation for cycling
systems. Computing in Science & Engineering 21(4), 7–21 (2019)
19. Pandey, S., Wu, L., Guru, S.M., Buyya, R.: A particle swarm optimization-based
heuristic for scheduling workflow applications in cloud computing environments.
In: 2010 24th IEEE International Conference on Advanced Information Networking
and Applications. pp. 400–407 (2010)
20. Qin, Y., Wang, H., Yi, S., Li, X., Zhai, L.: An energy-aware scheduling algorithm
for budget-constrained scientific workflows based on multi-objective reinforcement
learning. The Journal of Supercomputing 76(1), 455–480 (2020)
21. Rodriguez, M.A., Buyya, R.: Deadline based resource provisioningand scheduling
algorithm for scientific workflows on clouds. IEEE Transactions on Cloud Com-
puting 2(2), 222–235 (2014)
22. Salimans, T., Ho, J., Chen, X., Sidor, S., Sutskever, I.: Evolution strategies as
a scalable alternative to reinforcement learning. arXiv preprint arXiv:1703.03864
(2017)
23. Schulman, J., Levine, S., Abbeel, P., Jordan, M., Moritz, P.: Trust region policy
optimization. In: International Conference on Machine Learning (ICML). pp. 1889–
1897 (2015)
24. Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal policy
optimization algorithms. arXiv preprint arXiv:1707.06347 (2017)
25. Suresh Kumar, D., Jagadeesh Kannan, R.: Reinforcement learning-based controller
for adaptive workflow scheduling in multi-tenant cloud computing. The Interna-
tional Journal of Electrical Engineering & Education p. 0020720919894199 (2020)
26. Topcuoglu, H., Hariri, S., Min-You Wu: Performance-effective and low-complexity
task scheduling for heterogeneous computing. IEEE Transactions on Parallel and
Distributed Systems 13(3), 260–274 (2002)
27. Wang, Y., Liu, H., Zheng, W., Xia, Y., Li, Y., Chen, P., Guo, K., Xie, H.: Multi-
objective workflow scheduling with deep-q-network-based multi-agent reinforce-
ment learning. IEEE access 7, 39974–39982 (2019)
28. Wu, L., Garg, S.K., Versteeg, S., Buyya, R.: Sla-based resource provisioning for
hosted software-as-a-service applications in cloud computing environments. IEEE
Transactions on Services Computing 7(3), 465–485 (2014)
29. Wu, Q., Ishikawa, F., Zhu, Q., Xia, Y., Wen, J.: Deadline-constrained cost op-
timization approaches for workflow scheduling in clouds. IEEE Transactions on
Parallel and Distributed Systems 28(12), 3401–3412 (2017)
30. Xiaoyong, Y., Ying, L., Tong, J., Tiancheng, L., Zhonghai, W.: An analysis on
availability commitment and penalty in cloud sla. In: 2015 IEEE 39th Annual
Computer Software and Applications Conference. vol. 2, pp. 914–919 (2015)
31. Yang, Y., Chen, G., Ma, H., Zhang, M., Huang, V.: Budget and sla aware dynamic
workflow scheduling in cloud computing with heterogeneous resources. In: 2021
IEEE Congress on Evolutionary Computation (CEC). pp. 2141–2148. IEEE (2021)
32. Youn, C.H., Chen, M., Dazzi, P.: Cloud Broker and Cloudlet for Workflow Schedul-
ing. Springer (2017)