ProMo: A Probabilistic Model for Dynamic Load-Balanced Scheduling of Data Flows in Cloud Systems
Abstract
:1. Introduction
- The proposed model can be used independently with existing scheduling strategies for testing;
- The proposed model is simple, so its use does not add any overheads to a scheduling scheme;
- The model can be the basis for the development of a complete simulator of data flow scheduling algorithms, which will be used for testing different algorithms, topologies, and network parameters;
- The model can also become the basis for the development of a new scheduling strategy;
- It is a time interval-based model, so it is particularly useful in testing dynamic scheduling algorithms.
2. Related Work
2.1. Traditional Networking Schemes
2.2. Software-Defined Networking Schemes
3. The ProMo Model for Dynamic Flow-Scheduling
3.1. Three-Layer Fully-Populated Network
3.2. The ProMo Model
- The probability values can be obtained after collecting information regarding the system’s behavior. For example, to obtain the probability of no link collapse, information regarding the total time within a period (for example, one day or some hours) that links work properly is required. Furthermore, to have a probability of data corruption, it is necessary to keep track of how many times the links have received or sent corrupted data out of the total transmission it has taken over.
- The probability values used in the model are the sum of two probabilities per layer (that is, for all the layer links) mentioned in the first observation divided by two (average of the two probabilities).
- A new job arrives, with probability
- A layer completes processing of a job to forward it to another layer, with probability
- (C.i)
- No event occurs, that is, :
- (C.ii)
- One new job arrival:
- (C.iii)
- One job completed and transferred from to , or from to , or from to the end nodes:
- (C.iv)
- One job was interrupted due to link collapse or data corruption:
- λ1:
- The mean arrival rate of Layer 1 is determined by ; the arrival rate of the users’ jobs multiplied by a probability equal to one (assuming that the users always generate jobs) plus the mean arrival rate of the next Layer 2, multiplied by the probability of interrupted transmissions from Layer 1 to Layer 2 (due to corrupted data or collapsed links), .
- λ2:
- The mean arrival rate of Layer 2 is determined by the mean arrival rate of Layer 1, , multiplied by the probability of uninterrupted transmissions between Layers 1 and 2, , plus the mean arrival rate of the next Layer 3, multiplied by the probability of interrupted transmissions from Layer 2 to Layer 3 (due to corrupted data or collapsed links), .
- λ3:
- The mean arrival rate of Layer 3 is determined by the mean arrival rate of Layer 2, , multiplied by the probability of uninterrupted transmissions between Layers 2 and 3, , plus the mean arrival rate of the next Layer 4, multiplied by the probability of interrupted transmissions from Layer 4 to Layer 3 (due to corrupted data or collapsed links), .
- λ4:
- The mean arrival rate of Layer 4 is the product of the mean arrival rate of the upper layer multiplied by the probability of uninterrupted transmission from Layer 3 to Layer 4, .
3.3. An Illustrative Example
4. Simulation Analysis
4.1. Application of ProMo to the DLBS-FPN Scheduling Scheme
- The ProMo model was scheduled to work on network layers, not on single links, so the values obtained were average link values. For example, an average of 10% interrupted transfers was recorded in total for all the links connecting Layers 2 and 3 or 3 and 4. Thus, .
- Continuing the previous observation, the mean throughput values were average layer throughput values. To change these values to bytes/s, it was necessary to consider the average job size, record the number of bytes/s per link, and get the average. This approach will also lead to similar curves.
- The maximum point was the Layer 1 saturation point, beyond which the throughput will be dramatically reduced. In ProMo, this was shown by the fact that for f values beyond 1.31, the value of exceeded one, making negative (see Equation (13)). This is the reason that the curve was not plotted beyond the value of , in Figure 3.
- The bandwidth utilization line presented in [8] was somehow smoother compared to the one of Figure 4. The reason is because the results here were obtained on a per layer basis and not as the average values from the total number of links. Looking at the corresponding graph in [8], one can observe longer periods of stable bandwidth values compared to the line derived by the ProMo tester.
- The bandwidth utilization fell off as the load transmitted kept increasing with time, as also described for the DLBS scheme. This case was modeled by ProMo by an ever-decreased value of to avoid possible congestions in the links. In its turn, this caused a reduction of values, and thus, the average bandwidth utilization was reduced.
4.2. Application of ProMo to the DRB Scheduling Scheme
- The MRT values were on a “per user job basis”. To find the MRT values on a byte basis, it is necessary to use the average job size, but the behavior will be identical.
- The authors in [19] proposed a DRB probability-based threshold, which is used to improve the average latency. In correspondence with this, the value factored by at most f is an analogous parameter in the sense that, above this traffic value, there is not much to be done to prevent the MRT from falling-off.
5. Conclusions and Future Work
Funding
Conflicts of Interest
References
- Lu, J.; Li, D. Bias Correction in Small Sample from Big Data. IEEE Trans. Knowl. Data Eng. 2013, 25, 2658–2663. [Google Scholar] [CrossRef]
- Han, X.; Li, J.; Yang, D.; Wang, J. Efficient Skyline Computation on Big Data. IEEE Trans. Knowl. Data Eng. 2013, 25, 2521–2535. [Google Scholar] [CrossRef]
- Tantalaki, N.; Souravlas, S.; Roumeliotis, M. A review on Big Data real-time stream processing and its scheduling techniques. Int. J. Parallel Emerg. Distrib. Syst. 2019. [Google Scholar] [CrossRef]
- Souravlas, S.; Roumeliotis, M. Dynamic Load Balancing on All-to-All Personalized Communications Using the NNLB Principal. In Proceedings of the 2014 Sixth International Conference on Computational Intelligence, Communication Systems and Networks, Tetova, Macedonia, 27–29 May 2014; pp. 107–112. [Google Scholar]
- Souravlas, S.I.; Sifaleras, A. Network Load Balancing Using Modular Arithmetic Computations. In GeNeDis 2016; Springer: Berlin, Germany, 2017; Volume 988, pp. 271–280. [Google Scholar]
- Zhang, Q.; Zhani, M.F.; Yang, Y.; Boutaba, R.; Wong, B. PRISM: Fine-Grained Resource-Aware Scheduling for MapReduce. IEEE Trans. Cloud Comput. 2015, 99, 182–194. [Google Scholar] [CrossRef]
- Yuan, Y.; Wang, D.; Liu, J. Joint scheduling of MapReduce jobs with servers: Performance bounds and experiments. In Proceedings of the IEEE Conference on Computer Communications (IEEE INFOCOM 2014), Toronto, ON, Canada, 27 April–2 May 2014; pp. 2175–2183. [Google Scholar]
- Tang, F.; Yang, L.T.; Tang, C.; Li, J., Sr.; Guo, M. A Dynamical and Load-Balanced Flow Scheduling Approach for Big Data Centers in Clouds. IEEE Trans. Cloud Comput. 2018, 6, 915–928. [Google Scholar] [CrossRef]
- Cao, Z.Z.; Kodialam, M.; Lakshman, T.V. Joint Static and Dynamic Traffic Scheduling in Data Center Networks. In Proceedings of the IEEE Conference on Computer Communications 2014 (IEEE INFOCOM 2014), Toronto, ON, Canada, 27 April–2 May 2014; pp. 2445–2553. [Google Scholar]
- Congdon, P.T.; Mohapatra, P.; Farrens, M.; Akella, V. Simultaneously Reducing Latency and Power Consumption in OpenFlow Switches. IEEE/ACM Trans. Netw. 2014, 22, 1007–1020. [Google Scholar] [CrossRef]
- Misra, S.; Das, S.; Khatua, M.; Obaidat, M.S. QoS-Guaranteed Bandwidth Shifting and Redistribution in Mobile Cloud Environment. IEEE Trans. Cloud Comput. 2014, 2, 181–193. [Google Scholar] [CrossRef]
- Zhang, F.; Cao, J.; Hwang, K.; Li, K.; Khan, S.U. Adaptive Workflow Scheduling on Cloud Computing Platforms with Iterative Ordinal Optimization. IEEE Trans. Cloud Comput. 2014, 3, 156–168. [Google Scholar] [CrossRef]
- Rodriguez, M.A.; Buyya, R. Deadline Based Resource Provisioning and Scheduling Algorithm for Scientific Workflows on Clouds. IEEE Trans. Cloud Comput. 2014, 2, 222–235. [Google Scholar] [CrossRef]
- Shang, Y. Optimal Control Strategies for Virus Spreading in Inhomogeneous Epidemic Dynamics. Can. Math. Bull. 2012, 56, 621–629. [Google Scholar] [CrossRef]
- Shang, Y. Optimal Attack Strategies in a Dynamic Botnet Defense Model. Appl. Math. Inf. Sci. 2012, 6, 29–33. [Google Scholar]
- Malawski, M.; Juve, G.; Deelman, E.; Nabrzyski, J. Cost-and deadline-constrained provisioning for scientific workflow ensembles in iaas clouds. In Proceedings of the International Conference on High Performance Computing, Networking, Storage and Analysis, Salt Lake City, UT, USA, 10–16 November 2012; p. 22. [Google Scholar]
- Tsai, C.W.; Huang, W.C.; Chiang, M.H.; Chiang, M.C.; Yang, C.S. A Hyper-Heuristic Scheduling Algorithm for Cloud. IEEE Trans. Cloud Comput. 2014, 2, 236–249. [Google Scholar] [CrossRef]
- Neely, M.J. Distributed stochastic optimization via correlated scheduling. In Proceedings of the IEEE Conference on Computer Communications (IEEE INFOCOM 2014), Toronto, ON, Canada, 27 April–2 May 2014; pp. 2418–2426. [Google Scholar]
- Wang, S.; Luo, J.; Tong, B.K.B.; Wong, W.S. Randomized load-balanced routing for fat-tree networks. arXiv 2017, arXiv:1708.09135v1. [Google Scholar]
- Bianco, A.; Krishnamoorthi, V.; Li, N.; Giraudo, L. OpenFlow driven ethernet traffic analysis. In Proceedings of the 2014 IEEE International Conference on Communications (ICC), Sydney, Australia, 10–14 June 2014; pp. 3001–3006. [Google Scholar]
- Jin, H.; Pan, D.; Liu, J.; Pissinou, N. OpenFlow-Based Flow-Level Bandwidth Provisioning for CICQ Switches. IEEE Trans. Comput. 2013, 62, 1799–1812. [Google Scholar] [CrossRef]
- Chowdhury, M.; Zhong, Y.; Stoica, I. Efficient co-flow scheduling with Varys. In Proceedings of the 2014 ACM conference on SIGCOMM, Chicago, IL, USA, 17–22 August 2014; Volume 44, pp. 443–454. [Google Scholar]
- Ramos, R.M.; Martinello, M.; Rothenberg, E.C. SlickFlow: Resilient source routing in Data Center Networks unlocked by OpenFlow. In Proceedings of the 38th annual IEEE Conference on Local Computer Networks, Sydney, Australia, 21–24 October 2013; pp. 606–613. [Google Scholar]
- Bolla, R.; Bruschi, R.; Lombardo, C.; Podda, F. OpenFlow in the Small: A Flexible and Efficient Network Acceleration Framework for Multi-Core Systems. IEEE Trans. Netw. Serv. Manag. 2014, 11, 390–404. [Google Scholar] [CrossRef]
- Egilmez, H.E.; Civanlar, S.; Tekalp, A.M. An Optimization Framework for QoS-Enabled Adaptive Video Streaming Over OpenFlow Networks. IEEE Trans. Multimed. 2013, 15, 710–715. [Google Scholar] [CrossRef]
- Souravlas, S.; Sifaleras, A. Binary-tree based estimation of file requests for efficient data replication. IEEE Trans. Parallel Distrib. Syst. 2017, 28, 1839–1852. [Google Scholar] [CrossRef]
- Shang, Y. Deffuant model of opinion formation in one-dimensional multiplex networks. J. Phys. Math. Theor. 2015, 48, 395101. [Google Scholar] [CrossRef]
- Jackson, J.R. Jobshop-like Queueing Systems. Manag. Sci. 1963, 10, 131–142. [Google Scholar] [CrossRef]
Parameter Symbol | Meaning |
---|---|
Mean users’ job arrival rate per second | |
Layer i mean arrival rate, –4 (Layer 4 is for end nodes) | |
Layer i mean average service time | |
Probability that data transfers between Layers 1 and 2 is uninterrupted | |
(no collapsed links or corrupted data) | |
Similar to , for communication between Layers 2 and 3 | |
Similar to , for communication between Layer 3 and the end hosts | |
N | Number of user jobs |
Number of jobs in each of the layers |
MRT | Throughput (In Jobs/s) | |
---|---|---|
6.875 | 0.34 | 0.41 |
8.75 | 0.41 | 0.53 |
10.625 | 0.51 | 0.64 |
12.5 | 0.69 | 0.75 |
14.375 | 1.1 | 0.87 |
15 | 1.4 | 0.91 |
16.25 | 6 | 0.98 |
Time (in Seconds) | Load Transmitted (in Mbytes) | Bandwidth Utilization |
---|---|---|
0 | 380 | 0.88 |
350 | 458 | 0.73 |
700 | 517 | 0.65 |
1050 | 580 | 0.58 |
1400 | 640 | 0.52 |
1750 | 700 | 0.48 |
2100 | 760 | 0.44 |
2450 | 830 | 0.4 |
2800 | 920 | 0.36 |
3150 | 1024 | 0.32 |
3500 | 2048 | 0.16 |
3850 | 3072 | 0.1 |
4200 | 4608 | 0.09 |
4550 | 4915 | 0.07 |
Average Load Transmitted (in Jobs/s) | MRT |
---|---|
6 | 0.32 |
8 | 0.38 |
9 | 0.42 |
10 | 0.8 |
11 | 0.55 |
12 | 0.64 |
13 | 0.77 |
14 | 0.98 |
15 | 1.4 |
16 | 3.17 |
16.43 | 46.58 |
16.45 | 501 |
© 2019 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Souravlas, S. ProMo: A Probabilistic Model for Dynamic Load-Balanced Scheduling of Data Flows in Cloud Systems. Electronics 2019, 8, 990. https://doi.org/10.3390/electronics8090990
Souravlas S. ProMo: A Probabilistic Model for Dynamic Load-Balanced Scheduling of Data Flows in Cloud Systems. Electronics. 2019; 8(9):990. https://doi.org/10.3390/electronics8090990
Chicago/Turabian StyleSouravlas, Stavros. 2019. "ProMo: A Probabilistic Model for Dynamic Load-Balanced Scheduling of Data Flows in Cloud Systems" Electronics 8, no. 9: 990. https://doi.org/10.3390/electronics8090990
APA StyleSouravlas, S. (2019). ProMo: A Probabilistic Model for Dynamic Load-Balanced Scheduling of Data Flows in Cloud Systems. Electronics, 8(9), 990. https://doi.org/10.3390/electronics8090990