Improving Spark Application Throughput Via Memory Aware Task Co-location : A Mixture of Experts Approach

Sanz Marco, Vicent and Taylor, Ben and Porter, Barry Francis and Wang, Zheng (2017) Improving Spark Application Throughput Via Memory Aware Task Co-location : A Mixture of Experts Approach. In: Middleware '17 Proceedings of the 18th ACM/IFIP/USENIX Middleware Conference :. ACM, New York, pp. 95-108. ISBN 9781450347204

[thumbnail of middleware17-paper74]
PDF (middleware17-paper74)
middleware17_paper74.pdf - Accepted Version
Available under License Creative Commons Attribution-NonCommercial.

Download (14MB)


Data analytic applications built upon big data processing frameworks such as Apache Spark are an important class of applications. Many of these applications are not latency-sensitive and thus can run as batch jobs in data centers. By running multiple applications on a computing host, task co-location can significantly improve the server utilization and system throughput. However, effective task co-location is a non-trivial task, as it requires an understanding of the computing resource requirement of the co-running applications, in order to determine what tasks, and how many of them, can be co-located. State-of-the-art co-location schemes either require the user to supply the resource demands which are often far beyond what is needed; or use a one-size-fits-all function to estimate the requirement, which, unfortunately, is unlikely to capture the diverse behaviors of applications. In this paper, we present a mixture-of-experts approach to model the memory behavior of Spark applications. We achieve this by learning, off-line, a range of specialized memory models on a range of typical applications; we then determine at runtime which of the memory models, or experts, best describes the memory behavior of the target application. We show that by accurately estimating the resource level that is needed, a co-location scheme can effectively determine how many applications can be co-located on the same host to improve the system throughput, by taking into consideration the memory and CPU requirements of co-running application tasks. Our technique is applied to a set of representative data analytic applications built upon the Apache Spark framework. We evaluated our approach for system throughput and average normalized turnaround time on a multi-core cluster. Our approach achieves over 83.9% of the performance delivered using an ideal memory predictor. We obtain, on average, 8.69x improvement on system throughput and a 49% reduction on turnaround time over executing application tasks in isolation, which translates to a 1.28x and 1.68x improvement over a state-of-the-art co-location scheme for system throughput and turnaround time respectively.

Item Type:
Contribution in Book/Report/Proceedings
Additional Information:
© Owner/Author ACM, 2017. This is the author's version of the work. It is posted here for your personal use. Not for redistribution. The definitive Version of Record was published in Middleware '17 Proceedings of the 18th ACM/IFIP/USENIX Middleware Conference
ID Code:
Deposited By:
Deposited On:
04 Sep 2017 12:52
Last Modified:
19 Apr 2024 00:06