The Community for Technology Leaders
Green Image
Issue No. 04 - April (2014 vol. 63)
ISSN: 0018-9340
pp: 954-967
Qi Chen , Dept. of Comput. Sci., Peking Univ., Beijing, China
Cheng Liu , Dept. of Comput. Sci., Peking Univ., Beijing, China
Zhen Xiao , Dept. of Comput. Sci., Peking Univ., Beijing, China
MapReduce is a widely used parallel computing framework for large scale data processing. The two major performance metrics in MapReduce are job execution time and cluster throughput. They can be seriously impacted by straggler machines-machines on which tasks take an unusually long time to finish. Speculative execution is a common approach for dealing with the straggler problem by simply backing up those slow running tasks on alternative machines. Multiple speculative execution strategies have been proposed, but they have some pitfalls: (i) Use average progress rate to identify slow tasks while in reality the progress rate can be unstable and misleading, (ii) Cannot appropriately handle the situation when there exists data skew among the tasks, (iii) Do not consider whether backup tasks can finish earlier when choosing backup worker nodes. In this paper, we first present a detailed analysis of scenarios where existing strategies cannot work well. Then we develop a new strategy, maximum cost performance (MCP), which improves the effectiveness of speculative execution significantly. To accurately and promptly identify stragglers, we provide the following methods in MCP: (i) Use both the progress rate and the process bandwidth within a phase to select slow tasks, (ii) Use exponentially weighted moving average (EWMA) to predict process speed and calculate a task's remaining time, (iii) Determine which task to backup based on the load of a cluster using a cost-benefit model. To choose proper worker nodes for backup tasks, we take both data locality and data skew into consideration. We evaluate MCP in a cluster of 101 virtual machines running a variety of applications on 30 physical servers. Experiment results show that MCP can run jobs up to 39 percent faster and improve the cluster throughput by up to 44 percent compared to Hadoop-0.21.
Optimization, Silicon, Time factors, Algorithm design and analysis, Redundancy, Real-time systems, Indexes

Qi Chen, Cheng Liu and Zhen Xiao, "Improving MapReduce Performance Using Smart Speculative Execution Strategy," in IEEE Transactions on Computers, vol. 63, no. 4, pp. 954-967, 2014.
254 ms
(Ver 3.3 (11022016))