Tiled-MapReduce: Optimizing Resource Usages of Data-parallel Applications on Multicore with Tiling

被引:56
作者
Chen, Rong [1 ]
Chen, Haibo [1 ]
Zang, Binyu [1 ]
机构
[1] Fudan Univ, Parallel Proc Inst, Shanghai, Peoples R China
来源
PACT 2010: PROCEEDINGS OF THE NINETEENTH INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES | 2010年
关键词
MapReduce; Tiled-MapReduce; Tiling; Multicore; CORE;
D O I
10.1145/1854273.1854337
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The prevalence of chip multiprocessor opens opportunities of running data-parallel applications originally in clusters on a single machine with many cores. Map Reduce, a simple and elegant programming model to program large scale clusters, has recently been shown to be a promising alternative to harness the multicore platform. The differences such as memory hierarchy and communication patterns between clusters and multicore platforms raise new challenges to design and implement an efficient Map Reduce system on multicore. This paper argues that it is more efficient for MapReduce to iteratively process small chunks of data in turn than processing a large chunk of data at one time on shared memory multicore platforms. Based on the argument, we extend the general Map Reduce programming model with "tiling strategy", called Tiled-MapReduce (TMR). TMR partitions a large Map Reduce job into a number of small sub-jobs and iteratively processes one sub-job at a time with efficient use of resources; TMR finally merges the results of all sub-jobs for output. Based on Tiled-MapReduce, we design and implement several optimizing techniques targeting multicore, including the reuse of input and intermediate data structure among sub-jobs, a NUCA/NUMA-aware scheduler, and pipelining a sub-job's reduce phase with the successive sub-job's map phase, to optimize the memory, cache and CPU resources accordingly. We have implemented a prototype of Tiled-MapReduce based on Phoenix, an already highly optimized Map Reduce runtime for shared memory multiprocessors. The prototype, namely Ostrich, runs on an Intel machine with 16 cores. Experiments on four different types of benchmarks show that Ostrich saves up to 85% memory, causes less cache misses and makes more efficient uses of CPU cores, resulting in a speedup ranging from 1.2X to 3.3X.
引用
收藏
页码:523 / 534
页数:12
相关论文
共 29 条
[1]  
[Anonymous], P 2 ACM EUROPEAN C C, DOI DOI 10.1145/1272996.1273005
[2]  
[Anonymous], 2010, MITCSAILTR2010020
[3]  
[Anonymous], 2010, R PROJECT STAT COMPU
[4]  
Bialecki A., 2005, HADOOP FRAMEWORK RUN
[5]   Programming parallel algorithms [J].
Blelloch, GE .
COMMUNICATIONS OF THE ACM, 1996, 39 (03) :85-97
[6]   Thousand core chips-a technology perspective [J].
Borkar, Shekhar .
2007 44TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, VOLS 1 AND 2, 2007, :746-749
[7]  
BOYDWICKIZER S, 2008, P 8 USENIX S OP SYST, P43
[8]  
COLEMAN S, 1995, P ACM SIGPLAN C PROG, P279
[9]  
Condie T., 2010, P 7 USENIX S NETW SY
[10]  
Dean J, 2004, USENIX ASSOCIATION PROCEEDINGS OF THE SIXTH SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION (OSDE '04), P137