Cluster Computing and the Grid, IEEE International Symposium on (2010)
Melbourne, VIC, Australia
May 17, 2010 to May 20, 2010
Map-reduce framework has received a significant attention and is being used for programming both large-scale clusters and multi-core systems. While the high productivity aspect of map-reduce has been well accepted, it is not clear if the API results in efficient implementations for different subclasses of data-intensive applications. In this paper, we present a system MATE (Map-reduce with an Alternate API), that provides a high-level, but distinct API. Particularly, our API includes a programmer-managed reduction object, which results in lower memory requirements at runtime for many data-intensive applications. MATE implements this API on top of the Phoenix system, a multi-core map-reduce implementation from Stanford. We evaluate our system using three data mining applications, and compare its performance to that of both Phoenix and Hadoop. Our results show that for all the three applications, MATE outperforms Phoenix and Hadoop. Despite achieving good scalability, MATE also maintains the easy-to-use API of map-reduce. Overall, we argue that, our approach, which is based on the generalized reduction structure, provides an alternate high-level API, leading to more efficient and scalable implementations
Data-intensive computing, Multi-Core Architectures, Middleware, Map-Reduce
V. T. Ravi, G. Agrawal and W. Jiang, "A Map-Reduce System with an Alternate API for Multi-core Environments," Cluster Computing and the Grid, IEEE International Symposium on(CCGRID), Melbourne, VIC, Australia, 2010, pp. 84-93.