Cluster Computing and the Grid, IEEE International Symposium on (2008)
May 19, 2008 to May 22, 2008
Large scale production grids are a major case for autonomic computing. Following the classical definition of Kephart, an autonomic computing system should optimize its own behavior in accordance with high level guidance from humans. This central tenet of this paper is that the combination of utility functions and reinforcement learning (RL) can provide a general and efficient method for dynamically allocating grid resources in order to optimize the satisfaction of both end-users and participating institutions. The flexibility of an RL-based system allows to model the state of the grid,the jobs to be scheduled, and the high-level objectives of the various actors on the grid. RL-based scheduling can seamlessly adapt its decisions to changes in the distributions ofinter-arrival time, QoS requirements, and resource availability. Moreover, it requires minimal prior knowledge about thetarget environment, including user requests and infrastructure. Our experimental results, both on a synthetic workloadand a real trace, show that RL is not only a realistic alternative to empirical scheduler design, but is able to outperform them.
J. Perez, C. Loomis, B. K?gl and C. Germain-Renaud, "Grid Differentiated Services: A Reinforcement Learning Approach," 2008 8th International Symposium on Cluster Computing and the Grid (CCGRID '08)(CCGRID), Lyon, 2008, pp. 287-294.