A Speed-Up Technique for an Auto-Memoization Processor by Collectively Reusing Continuous Iterations
2010 First International Conference on Networking and Computing (2010)
Higashi, Hiroshima Japan
Nov. 17, 2010 to Nov. 19, 2010
DOI Bookmark: http://doi.ieeecomputersociety.org/10.1109/IC-NC.2010.46
We have proposed an auto-memoization processor based on computation reuse, and merged it with speculative multithreading based on value prediction into a parallel early computation. In the past model, the parallel early computation detects each iteration of loops as a reusable block. This paper proposes a new parallel early computation model, which integrates multiple continuous iterations into a reusable block automatically and dynamically without modifing executable binaries. We also propose a model for automatically detecting how many iterations should be integrated into one reusable block. Our model reduces the overhead of computation reuse, and further exploits reuse tables. The result of the experiment with SPEC CPU95 FP suite benchmarks shows that the new model improves the maximum speedup from 40.5% to 57.6%, and the average speedup from 15.0% to 26.0%.
Y. Nakashima, T. Ikegaya, H. Matsuo and T. Tsumura, "A Speed-Up Technique for an Auto-Memoization Processor by Collectively Reusing Continuous Iterations," 2010 First International Conference on Networking and Computing(ICNC), Higashi, Hiroshima Japan, 2010, pp. 63-70.