Issue No. 02 - July-Dec. (2012 vol. 11)
DOI Bookmark: http://doi.ieeecomputersociety.org/10.1109/L-CA.2012.2
Justin Meza , Carnegie Mellon University, Pittsburgh
Jichuan Chang , Hewlett-Packard Labs, Palo Alto
HanBin Yoon , Carnegie Mellon University, Pittsburgh
Onur Mutlu , Carnegie Mellon University, Pittsburgh
Parthasarathy Ranganathan , Hewlett-Packard Labs, Palo Alto
Hybrid main memories composed of DRAM as a cache to scalable non-volatile memories such as phase-change memory (PCM) can provide much larger storage capacity than traditional main memories. A key challenge for enabling high-performance and scalable hybrid memories, though, is efficiently managing the metadata (e.g., tags) for data cached in DRAM at a fine granularity. Based on the observation that storing metadata off-chip in the same row as their data exploits DRAM row buffer locality, this paper reduces the overhead of fine-granularity DRAM caches by only caching the metadata for recently accessed rows on-chip using a small buffer. Leveraging the flexibility and efficiency of such a fine-granularity DRAM cache, we also develop an adaptive policy to choose the best granularity when migrating data into DRAM. On a hybrid memory with a 512MB DRAM cache, our proposal using an 8KB on-chip buffer can achieve within 6% of the performance of, and 18% better energy efficiency than, a conventional 8MB SRAM metadata store, even when the energy overhead due to large SRAM metadata storage is not considered.
Random access memory, Phase change materials, System-on-a-chip, Buffer storage, Bandwidth, Memory management, Cache memory, non-volatile memories, Random access memory, Phase change materials, System-on-a-chip, Buffer storage, Bandwidth, Memory management, Indexes, hybrid main memories, Cache memories, tag storage
O. Mutlu, P. Ranganathan, J. Chang, H. Yoon and J. Meza, "Enabling Efficient and Scalable Hybrid Memories Using Fine-Granularity DRAM Cache Management," in IEEE Computer Architecture Letters, vol. 11, no. , pp. 61-64, 2012.