Affiliation:
1. The University of Texas, & IBM Corp., Austin, TX, USA
2. The University of Texas, Austin, TX, USA
3. IBM, Yorktown Heights, NY, USA
Abstract
In computer architecture, caches have primarily been viewed as a means to hide memory latency from the CPU. Cache policies have focused on anticipating the CPU's data needs, and are mostly oblivious to the main memory. In this paper, we demonstrate that the era of many-core architectures has created new main memory bottlenecks, and mandates a new approach: coordination of cache policy with main memory characteristics. Using the cache for memory optimization purposes, we propose a
Virtual Write Queue
which dramatically expands the memory controller's visibility of processor behavior, at low implementation overhead. Through memory-centric modification of existing policies, such as scheduled writebacks, this paper demonstrates that performance limiting effects of highly-threaded architectures can be overcome. We show that through awareness of the physical main memory layout and by focusing on writes, both read and write average latency can be shortened, memory power reduced, and overall system performance improved. Through full-system cycle-accurate simulations of SPEC cpu2006, we demonstrate that the proposed
Virtual Write Queue
achieves an average 10.9% system-level throughput improvement on memory-intensive workloads, along with an overall reduction of 8.7% in memory power across the whole suite.
Publisher
Association for Computing Machinery (ACM)
Reference25 articles.
1. Read prediction algorithm to provide low latency reads with SDRAM cache;Borkenhagen J.;US Patent,2004
2. 3
. P. Glaskowsky "High-end server chips breaking records " http://news.cnet.com/8301--13512_3-10321740-23.html Aug. 2009. 3. P. Glaskowsky "High-end server chips breaking records " http://news.cnet.com/8301--13512_3-10321740-23.html Aug. 2009.
3. J. Hruska "Nehalem by the numbers: The Ars review " http://arstechnica.com/hardware/reviews/2008/11/nehalem-launch-review.ars/3. J. Hruska "Nehalem by the numbers: The Ars review " http://arstechnica.com/hardware/reviews/2008/11/nehalem-launch-review.ars/3.
Cited by
42 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. CABARRE: Request Response Arbitration for Shared Cache Management;ACM Transactions on Embedded Computing Systems;2023-09-09
2. Efficient Resource Scheduling for Runtime Reconfigurable Systems on FPGAs;2023 33rd International Conference on Field-Programmable Logic and Applications (FPL);2023-09-04
3. Write-Light Cache for Energy Harvesting Systems;Proceedings of the 50th Annual International Symposium on Computer Architecture;2023-06-17
4. Adaptively Reduced DRAM Caching for Energy-Efficient High Bandwidth Memory;IEEE Transactions on Computers;2022-10-01
5. DR-STRaNGe: End-to-End System Design for DRAM-based True Random Number Generators;2022 IEEE International Symposium on High-Performance Computer Architecture (HPCA);2022-04