Affiliation:
1. Advanced Micro Devices, Inc.
2. Advanced Micro Devices, Inc. and University of California, Santa Barbara
Abstract
High-performance computing, enterprise, and datacenter servers are driving demands for higher total memory capacity as well as memory performance. Memory "cubes" with high per-package capacity (from 3D integration) along with high-speed point-to-point interconnects provide a scalable memory system architecture with the potential to deliver both capacity and performance. Multiple such cubes connected together can form a "Memory Network" (MN), but the design space for such MNs is quite vast, including multiple topology types and multiple memory technologies per memory cube.
In this work, we first analyze several MN topologies with different mixes of memory package technologies to understand the key tradeoffs and bottlenecks for such systems. We find that most of a MN's performance challenges arise from the interconnection network that binds the memory cubes together. In particular, arbitration schemes used to route through MNs, ratio of NVM to DRAM, and specific topologies used have dramatic impact on performance and energy results. Our initial analysis indicates that introducing non-volatile memory to the MN presents a unique tradeoff between memory array latency and network latency. We observe that placing NVM cubes in a specific order in the MN improves performance by reducing the network size/diameter up to a certain NVM to DRAM ratio. Novel MN topologies and arbitration schemes also provide performance and energy deltas by reducing the hop count of requests and response in the MN. Based on our analyses, we introduce three techniques to address MN latency issues: (1) Distance-based arbitration scheme to improve queuing latencies throughout the network, (2) skip-list topology, derived from the classic data structure, to improve network latency and link usage, and (3) the MetaCube, a denser memory cube that leverages advanced packaging technologies to improve latency by reducing MN size.
Publisher
Association for Computing Machinery (ACM)
Reference40 articles.
1. The Machine: A new kind of computer. https://www.labs.hpe.com/the-machine. The Machine: A new kind of computer. https://www.labs.hpe.com/the-machine.
2. Inc. Advanced Micro Devices. AMD SDK. http://developer.amd.com/tools-and-sdks. Inc. Advanced Micro Devices. AMD SDK. http://developer.amd.com/tools-and-sdks.
3. GARNET: A detailed on-chip network model inside a full-system simulator
4. A scalable processing-in-memory accelerator for parallel graph processing
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. A traffic-aware memory-cube network using bypassing;Microprocessors and Microsystems;2022-04
2. Innovations in the Memory System;Synthesis Lectures on Computer Architecture;2019-09-10