Abstract
Processing systems are in continuous evolution thanks to the constant technological advancement and architectural progress. Over the years, computing systems have become more and more powerful, providing support for applications, such as Machine Learning, that require high computational power. However, the growing complexity of modern computing units and applications has had a strong impact on power consumption. In addition, the memory plays a key role on the overall power consumption of the system, especially when considering data-intensive applications. These applications, in fact, require a lot of data movement between the memory and the computing unit. The consequence is twofold: Memory accesses are expensive in terms of energy and a lot of time is wasted in accessing the memory, rather than processing, because of the performance gap that exists between memories and processing units. This gap is known as the memory wall or the von Neumann bottleneck and is due to the different rate of progress between complementary metal–oxide semiconductor (CMOS) technology and memories. However, CMOS scaling is also reaching a limit where it would not be possible to make further progress. This work addresses all these problems from an architectural and technological point of view by: (1) Proposing a novel Configurable Logic-in-Memory Architecture that exploits the in-memory computing paradigm to reduce the memory wall problem while also providing high performance thanks to its flexibility and parallelism; (2) exploring a non-CMOS technology as possible candidate technology for the Logic-in-Memory paradigm.
Subject
Electrical and Electronic Engineering,Mechanical Engineering,Control and Systems Engineering
Reference56 articles.
1. 2013 International Technology Roadmap for Semiconductors (ITRS)
http://www.itrs2.net/2013-itrs.html
2. 2009 International Technology Roadmap for Semiconductors (ITRS)
https://www.semiconductors.org/wp-content/uploads/2018/09/Interconnect.pdf
3. Design and Analysis of 3D-MAPS (3D Massively Parallel Processor with Stacked Memory)
4. A 3D-stacked logic-in-memory accelerator for application-specific data intensive computing
5. A scalable processing-in-memory accelerator for parallel graph processing
Cited by
41 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献