Affiliation:
1. University of California, Riverside, CA
Abstract
Instruction caches have traditionally been used to improve software performance. Recently, several tiny instruction cache designs, including filter caches and dynamic loop caches, have been proposed to instead reduce software power. We propose several new tiny instruction cache designs, including preloaded loop caches, and one-level and two-level hybrid dynamic/preloaded loop caches. We evaluate the existing and proposed designs on embedded system software benchmarks from both the Powerstone and MediaBench suites, on two different processor architectures, for a variety of different technologies. We show on average that filter caching achieves the best instruction fetch energy reductions of 60--80%, but at the cost of about 20% performance degradation, which could also affect overall energy savings. We show that dynamic loop caching gives good instruction fetch energy savings of about 30%, but that if a designer is able to profile a program, preloaded loop caching can more than double the savings. We describe automated methods for quickly determining the best loop cache configuration, methods useful in a core-based design flow.
Publisher
Association for Computing Machinery (ACM)
Subject
Hardware and Architecture,Software
Reference27 articles.
1. Selective cache ways: On-demand cache resource allocation;Albonesi D. H.;Journal of Instruction Level Parallelism.,2000
2. Artisan. http://www.artisan.com. Artisan. http://www.artisan.com.
3. Instruction buffering to reduce power in processors for signal processing
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献