TriCache: A User-Transparent Block Cache Enabling High-Performance Out-of-Core Processing with In-Memory Programs

Author:

Feng Guanyu1ORCID,Cao Huanqi1ORCID,Zhu Xiaowei2ORCID,Yu Bowen1ORCID,Wang Yuanwei1ORCID,Ma Zixuan1ORCID,Chen Shengqi1ORCID,Chen Wenguang1ORCID

Affiliation:

1. Tsinghua University, Beijing, Beijing, China

2. Ant Group, Beijing, Beijing, China

Abstract

Out-of-core systems rely on high-performance cache sub-systems to reduce the number of I/O operations. Although the page cache in modern operating systems enables transparent access to memory and storage devices, it suffers from efficiency and scalability issues on cache misses, forcing out-of-core systems to design and implement their own cache components, which is a non-trivial task. This study proposes TriCache, a cache mechanism that enables in-memory programs to efficiently process out-of-core datasets without requiring any code rewrite. It provides a virtual memory interface on top of the conventional block interface to simultaneously achieve user transparency and sufficient out-of-core performance. A multi-level block cache design is proposed to address the challenge of per-access address translations required by a memory interface. It can exploit spatial and temporal localities in memory or storage accesses to render storage-to-memory address translation and page-level concurrency control adequately efficient for the virtual memory interface. Our evaluation shows that in-memory systems operating on top of TriCache can outperform Linux OS page cache by more than one order of magnitude, and can deliver performance comparable to or even better than that of corresponding counterparts designed specifically for out-of-core scenarios.

Funder

NSFC

National Key Research and Development Plan of China

Publisher

Association for Computing Machinery (ACM)

Subject

Hardware and Architecture

Reference57 articles.

1. Marcos K. Aguilera, Nadav Amit, Irina Calciu, Xavier Deguillard, Jayneel Gandhi, Stanko Novakovic, Arun Ramanathan, Pratap Subrahmanyam, Lalith Suresh, Kiran Tati, et al. 2018. Remote regions: A simple abstraction for remote memory. In Proceedings of the 2018 USENIX Annual Technical Conference (USENIX’18). 775–787.

2. Nadav Amit. 2017. Optimizing the TLB shootdown algorithm with page access tracking. In Proceedings of the 2017 USENIX Annual Technical Conference (USENIX ATC’17). 27–39. https://www.usenix.org/conference/atc17/technical-sessions/presentation/amit.

3. Nadav Amit, Amy Tai, and Michael Wei. 2020. Don’t shoot down TLB shootdowns! In Proceedings of the 15th European Conference on Computer Systems (EuroSys’20). ACM, New York, NY, 1–14. 10.1145/3342195.3387518

4. Timo Bingmann Michael Axtmann Emanuel Jöbstl Sebastian Lamm Huyen Chau Nguyen Alexander Noe Sebastian Schlag Matthias Stumpp Tobias Sturm and Peter Sanders. 2016. Thrill: High-performance algorithmic distributed batch data processing with C++. arxiv:cs.DC/1608.05634 (2016).

5. Paolo Boldi, Marco Rosa, Massimo Santini, and Sebastiano Vigna. 2011. Layered label propagation: A multiresolution coordinate-free ordering for compressing social networks. In Proceedings of the 20th International Conference on World Wide Web (WWW’11). ACM, New York, NY, 587–596. DOI:10.1145/1963405.1963488

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Towards Survivable In-Memory Stores with Parity Coded NVRAM;2023 IEEE 22nd International Conference on Trust, Security and Privacy in Computing and Communications (TrustCom);2023-11-01

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3