An Asynchronous Compaction Acceleration Scheme for Near-Data Processing-enabled LSM-Tree-based KV Stores

Author:

Sun Hui1,Lou Bendong1,Zhao Chao1,Kong Deyan1,Zhang Chaowei2,Huang Jianzhong3,Yue Yinliang4,Qin Xiao5

Affiliation:

1. Anhui University, China

2. Yangzhou University, China

3. Huazhong University of Science and Technology, China

4. Zhongguancun Laboratory, China

5. Auburn University, USA

Abstract

LSM-tree-based key-value stores (KV stores) convert random-write requests to sequence-write ones to achieve high I/O performance. Meanwhile, compaction operations in KV stores update SSTables in forms of reorganizing low-level data components to high-level ones, thereby guaranteeing an orderly data layout in each component. Repeated writes caused by compaction ( a.k.a, write amplification) impacts I/O bandwidth and overall system performance. Near-data processing (NDP) is one of effective approaches to addressing this write-amplification issue. Most NDP-based techniques adopt synchronous parallel schemes to perform a compaction task on both the host and its NDP-enabled device. In synchronous parallel compaction schemes, the execution time of compaction is determined by a subsystem that has lower compaction performance coupled by under-utilized computing resources in a NDP framework. To solve this problem, we propose an asynchronous parallel scheme named PStore to improve the compaction performance in KV stores. In PStore, we designed a multi-tasks queue and three priority-based scheduling methods. PStore elects proper compaction tasks to be offloaded in host- and device-side compaction modules. Our proposed cross-leveled compaction mechanism mitigates write amplification induced by asynchronous compaction. PStore featured with the asynchronous compaction mechanism fully utilizes computing resources in both host and device-side subsystems. Compared with the two popular synchronous compaction modes based on KV stores (TStore and LevelDB), our PStore immensely improves the throughput by up to a factor of 14 and 10.52 with an average of a factor of 2.09 and 1.73, respectively.

Publisher

Association for Computing Machinery (ACM)

Subject

Hardware and Architecture,Software

Reference44 articles.

1. AAEON. 2021. RICO-3399. https://newdata.aaeon.com.tw/DOWNLOAD/2014%20datasheet/Boards/RICO-3399.pdf AAEON. 2021. RICO-3399. https://newdata.aaeon.com.tw/DOWNLOAD/2014%20datasheet/Boards/RICO-3399.pdf

2. Near-Data Processing: Insights from a MICRO-46 Workshop

3. Helen  HW Chan , Chieh- Jan Mike Liang , Yongkun Li , Wenjia He , Patrick  PC Lee , Lianjie Zhu , Yaozu Dong , Yinlong Xu , Yu Xu , Jin Jiang , et al . 2018 . Hashkv : Enabling efficient updates in {KV} storage via hashing. In 2018 {USENIX} Annual Technical Conference ( {USENIX}{ATC} 18). 1007–1019. Helen HW Chan, Chieh-Jan Mike Liang, Yongkun Li, Wenjia He, Patrick PC Lee, Lianjie Zhu, Yaozu Dong, Yinlong Xu, Yu Xu, Jin Jiang, et al. 2018. Hashkv: Enabling efficient updates in {KV} storage via hashing. In 2018 {USENIX} Annual Technical Conference ({USENIX}{ATC} 18). 1007–1019.

4. Bigtable

5. Benjamin  Y Cho , Won Seob Jeong , Doohwan Oh , and Won Woo Ro . 2013 . Xsd: Accelerating mapreduce by harnessing the gpu inside an ssd . In Proceedings of the 1st Workshop on Near-Data Processing. Benjamin Y Cho, Won Seob Jeong, Doohwan Oh, and Won Woo Ro. 2013. Xsd: Accelerating mapreduce by harnessing the gpu inside an ssd. In Proceedings of the 1st Workshop on Near-Data Processing.

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Coordinating Compaction Between LSM-Tree Based Key-Value Stores for Edge Federation;2024 IEEE 17th International Conference on Cloud Computing (CLOUD);2024-07-07

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3