Similarity Search in the Blink of an Eye with Compressed Indices

Author:

Aguerrebere Cecilia1,Bhati Ishwar Singh2,Hildebrand Mark2,Tepper Mariano2,Willke Theodore2

Affiliation:

1. Intel Labs, Santa Clara, California

2. Intel Labs, Hillsboro, Oregon

Abstract

Nowadays, data is represented by vectors. Retrieving those vectors, among millions and billions, that are similar to a given query is a ubiquitous problem, known as similarity search, of relevance for a wide range of applications. Graph-based indices are currently the best performing techniques for billion-scale similarity search. However, their random-access memory pattern presents challenges to realize their full potential. In this work, we present new techniques and systems for creating faster and smaller graph-based indices. To this end, we introduce a novel vector compression method, Locally-adaptive Vector Quantization (LVQ), that uses per-vector scaling and scalar quantization to improve search performance with fast similarity computations and a reduced effective bandwidth, while decreasing memory footprint and barely impacting accuracy. LVQ, when combined with a new high-performance computing system for graph-based similarity search, establishes the new state of the art in terms of performance and memory footprint. For billions of vectors, LVQ outcompetes the second-best alternatives: (1) in the low-memory regime, by up to 20.7x in throughput with up to a 3x memory footprint reduction, and (2) in the high-throughput regime by 5.8x with 1.4x less memory.

Publisher

Association for Computing Machinery (ACM)

Subject

General Earth and Planetary Sciences,Water Science and Technology,Geography, Planning and Development

Reference60 articles.

1. Cecilia Aguerrebere Ishwar Bhati Mark Hildebrand Mariano Tepper and Ted Willke. 2023. Similarity search in the blink of an eye with compressed indices. arXiv:2304.04759 [cs.LG] Cecilia Aguerrebere Ishwar Bhati Mark Hildebrand Mariano Tepper and Ted Willke. 2023. Similarity search in the blink of an eye with compressed indices. arXiv:2304.04759 [cs.LG]

2. Approximate nearest neighbors and the fast Johnson-Lindenstrauss transform

3. Quicker ADC : Unlocking the Hidden Potential of Product Quantization With SIMD

4. Cache locality is not enough

5. Sunil Arya and David M Mount . 1993 . Approximate nearest neighbor queries in fixed dimensions . In Proceedings of the fourth annual ACM-SIAM symposium on Discrete algorithms , Vol. 93 . 271--280. Sunil Arya and David M Mount. 1993. Approximate nearest neighbor queries in fixed dimensions. In Proceedings of the fourth annual ACM-SIAM symposium on Discrete algorithms, Vol. 93. 271--280.

Cited by 5 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search;Proceedings of the ACM on Management of Data;2024-05-29

2. Reinforcement Learning Infused MAC for Adaptive Connectivity;2024 IEEE Wireless Communications and Networking Conference (WCNC);2024-04-21

3. Optimizing Resource Utilization Using Vector Databases in Green Internet of Things;2023 IEEE Globecom Workshops (GC Wkshps);2023-12-04

4. Lossy Compression Options for Dense Index Retention;Proceedings of the Annual International ACM SIGIR Conference on Research and Development in Information Retrieval in the Asia Pacific Region;2023-11-26

5. Similarity Search in the Blink of an Eye with Compressed Indices;Proceedings of the VLDB Endowment;2023-07

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3