Affiliation:
1. National University of Singapore
2. University of Electronic Science and Technology of China
3. Zhejiang University
4. Beijing Institute of Technology
Abstract
Big data analytics is gaining massive momentum in the last few years. Applying machine learning models to big data has become an implicit requirement or an expectation for most analysis tasks, especially on high-stakes applications. Typical applications include sentiment analysis against reviews for analyzing on-line products, image classification in food logging applications for monitoring user's daily intake, and stock movement prediction. Extending traditional database systems to support the above analysis is intriguing but challenging. First, it is almost impossible to implement all machine learning models in the database engines. Second, expert knowledge is required to optimize the training and inference procedures in terms of efficiency and effectiveness, which imposes heavy burden on the system users. In this paper, we develop and present a system, called Rafiki, to provide the training and inference service of machine learning models. Rafiki provides distributed hyper-parameter tuning for the training service, and online ensemble modeling for the inference service which trades off between latency and accuracy. Experimental results confirm the efficiency, effectiveness, scalability and usability of Rafiki.
Subject
General Earth and Planetary Sciences,Water Science and Technology,Geography, Planning and Development
Cited by
50 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. SPOT: Structure Patching and Overlap Tweaking for Effective Pipelining in Privacy-Preserving MLaaS with Tiny Clients;2024 IEEE 44th International Conference on Distributed Computing Systems (ICDCS);2024-07-23
2. Loki: A System for Serving ML Inference Pipelines with Hardware and Accuracy Scaling;Proceedings of the 33rd International Symposium on High-Performance Parallel and Distributed Computing;2024-06-03
3. Biathlon: Harnessing Model Resilience for Accelerating ML Inference Pipelines;Proceedings of the VLDB Endowment;2024-06
4. LBSC: A Cost-Aware Caching Framework for Cloud Databases;2024 IEEE 40th International Conference on Data Engineering (ICDE);2024-05-13
5. A Multi-Task Learning Framework for Reading Comprehension of Scientific Tabular Data;2024 IEEE 40th International Conference on Data Engineering (ICDE);2024-05-13