Affiliation:
1. School of Physics,
The University of Melbourne, Parkville, 3010 VIC, Australia.
2. Data61, CSIRO, Clayton, 3168 VIC, Australia.
Abstract
Quantum machine learning (QML) is emerging as an application of quantum computing with the potential to deliver quantum advantage, but its realization for practical applications remains impeded by challenges. Among these, a key barrier is the computationally expensive task of encoding classical data into a quantum state, which could erase any prospective speedups over classical algorithms. In this study, we implement methods for the efficient preparation of quantum states representing encoded image data using variational, genetic, and matrix product state algorithms. Our results show that these methods can approximately prepare states to a level suitable for QML using circuits 2 orders of magnitude shallower than a standard state preparation implementation, thus drastically reducing circuit depth and gate count without unduly sacrificing classification accuracy. Additionally, the QML models trained and evaluated on approximately encoded data displayed an increased robustness to adversarially generated input data perturbations. This partial alleviation of adversarial vulnerability, made possible by the retention of the meaningful large-scale features of the data despite the “drowning out” of adversarial perturbations, constitutes a considerable benefit for approximate state preparation in addition to lessening the requirements of the quantum hardware. Our results, based on simulations and experiments on IBM quantum devices, highlight a promising pathway for the future implementation of accurate and robust QML models on complex datasets relevant for practical applications, bringing the possibility of NISQ-era QML advantage closer to reality.
Funder
Australian Army Quantum Technology Challenge
Australian Research Council
Publisher
American Association for the Advancement of Science (AAAS)
Reference84 articles.
1. Bubeck S Chandrasekaran V Eldan R Gehrke J Horvitz E Kamar E Lee P Lee YT Li Y Lundberg S et al. Sparks of artificial general intelligence: Early experiments with GPT-4. arXiv. 2023. https://doi.org/10.48550/arXiv.2303.12712
2. Chen M Tworek J Jun H Yuan Q de Oliveira Pinto HP Kaplan J Edwards H Burda Y Joseph N Brockman G et al. Evaluating large language models trained on code. arXiv. 2021. https://doi.org/10.48550/arXiv.2107.03374
3. Rombach R Blattmann A Lorenz D Esser P Ommer B. High-resolution image synthesis with latent diffusion models. Paper presented at: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2022 June 18–24; New Orleans LA USA.
4. Radford A Radford A Kim JW Xu T Brockman G McLeavey C Sutskever I. Robust speech recognition via large-scale weak supervision. arXiv. 2022. https://doi.org/10.48550/arXiv.2212.04356
5. Ho J Chan W Saharia C Whang J Gao R Gritsenko A Kingma DP Poole B Norouzi M Fleet DJ et al. Imagen video: High definition video generation with diffusion models. arXiv. 2022. https://doi.org/10.48550/arXiv.2210.02303