Abstract
AbstractReinforcement learning has witnessed recent applications to a variety of tasks in quantum programming. The underlying assumption is that those tasks could be modeled as Markov decision processes (MDPs). Here, we investigate the feasibility of this assumption by exploring its consequences for single-qubit quantum state preparation and gate compilation. By forming discrete MDPs, we solve for the optimal policy exactly through policy iteration. We find optimal paths that correspond to the shortest possible sequence of gates to prepare a state or compile a gate, up to some target accuracy. Our method works in both the absence and presence of noise and compares favorably to other quantum compilation methods, such as the Ross–Selinger algorithm. This work provides theoretical insight into why reinforcement learning may be successfully used to find optimally short gate sequences in quantum programming.
Publisher
Springer Science and Business Media LLC
Subject
Computational Theory and Mathematics,Computer Networks and Communications,Statistical and Nonlinear Physics,Computer Science (miscellaneous)
Reference56 articles.
1. Norvig, P. & Russell, S. J. Artificial Intelligence: A Modern Approach (Pearson Education Limited, 2016).
2. Shalev-Schwartz, S. & Ben-David, S. Understanding Machine Learning: From Theory to Algorithms (Cambridge University Press, 2014).
3. Goodfellow, I., Bengio, Y. & Courville, A. Deep Learning (MIT Press, 2016).
4. Sutton, R. S. & Barto, A. G. Reinforcement Learning: An Introduction, 2nd edn (Bradford Books, Cambridge, MA, 2018).
5. Bellman, R. On the theory of dynamic programming. Proc. Natl. Acad. Sci. USA 38, 716 (1952).