Abstract
AbstractWith the increased focus on quantum circuit learning for near-term applications on quantum devices, in conjunction with unique challenges presented by cost function landscapes of parametrized quantum circuits, strategies for effective training are becoming increasingly important. In order to ameliorate some of these challenges, we investigate a layerwise learning strategy for parametrized quantum circuits. The circuit depth is incrementally grown during optimization, and only subsets of parameters are updated in each training step. We show that when considering sampling noise, this strategy can help avoid the problem of barren plateaus of the error surface due to the low depth of circuits, low number of parameters trained in one step, and larger magnitude of gradients compared to training the full circuit. These properties make our algorithm preferable for execution on noisy intermediate-scale quantum devices. We demonstrate our approach on an image-classification task on handwritten digits, and show that layerwise learning attains an 8% lower generalization error on average in comparison to standard learning schemes for training quantum circuits of the same size. Additionally, the percentage of runs that reach lower test errors is up to 40% larger compared to training the full circuit, which is susceptible to creeping onto a plateau during training.
Publisher
Springer Science and Business Media LLC
Subject
Applied Mathematics,Artificial Intelligence,Computational Theory and Mathematics,Theoretical Computer Science,Software
Reference48 articles.
1. Arute F, Arya K, Babbush R, Bacon D, Bardin JC, Barends R, Biswas R, Boixo S, Brandao Fernando GSL, Buell DA et al (2019) Nature 574(7779):505–510
2. Benedetti M, Garcia-Pintos D, Perdomo O, Leyton-Ortega V, Nam Y, Perdomo-Ortiz A (2019a) npj Quantum Inf 5(1):45. https://doi.org/10.1038/s41534-019-0157-8, http://www.nature.com/articles/s41534-019-0157-8
3. Benedetti M, Grant E, Wossnig L, Severini S (2019b) New J Phys 21(4):043023. https://doi.org/10.1088/1367-2630/ab14b5, http://stacks.iop.org/1367-2630/21/i=4/a=043023?key=crossref.0b5ab94ed3e2ea2943830f1d0073c780
4. Bengio Y, Bengio Y, Lamblin P, Popovici D, Larochelle H (2007) Greedy layer-wise training of deep networks. Advances in Neural Information Processing. http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.70.2022
5. Boixo S, Isakov S V, Smelyanskiy VN, Babbush R, Ding N, Jiang Z, Bremner MJ, Martinis JM, Neven H (2018) Nat Phys 14(6):595–600. https://doi.org/10.1038/s41567-018-0124-x, http://www.nature.com/articles/s41567-018-0124-x
Cited by
171 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献